Back to Blog

Northeastern University: Foundations of Large Language Models

Jeremy WeaverJanuary 27, 2025
Premium

Summary: The content explores foundational methods and advanced techniques in large language model development, including pre-training, generative architectures like Transformers, scaling strategies, alignment through reinforcement learning and instruction fine-tuning, and various prompting methods.

Northeastern University: Foundations of Large Language Models



Summary of https://arxiv.org/pdf/2501.09223

Detail foundational concepts and advanced techniques in large language model (LLM) development. It covers pre-training methods, including masked language modeling and discriminative training, and explores generative model architectures like Transformers.

The text also examines scaling LLMs for size and context length, along with alignment strategies such as reinforcement learning from human feedback (RLHF) and instruction fine-tuning.

Finally, it discusses prompting techniques, including chain-of-thought prompting and prompt optimization methods to improve LLM performance and alignment with human preferences.