SYSTEM Cited by 1 source
Llama 2¶
Llama 2 is Meta's second-generation open-weight foundation model family, released July 2023 in 7B / 13B / 70B sizes, with chat variants produced via supervised fine-tuning (SFT) + reinforcement learning from human feedback (RLHF). Meta's open release was a pivotal moment for public access to competitive-tier LLM weights.
Significance on this wiki¶
- Base of Meta's internal web-monorepo RCA system (2024-06; sources/2024-08-23-meta-leveraging-ai-for-efficient-incident-response) — the 7B variant was fine-tuned via continued pre-training on internal wikis/Q&As/code followed by RCA-specific SFT.
- Predecessor of systems/llama-3 / systems/llama-3-1 — the subsequent generations that this wiki documents in more architectural depth (e.g. trained on Meta's 24K-GPU H100 RoCE + InfiniBand clusters per sources/2024-06-12-meta-how-meta-trains-large-language-models-at-scale).
Llama 2's full architectural disclosures (Transformer variant, training data composition, RLHF process) come from the Llama 2 paper + blog, not yet ingested on this wiki; this page is a stub oriented toward the wiki's internal cross-reference needs.
Training-pipeline shape (as reused by downstream teams)¶
Llama 2 is a base model (autoregressive next-token) in its non-chat form. The 2024 Meta RCA team's adaptation recipe was:
- Continued pre-training on internal Meta corpora ("limited and approved internal wikis, Q&As, and code").
- Mixed SFT combining Llama 2's original SFT data + internal context + a dedicated RCA SFT dataset (~5,000 instruction-tuning examples).
- A second SFT round to produce logprob-rankable ordered lists as output.
Seen in¶
- sources/2024-08-23-meta-leveraging-ai-for-efficient-incident-response — base of Meta's RCA ranker.
Related¶
- systems/llama-3 / systems/llama-3-1 — successor generations.
- systems/meta-rca-system — the production system built on top.
- concepts/continued-pretraining / concepts/supervised-fine-tuning — adaptation techniques.
- companies/meta