Large language models: Section catalog
Jump to navigation
Jump to search
A catalog of articles from the Systems Analysis Wiki on the topic of Large Language Models (large language model, LLM).
Large Language Models (LLM)
- Large Language Models
- Theoretical Foundations of LLM
- LLM Architectures
- Transformer Architecture
- Encoder
- Decoder
- Encoder-only
- Decoder-only
- Encoder-Decoder
- Tokenization
- Token
- Embedding
- Context Window
- Training Large Language Models
- Pre-training
- Fine-tuning
- In-Context Learning
- Top-p
- Top-k
- Temperature (LLM)
- Hallucinations and Incorrect LLM Responses
- Data Distortion and Bias
- Contextual Forgetting
- Generation Bias
- Mixture-of-Experts (MoE)
- Reducing LLM Errors
- Optimizing LLM Usage Costs
- Open-weight and Closed-weight Models
- Constitutional AI
- Explainable AI
- RLHF
- Direct Preference Optimization
- Low‑Rank Adaptation (LoRA)
- PEFT
- Vector Databases
- Multimodal LLMs
- Jailbreaks
- FlashAttention
- FlashAttention-2
- FlashAttention-3
- Stop sequences
- Synthetic Data Generation
- Multimodal Reasoning
- Stochastic Parrot
Catalog of Large Language Models (LLM)
- T5
- LaMDA
- PaLM
- BERT
- Chinchilla
- Huawei PanGu
- IBM Granite
- BLOOM
- Mixtral
- DBRX
- GPT
- Claude
- Gemma
- Gemini
- LLaMA
- Mistral
- DeepSeek
- Grok
- Qwen
- Phi
- Jais
- Jamba
- Cohere
- Falcon
- Perplexity
- YandexGPT
- Huggingface
- OpenAI Large Language Models
- Google Large Language Models
- Large Language Models: Catalog of Models
Prompt Engineering (LLM)
- Prompt
- Prompt Engineering
- Prompt and Context
- Basic Prompt Engineering Techniques
- Retrieval‑Augmented Generation (RAG)
- Chain-of-Thought Prompting
- Few-shot and Zero-shot
- Role Prompting
- Tree of Thoughts
- Self‑refine prompting
- Self-consistency prompting
- Meta Prompting
- Multi‑agent prompting
- Prompt compression
- Program of Thoughts Prompting
- Generated Knowledge Prompting
- Multimodal CoT Prompting
- Graph-of-Thoughts
- Chain-of-Verification
- Toolformer
- Least-to-Most Prompting
- Automatic Prompt Engineer (APE)
- ReAct Prompting
- Function Calling
- RAG Patterns
- GraphRAG
- MM-RAG (Multimodal RAG)
- Hypothetical Document Expansion
- Hybrid Retrieval
- Packaging & Context Handling
- Prompt Engineering: Section Catalog
AI Agents (LLM)
- AI Agent
- Agentic workflows
- Multi‑Agent Frameworks
- LangChain
- AutoGPT
- Multi-Agent Debate
Evaluation and Metric Comparison (LLM)
- LLM Evaluation
- LLM Quality Metrics
- Perplexity
- BLEU
- ROUGE
- BERTScore
- METEOR
- MAUVE
- LLM‑as‑a‑Judge
Benchmarks and Datasets (LLM)
- LLM Benchmarks
- MMLU benchmark
- HellaSwag benchmark
- HumanEval benchmark
- TruthfulQA benchmark
- MT-Bench benchmark
- GLUE benchmark
- SuperGLUE
- Humanity's Last Exam
- GSM8K (Grade School Math 8K)
- WinoGrande benchmark
- AgentHarm
- SafetyBench
- SWE-bench
- BIG-bench
- MATH benchmark
- FLORES‑200
- RealToxicityPrompts
- PromptRobust
- BOLD
- BBQ
- LMArena
- ELO Ranking of Models