Large language models: Section catalog

From Systems Analysis Wiki
Jump to navigation Jump to search

A catalog of articles from the Systems Analysis Wiki on the topic of Large Language Models (large language model, LLM).

Large Language Models (LLM)

  • Large Language Models
  • Theoretical Foundations of LLM
  • LLM Architectures
  • Transformer Architecture
  • Encoder
  • Decoder
  • Encoder-only
  • Decoder-only
  • Encoder-Decoder
  • Tokenization
  • Token
  • Embedding
  • Context Window
  • Training Large Language Models
  • Pre-training
  • Fine-tuning
  • In-Context Learning
  • Top-p
  • Top-k
  • Temperature (LLM)
  • Hallucinations and Incorrect LLM Responses
  • Data Distortion and Bias
  • Contextual Forgetting
  • Generation Bias
  • Mixture-of-Experts (MoE)
  • Reducing LLM Errors
  • Optimizing LLM Usage Costs
  • Open-weight and Closed-weight Models
  • Constitutional AI
  • Explainable AI
  • RLHF
  • Direct Preference Optimization
  • Low‑Rank Adaptation (LoRA)
  • PEFT
  • Vector Databases
  • Multimodal LLMs
  • Jailbreaks
  • FlashAttention
  • FlashAttention-2
  • FlashAttention-3
  • Stop sequences
  • Synthetic Data Generation
  • Multimodal Reasoning
  • Stochastic Parrot

Catalog of Large Language Models (LLM)

  • T5
  • LaMDA
  • PaLM
  • BERT
  • Chinchilla
  • Huawei PanGu
  • IBM Granite
  • BLOOM
  • Mixtral
  • DBRX
  • GPT
  • Claude
  • Gemma
  • Gemini
  • LLaMA
  • Mistral
  • DeepSeek
  • Grok
  • Qwen
  • Phi
  • Jais
  • Jamba
  • Cohere
  • Falcon
  • Perplexity
  • YandexGPT
  • Huggingface
  • OpenAI Large Language Models
  • Google Large Language Models
  • Large Language Models: Catalog of Models

Prompt Engineering (LLM)

AI Agents (LLM)

Evaluation and Metric Comparison (LLM)

  • LLM Evaluation
  • LLM Quality Metrics
  • Perplexity
  • BLEU
  • ROUGE
  • BERTScore
  • METEOR
  • MAUVE
  • LLM‑as‑a‑Judge

Benchmarks and Datasets (LLM)

  • LLM Benchmarks
  • MMLU benchmark
  • HellaSwag benchmark
  • HumanEval benchmark
  • TruthfulQA benchmark
  • MT-Bench benchmark
  • GLUE benchmark
  • SuperGLUE
  • Humanity's Last Exam
  • GSM8K (Grade School Math 8K)
  • WinoGrande benchmark
  • AgentHarm
  • SafetyBench
  • SWE-bench
  • BIG-bench
  • MATH benchmark
  • FLORES‑200
  • RealToxicityPrompts
  • PromptRobust
  • BOLD
  • BBQ
  • LMArena
  • ELO Ranking of Models