Tag: LLM quality

Explore how tokenizer design choices, vocabulary size, and algorithms like BPE and Unigram impact LLM accuracy, memory usage, and numerical reasoning.

Recent-posts

Compressed LLM Evaluation: Essential Protocols for 2026

Compressed LLM Evaluation: Essential Protocols for 2026

Feb, 5 2026

How Training Duration and Token Counts Affect LLM Generalization

How Training Duration and Token Counts Affect LLM Generalization

Dec, 17 2025

Preventing Catastrophic Forgetting During LLM Fine-Tuning: Techniques That Work

Preventing Catastrophic Forgetting During LLM Fine-Tuning: Techniques That Work

Apr, 1 2026

Procurement Checklists for Vibe Coding Tools: Security and Legal Terms You Can't Ignore

Procurement Checklists for Vibe Coding Tools: Security and Legal Terms You Can't Ignore

Jan, 21 2026

Prompt Sensitivity in Large Language Models: Why Small Word Changes Change Everything

Prompt Sensitivity in Large Language Models: Why Small Word Changes Change Everything

Oct, 12 2025