Tag: BPE

Despite the rise of massive language models, tokenization remains essential for accuracy, efficiency, and cost control. Learn why subword methods like BPE and SentencePiece still shape how LLMs understand language.

Recent-posts

Latency Optimization for Large Language Models: Streaming, Batching, and Caching

Latency Optimization for Large Language Models: Streaming, Batching, and Caching

Aug, 1 2025

Domain-Specialized Large Language Models: Code, Math, and Medicine

Domain-Specialized Large Language Models: Code, Math, and Medicine

Oct, 3 2025

Why Tokenization Still Matters in the Age of Large Language Models

Why Tokenization Still Matters in the Age of Large Language Models

Sep, 21 2025

Caching and Performance in AI-Generated Web Apps: Where to Start

Caching and Performance in AI-Generated Web Apps: Where to Start

Dec, 14 2025

How Training Duration and Token Counts Affect LLM Generalization

How Training Duration and Token Counts Affect LLM Generalization

Dec, 17 2025