supervised-fine-tuning

How OpenAI’s o3, Grok 3, DeepSeek R1, Gemini 2.0, and Claude 3.7 Differ in Their Reasoning Approaches

Large language models (LLMs) are rapidly evolving from easy text prediction systems into advanced reasoning engines able to tackling complex challenges. Initially designed to predict the following word in a sentence, these models have...

Unraveling Large Language Model Hallucinations

Introduction In a YouTube video titled , former Senior Director of AI at Tesla, Andrej Karpathy discusses the psychology of Large Language Models (LLMs) as emergent cognitive effects of the training pipeline. This text is inspired by his...

LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs

Current long-context large language models (LLMs) can process inputs as much as 100,000 tokens, yet they struggle to generate outputs exceeding even a modest length of two,000 words. Controlled experiments reveal that the model’s...

MoRA: High-Rank Updating for Parameter-Efficient Fantastic-Tuning

Owing to its robust performance and broad applicability when put next to other methods, LoRA or Low-Rank Adaption is some of the popular PEFT or Parameter Efficient Fantastic-Tuning methods for fine-tuning a big language...

RAFT – A High quality-Tuning and RAG Approach to Domain-Specific Query Answering

Because the applications of enormous language models expand into specialized domains, the necessity for efficient and effective adaptation techniques becomes increasingly crucial. Enter RAFT (Retrieval Augmented High quality Tuning), a novel approach that mixes...

Recent posts

Popular categories

ASK ANA