large language model

LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs

Current long-context large language models (LLMs) can process inputs as much as 100,000 tokens, yet they struggle to generate outputs exceeding even a modest length of two,000 words. Controlled experiments reveal that the model’s...

The Most Powerful Open Source LLM Yet: Meta LLAMA 3.1-405B

Memory Requirements for Llama 3.1-405BRunning Llama 3.1-405B requires substantial memory and computational resources:GPU Memory: The 405B model can utilize as much as 80GB of GPU memory per A100 GPU for efficient inference. Using Tensor...

Acrylic, Jonathan ‘ALLM’ Tops Open Source ‘Tiger Leaderboard’

Artificial intelligence (AI) specialist Acrylic (CEO Park Oe-jin) announced on the first that its large language model (LLM) 'Jonathan Allm' ranked first within the open source category on the 'Tiger Leaderboard' operated by Weight...

Redefining Search: How Emerging Conversational Engines Overcome Outdated LLMs and Context-Less Traditional Search Engines

The appearance of conversational serps is redefining how we retrieve information online, shifting from traditional keyword searches to more natural, conversational interactions. By combining large language models (LLMs) with real-time web data, these recent...

MakinaRocks “Demand for LLM in industrial sites is skyrocketing… Will develop into a representative industrial AI smarter than ChatGPT”

MakinaRocks (CEO Seong-ho Yoon) is an organization representing the domestic ‘industrial artificial intelligence (AI)’ field. It was the one company in Korea to be included within the ‘Top 100 AI Firms’ announced by CB...

Crowdworks “Self-constructed high-quality data has definite market value”

Currently, scale AI is gaining momentum in the US. A month ago, this company attracted $1 billion value of investment at a worth of $13.8 billion (roughly 19 trillion won) from big tech...

MoRA: High-Rank Updating for Parameter-Efficient Fantastic-Tuning

Owing to its robust performance and broad applicability when put next to other methods, LoRA or Low-Rank Adaption is some of the popular PEFT or Parameter Efficient Fantastic-Tuning methods for fine-tuning a big language...

Deepnoid ranks 1st in 'Open Ko-LLM Leaderboard'… Average increase by 1 point in 1 week

Deepnoid (CEO Choi Woo-sik), Korea's first-generation medical artificial intelligence (AI) specialist, won the 'Open Ko-LLM Leaderboard' hosted by the Korea Intelligence and Information Society Agency (NIA) and Upstage. It was announced on the fifteenth...

Recent posts

Popular categories

ASK DUKE