#llm-efficiency

[ follow ]
fromYcombinator
1 week ago
Roam Research

Show HN: LLMs consume 5.4x less mobile energy than ad-supported web search | Hacker News

Mobile LLM sessions consume 5.4x less energy than ad-supported web searches, with programmatic ads causing significant battery drain.
fromTechzine Global
3 months ago

DeepSeek breakthrough gives LLMs the highways it has long needed

As LLMs cannot grow infinitely large but do improve with size, researchers must find ways to make the technology effective at smaller scales. One well-known method is Mixture-of-Experts, where an LLM activates only a portion of itself to generate a response (text, photo, video) based on a prompt. This makes a larger model effectively smaller and faster during operation. mHC promises to be even more fundamental. It offers the chance to increase model complexity without the pain points of the past.
Artificial intelligence
Science
fromTechzine Global
5 months ago

Once again, DeepSeek suggests AI can be done much more efficiently

Feeding LLMs images of words (pixels) enables far more efficient processing, reducing model size, data footprint, and compute compared with raw word sequences.
[ Load more ]