#model-optimization

[ follow ]
Artificial intelligence
fromArs Technica
1 week ago

Microsoft's "1bit" AI model runs on a CPU only, while matching larger systems

BitNet b1.58 provides high efficiency in memory and operation while maintaining comparable performance to larger models.
#machine-learning
Artificial intelligence
fromHackernoon
1 month ago

Rethinking AI Quantization: The Missing Piece in Model Efficiency | HackerNoon

Quantum strategies optimize LLM precision while balancing accuracy and effectiveness through methods like post-training quantization and quantization-aware training.
fromHackernoon
1 month ago
SOMA, SF

Mamba: A Generalized Sequence Model Backbone for AI | HackerNoon

Selective State Space Models enhance performance on discrete data but can hinder efficiency on continuous tasks.
Artificial intelligence
fromHackernoon
1 month ago

Rethinking AI Quantization: The Missing Piece in Model Efficiency | HackerNoon

Quantum strategies optimize LLM precision while balancing accuracy and effectiveness through methods like post-training quantization and quantization-aware training.
fromHackernoon
1 month ago
SOMA, SF

Mamba: A Generalized Sequence Model Backbone for AI | HackerNoon

Selective State Space Models enhance performance on discrete data but can hinder efficiency on continuous tasks.
more#machine-learning
Scala
fromHackernoon
1 month ago

The Hidden Power of "Cherry" Parameters in Large Language Models | HackerNoon

Parameter heterogeneity in LLMs shows that a small number of parameters greatly influence performance, leading to the development of the CherryQ quantization method.
fromHackernoon
3 months ago
Miscellaneous

Wonder3D: 3D Generative Models and Multi-View Diffusion Models | HackerNoon

Utilizing 2D diffusion models facilitates improved 3D asset generation and generalization due to limitations in 3D datasets.
fromHackernoon
1 year ago
Data science

Meet The AI Tag-Team Method That Reduces Latency in Your Model's Response | HackerNoon

Speculative decoding efficiently enhances AI inference in NLP by balancing speed and quality.
fromHackernoon
1 year ago
Data science

The Most Detailed Guide On MLOps: Part 2 | HackerNoon

MLOps involves managing artifacts like data, models, and code for efficient machine learning processes.
[ Load more ]