#transformer-models

[ follow ]
Mobile UX
fromInfoQ
4 weeks ago

Gemma 3n Introduces Novel Techniques for Enhanced Mobile AI Inference

Gemma 3n enhances mobile AI applications with improved performance and efficiency through techniques like per-layer embeddings and transformer nesting.
fromHackernoon
1 year ago

How Do You Train an AI to Understand Time? With a Giant Pile of Data. | HackerNoon

The Time Series Pile consolidates diverse public time series datasets to enhance model pre-training capabilities for time series analysis.
fromHackernoon
10 months ago

Even AI Needs Glasses: When Space Images Get Too Fuzzy to Fix | HackerNoon

Transformers enhance astronomical image restoration but struggle with high noise levels.
fromThegreenplace
3 months ago

Sparsely-gated Mixture Of Experts (MoE)

The feed forward layer in transformer models is crucial for reasoning on token relationships, often housing most of the model's weights due to its larger dimensionality.
Marketing tech
Artificial intelligence
fromHackernoon
1 year ago

New AI Method Lets Models Decide What to Think About | HackerNoon

Mixture-of-Depths Transformers improve efficiency in transformer architectures by dynamically allocating computational resources.
fromHackernoon
1 year ago

Evaluating the Performance of vLLM: How Did It Do? | HackerNoon

vLLM was tested using various Transformer-based large language models to evaluate its performance under load.
Data science
fromHackernoon
1 year ago

Where does In-context Translation Happen in Large Language Models: Inference Efficiency | HackerNoon

Identifying task recognition in transformer models enables significant inference speed-ups.
[ Load more ]