#transformer-models

[ follow ]
fromThegreenplace
1 week ago
Marketing tech

Sparsely-gated Mixture Of Experts (MoE)

The feed-forward layer in transformers plays a vital role in processing relationships between tokens.
#machine-learning
fromHackernoon
11 months ago
Miscellaneous

Using Large Language Models for Zero-Shot Video Generation: A VideoPoet Case Study | HackerNoon

VideoPoet synthesizes high-quality videos using a transformer model that integrates multiple conditioning signals across various modalities.
fromHackernoon
11 months ago
Miscellaneous

Using Large Language Models for Zero-Shot Video Generation: A VideoPoet Case Study | HackerNoon

VideoPoet synthesizes high-quality videos using a transformer model that integrates multiple conditioning signals across various modalities.
more#machine-learning
fromHackernoon
3 months ago
Miscellaneous

RNNs vs. Transformers: Innovations in Scalability and Efficiency | HackerNoon

RNNs can be efficiently scaled and trained, providing competitive alternatives to Transformer models for certain applications.
#memory-management
fromHackernoon
1 year ago
Data science

Evaluating the Performance of vLLM: How Did It Do? | HackerNoon

vLLM was tested using various Transformer-based large language models to evaluate its performance under load.
fromHackernoon
1 year ago
Miscellaneous

The Generation and Serving Procedures of Typical LLMs: A Quick Explanation | HackerNoon

Transformer-based language models use autoregressive approaches for token sequence probability modeling.
fromHackernoon
1 year ago
Miscellaneous

Batching Techniques for LLMs | HackerNoon

Batching improves compute utilization for LLMs, but naive strategies can cause delays and waste resources. Fine-grained batching techniques offer a solution.
fromHackernoon
1 year ago
Data science

Evaluating the Performance of vLLM: How Did It Do? | HackerNoon

vLLM was tested using various Transformer-based large language models to evaluate its performance under load.
fromHackernoon
1 year ago
Miscellaneous

The Generation and Serving Procedures of Typical LLMs: A Quick Explanation | HackerNoon

Transformer-based language models use autoregressive approaches for token sequence probability modeling.
fromHackernoon
1 year ago
Miscellaneous

Batching Techniques for LLMs | HackerNoon

Batching improves compute utilization for LLMs, but naive strategies can cause delays and waste resources. Fine-grained batching techniques offer a solution.
more#memory-management
fromHackernoon
1 year ago
Miscellaneous

Memory Challenges in LLM Serving: The Obstacles to Overcome | HackerNoon

LLM serving throughput is limited by GPU memory capacity, especially due to large KV cache demands.
fromHackernoon
9 months ago
Data science

Where does In-context Translation Happen in Large Language Models: Inference Efficiency | HackerNoon

Identifying task recognition in transformer models enables significant inference speed-ups.
Artificial intelligence
fromTheregister
10 months ago

Etched scores $120M for an ASIC built for transformer models

Etched is developing an inference chip, Sohu, specialized in serving transformer models, claiming a 20x performance advantage over Nvidia's H100 by focusing on a specific type of AI model.
[ Load more ]