#ai-inference

[ follow ]
#amd
fromInfoQ
2 months ago

Google Enhances LiteRT for Faster On-Device Inference

LiteRT, previously TensorFlow Lite, enhances on-device ML inference by simplifying GPU and NPU integration, achieving up to 25x speed improvements and lower power usage.
Artificial intelligence
fromTechzine Global
2 months ago

Red Hat lays foundation for AI inferencing: Server and llm-d project

AI inferencing is crucial for unlocking the full potential of artificial intelligence, as it enables models to apply learned knowledge to real-world situations.
Artificial intelligence
Artificial intelligence
fromIT Pro
3 months ago

'TPUs just work': Why Google Cloud is betting big on its custom chips

Google's seventh generation TPU, 'Ironwood', aims to lead in AI workload efficiency and cost-effectiveness.
TPUs were developed with a cohesive hardware-software synergy, enhancing their utility for AI applications.
#nvidia
Silicon Valley
fromBusiness Insider
4 months ago

2 reasons why Nvidia's Jensen Huang isn't worried

Nvidia CEO Jensen Huang is confident in sustained demand for Nvidia chips due to new powerful GPUs and an industry shift towards AI inference.
[ Load more ]