Meta will offer its Llama AI model as an API too
Briefly

Meta has introduced a preview version of an API for its Llama large language models, transitioning from open-source models to an enterprise service. This initiative addresses key enterprise concerns, specifically vendor lock-in, allowing businesses to maintain control over their models. The API supports various languages and provides tools for fine-tuning, while being compatible with OpenAI’s SDK. Moreover, partnerships with AI chip manufacturers aim to enhance performance, challenging competitors like OpenAI by focusing on inference costs and system efficiency in AI production applications.
We want to make it even easier for you to quickly start building with Llama, while also giving you complete control over your models and weights without being locked into an API.
They're shifting the battlefield from model quality alone to inference cost, openness, and hardware advantage.
The solution includes tools for fine-tuning and evaluation, enabling developers to create custom versions of the new Llama 3.3 8B model - potentially reducing costs while improving performance for specific use cases.
Meta will collaborate with AI chip makers Cerebras and Groq to improve inferencing speed, a critical factor in production AI applications.
Read at InfoWorld
[
|
]