
"Seizing upon a shift in the field of open-source artificial intelligence, chip giant Nvidia, whose processors dominate AI, has unveiled the third generation of its Nemotron family of open-source large language models. The new Nemotron 3 family scales the technology from what had been one-billion-parameter and 340-billion-parameter models, the number of neural weights, to three new models, ranging from 30 billion for Nano, 100 billion for Super, and 500 billion for Ultra."
"The Nano model, available now on the HuggingFace code hosting platform, increases the throughput in tokens per second by four times and extends the context window -- the amount of data that can be manipulated in the model's memory -- to one million tokens, seven times as large as its predecessor. Nvidia emphasized that the models aim to address several concerns for enterprise users of generative AI, who are concerned about accuracy, as well as the rising cost of processing an increasing number of tokens each time AI makes a prediction."
""With Nemotron 3, we are aiming to solve those problems of openness, efficiency, and intelligence," said Kari Briski, vice president of generative AI software at Nvidia, in an interview with ZDNET before the release. "This year alone, we had the most contributions and repositories on HuggingFace," she told me."
Nvidia released Nemotron 3, the third generation of its open-source large language models, in three sizes: Nano (30B), Super (100B) and Ultra (500B). The Nano model is available on HuggingFace and delivers four-times higher token throughput and a one-million-token context window, seven times larger than its predecessor. Nemotron 3 focuses on improving accuracy and reducing the cost of processing larger token volumes for enterprise generative-AI use. Nvidia plans to ship Super in January and Ultra in March or April. Reports suggest Meta is leaning away from open-source, and Nvidia asserts greater openness and data transparency while seeking increased enterprise adoption.
Read at ZDNET
Unable to calculate read time
Collection
[
|
...
]