Everything you need to know about OpenAI's new open weight AI models, including price, performance, and where you can access them
Briefly

Everything you need to know about OpenAI's new open weight AI models, including price, performance, and where you can access them
"OpenAI has released two open weight LLMs, gpt-oss-120b and gpt-oss-20b, which can perform similarly to recent small language models on accessible hardware."
"Both models utilize a mixture of experts approach, allowing for faster inference and less costly pre-training through selective activation of parameters."
"The larger model, gpt-oss-120b, has 117 billion parameters and outperforms o4-mini in health and expert questioning benchmarks, though lags slightly in code completion."
"Sam Altman stated that gpt-oss is a state-of-the-art open-weights reasoning model, emphasizing its strong real-world performance and usability on local systems."
OpenAI introduced the gpt-oss-120b and gpt-oss-20b models, which are open weight large language models designed to run on smaller hardware. The gpt-oss-120b model features 117 billion parameters and can operate on a single A100 GPU, while the gpt-oss-20b has 21 billion parameters and can run with just 16GB of memory. Both models employ a mixture of experts approach for efficiency, performing comparably to the o4-mini model in various benchmarks. These models are intended to enhance transparency and safety in AI technology.
Read at IT Pro
Unable to calculate read time
[
|
]