<RETURN_TO_BASE

OpenAI Unveils Groundbreaking Open-Weight LLMs: gpt-oss-120B and gpt-oss-20B for High-End Laptops and Phones

OpenAI introduces two powerful open-weight language models, gpt-oss-120B and gpt-oss-20B, allowing users to run advanced AI locally on laptops and phones with full customization and privacy.

Two Revolutionary Open-Weight Language Models from OpenAI

OpenAI has just transformed the AI landscape by releasing two open-weight language models: gpt-oss-120B and gpt-oss-20B. These models are fully downloadable, inspectable, fine-tunable, and runnable on personal hardware. This unprecedented move ushers in a new era of openness, customization, and computational power.

Why This Release Matters

Traditionally, OpenAI has been known for impressive models but limited access to their inner workings. On August 5, 2025, this changed dramatically. Both models are released under the Apache 2.0 license, allowing commercial and experimental use without restrictions. Users can now run these OpenAI-grade models locally, bypassing cloud limitations and gaining full control.

Meet gpt-oss-120B

  • Size: 117 billion parameters with 5.1 billion active parameters per token via Mixture-of-Experts (MoE) technology.
  • Performance: Comparable or superior to OpenAI’s o4-mini in real-world benchmarks.
  • Hardware Requirements: Runs on a single high-end GPU like Nvidia H100 or 80GB-class cards, no need for server farms.
  • Capabilities: Supports chain-of-thought reasoning and agentic functions, suitable for research automation, technical writing, and code generation.
  • Customization: Adjustable reasoning effort (low, medium, high) for balancing power and resource use.
  • Context Length: Can process up to 128,000 tokens, enough to read entire books.
  • Fine-Tuning: Designed for easy local customization and private inference with no rate limits or data privacy concerns.

Meet gpt-oss-20B

  • Size: 21 billion parameters with 3.6 billion active parameters per token, also using MoE.
  • Performance: Positioned between o3-mini and o4-mini models, making it a top-tier small model.
  • Hardware Requirements: Runs on consumer laptops with as little as 16GB RAM and even on phones.
  • Mobile Optimization: Supports low-latency on-device AI for smartphones (including Qualcomm Snapdragon) and edge devices.
  • Agentic Features: Can use APIs, generate structured outputs, and execute Python code dynamically.

Advanced Technology Behind the Models

Both models leverage Mixture-of-Experts architecture, activating only a few expert subnetworks per token. This provides extremely high parameter counts while keeping memory usage and inference times low. Additionally, MXFP4 quantization reduces memory footprint without accuracy loss, enabling the 120B model to fit on a single advanced GPU and the 20B model to run on laptops and mobile devices.

Practical Applications and Impact

  • For Enterprises: Enables on-premises deployment ensuring data privacy and compliance, ideal for sectors like finance, healthcare, and law.
  • For Developers: Offers freedom to experiment, fine-tune, and extend AI capabilities with no API limits or additional costs.
  • For the Community: Models are readily available on platforms like Hugging Face and Ollama, allowing quick download and deployment.

How GPT-OSS Compares

The gpt-oss-120B is the first open-weight model freely available that rivals commercial top-tier models. The 20B model brings powerful local AI to consumer devices, accelerating innovation in on-device AI applications.

A New Era of Open AI

OpenAI’s GPT-OSS models invite researchers, developers, and enterprises to fully engage with state-of-the-art AI. This release is not just about usage but about building, iterating, and evolving AI together.

For more information, check out the GPT-OSS technical blog, GitHub tutorials, and join the community on Twitter and ML SubReddit.

🇷🇺

Сменить язык

Читать эту статью на русском

Переключить на Русский