Close Menu
Elon Musk Monitor
  • Home
  • Elon Musk
  • AI
  • Cybertruck
    • DOGE & Cryptocurrency
    • Financial & Business
  • Grok
    • Hyperloop & Urban Mobility
    • Innovations & Future Projects
  • Mars Colonization
  • Neuralink
    • Philanthropy & Humanitarian Efforts
    • Public Perception & Cultural Impact
    • SolarCity & Renewable Energy
  • SpaceX
  • Starlink
  • Tesla
    • The Boring Company
  • X

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

What's Hot

NASA sets new potential launch date for Ax-4 mission to ISS

June 15, 2025

Bitcoin Tests Critical $104K Support – Eyes On $97K If It Breaks

June 15, 2025

Boeing Dreamliner crash, military escalations darken Paris Air Show

June 15, 2025
Facebook X (Twitter) Instagram
Elon Musk Monitor
  • Home
  • About Us
  • Advertise with Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
  • Home
  • Elon Musk
  • AI
  • Cybertruck
    • DOGE & Cryptocurrency
    • Financial & Business
  • Grok
    • Hyperloop & Urban Mobility
    • Innovations & Future Projects
  • Mars Colonization
  • Neuralink
    • Philanthropy & Humanitarian Efforts
    • Public Perception & Cultural Impact
    • SolarCity & Renewable Energy
  • SpaceX
  • Starlink
  • Tesla
    • The Boring Company
  • X
Elon Musk Monitor
Home » Researchers Create a Low-Cost Open-Source AI Model to Analyse How OpenAI’s o1 Reasons
Grok

Researchers Create a Low-Cost Open-Source AI Model to Analyse How OpenAI’s o1 Reasons

elonmuskBy elonmuskFebruary 9, 2025No Comments3 Mins Read
Share Facebook Twitter Pinterest LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email Copy Link


Researchers from Stanford University and Washington University have developed an open-source artificial intelligence (AI) model that is comparable in performance to OpenAI’s o1 model. The main objective of the researchers was not to create a powerful reasoning-focused model but to understand how the San Francisco-based AI firm instructed its o1 series models to perform test time scaling. Notably, the researchers were able to showcase the methodology and replicate the model’s behaviour at an extremely low cost while using far fewer compute resources.

Researchers Develop S1-32B AI Model

The researchers detailed the methodology and process of developing the model in a study published in the pre-print journal arXiv. The process involved creating a synthetic dataset from a different AI model and using several new techniques such as ablation and supervised fine-tuning (SFT). The model is available in a GitHub listing.

It should be noted that the AI model was not built from scratch. The developers used the Qwen2.5-32B-Instruct and distilled it to create the s1-32B large language model (LLM). Released in September 2024, the model is capable but given its size and lack of reasoning capabilities, it cannot match up to OpenAI’s o1.

During the process, the researchers used the Gemini Flash Thinking application processing interface (API) to generate reasoning traces and responses. A total of 59,000 triplets of questions, reasoning traces (the chain of thought or CoT), and responses were extracted from the API. A dataset called the s1K was then created by selecting 1,000 high-quality, diverse, and difficult questions as well as the reasoning traces and the responses.

After creating the s1K dataset, the researchers performed supervised fine-tuning on the Qwen2.5-32B-Instruct model. For this, basic fine-tuning hyperparameters were used. The distillation process took 26 minutes of training on 16 Nvidia H100 GPUs.

Till this point, the researchers had no idea how OpenAI trained the models to “think” and how it managed to stop the thinking process. Without this, a model runs the risk of overthinking indefinitely as it second-guesses its output wasting valuable processing power.

While fine-tuning the model, the researcher found something interesting. They found that they could manipulate the inference time by adding and XML tags. Once a model reaches the end tag, it is told to change its voice to an authoritative tone for the final answer. Notably, inference time is the near real-time responses that a typical AI model generates. Anything more than this would require careful manipulation of the code.

With the s1-32B model, the researchers added a “wait” command to force it to think beyond the usual inference period. Once added, the model began second-guessing and verifying its output. Then, the tag was used to either shorten this test time scaling phase or lengthen it.

Then, the researchers also experimented with several other phrases such as “alternatively”, and “hmm”, but found that the best performance metrics were achieved when using the “wait” tag. By bringing the model close to the performance of o1, the researchers claim that this might be the method used by OpenAI to fine-tune its reasoning models.

A TechCrunch report claims that the researchers were able to create the s1-32B AI model under $50 (roughly Rs. 4,380), highlighting that creating a post-training structure for reasoning models can be done at an extremely low cost.



Source link

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
elonmusk
  • Website

Related Posts

Android 16 QPR1 Beta 2 Update for Pixel Reportedly Brings New Launch Animation for Gemini Overlay

June 15, 2025

Google, Scale AI’s Largest Customer, Said to Plan Split After Meta Deal

June 14, 2025

Meta AI Discovery Feed Is Reportedly Filled With Users’ Seemingly Private Chats

June 13, 2025
Leave A Reply Cancel Reply

Don't Miss
Cybertruck

Tesla Cybertruck police truck donor revealed

A batch of Tesla Cybertrucks were recently revealed to be a donation to the Las…

Tesla upgrades its ridiculous Cybertruck wiper after owners report issue

February 27, 2025

Tesla Cybertruck contract with State Dept. may have been modified after Biden admin

February 26, 2025

This Tesla Cybertruck feature helped it earn a ‘Best Tech’ award

February 25, 2025
Top Posts

Bitcoin Tests Critical $104K Support – Eyes On $97K If It Breaks

June 15, 2025

Inverse Head And Shoulders Signals Quiet Surge Ahead

June 15, 2025

On-Chain Analysis Disputes BTC’s $50K Crash

June 15, 2025

JPMorgan Unveils Record Profits in Q1 Analysis

June 15, 2025

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

About Us
About Us

Welcome to Elon Musk Monitor, your go-to source for comprehensive, up-to-date information on the life, work, and innovations of one of the most influential figures in the world today—Elon Musk. Our mission is to keep you informed about Musk’s ventures and projects, ranging from electric vehicles to space exploration, and everything in between. Whether you’re a tech enthusiast, investor, or simply curious about Musk’s impact on the world, we’ve got you covered.

Facebook X (Twitter) Pinterest YouTube WhatsApp
Our Picks

Bitcoin Tests Critical $104K Support – Eyes On $97K If It Breaks

June 15, 2025

Inverse Head And Shoulders Signals Quiet Surge Ahead

June 15, 2025

On-Chain Analysis Disputes BTC’s $50K Crash

June 15, 2025
Most Popular

How I met my partner on X/Twitter

February 8, 2025

DOGE staffer resigns after racist posts uncovered. Elon Musk might bring him back.

February 9, 2025

OpenAI accuses DeepSeek of stealing data, internet digs into the ‘irony’

February 9, 2025
  • Home
  • About Us
  • Advertise with Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 elonmuskmonitor. Designed by elonmuskmonitor.

Type above and press Enter to search. Press Esc to cancel.