Please Select Your Location
Australia
Österreich
België
Canada
Canada - Français
中国
Česká republika
Denmark
Deutschland
France
HongKong
Iceland
Ireland
Italia
日本
Korea
Latvija
Lietuva
Lëtzebuerg
Malta
المملكة العربية السعودية (Arabic)
Nederland
New Zealand
Norge
Polska
Portugal
Russia
Saudi Arabia
Southeast Asia
Suisse
Suomi
Sverige
台灣
Ukraine
United Kingdom
United States
Please Select Your Location
België
Česká republika
Denmark
Iceland
Ireland
Italia
Latvija
Lietuva
Lëtzebuerg
Malta
Nederland
Norge
Polska
Portugal
Suisse
Suomi
Sverige
<< Back to Blog

Study Finds 'Beginner Mode' Training Enhances AI Performance

VIVE POST-WAVE Team • Feb. 26, 2025

4-minute read

Does AI training always need to be as close to reality as possible? A recent study by MIT and Harvard suggests that might not be the case! If you've ever played Pac-Man or any challenging game, you know how tough it can be on hard mode. Enemies are unpredictable, variables increase, and there's little room for error. Jumping straight into hard mode as a newbie often means game over before you even learn the rules or grasp the control mechanisms.

AI training faces a similar scenario. Traditionally, scientists believed AI should learn in "hard mode" environments that mimic the real world to prepare for future challenges. For instance, if we want a household robot to clean a kitchen, the intuitive approach is to train it in a simulated kitchen, not a factory. However, the research team discovered something surprising—training AI in "beginner mode" first actually helps it perform better in chaotic real-world situations!

This phenomenon is known as the "Indoor Training Effect," which challenges conventional AI training wisdom.

What is the "Indoor Training Effect"?

 

 

Think about learning to drive. Would you start on busy city streets or practice in a driving school lot first? The answer is clear—master the basics in a controlled environment before hitting the road.

The "Indoor Training Effect" is like a driving school for AI. Instead of throwing AI into a storm right away, it’s better to let it master core skills in a calm, controlled setting. Once AI has these skills down, it can handle the chaos and unpredictability of the real world more effectively.

How did they prove their hypothesis? Let AI Play Games

To test this idea, the research team used the classic game Pac-Man, setting up two different training methods based on the level of "noise."

Here, "noise" refers to random or unpredictable factors added to the game by researchers, making the environment more complex. Ghosts might suddenly move to unexpected positions. In reinforcement learning, these random factors are controlled by a "transition function"—determining the probability of the environment shifting to the next state when AI takes action. More noise in the transition function means greater challenges for AI.

1. Training Directly in a "High-Noise Environment" (Hard Mode)

Ghosts move unpredictably, and AI must navigate a constantly changing environment. It's like throwing a newbie player into the hardest level, facing frequent surprises from the start.

2. Training First in a "Low-Noise Environment" (Easy Mode)

Ghost movements are more predictable, allowing AI to focus on learning rules and strategies, similar to players starting in easy mode to grasp the game rhythm before tackling harder levels.

Pac-Man AI Training

AI playing Pac-Man. (Source: Atari)

The results showed that AI performed better when it first learned in easy mode before being tested in hard mode. Without distractions, AI could focus on learning the game's basic rules, like avoiding ghosts and eating pellets, which is the core concept of the "Indoor Training Effect."

Harvard researcher Spandan Madan noted, "We initially thought AI should train under conditions similar to the final test environment for optimal performance. But the results surprised us, showing AI learns better in simpler environments, offering a new perspective."

Applications of the "Indoor Training Effect"

Before the "Indoor Training Effect," AI training often aimed for high-fidelity environments close to reality or used "domain randomization" to introduce various random variables during simulation, ensuring AI could learn in different scenarios and maintain stable performance in real environments.

For example, Tesla's autonomous driving training relies heavily on real-world data, not simulations. With cameras and sensors on millions of Tesla vehicles worldwide, the system collects billions of image data daily, processed by the Dojo supercomputer for end-to-end learning, directly adapting AI to real-world variables like sudden road conditions and pedestrian crossings. While this method teaches AI the variability of real environments, it’s costly and challenging for AI to grasp basic rules in early learning stages.

Tesla AI Training

Tesla's autonomous system training process. (Source: Tesla AI)

However, experiments validating the "Indoor Training Effect" suggest that instead of immediately testing self-driving cars in bustling cities, it’s better to start in "low-noise" closed test tracks or simulation systems to learn basic driving, turning, and obstacle avoidance. Once AI masters core path planning and sensing skills, gradually introduce variables like jaywalking pedestrians and unpredictable traffic. This approach not only reduces experimental risks but also helps AI effectively grasp traffic rules.

Similarly, collaborative robots in factories might struggle to perform basic tasks like "picking up materials" if placed directly on busy production lines with numerous staff and robotic arms. Training in a "simple mode" test room first, where robots practice sensing object positions and adjusting grip strength, then gradually adding more noise (like moving personnel and irregularly placed parts), significantly improves learning efficiency and safety.

This research offers a fresh perspective on AI training strategies. Instead of rushing to simulate all real-world complexities, let AI build a solid foundation in a "beginner's village" before gradually facing real-world chaos and challenges. Perhaps in the future, not just household robots and self-driving cars, but any field requiring AI intervention can benefit from this "Indoor Training Effect."

Ultimately, designing the "ideal beginner's village" might be a crucial next step in AI development: how to quickly accumulate experience in a controlled environment and smoothly handle the ever-changing real world? This could be the key to AI's continued advancement from the lab to real society.