MechanizeWork Profile Banner
Mechanize Profile
Mechanize

@MechanizeWork

Followers
4K
Following
0
Media
19
Statuses
33

We're a software company building RL environments to power the full automation of the economy

San Francisco, CA
Joined April 2025
Don't wanna be here? Send us removal request.
@MechanizeWork
Mechanize
3 months
Today we’re announcing Mechanize, a startup focused on developing virtual work environments, benchmarks, and training data that will enable the full automation of the economy. We will achieve this by creating simulated environments and evaluations that capture the full scope of.
86
57
644
@MechanizeWork
Mechanize
3 days
The bitter lesson of the past decades is that the best algorithms are discovered by applying massive compute for search & learning. Progress in architectures has been incremental, while data-driven innovations have been highly impactful. We expect this trend to continue.
Tweet media one
@garybasin
Gary Basin
3 days
@vbot30000 @MechanizeWork the architecture isn't the bottleneck, it's the data.
1
0
34
@MechanizeWork
Mechanize
3 days
Read the blog post here:.
0
1
35
@MechanizeWork
Mechanize
3 days
Current RL methods like verifiable rewards can teach models to solve neat puzzles or prove theorems. But real-world tasks aren’t neatly packaged. To build genuinely capable AIs, we need richer RL environments, ones that capture the messiness of reality and reward good judgment.
Tweet media one
2
2
61
@MechanizeWork
Mechanize
3 days
Despite being trained on more compute than GPT-3, AlphaGo Zero could only play Go, while GPT-3 could write essays, code, translate languages, and assist with countless other tasks. That gap shows that what you train on matters. Rich RL environments are now the bottleneck.
Tweet media one
4
17
255
@MechanizeWork
Mechanize
4 days
xAI is probably the first to spend as much compute on RL as on pretraining. The easy gains from shifting compute to RL are now gone. With this arbitrage closed RL scaling will slow. Progress will now come from the quality and realism of RL environments rather than mere scaling.
Tweet media one
8
8
178
@MechanizeWork
Mechanize
4 days
Read the blog post here:
0
0
19
@MechanizeWork
Mechanize
4 days
The future is building software, not curating static datasets. Today’s AI systems learn best by interacting with digital environments, attempting tasks, and learning from outcomes. This demands dedicated, full-time specialists with strong expertise, not outsourced contractors.
Tweet media one
1
1
28
@MechanizeWork
Mechanize
4 days
Previously, AI progress relied heavily on monotonous, low-skill labeling from third-party contractors producing basic text, visual, and audio data at scale. But models have outgrown simple tasks, demanding richer context and deeper expertise. The era of sweatshop data is over.
Tweet media one
9
8
188
@MechanizeWork
Mechanize
5 days
Read the blog post here:
0
0
7
@MechanizeWork
Mechanize
5 days
We call this approach replication training: tasking AI models with exactly replicating existing software using clear specs and references. This trains models to read precisely, execute reliably, and demonstrate resilience in complex, long-horizon tasks.
Tweet media one
1
0
8
@MechanizeWork
Mechanize
5 days
Imagine if pretraining a language model meant manually creating the entire training corpus. Clearly, this would be impractical. Instead, we leverage vast content available online. We expect RL will follow a similar path, replicating abundant existing software as RL tasks.
Tweet media one
3
2
64
@MechanizeWork
Mechanize
7 days
Beyond this point, we expect models trained via RL to acquire powerful, task-agnostic, few-shot capabilities on tasks that currently require painstaking, task-specific training, just as GPT-3 unlocked few-shot capabilities for language.
2
0
16
@MechanizeWork
Mechanize
7 days
How much bigger must RL get to have a GPT-3 moment?. We expect this will soon require roughly 10,000 years of cumulative human-equivalent task time, comparable to GTA V or major operating systems.
Tweet media one
2
4
56
@MechanizeWork
Mechanize
22 days
Read the blog post here:.
0
2
23
@MechanizeWork
Mechanize
22 days
Replication tasks target critical skills current models lack: accurately interpreting instructions, reliably recovering from mistakes, and sustaining precise execution on tasks that humans take months to complete, moving us closer to reliable, capable AI agents.
Tweet media one
1
0
28
@MechanizeWork
Mechanize
22 days
Replication training naturally extends existing AI trends by building tasks directly from abundant human-generated data already available online.
Tweet media one
1
0
25
@MechanizeWork
Mechanize
22 days
We are proposing a new AI paradigm called replication training: tasking AIs to precisely replicate existing software. Abundant internet text unlocked powerful language models. Similarly, abundant software available today will enable massive-scale, efficient RL training.
Tweet media one
9
32
405
@MechanizeWork
Mechanize
23 days
Read the blog post here:
1
1
36
@MechanizeWork
Mechanize
23 days
Before GPT-3, achieving good performance required specialized fine-tuning for each task. Today's RL is similar: models need to be carefully trained to handle tasks like deep research, web search, or coding. But we think RL will soon have its GPT-3 moment.
Tweet media one
11
49
477
@MechanizeWork
Mechanize
1 month
RT @tamaybes: Most AI labs talk about merely “augmenting” humans at work. They say this because AI currently falls short, not out of some d….
0
17
0