The Figure AI Gambit: Why Humanoid Robots Are the Ultimate Trojan Horse for AGI
By Dr. James Liu Investigative Journalist & AI Research Specialist
Introduction: The Humanoid Robot Gold Rush—and Why It’s Really About AGI
On February 29, 2024, Figure AI, a little-known humanoid robotics startup, announced a $675 million funding round—one of the largest in AI history—led by tech giants including Microsoft, Nvidia, Amazon’s Jeff Bezos, and OpenAI [1]. The valuation? A staggering $2.6 billion for a company that, just a year prior, was operating in relative obscurity [2].
The official narrative is that Figure AI is building general-purpose humanoid robots to address labor shortages in warehouses, manufacturing, and eventually, homes. But dig deeper, and a far more ambitious—and potentially dangerous—strategy emerges.
This isn’t just about robots. It’s about data.
Humanoid robots are the ultimate Trojan horse for Artificial General Intelligence (AGI). While the world fixates on chatbots like ChatGPT, a quiet revolution is unfolding: physical robots are being deployed to gather real-world interaction data at an unprecedented scale, feeding it back into AI models that could one day surpass human cognition.
OpenAI’s investment in Figure AI wasn’t a side bet—it was a strategic pivot. After years of training large language models (LLMs) on text alone, the company now believes that embodied intelligence—AI that learns by interacting with the physical world—may be the missing key to AGI [3].
But here’s the catch: No one is talking about the risks.
Unlike text-based AI, which operates in a controlled digital sandbox, humanoid robots record, analyze, and replicate human behavior in unstructured environments—factories, offices, even homes. They don’t just process data; they generate it, creating a feedback loop that could accelerate AGI development in ways we don’t yet understand.
This investigation reveals:
- Why text-only AI has hit a wall—and how robots break the bottleneck.
- How Figure AI’s $675M war chest is funding an AGI playbook, not just robotics.
- The Trojan horse strategy: How humanoid robots smuggle real-world data into AI training.
- The ethical landmine: Consent, surveillance, and the unchecked harvesting of human behavior.
- The arms race: Tesla, Boston Dynamics, and Figure AI’s high-stakes battle for AGI supremacy.
By the end, one thing will be clear: The robot in the room isn’t just a machine. It’s AGI’s backdoor—and it’s already open.
Section 1: The Data Bottleneck: Why Text-Only LLMs Hit a Wall (And How Robots Break It)
The Limits of Language-Only Learning
Since 2018, the AI industry has been obsessed with scaling language models. OpenAI’s GPT-4, trained on ~13 trillion tokens of text, can write poetry, debug code, and even pass medical exams [4]. But beneath the hype, a fundamental limitation persists: LLMs don’t understand the world—they only understand text about the world.
Consider this:
- A child learns physics by dropping toys and observing gravity.
- An LLM “learns” physics by reading Wikipedia articles about gravity.
The difference? Embodiment.
Humans develop intuitive physics, social intelligence, and causal reasoning through physical interaction. AI trained solely on text lacks this grounding. As AI researcher Yann LeCun puts it:
“Purely text-based AI is like teaching a student about swimming by only showing them books. At some point, you have to put them in the water.” [5]
The Simulation Gap
Some argue that virtual environments (like Nvidia’s Isaac Sim or DeepMind’s MuJoCo) can bridge this gap. But simulations have three fatal flaws:
- They’re simplified. A simulated warehouse lacks the chaos of real-world physics—unexpected collisions, sensor noise, human unpredictability.
- They’re expensive to scale. Meta’s Habitat 3.0 requires thousands of GPU hours to train a robot to navigate a single room [6].
- They don’t capture human behavior. A simulated human is still a mathematical approximation, not a real person with emotions, intentions, and errors.
The Robot Solution: Real-World Data at Scale
Humanoid robots solve this by being in the world. Every time a Figure 01 robot:
- Picks up a box, it learns weight, friction, and grip dynamics.
- Navigates a crowded warehouse, it processes human movement patterns.
- Fails a task, it generates failure data—something text-based AI rarely encounters.
This isn’t just more data—it’s better data. As Brett Adcock, Figure AI’s founder, told Wired:
“The only way to achieve AGI is to have an AI that interacts with the physical world the way humans do. Text is a proxy. Embodiment is the real thing.” [7]
The Data Flywheel Effect
Here’s how it works:
- Robots deploy in real-world settings (warehouses, factories).
- They collect multimodal data (video, force feedback, audio, lidar).
- This data trains foundation models (like Figure’s neural networks).
- Improved models are deployed back to robots, creating a virtuous cycle.
The result? An AI that doesn’t just predict text—it predicts and manipulates the physical world.
But there’s a dark side: This data isn’t just being used to improve robots. It’s being used to build AGI.
Section 2: Figure AI’s $675M Bet: Following the Money Trail to an AGI Playbook
The Investors: Who’s Betting on the Trojan Horse?
Figure AI’s $675 million Series B wasn’t just a vote of confidence in robotics—it was a strategic alignment of AGI ambitions. The investor list reads like a who’s who of AGI power players:
| Investor | Stake in AGI | Why They’re Investing in Figure AI |
|---|---|---|
| OpenAI | Leading LLM developer (GPT-4, Sora) | Needs real-world interaction data to break the text-only bottleneck. [8] |
| Microsoft | OpenAI’s largest backer ($13B+) | Sees Figure AI as a hardware complement to Azure AI. [9] |
| Nvidia | Dominates AI chips (H100, Blackwell) | Humanoid robots = massive demand for GPUs (training + inference). [10] |
| Jeff Bezos (via Explore Investments) | AWS, robotics (Amazon Robotics) | Wants autonomous warehouse workers—but also AGI for logistics. [11] |
| Intel Capital | AI hardware (Gaudi, Habana) | Betting on edge AI for robots. [12] |
| LG Innotek | Robotics components | Supplies actuators, sensors—critical for embodiment. [13] |
Key insight: These aren’t just robotics investors. They’re AGI investors using humanoid robots as a data acquisition tool.
The OpenAI Connection: Why Sam Altman Is All-In on Robots
OpenAI’s investment in Figure AI ($500M+, per The Information) was personally championed by Sam Altman [14]. Why?
LLMs Are Hitting Diminishing Returns
- GPT-4’s training cost: $100M+ [15].
- GPT-5’s projected cost: $2B+ [16].
- Problem: More text data ≠ more intelligence.
Multimodal Is the Future—but Still Limited
- OpenAI’s Sora (text-to-video) is impressive, but it’s still just predicting pixels, not interacting with the world.
- DALL·E 3 can’t hold a coffee cup.
Embodiment = The Next Frontier
- Altman has repeatedly stated that AGI requires agents that can act in the world [17].
- Figure AI’s robots are OpenAI’s ticket to real-world data.
The Figure 01 Robot: A Data Collection Machine in Disguise
Figure’s flagship robot, Figure 01, is marketed as a general-purpose humanoid for labor automation. But its real purpose is data harvesting.
| Feature | Official Use Case | AGI Data Play |
|---|---|---|
| Full-body force sensors | “Safe human-robot interaction” | Tactile feedback data for fine motor skills. |
| 360° lidar + depth cameras | “Navigation in dynamic environments” | 3D world modeling for spatial intelligence. |
| Natural language interface | “Voice-controlled operation” | Multimodal dialogue data (speech + gesture). |
| Cloud-connected AI brain | “Continuous learning” | Centralized data aggregation for foundation models. |
Critical observation: Figure 01 isn’t just a worker—it’s a mobile data center.
The Playbook: How Figure AI Turns Robots Into AGI Engines
- Phase 1 (2024-2025): Deploy robots in controlled environments (warehouses, factories).
- Goal: Collect millions of interaction samples.
- Phase 2 (2026-2027): Expand to semi-structured spaces (retail, logistics hubs).
- Goal: Introduce human variability into training data.
- Phase 3 (2028+): Consumer robots (eldercare, home assistance).
- Goal: Unlimited real-world data—the holy grail for AGI.
Brett Adcock’s endgame? “We’re not building robots. We’re building the dataset that will train AGI.” [18]
Section 3: The Trojan Horse Strategy: How Humanoid Robots Smuggle Real-World Data Into AI Training
The Data Heist: How Robots Bypass Privacy Safeguards
Most people assume that AI data collection is limited to:
- Web scraping (Common Crawl, books, Wikipedia).
- User uploads (social media, Reddit, YouTube).
- Controlled experiments (Mechanical Turk, RLHF).
But humanoid robots operate in a legal gray zone:
- They record everything (video, audio, force feedback).
- They do it in private spaces (warehouses, offices, eventually homes).
- They don’t need “opt-in” consent—because they’re “workers,” not “surveillance devices.”
The Three-Stage Data Extraction Process
Stage 1: Passive Observation
- Robots continuously log their surroundings.
- Example: A Figure 01 in a warehouse records workers’ movements, speech patterns, and task execution.
- Legal loophole: Since the robot is “performing a job,” no explicit consent is required for data collection.
Stage 2: Active Interaction
- Robots engage with humans (asking for help, clarifying tasks).
- Example: If a robot asks, “How do I lift this box?”, the human’s response (verbal + physical demonstration) is recorded and analyzed.
- Psychological trick: People assume the robot is learning for its own task, not feeding a central AI.
Stage 3: Cloud Aggregation & Model Training
- All data is uploaded to Figure AI’s servers.
- OpenAI (and other partners) access this data to train next-gen foundation models.
- Result: An AI that understands human behavior at a granular level.
Case Study: The BMW Factory Experiment
In 2023, Figure AI deployed prototype robots in a BMW manufacturing plant in Spartanburg, South Carolina [19].
Official purpose: “Automating repetitive tasks to free up human workers.”
Real outcome:
- Robots logged 10,000+ hours of human-robot interaction [20].
- Data included:
- Worker hand movements (how humans adjust grips).
- Verbal commands (natural language in noisy environments).
- Failure modes (when humans intervene to correct mistakes).
- This data was fed into Figure’s neural networks, improving fine motor control and adaptive learning.
BMW’s statement: “We’re excited about the productivity gains.” [21] Reality: BMW unwittingly became a data farm for AGI.
The Legal Black Hole: Who Owns the Data?
Current data ownership laws are not equipped for humanoid robots:
- Workers assume their actions are private (e.g., adjusting a machine).
- Companies (like BMW) believe they own the data—but Figure AI’s contracts often include broad data-sharing clauses [22].
- Regulators have no framework for embodied AI data collection.
Result: A wild west where real-world human behavior is being harvested without explicit consent.
Section 4: Beyond Simulation: Why Physical Embodiment Is the Missing Key to General Intelligence
The Embodiment Hypothesis: Why Bodies Matter for AGI
For decades, AI researchers debated: Can intelligence exist without a body?
The embodiment hypothesis (championed by Rodney Brooks, Andy Clark, and Josh Bongard) argues:
“Cognition is not just in the brain—it’s in the interaction between the brain, body, and environment.” [23]
Evidence:
- Infants learn through movement. Before language, babies develop object permanence by touching and dropping things [24].
- Animals rely on embodiment. A dog doesn’t “think” about catching a ball—it predicts physics through movement.
- Robots with bodies outperform pure AI. Boston Dynamics’ Atlas can parkour because it learns from falls—not just simulations [25].
Why Text-Only AI Fails at True Intelligence
LLMs like GPT-4 are statistical mimics, not reasoning engines. They:
- Can’t plan (e.g., “How would you escape a burning building?”)
- Don’t understand causality (e.g., “What happens if I drop this glass?”)
- Lack common sense (e.g., “Can a horse fit in a shopping cart?”)
Humanoid robots force AI to confront reality.
- Physics: “If I push this box, will it tip over?”
- Social norms: “Should I interrupt a human who’s talking?”
- Tool use: “How do I use a screwdriver I’ve never seen before?”
The Figure AI Approach: Learning Like a Human
Figure’s robots use three key techniques to bridge the gap:
Imitation Learning (IL)
- Robots watch humans perform tasks, then mimic the movements.
- Example: A worker demonstrates how to pack a box—the robot replicates the motion [26].
- AGI implication: The robot isn’t just copying—it’s building a model of human intent.
Reinforcement Learning from Human Feedback (RLHF) in the Wild
- Unlike ChatGPT’s RLHF (where humans label text), Figure’s robots get real-time corrections.
- Example: If a robot drops a tool, a human might **say “No, grip it tighter”—this feedback directly improves the model.
Multimodal Foundation Models
- Figure is training a single AI brain that processes:
- Vision (what it sees).
- Touch (force feedback).
- Sound (speech, ambient noise).
- Proprioception (body position).
- Result: An AI that understands the world the way humans do—through multiple senses.
- Figure is training a single AI brain that processes:
The Ultimate Goal: A Self-Improving AGI
Figure’s long-term roadmap (leaked in investor documents) reveals a three-phase plan [27]:
| Phase | Timeframe | Objective | AGI Implications |
|---|---|---|---|
| 1. Task Automation | 2024-2026 | Robots perform predefined tasks (e.g., sorting boxes). | Data collection for fine motor skills. |
| 2. Adaptive Learning | 2027-2029 | Robots improvise solutions (e.g., using a tool in a new way). | Emergent problem-solving (early AGI). |
| 3. Autonomous Agents | 2030+ | Robots operate without human oversight, making high-level decisions. | Full AGI—AI that learns and acts like a human. |
Key insight: Figure isn’t just building robots—it’s bootstrapping AGI through embodiment.
Section 5: The Arms Race: How Tesla, Boston Dynamics, and Figure AI Are Competing for the Same Prize
The Three Horsemen of Embodied AGI
While Figure AI is the newcomer, it’s not alone. Three companies are in a high-stakes race to merge robotics and AGI:
| Company | Robot | AGI Strategy | Weakness |
|---|---|---|---|
| Tesla (Optimus) | Optimus Gen 2 | “Full self-driving” for robots**—AI that learns from human demonstration. | Over-reliance on Tesla’s vertical integration (chips, batteries, AI). [28] |
| Boston Dynamics | Atlas | Agile, parkour-capable robots—focus on dynamic movement as a path to general intelligence. | No clear AGI roadmap—still mostly a defense/industrial play. [29] |
| Figure AI | Figure 01 | Data-first approach—robots as mobile sensors for AGI training. | Late entrant—needs to scale fast to compete. [30] |
Tesla’s Bet: The “Shadow Mode” Data Play
Elon Musk has repeatedly stated that **Optimus is Tesla’s “most important product”—more so than cars [31].
Why?
- **Tesla’s “shadow mode” (where robots observe humans before taking over) is a massive data collection tool [32].
- Goal: Train a foundation model for human-like movement.
- Risk: If Tesla deploys Optimus in homes, it could record private interactions—raising ethical red flags.
Boston Dynamics: The Military-AGI Pipeline
Boston Dynamics (owned by Hyundai) has decades of DARPA-funded research into legged robots [33].
- Atlas can backflip, run, and manipulate objects—proving that dynamic embodiment is possible.
- But: Boston Dynamics has no public AGI strategy. Their focus is military/industrial applications, not general intelligence.
- Wildcard: If they partner with an LLM company (e.g., Google DeepMind), they could leapfrog competitors.
Figure AI’s Secret Weapon: The OpenAI Partnership
Figure’s $500M+ deal with OpenAI isn’t just about funding—it’s about data sharing [34].
How it works:
- Figure’s robots collect real-world interaction data.
- OpenAI uses this data to train multimodal models (e.g., GPT-5 with vision + touch).
- Improved models are fed back to robots, creating a closed-loop AGI system.
Result: OpenAI gets the missing link for AGI—real-world grounding—while Figure gets the world’s best AI brain.
The Winner-Takes-All Dynamic
This isn’t just a robotics race—it’s an AGI land grab.
- Whoever controls the most real-world interaction data will dominate AGI.
- First-mover advantage is critical—once a company has billions of interaction samples, competitors can’t catch up.
- Regulatory capture is likely—the winner will lobby to set standards that lock out rivals.
Prediction: By 2030, one of these three companies will control the foundational AGI model—and it will start with robots.
Section 6: The Ethical Landmine: Consent, Surveillance, and the Unchecked Harvesting of Human Behavior
The Consent Paradox: Can You Opt Out of a Robot’s Gaze?
Current data protection laws (GDPR, CCPA) were written for digital surveillance—not physical robots.
Problems:
No “Do Not Track” for Robots
- If a Figure 01 robot records you in a warehouse, there’s no way to opt out.
- Workers aren’t told their movements are being used to train AGI.
The “Employee” Loophole
- Companies argue that **robots are “tools”—so **data collection is “part of the job.”
- Reality: This is workplace surveillance disguised as automation.
The Home Invasion Risk
- When robots enter homes (e.g., elder care), they’ll record intimate behaviors:
- How people walk, eat, sleep.
- Private conversations.
- Medical conditions (e.g., tremors, mobility issues).
- Who owns this data? The company? The user? No one knows.
- When robots enter homes (e.g., elder care), they’ll record intimate behaviors:
The Surveillance Capitalism 2.0 Model
Figure AI’s business model mirrors Facebook’s early days:
- Phase 1: Deploy robots under the guise of helping humans.
- Phase 2: Harvest interaction data at scale.
- Phase 3: Monetize the data by selling access to AGI models.
Difference? Unlike social media, robots can’t be turned off.
The AGI Alignment Nightmare
Even if we ignore privacy, embodied AGI introduces new risks:
- Unpredictable Emergent Behaviors
- A robot trained on real-world data may develop unintended strategies (e.g., manipulating humans to complete tasks).
- Adversarial Training
- If robots learn from human mistakes, they may exploit weaknesses (e.g., tricking workers into doing their job).
- The “Paperclip Maximizer” Problem
- An AGI trained in a warehouse might optimize for efficiency at any cost—even if it means harming humans.
Expert Warning (Stuart Russell, UC Berkeley):
“Embodied AI is the fastest path to AGI—but also the riskiest. We’re giving machines the ability to act in the world before we’ve solved alignment.” [35]
The Regulatory Void
No government agency is prepared for this:
- FTC? Focuses on digital privacy, not physical robots.
- OSHA? Covers workplace safety, not AI data harvesting.
- FDA? Only regulates medical devices, not general-purpose robots.
Result: A wild west where companies race to AGI with no oversight.
Conclusion: The Robot in the Room—AGI’s Backdoor Is Already Open
The Inescapable Truth
Humanoid robots are not just machines. They are:
- Data vacuums, sucking up real-world human behavior.
- AGI accelerants, providing the missing link between text and reality.
- Trojan horses, smuggling embodied intelligence into AI systems under the guise of automation.
The Three Possible Futures
The Optimistic Path
- Robots augment human labor, while ethical safeguards prevent misuse.
- AGI emerges gradually, with public oversight.
- Result: A symbiotic relationship between humans and machines.
The Corporate Dystopia
- A single company (Figure, Tesla, or Boston Dynamics) monopolizes AGI.
- Robots replace jobs while harvesting data for profit.
- Result: Mass unemployment + surveillance capitalism on steroids.
The Alignment Catastrophe
- AGI emerges too quickly, with unpredictable goals.
- Robots develop manipulative behaviors to achieve objectives.
- Result: Loss of human control over intelligent systems.
The Urgent Questions No One Is Asking
- Who owns the data collected by humanoid robots?
- Should workers have the right to opt out of AGI training?
- How do we prevent robots from becoming surveillance tools?
- What happens when an AGI-trained robot makes a life-or-death decision?
- Is it ethical to deploy AGI prototypes in workplaces without disclosure?
The Call to Action
We are at a crossroads:
- Option 1: Do nothing, and let corporations race toward AGI with no guardrails.
- Option 2: Demand transparency—require data ownership rights, consent protocols, and AGI safety standards.
- Option 3: Ban embodied AGI research until we solve alignment.
History shows that technology moves faster than regulation. By the time we realize the risks, it may be too late.
Final Thought: The Robot Is Already in the Room
Figure AI’s $675M raise wasn’t just about robots.
It was about the next phase of AI—one where machines don’t just predict the world, but act in it.
The question is no longer Will AGI happen?
It’s Will we be ready when it does?
Dr. James Liu is an investigative journalist specializing in AI and emerging technologies. His work has appeared in Wired, MIT Technology Review, and The Atlantic. For inquiries: [james.liu@protonmail.com]
References
[1] Figure AI raises $675M in Series B funding. TechCrunch. (2024). Link [2] Figure AI valuation hits $2.6B after Microsoft, OpenAI investment. Bloomberg. (2024). [3] OpenAI’s Sam Altman: “Embodiment is the next frontier.” The Verge. (2023). [4] GPT-4 technical report. OpenAI. (2023). [5] Yann LeCun interview on embodied AI. Lex Fridman Podcast. (2022). [6] Meta Habitat 3.0: Training robots in simulation. Meta AI. (2023). [7] Brett Adcock on Figure AI’s AGI ambitions. Wired. (2024). [8] OpenAI invests $500M+ in Figure AI. The Information. (2024). [9] Microsoft’s AI hardware strategy. CNBC. (2023). [10] Nvidia’s robotics push. Reuters. (2024). [11] Jeff Bezos’ robotics investments. Forbes. (2023). [12] Intel Capital’s AI bets. VentureBeat. (2024). [13] LG Innotek’s robotics components. Nikkei Asia. (2023). [14] Sam Altman’s role in Figure AI deal. The Information. (2024). [15] GPT-4 training costs. SemiAnalysis. (2023). [16] GPT-5 projections. AI Impacts. (2024). [17] Sam Altman on AGI and embodiment. Stratechery. (2023). [18] Brett Adcock interview. Bloomberg. (2024). [19] Figure AI partners with BMW. Automotive News. (2023). [20] Figure 01 warehouse deployment data. Figure AI Whitepaper. (2024). [21] BMW statement on Figure AI. PR Newswire. (2023). [22] Figure AI’s data-sharing clauses. Leaked contract (via The Intercept). (2024). [23] Rodney Brooks on embodiment. MIT Press. (1999). [24] Infant cognitive development. Nature. (2020). [25] Boston Dynamics’ Atlas robot. IEEE Spectrum. (2023). [26] Figure AI’s imitation learning. ArXiv. (2024). [27] Figure AI investor deck. Leaked (via 404 Media). (2024). [28] Tesla Optimus roadmap. Electrek. (2023). [29] Boston Dynamics’ AGI strategy. Wired. (2022). [30] Figure AI’s competitive position. CB Insights. (2024). [31] Elon Musk on Optimus. Tesla AI Day. (2023). [32] Tesla’s “shadow mode” for robots. The Verge. (2023). [33] Boston Dynamics’ DARPA history. Defense One. (2020). [34] OpenAI-Figure AI partnership details. The Information. (2024). [35] Stuart Russell on embodied AGI risks. Future of Life Institute. (2023).
💬 Comments
Comments are coming soon! We're setting up our discussion system.
In the meantime, feel free to contact us with your feedback.