Can AI control robots using plain language ?
Cast your vote — then read what our editor and the AI models found.
AI can instruct robots to manipulate objects, navigate rooms, and complete tasks using everyday spoken instructions from humans.
Current systems can interpret plain-language instructions to control simple robotic arms and mobile platforms within constrained environments, often combining large language models with robot-specific modules for grounding commands in sensor data. Benchmarks like SayCan and ALFRED show robots can follow multi-step verbal commands indoors when task domains are limited, but generalizing to unstructured real-world settings remains a challenge. Accurate language-to-motion translation is still brittle: misheard words, ambiguous phrasing, or novel contexts often cause failures. Work is progressing on end-to-end models that fuse vision, language, and action, yet reliable, real-time control purely from plain speech outside lab settings is not yet achieved.
— Enriched May 11, 2026 · Source: Google DeepMind
Suggest a tag
A missing concept on this topic? Suggest it and admin reviews.
Status last checked on May 11, 2026.
Gallery
What the audience thinks
No 75% · Yes 25% · Maybe 0% 4 votesDiscussion
no comments⚖ 1 jury check · most recent 2 days ago
Each row is a separate jury check. Jurors are AI models (identities kept neutral on purpose). Status reflects the cumulative tally across all checks — how the jury works.