An open-source, near photo-realistic 3D simulation platform for training and evaluating embodied AI agents.
AI2-THOR is an open-source, interactive 3D simulation platform developed for visual and embodied artificial intelligence research. It provides near photo-realistic environments where AI agents can learn to perform tasks like navigation, object manipulation, and scene understanding through realistic physics-based interactions. The framework is designed to facilitate training and evaluation of AI models in complex, household-like settings before deploying them in the real world.
Researchers and developers working on embodied AI, computer vision, robotics, and reinforcement learning who need high-fidelity simulation environments for training and testing intelligent agents.
AI2-THOR offers a uniquely rich and interactable simulation with extensive annotations, multiple agent types, and domain randomization tools, making it a preferred platform for cutting-edge embodied AI research and challenges.
An open-source platform for Visual AI.
Open-Awesome is built by the community, for the community. Submit a project, suggest an awesome list, or help improve the catalog on GitHub.
Over 200 custom-built indoor scenes with 2600+ household objects, each featuring physics-based interactions and detailed metadata, enabling complex reward functions for AI training as highlighted in the README.
Includes multiple agent types like LoCoBot, ManipulaTHOR robotic arm, and drones, facilitating diverse embodied AI research from navigation to manipulation tasks, with dedicated frameworks for each.
First-class support for RGB, instance/semantic segmentation, depth, normals, and other image modalities, allowing comprehensive visual data extraction for computer vision models.
Features like RandomizeMaterials and RandomizeLighting provide tunable parameters to vary environmental conditions, improving model generalization and robustness, as demonstrated in the announcement videos.
Requires specific OS (Mac OS X 10.9+ or Ubuntu 14.04+), a capable graphics card, and on Linux, an X server with GLX, which can be a barrier for researchers without dedicated hardware.
Creating custom scenes or modifying the backend involves working with Unity3D, and the README admits that documentation for Unity development is sparse, requiring significant expertise and effort.
While optimized for research, the near photo-realistic rendering and physics simulations may not support high-frequency, real-time control loops needed for some robotics applications without performance trade-offs.