Stars
[CVPR 2025] The offical Implementation of "Universal Actions for Enhanced Embodied Foundation Models"
"MimicPlay: Long-Horizon Imitation Learning by Watching Human Play" code repository
[Lumina Embodied AI Community] 具身智能技术指南 Embodied-AI-Guide
Code for Point Policy: Unifying Observations and Actions with Key Points for Robot Manipulation
Production-tested AI infrastructure tools for efficient AGI development and community-driven innovation
Official github repo for SafeDialBench, a comprehensive multi-turn dialogue benchmark to evaluate LLMs' safety.
✨✨Long-VITA: Scaling Large Multi-modal Models to 1 Million Tokens with Leading Short-Context Accuracy
The Next Step Forward in Multimodal LLM Alignment
RDT-1B: a Diffusion Foundation Model for Bimanual Manipulation
Code for "Policy Decorator: Model-Agnostic Online Refinement for Large Policy Model"
[RA-L 2023] EasyHeC: Accurate and Automatic Hand-eye Calibration via Differentiable Rendering and Space Exploration
GRAPE: Guided-Reinforced Vision-Language-Action Preference Optimization
Stanford-ILIAD / openvla-mini
Forked from openvla/openvlaOpenVLA: An open-source vision-language-action model for robotic manipulation.
Official repo of VLABench, a large scale benchmark designed for fairly evaluating VLA, Embodied Agent, and VLMs.
Implementation of soft prompt tuning from scratch
Latest Advances on Vison-Language-Action Models.
Google DeepMind's software stack for physics-based simulation and Reinforcement Learning environments, using MuJoCo.
Code for the paper "3D Diffuser Actor: Policy Diffusion with 3D Scene Representations"
Automated, hardware-independent Hand-Eye Calibration
A curated list of 3D Vision papers relating to Robotics domain in the era of large models i.e. LLMs/VLMs, inspired by awesome-computer-vision, including papers, codes, and related websites
🆓免费的 ChatGPT 镜像网站列表,持续更新。List of free ChatGPT mirror sites, continuously updated.