-
The University of Adelaide
- Adelaide, Australia
- https://yanyuanqiao.github.io/
- in/yanyuan-qiao
- @YanyuanQiao
Stars
openvla / openvla
Forked from TRI-ML/prismatic-vlmsOpenVLA: An open-source vision-language-action model for robotic manipulation.
[TMLR 2024] repository for VLN with foundation models
[RSS 2024] NaVid: Video-based VLM Plans the Next Step for Vision-and-Language Navigation
Vision-Language Navigation Benchmark in Isaac Lab
This is the first released survey paper on hallucinations of large vision-language models (LVLMs). To keep track of this field and continuously update our survey, we maintain this repository of rel…
The official code repository for PRMBench.
The simplest, fastest repository for training/finetuning medium-sized GPTs.
A generative world for general-purpose robotics & embodied AI learning.
[ACL 24] The official implementation of MapGPT: Map-Guided Prompting with Adaptive Path Planning for Vision-and-Language Navigation.
A library for generative social simulation
[ECCV 2024] Official PyTorch implementation code for realizing the technical part of Mixture of All Intelligence (MoAI) to improve performance of numerous zero-shot vision language tasks.
Official GitHub Repository for Paper "Bridging Zero-shot Object Navigation and Foundation Models through Pixel-Guided Navigation Skill", ICRA 2024
Official code and checkpoint release for mobile robot foundation models: GNM, ViNT, and NoMaD.
Official implementation of Sim-to-Real Transfer via 3D Feature Fields for Vision-and-Language Navigation (CoRL'24).
RoboCasa: Large-Scale Simulation of Everyday Tasks for Generalist Robots
[RSS 2024] 3D Diffusion Policy: Generalizable Visuomotor Policy Learning via Simple 3D Representations
[ICCV 2023 Oral]: Scaling Data Generation in Vision-and-Language Navigation