Starred repositories
A biblatex implementation of the GB/T7714-2015 bibliography style || GB/T 7714-2015 参考文献著录和标注的biblatex样式包
CVPR 2024 Papers Autonomous Driving
[CVPR 2024 Highlight] GenAD: Generalized Predictive Model for Autonomous Driving & Foundation Models in Autonomous System
[CVPR 2023 Highlight] LaserMix for Semi-Supervised LiDAR Semantic Segmentation
[CVPR 2024 Highlight] OpenESS: Event-Based Semantic Scene Understanding with Open Vocabularies
This project aims to collect the latest "call for reviewers" links from various top CS/ML/AI conferences/journals
Evaluating LLMs' multi-round chatting capability via assessing conversations generated by two LLM instances.
Examples using the Deep Search functionalities
FINDER - FInding key players in complex Networks through DEep Reinforcement learning (Nature Machine Intelligence)
This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.
SuperCLUE: 中文通用大模型综合性基准 | A Benchmark for Foundation Models in Chinese
西安交通大学学位论文模板(LaTeX)(适用硕士、博士学位)An official LaTeX template for Xi'an Jiaotong University degree thesis (Chinese and English)
All-in-One Development Tool based on PaddlePaddle(飞桨低代码开发工具)
a state-of-the-art-level open visual language model | 多模态预训练模型
This repo contains the code for paper "Learning naturalistic driving environment with statistical realism"
Omniverse Kit Extension Template
Reinforcement Learning Environments for Omniverse Isaac Gym
Modular reinforcement learning library (on PyTorch and JAX) with support for NVIDIA Isaac Gym, Omniverse Isaac Gym and Isaac Lab
InvAgent: A LLM-based Multi-Agent System for Inventory Management in Supply Chains
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use th…
🔥🔥🔥Latest Papers, Codes and Datasets on Vid-LLMs.
[CVPR 2024 Highlight🔥] Chat-UniVi: Unified Visual Representation Empowers Large Language Models with Image and Video Understanding
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
Official Repository of paper VideoGPT+: Integrating Image and Video Encoders for Enhanced Video Understanding