Stars
A free, fast, and reliable Open Source CDN for npm, GitHub, Javascript, and ESM
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
finetune qwen vl, both local and distributed model.
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, DeepSeek, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discr…
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
[Findings of EMNLP 2024] AdaMoE: Token-Adaptive Routing with Null Experts for Mixture-of-Experts Language Models
In-context KV-Cache Eviction for LLMs via Attention-Gate
Qwen2.5-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model
✨✨Latest Advances on Multimodal Large Language Models
Codes for paper "Three Steps to Multimodal Trajectory Prediction: Modality Clustering, Classification and Synthesis", "Human Trajectory Prediction with Momentary Observation" and "Stimulus Verifica…
[RSS 2024] Code for "Multimodal Diffusion Transformer: Learning Versatile Behavior from Multimodal Goals" for CALVIN experiments with pre-trained weights
Janus-Series: Unified Multimodal Understanding and Generation Models
[IEEE T-PAMI 2024] All you need for End-to-end Autonomous Driving
State-of-the-art bilingual open-sourced Math reasoning LLMs.
Reimplementation of GR-1, a generalized policy for robotics manipulation.
DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models
Official PyTorch Implementation for the "Recovering the Pre-Fine-Tuning Weights of Generative Models" paper (ICML 2024).
SIFT: Grounding LLM Reasoning in Contexts via Stickers
Code for "AnyGPT: Unified Multimodal LLM with Discrete Sequence Modeling"
Official repo of EmbodiedBench, a comprehensive benchmark designed to evaluate MLLMs as embodied agents.