-
National Taiwan University
- Taipei, Taiwan
-
12:53
(UTC +08:00) - xjchen.tech
- @xjchen_ntu
- in/jun-ntu
Lists (1)
Sort Name ascending (A-Z)
Starred repositories
Efficient Triton Kernels for LLM Training
Making data higher-quality, juicier, and more digestible for foundation models! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷为大模型提供更高质量、更丰富、更易”消化“的数据!
Get up and running with Llama 3.3, Mistral, Gemma 2, and other large language models.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
本项目是作者们根据个人面试和经验总结出的自然语言处理(NLP)面试准备的学习笔记与资料,该资料目前包含 自然语言处理各领域的 面试题积累。
A statistical toolkit for scientific discovery using machine learning
Codes and packages for the paper titled Evaluating Retrieval Quality in Retrieval-Augmented Generation.
Your AI second brain. Self-hostable. Get answers from the web or your docs. Build custom agents, schedule automations, do deep research. Turn any online or local LLM into your personal, autonomous …
Research and development (R&D) is crucial for the enhancement of industrial productivity, especially in the AI era, where the core aspects of R&D are mainly focused on data and models. We are commi…
This is the Plato Research Dialogue System, a flexible platform for developing conversational AI agents.
LLM based autonomous agent that conducts local and web research on any topic and generates a comprehensive report with citations.
[KDD24-ADS] R-Eval: A Unified Toolkit for Evaluating Domain Knowledge of Retrieval Augmented Large Language Models
Source code and dataset for "Difficulty-controllable Multi-hop Question Generation From Knowledge Graphs"
Code and implementations for the paper "AgentGym: Evolving Large Language Model-based Agents across Diverse Environments" by Zhiheng Xi et al.
The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.
A framework for serving and evaluating LLM routers - save LLM costs without compromising quality!
TensorZero creates a feedback loop for optimizing LLM applications — turning production data into smarter, faster, and cheaper models.
ColBERT: state-of-the-art neural search (SIGIR'20, TACL'21, NeurIPS'21, NAACL'22, CIKM'22, ACL'23, EMNLP'23)
A compilation of the best multi-agent papers
[EMNLP 2024: Demo Oral] RAGLAB: A Modular and Research-Oriented Unified Framework for Retrieval-Augmented Generation
Temporal Action Detection & Weakly Supervised Temporal Action Detection & Temporal Action Proposal Generation
Code for LAMOL: LAnguage MOdeling for Lifelong Language Learning
Active Learning Awesome Paper
This is the code repo for our paper "Autonomously Knowledge Assimilation and Accommodation through Retrieval-Augmented Agents".
PyTorch implementation of VALL-E(Zero-Shot Text-To-Speech), Reproduced Demo https://lifeiteng.github.io/valle/index.html
Code for the EMNLP 2024 paper "Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps"