-
Tencent WeChat AI
- Beijing
Stars
DRT-o1: Optimized Deep Reasoning Translation via Long Chain-of-Thought
Code for paper "Patch-Level Training for Large Language Models"
[NeurIPS 2022] "A Win-win Deal: Towards Sparse and Robust Pre-trained Language Models", Yuanxin Liu, Fandong Meng, Zheng Lin, Jiangnan Li, Peng Fu, Yanan Cao, Weiping Wang, Jie Zhou
Making large AI models cheaper, faster and more accessible
Code and Data for the ACL22 main conference paper "MSCTD: A Multimodal Sentiment Chat Translation Dataset"
EMNLP 2022: ClidSum: A Benchmark Dataset for Cross-Lingual Dialogue Summarization
A benchmark for the task of translation suggestion
PatrickStar enables Larger, Faster, Greener Pretrained Models for NLP and democratizes AI for everyone.
《机器翻译:基础与模型》肖桐 朱靖波 著 - Machine Translation: Foundations and Models
a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.
Code for the paper: GCDT: A Global Context Enhanced Deep Transition Architecture for Sequence Labeling
fandongmeng / RSI-NAT
Forked from ictnlp/RSI-NATSource code for "Retrieving Sequential Information for Non-Autoregressive Neural Machine Translation"