-
HangzhouDianzi university
- CHINA
- http://kid177.github.io/
Stars
The official GitHub page for the survey paper "A Survey of Large Language Models".
Curated tutorials and resources for Large Language Models, AI Painting, and more.
A series of large language models trained from scratch by developers @01-ai
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
The ChatGPT Retrieval Plugin lets you easily find personal or work documents by asking questions in natural language.
Graphormer is a general-purpose deep learning backbone for molecular modeling.
Avatars for Zoom, Skype and other video-conferencing apps.
Transformer with Untied Positional Encoding (TUPE). Code of paper "Rethinking Positional Encoding in Language Pre-training". Improve existing models like BERT.
Apollo: An Adaptive Parameter-wise Diagonal Quasi-Newton Method for Nonconvex Stochastic Optimization
A library for efficient similarity search and clustering of dense vectors.
Code for: S.R. Qasim, H. Mahmood, and F. Shafait, Rethinking Table Recognition using Graph Neural Networks (2019)
Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
[ACL 2020] DeFormer: Decomposing Pre-trained Transformers for Faster Question Answering
XLNet: Generalized Autoregressive Pretraining for Language Understanding
Code for the paper "Language Models are Unsupervised Multitask Learners"
TensorFlow implementation of Pointer Networks
BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)
A general-purpose encoder-decoder framework for Tensorflow
Collection of scripts to aggregate image data for the purposes of training an NSFW Image Classifier
Code and model for the paper "Improving Language Understanding by Generative Pre-Training"
A TensorFlow Implementation of the Transformer: Attention Is All You Need