Stars
- All languages
- Assembly
- Astro
- C
- C#
- C++
- CSS
- Crystal
- Cuda
- Cython
- Dart
- Dockerfile
- Elixir
- Elm
- Go
- HTML
- Haskell
- JSON
- Java
- JavaScript
- Julia
- Jupyter Notebook
- Kotlin
- Lua
- MDX
- MLIR
- Markdown
- Nix
- OCaml
- Objective-C
- OpenEdge ABL
- PHP
- PLSQL
- Perl
- Processing
- Python
- ReScript
- Red
- Roff
- Ruby
- Rust
- SCSS
- Scala
- Shell
- Svelte
- Swift
- TeX
- TypeScript
- V
- Verilog
- Vim Snippet
- Vue
- Yacc
A Datacenter Scale Distributed Inference Serving Framework
GISTEmbed: Guided In-sample Selection of Training Negatives for Text Embeddings
A Docker-powered service for PDF document layout analysis. This service provides a powerful and flexible PDF analysis service. The service allows for the segmentation and classification of differen…
Negima is a Python package to extract phrases in Japanese text by using the part-of-speeches based rules you defined.
OCR, layout analysis, reading order, table recognition in 90+ languages
A daily digest web app that scrapes and summarizes blogs, Reddit threads, GitHub trending, and Hacker-News-trending articles all in one place.
A blazing fast inference solution for text embeddings models
Official implementation of "TAID: Temporally Adaptive Interpolated Distillation for Efficient Knowledge Transfer in Language Models"
A simple yet powerful tool to turn traditional container/OS images into unprivileged sandboxes.
AJIMEE-Bench (Advanced Japanese IME Evaluation Benchmark)
Hackable and optimized Transformers building blocks, supporting a composable construction.
Bringing BERT into modernity via both architecture changes and scaling
SONAR, a new multilingual and multimodal fixed-size sentence embedding space, with a full suite of speech and text encoders and decoders.
Scripts for creating a Japanese-English parallel corpus and training NMT models
Yomitoku is an AI-powered document image analysis package designed specifically for the Japanese language.
A Python package for intrinsic dimension estimation
Training and evaluation code for the paper "Headless Language Models: Learning without Predicting with Contrastive Weight Tying" (https://arxiv.org/abs/2309.08351)