-
Technion
- mttk.github.io
- @mtutek
- @mtutek.bsky.social
Stars
A benchmark with locally sourced multilingual questions for 31 languages.
Toolkit for linearizing PDFs for LLM datasets/training
An AI-powered research assistant that performs iterative, deep research on any topic by combining search engines, web scraping, and large language models. The goal of this repo is to provide the si…
A bibliography and survey of the papers surrounding o1
A unified interface for computing surprisal (log probabilities) from language models! Supports neural, symbolic, and black-box API models.
A professionally curated list of awesome Conformal Prediction videos, tutorials, books, papers, PhD and MSc theses, articles and open-source libraries.
ColBERT: state-of-the-art neural search (SIGIR'20, TACL'21, NeurIPS'21, NAACL'22, CIKM'22, ACL'23, EMNLP'23)
Dermatology ddx dataset, Jax implementations of Monte Carlo conformal prediction, plausibility regions and statistical annotation aggregation from our recent work on uncertain ground truth (TMLR'23…
The official PyTorch implementation of Google's Gemma models
Run code inference-only benchmarks quickly using vLLM
Code Prompting Elicits Conditional Reasoning Abilities in Text+Code LLMs. EMNLP 2024
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
Large-scale multi-document summarization dataset and code
Enhancing small language models with LLM generated counterfactuals.
Codes and files for the paper Are Emergent Abilities in Large Language Models just In-Context Learning
Code associated with NLPeer: A unified resource for the study of peer review
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
INCEpTION provides a semantic annotation platform offering intelligent annotation assistance and knowledge management.
Hackable and optimized Transformers building blocks, supporting a composable construction.
Generate automated tests for your Node.js app via LLMs without developers having to write a single line of code.