Stars
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
中文LLaMA-2 & Alpaca-2大模型二期项目 + 64K超长上下文模型 (Chinese LLaMA-2 & Alpaca-2 LLMs with 64K long context models)
Sparsity-aware deep learning inference runtime for CPUs
A Python package for extending the official PyTorch that can easily obtain performance on Intel platform
AWS Deep Learning Containers are pre-built Docker images that make it easier to run popular deep learning frameworks and tools on AWS.
TorchBench is a collection of open source benchmarks used to evaluate PyTorch performance.
AWS ParallelCluster is an AWS supported Open Source cluster management tool to deploy and manage HPC clusters in the AWS cloud.
Intel® AI Reference Models: contains Intel optimizations for running deep learning workloads on Intel® Xeon® Scalable processors and Intel® Data Center GPUs
AI Accelerator Benchmark focuses on evaluating AI Accelerators from a practical production perspective, including the ease of use and versatility of software and hardware.
This is an all-in-one client for EC2 Instance Connect that handles key brokerage and establishing connection to EC2 Instances through an interface near-identical to standard system ssh, sftp, and o…
scikit-learn_bench benchmarks various implementations of machine learning algorithms across data analytics frameworks. It currently support the scikit-learn, DAAL4PY, cuML, and XGBoost frameworks f…
philschmid / llmperf
Forked from ray-project/llmperfLLMPerf is a library for validating and benchmarking LLMs
Automated TPC-DS and TPC-H benchmark for Apache Hive LLAP
Automated KRAI X workflows for reproducing MLPerf Inference submissions