Neural Magic's Projects
An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.
Hackathon 2022
CLIP-like model evaluation
TODO: Finalize name
🤗 The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools
Sparsity-aware deep learning inference runtime for CPUs
Repo for building and packaging a 1-click app for DigitalOcean
Top-level directory for documentation and general content
Notebooks using the Neural Magic libraries 📓
woop wooop
Reference implementations of MLPerf™ inference benchmarks
⚡ Building applications with LLMs through composability ⚡
A framework for few-shot evaluation of autoregressive language models.
Neural Magic GHA
An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.
Neural Magic Docker
A high-throughput and memory-efficient inference and serving engine for LLMs
PyTorch image models, scripts, pretrained weights -- ResNet, ResNeXT, EfficientNet, EfficientNetV2, NFNet, Vision Transformer, MixNet, MobileNet-V3/V2, RegNet, DPN, CSPNet, and more
Framework agnostic sliced/tiled inference + interactive ui + error analysis plots
Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models
Neural network model repository for highly sparse and sparse-quantized models with matching sparsification recipes
ML model optimization product to accelerate inference.
LLM training code for MosaicML foundation models
🤗Transformers: State-of-the-art Natural Language Processing for Pytorch and TensorFlow 2.0.
Supercharge Your Model Training
LLM training code for MosaicML foundation models