sirius93123 Goto Github PK
Type: User
Type: User
Library for specialized dense and sparse matrix operations, and deep learning primitives.
Fast inference from large lauguage models via speculative decoding
The LLVM Project is a collection of modular and reusable compiler and toolchain technologies. Note: the repository does not accept github pull requests at this moment. Please submit your patches at http://reviews.llvm.org.
FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.
Master Thesis on Bayesian Convolutional Neural Network using Variational Inference
Implemented the max pool filter in CUDA using the in built library and using shared memory
Assembler for NVIDIA Maxwell architecture
Ongoing research training transformer language models at scale, including: BERT & GPT-2
Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.
OpenMMLab Computer Vision Foundation
OpenMMLab Detection Toolbox and Benchmark
mobilev2-yolov5s剪枝、蒸馏,支持ncnn,tensorRT部署。ultra-light but better performence!
model compression and deploy. compression: 1、quantization: quantization-aware-training, 16/8/4/2-bit(dorefa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、ternary/binary(twn/bnn/xnor-net); post-training-quantization, 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization folding for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
Collection of algorithms for approximating Fisher Information Matrix for Natural Gradient (and second order method in general)
Assembler and Decompiler for NVIDIA (Maxwell Pascal Volta Turing Ampere) GPUs.
PyTorch*-based Neural Network Compression Framework for enhanced OpenVINO™ inference
A flexible and efficient deep neural network (DNN) compiler that generates high-performance executable from a DNN model description.
An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
CUDA Kernel Benchmarking Library
Enabling on-the-fly manipulations with LLVM IR code of CUDA sources
[ICLR 2020] Once for All: Train One Network and Specialize it for Efficient Deployment
one-shot-tuner
Open standard for machine learning interoperability
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.