Skip to content

Latest commit

 

History

History
38 lines (27 loc) · 1.54 KB

File metadata and controls

38 lines (27 loc) · 1.54 KB

Mini-Inference Engine

English | 简体中文 | Docs

CI Docs License: MIT CUDA C++ CMake

Mini-Inference Engine is a CUDA GEMM optimization learning project that packages progressive matrix multiplication kernels, a lightweight inference runtime, and profiling-oriented experimentation into one repository.

Repository Overview

  • CUDA kernels and runtime headers in src/ and include/
  • Technical docs under docs/
  • Benchmarks and demos in benchmarks/
  • GitHub Pages site for documentation entry, reading paths, and project updates

Quick Start

cmake --preset release
cmake --build --preset release
./build/release/benchmark
./build/release/tests

Docs

  • Project docs: https://lessup.github.io/mini-inference-engine/
  • Site home explains what to read first for architecture, optimization, and API details
  • See CONTRIBUTING.md for contribution workflow

License

MIT License