#machine-learning #ai #cpu-gpu

ort

A safe Rust wrapper for ONNX Runtime 1.17 - Optimize and Accelerate Machine Learning Inferencing

32 releases (stable)

new 2.0.0-rc.2 Apr 27, 2024
2.0.0-rc.1 Mar 28, 2024
2.0.0-rc.0 Feb 6, 2024
2.0.0-alpha.4 Dec 28, 2023
1.13.1 Nov 27, 2022

#5 in Machine learning

Download history 2882/week @ 2024-01-11 2210/week @ 2024-01-18 2117/week @ 2024-01-25 2118/week @ 2024-02-01 2613/week @ 2024-02-08 3024/week @ 2024-02-15 3075/week @ 2024-02-22 3091/week @ 2024-02-29 3533/week @ 2024-03-07 4792/week @ 2024-03-14 4840/week @ 2024-03-21 5631/week @ 2024-03-28 4581/week @ 2024-04-04 6206/week @ 2024-04-11 6002/week @ 2024-04-18 5301/week @ 2024-04-25

23,113 downloads per month
Used in 36 crates (25 directly)

MIT/Apache

505KB
10K SLoC


Coverage Results Crates.io Open Collective backers and sponsors
Crates.io ONNX Runtime

ort is an (unofficial) ONNX Runtime 1.17 wrapper for Rust based on the now inactive onnxruntime-rs. ONNX Runtime accelerates ML inference on both CPU & GPU.

πŸ“– Documentation

πŸ€” Support

πŸ’– Projects using ort

Open a PR to add your project here 🌟

  • Twitter uses ort to serve homepage recommendations to hundreds of millions of users.
  • Bloop uses ort to power their semantic code search feature.
  • pyke Diffusers uses ort for efficient Stable Diffusion image generation on both CPUs & GPUs.
  • edge-transformers uses ort for accelerated transformer model inference at the edge.
  • Ortex uses ort for safe ONNX Runtime bindings in Elixir.
  • Supabase uses ort to remove cold starts for their edge functions.
  • Lantern uses ort to provide embedding model inference inside Postgres.

🌠 Sponsor ort


Dependencies

~2–15MB
~183K SLoC