20 Repositories
Rust inference Libraries
Infer a JSON schema from example data, produce nonsense synthetic data (drivel) according to the schema
drivel drivel is a command-line tool written in Rust for inferring a schema from an example JSON (or JSON lines) file, and generating synthetic data (
Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.
candle-vllm Efficient platform for inference and serving local LLMs including an OpenAI compatible API server. Features OpenAI compatible API server p
LLaMA2 port for Rust inspired by llama2.c
llama2-rs LLaMA2 port for Rust inspired by llama2.c. TODOs: Implement loading of the model Implement forward pass Implement generation Implement token
`dfx new --type=rust` + burn-rs MNIST web inference example
ic-mnist The frontend provides a canvas where users can draw a digit. The drawn digit is then sent to the backend canister running burn-rs for inferen
LLaMa 7b with CUDA acceleration implemented in rust. Minimal GPU memory needed!
LLaMa 7b in rust This repo contains the popular LLaMa 7b language model, fully implemented in the rust programming language! Uses dfdx tensors and CUD
Run LLaMA inference on CPU, with Rust ๐ฆ๐๐ฆ
LLaMA-rs Do the LLaMA thing, but now in Rust ๐ฆ ๐ ๐ฆ Image by @darthdeus, using Stable Diffusion LLaMA-rs is a Rust port of the llama.cpp project. Th
Rust+OpenCL+AVX2 implementation of LLaMA inference code
RLLaMA RLLaMA is a pure Rust implementation of LLaMA large language model inference.. Supported features Uses either f16 and f32 weights. LLaMA-7B, LL
Run LLaMA inference on CPU, with Rust ๐ฆ๐๐ฆ
LLaMA-rs Do the LLaMA thing, but now in Rust ๐ฆ ๐ ๐ฆ Image by @darthdeus, using Stable Diffusion LLaMA-rs is a Rust port of the llama.cpp project. Th
pyke Diffusers is a modular Rust library for optimized Stable Diffusion inference ๐ฎ
pyke Diffusers is a modular Rust library for pretrained diffusion model inference to generate images, videos, or audio, using ONNX Runtime as a backen
Using OpenAI Codex's "davinci-edit" Model for Gradual Type Inference
OpenTau: Using OpenAI Codex for Gradual Type Inference Current implementation is focused on TypeScript Python implementation comes next Requirements r
A statically-typed, interpreted programming language, with generics and type inference
Glide A programming language. Currently, this includes: Static typing Generics, with monomorphization Type inference on function calls func identityT
Wonnx - a GPU-accelerated ONNX inference run-time written 100% in Rust, ready for the web
Wonnx is a GPU-accelerated ONNX inference run-time written 100% in Rust, ready for the web. Supported Platforms (enabled by wgpu) API Windows Linux &
An implementation of a predicative polymorphic language with bidirectional type inference and algebraic data types
Vinilla Lang Vanilla is a pure functional programming language based on System F, a classic but powerful type system. Merits Simple as it is, Vanilla
A fusion of OTP lib/dialyzer + lib/compiler for regular Erlang with type inference
Typed ERLC The Problem I have a dream, that one day there will be an Erlang compiler, which will generate high quality type-correct code from deduced
Tiny, no-nonsense, self-contained, Tensorflow and ONNX inference
Sonos' Neural Network inference engine. This project used to be called tfdeploy, or Tensorflow-deploy-rust. What ? tract is a Neural Network inference
Orkhon: ML Inference Framework and Server Runtime
Orkhon: ML Inference Framework and Server Runtime Latest Release License Build Status Downloads Gitter What is it? Orkhon is Rust framework for Machin
Snips NLU rust implementation
Snips NLU Rust Installation Add it to your Cargo.toml: [dependencies] snips-nlu-lib = { git = "https://github.com/snipsco/snips-nlu-rs", branch = "mas
Orkhon: ML Inference Framework and Server Runtime
Orkhon: ML Inference Framework and Server Runtime Latest Release License Build Status Downloads Gitter What is it? Orkhon is Rust framework for Machin
Tiny, no-nonsense, self-contained, Tensorflow and ONNX inference
Sonos' Neural Network inference engine. This project used to be called tfdeploy, or Tensorflow-deploy-rust. What ? tract is a Neural Network inference
A static, type inferred and embeddable language written in Rust.
gluon Gluon is a small, statically-typed, functional programming language designed for application embedding. Features Statically-typed - Static typin