You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Rustane builds on the work of many researchers and open-source projects. This documents every significant source that informed the architecture, research, and implementation.
Clean Rust FFI to private AppleNeuralEngine.framework via objc2. GPT-2 inference example. 2,567 LOC across 5 key files. Our base for ane-bridge (will vendor and extend for training).
Pure Rust Metal LLM engine. 40+ Metal shaders, fused MLP epilogue, quantized dispatch hierarchy, speculative decoding agent module. Closest to production Rust edge AI on M-series (no ANE). Informs our metal-decode crate design.
Rust ML framework with Metal + CUDA backends. Hardcoded 3-variant Storage/Device enums make direct ANE integration impractical — but CUDA backend via cudarc is our Jetson deployment path. SafeTensors + GGML quantization support.
Swift + MetalRT inference CLI. 658 tok/s MetalRT benchmarks on M-series. Voice/agent pipeline. Useful for comparing our ANE inference tok/s against GPU-only approaches on the same hardware.
Production agent SDKs (iOS/Android). Screen reindexing, /no_think prompting, inference guards. Language-agnostic patterns we'll port to Rust for the agent loop — especially relevant for Jetson drone/sat deployment.
MLX port of autoresearch for Apple Silicon. Muon+AdamW optimizer, 241 autonomous experiments, val_bpb 1.664→1.266. Provides our validation baseline and architecture exploration data.