
What is local.ai
Local AI is a native application designed to simplify the process of experimenting with AI models offline and in private. It requires no GPU and is free and open-source. The app supports CPU inferencing, model management, and digest verification, making it a versatile tool for AI enthusiasts and developers.
How to Use local.ai
- Download and install the local.ai app on your system.
- Load your desired AI model into the app.
- Start an inference session with just 2 clicks.
- Use the app's features to manage models, verify their integrity, and run inferencing tasks.
Features of local.ai
-
CPU Inferencing
Supports CPU-based inferencing, adapting to the available threads for efficient processing.
-
Model Management
Centralized management of AI models, including resumable and concurrent downloads, usage-based sorting, and directory agnostic storage.
-
Digest Verification
Ensures the integrity of downloaded models using BLAKE3 and SHA256 digest computation, with features like known-good model API and license usage chips.
-
Inferencing Server
Start a local streaming server for AI inferencing with just 2 clicks, featuring a quick inference UI and support for writing to .mdx files.