Run any AI model on any device
Secure, fast & open-source
Run text, vision, and speech models locally.
Pick any model from Hugging Face.
import "dart:io"
final chat = await nobodywho.Chat.fromPath(modelPath: "./model.gguf");
final response = await chat.ask("What is the capital of denmark?").completed();
// The capital of Denmark is Copenhagen.
from nobodywho import Chat
chat = Chat("./model.gguf")
response: str = chat.ask("What is the capital of denmark?").completed()
# The capital of Denmark is Copenhagen.
extends NobodyWhoChat
func _ready():
self.model_node = get_node("../ChatModel")
self.response_updated.connect(_on_response_updated)
self.response_finished.connect(_on_response_finished)
self.start_worker()
self.ask("What is the capital of denmark?")
func _on_response_finished(response):
print(response)
# The capital of Denmark is Copenhagen.
Run anywhere, no servers required.
On-device inference engine for smartwatch apps.
Fast & Secure inference
Performance
Fast on any hardware
Optimized kernels for Metal, CUDA, and Vulkan. Get high-performance inference even on integrated graphics and mobile chips.
Privacy
Keep it private
Everything stays on your machine, no cloud logging, no exceptions. Your prompts never leave the memory of your local GPU.
Offline
No internet needed
Run models in flight or in secure environments.
Any model
Thousands of LLMs
Direct support for GGUF format.
$0
Free
Completely free, no API keys, or usage fees.
European open-source AI
The library is open source under EUPL 1.2 and is free for both individuals and companies.