Run any AI model on any device

Secure, fast & open-source

Run text, vision, and speech models locally.

Pick any model from Hugging Face.

Mistral
Qwen
Open AI
Gemini
Deepseek
import "dart:io"

final chat = await nobodywho.Chat.fromPath(modelPath: "./model.gguf");
final response = await chat.ask("What is the capital of denmark?").completed();

// The capital of Denmark is Copenhagen.
from nobodywho import Chat

chat = Chat("./model.gguf")
response: str = chat.ask("What is the capital of denmark?").completed()

# The capital of Denmark is Copenhagen.
extends NobodyWhoChat

func _ready():
    self.model_node = get_node("../ChatModel")

    self.response_updated.connect(_on_response_updated)
    self.response_finished.connect(_on_response_finished)

    self.start_worker()

    self.ask("What is the capital of denmark?")

func _on_response_finished(response):
    print(response)

# The capital of Denmark is Copenhagen.

Run anywhere, no servers required.

On-device inference engine for smartwatch apps.

Apple Android Windows Linux

Fast & Secure inference

Bolt

Performance

Fast on any hardware

Optimized kernels for Metal, CUDA, and Vulkan. Get high-performance inference even on integrated graphics and mobile chips.

Lock

Privacy

Keep it private

Everything stays on your machine, no cloud logging, no exceptions. Your prompts never leave the memory of your local GPU.

Offline

Offline

No internet needed

Run models in flight or in secure environments.

Any model

Thousands of LLMs

Direct support for GGUF format.

Bank notes

$0

Free

Completely free, no API keys, or usage fees.

European open-source AI

The library is open source under EUPL 1.2 and is free for both individuals and companies.