Private Assistant

ON-DEVICE · LOCAL INFERENCE
Model: Llama 3 8B Quantization: Q4_K_M Backend: llama.cpp Cloud calls: 0 Endpoint: localhost
All inference runs locally. No data leaves this machine.
Network Activity 0 requests
GET /index.html 200 local