Private
Assistant
ON-DEVICE · LOCAL INFERENCE
Model:
Llama 3 8B
Quantization:
Q4_K_M
Backend:
llama.cpp
Cloud calls:
0
Endpoint:
localhost
All inference runs locally. No data leaves this machine.
Send
Network Monitor
Network Activity
0 requests
GET
/index.html
200 local