The AI Assistant must be installed before you can use it. Enable it during the Easy Setup wizard or install it from Settings → Apps.
What it can do
Answer questions
Ask about any topic — history, science, practical skills, or current events up to the model’s training cutoff.
Explain concepts
Get plain-language explanations of complex or technical subjects.
Help with writing
Draft, edit, proofread, or reformat text for any purpose.
Brainstorm ideas
Work through problems, generate options, or think out loud with an interactive partner.
How to access
Open AI Chat from the Command Center home screen, or navigate directly to/chat.
How to install
Open the setup wizard or Apps page
If this is your first time setting up N.O.M.A.D., use the Easy Setup wizard — the AI Assistant is one of the capabilities you can enable on the first step.To install it later, go to Settings → Apps.
Install the AI Assistant
Select the AI Assistant and start the installation. N.O.M.A.D. will pull the required Ollama container and configure it automatically.
GPU acceleration
Local AI inference is computationally intensive. On CPU only, expect roughly 10–15 tokens per second. With an NVIDIA GPU and the NVIDIA Container Toolkit installed, that typically improves to 100+ tokens per second — a 10–20x improvement in response speed.How to enable GPU acceleration
Install an NVIDIA GPU
If your server does not already have a compatible NVIDIA GPU, install one. More VRAM allows you to run larger, more capable models.
Install the NVIDIA Container Toolkit on the host
Follow the official NVIDIA Container Toolkit installation guide on your server’s operating system.
Supported models
The AI Assistant supports any model available through Ollama. You can browse, download, and manage models from Settings → Models (/settings/models). The full catalog of available models is listed at ollama.com/library.
Model size affects both speed and capability:
- Smaller models (1B–3B parameters) — fastest responses, good for simple Q&A
- Mid-size models (7B–8B parameters) — balanced speed and quality, recommended for most use cases
- Larger models (14B–70B+ parameters) — highest quality, require significantly more RAM and benefit most from GPU acceleration
AI model files range from approximately 1–40GB depending on the model. Plan your storage accordingly.
Tips for best results
Be specific in your questions
Be specific in your questions
Vague prompts produce vague answers. Instead of “tell me about plants”, try “what vegetables grow well in partial shade in a temperate climate?”. The more context you provide, the more useful the response.
Use follow-up questions
Use follow-up questions
The AI maintains context within a conversation session. If the first response isn’t quite right, ask it to clarify, expand, or approach the topic differently.
Upload relevant documents to the Knowledge Base
Upload relevant documents to the Knowledge Base
If you’re asking about a specific manual, procedure, or document, upload it to the Knowledge Base first. The AI will find and reference relevant sections automatically.
Start a new session for unrelated topics
Start a new session for unrelated topics
Use a larger model for complex tasks
Use a larger model for complex tasks
If you need detailed analysis, nuanced writing, or complex reasoning, download and select a larger model. If responses are too slow, switch to a smaller model for day-to-day questions.
Related
Knowledge Base
Upload documents so the AI can reference them when answering your questions.
Apps
Install, manage, and Force Reinstall the AI Assistant from the Apps page.