π§ ChatΒΆ
ποΈ AI Setup Timeline on TrueNAS SCALE (LLaMA + Open WebUI + Cloud APIs)ΒΆ
π₯οΈ 1. System OverviewΒΆ
- Running on TrueNAS SCALE
- Hardware:
- π§ AMD Ryzen 5 PRO 4650G (6C/12T)
- πΎ 32GB RAM
- π§© GIGABYTE B550I AORUS PRO AX motherboard
- No discrete GPU (using integrated Vega graphics)
- AI services hosted via app containers on TrueNAS
π§ 2. Install Local LLM Runtime (Ollama)ΒΆ
- Ollama was used to run local language models like:
phi4-mini,gemma:4b,mistral,llama2- The API became accessible at:
http://localhost:30068
(after deploying the app on TrueNAS SCALE)
Setup Notes: - Ollama on TrueNAS SCALE wasnβt plug-and-play β I had to open the app logs to find the public API key required to connect it to Open WebUI. - I allocated 6 CPUs and 20GB RAM to the app container.
Performance Reality:
- While models like mistral and llama2 technically ran, performance was underwhelming:
- Some responses took 25+ seconds or didnβt complete
- Usage was CPU-bound, even with aggressive RAM allocation
Final Local Model Choices (based on speed + practicality): - β phi4-mini β Reasonably fast, smart, low memory usage - β gemma:4b β Also fast and stable for general use
These are now my go-to free models for everyday tasks. Gemma is more verbose, but I like it.
π 3. Install Open WebUIΒΆ
- Open WebUI is a sleek browser-based interface for chatting with AI models.
- It connects directly to your Ollama backend and offers:
- β Multi-user support (great for sharing with family)
- π Chat history and per-user settings
- π Easy model switching between local and API models
- Hosted directly on the TrueNAS SCALE box.
Note:
I was able to add family members using group permissions, and I limited which models each could access β great for keeping things simple and budget-conscious.
βοΈ 4. Add Cloud API ModelsΒΆ
- Connected Open WebUI to OpenAI (GPT-4o, GPT-3.5) and Anthropic (Claude 3.5/3.7) by adding API keys.
- Cloud models are only used when local ones arenβt enough β keeping costs in check.
π API Setup Notes:ΒΆ
- π§ OpenAI:
- Added $5 to my OpenAI account and retrieved an API key from platform.openai.com/api-keys
-
Important: Check the pricing page first to avoid surprises
-
π€ Anthropic (Claude):
- Added $15 via Anthropic Console
- Integration required a custom provider pipeline to get Claude working: anthropic_manifold_pipeline.py
β Chosen Cloud Models (Based on Cost-Effectiveness):ΒΆ
| Model | Why I Chose It |
|---|---|
anthropic/claude-3.7-sonnet |
Great quality, lower cost than GPT-4 |
gpt-4o-mini |
Smart & responsive at a good price |
anthropic/claude-3.5-haiku |
Super cheap and fast for general tasks |
o1-mini |
Lightweight experimental model |
gpt-3.5-turbo |
Cheap fallback with decent capability |