Skip to content

🧠 Chat¢

πŸ—‚οΈ AI Setup Timeline on TrueNAS SCALE (LLaMA + Open WebUI + Cloud APIs)ΒΆ

πŸ–₯️ 1. System OverviewΒΆ

  • Running on TrueNAS SCALE
  • Hardware:
  • 🧠 AMD Ryzen 5 PRO 4650G (6C/12T)
  • πŸ’Ύ 32GB RAM
  • 🧩 GIGABYTE B550I AORUS PRO AX motherboard
  • No discrete GPU (using integrated Vega graphics)
  • AI services hosted via app containers on TrueNAS

🧠 2. Install Local LLM Runtime (Ollama)¢

  • Ollama was used to run local language models like:
  • phi4-mini, gemma:4b, mistral, llama2
  • The API became accessible at:
    http://localhost:30068
    (after deploying the app on TrueNAS SCALE)

Setup Notes: - Ollama on TrueNAS SCALE wasn’t plug-and-play β€” I had to open the app logs to find the public API key required to connect it to Open WebUI. - I allocated 6 CPUs and 20GB RAM to the app container.

Performance Reality: - While models like mistral and llama2 technically ran, performance was underwhelming: - Some responses took 25+ seconds or didn’t complete - Usage was CPU-bound, even with aggressive RAM allocation

Final Local Model Choices (based on speed + practicality): - βœ… phi4-mini β€” Reasonably fast, smart, low memory usage - βœ… gemma:4b β€” Also fast and stable for general use

These are now my go-to free models for everyday tasks. Gemma is more verbose, but I like it.


🌐 3. Install Open WebUI¢

  • Open WebUI is a sleek browser-based interface for chatting with AI models.
  • It connects directly to your Ollama backend and offers:
  • βœ… Multi-user support (great for sharing with family)
  • πŸ•“ Chat history and per-user settings
  • πŸ”€ Easy model switching between local and API models
  • Hosted directly on the TrueNAS SCALE box.

Note:
I was able to add family members using group permissions, and I limited which models each could access β€” great for keeping things simple and budget-conscious.


☁️ 4. Add Cloud API Models¢

  • Connected Open WebUI to OpenAI (GPT-4o, GPT-3.5) and Anthropic (Claude 3.5/3.7) by adding API keys.
  • Cloud models are only used when local ones aren’t enough β€” keeping costs in check.

πŸ” API Setup Notes:ΒΆ


βœ… Chosen Cloud Models (Based on Cost-Effectiveness):ΒΆ

Model Why I Chose It
anthropic/claude-3.7-sonnet Great quality, lower cost than GPT-4
gpt-4o-mini Smart & responsive at a good price
anthropic/claude-3.5-haiku Super cheap and fast for general tasks
o1-mini Lightweight experimental model
gpt-3.5-turbo Cheap fallback with decent capability