A complete and budget-friendly way to build a private, low-cost, always-available LLM server with a Raspberry Pi 5, Ollama inference engine, Tailscale for secure remote access, and Chatbox as a desktop/mobile client.
Notes tagged with ollama
ollama
A comprehensive guide to running Large Language Models locally on affordable single-board computers, exploring hardware options, performance benchmarks, and practical setup for private and cost-effective AI.
A guide to deciphering model names for better AI Engineering decisions
Learn how to effectively control the behavior of Large Language Models (LLMs) using inference parameters like temperature, top-p, and more. This guide provides practical examples with Python and LangChain.
Learn how to create precise, structured prompts for AI agents using the CO-STAR framework, enhancing reliability and performance in LLM applications.