Self-Hosting Fish Audio on Strix Halo
Running Fish Audio's 4B parameter S2-Pro text-to-speech model locally on an AMD Strix Halo integrated GPU via ROCm and Podman.
read more →Running Fish Audio's 4B parameter S2-Pro text-to-speech model locally on an AMD Strix Halo integrated GPU via ROCm and Podman.
read more →A Telegram bot backed by the Pi coding agent with autonomous scheduling, persistent memory, cross-session search, and a web dashboard.
read more →A browser-based AI music generation tool powered by ACE-Step, ported to Linux for local generation on AMD Strix Halo hardware.
read more →A full-stack quiz platform that turns markdown files and YouTube transcripts into mixed-format quizzes with AI grading, contextual chat, and performance analytics.
read more →A modular collection of services for building a personal AI agent — tool use, memory, browser automation, TTS, and multi-platform chat interfaces.
read more →Snap a photo of a handwritten list, OCR it with a local vision model, and print a formatted checklist on a thermal receipt printer.
read more →A Dockerfile and docker-compose setup for running llama.cpp with its Python bindings in a container, because finding a working one shouldn't be this hard.
read more →An all-in-one Docker container that bundles LLMs, embeddings, vision, and TTS into a single unified inference server.
read more →Generating weekly Spotify Wrapped-style reports from browser history using local models and Browser Recall data.
read more →Building a custom speech-to-text solution for Linux Wayland users using NVIDIA's Canary model, Silero VAD, and ydotool.
read more →Building a personal browser history search engine with full-text recall, inspired by Microsoft's Recall and rewind.ai.
read more →Building a natural language voice controller for the Linux desktop using Qt6, a tiny 1.7B LLM, and a clever vector embedding trick for tool calling.
read more →Building a private, browser-based voice assistant using WebAssembly, Moonshine STT, Piper TTS, and local LLMs.
read more →Building a fully offline voice interface for LLMs using WebAssembly — VAD, speech-to-text, and text-to-speech all running client-side.
read more →