- Switch whisper transcription from faster-whisper (CPU) to lightning-whisper-mlx (GPU) - Fix word_timestamps hanging, use ffprobe for accurate duration - Add Cloudflare Pages Worker for SignalWire voicemail fallback when server offline - Add voicemail sync on startup, delete tracking, save feature - Add /feed RSS proxy to _worker.js (was broken by worker taking over routing) - Redesign website hero section: ghost buttons, compact phone, plain text links - Rewrite caller prompts for faster point-getting and host-following - Expand TOPIC_CALLIN from ~250 to 547 entries across 34 categories - Add new categories: biology, psychology, engineering, math, geology, animals, work, money, books, movies, relationships, health, language, true crime, drunk/high/unhinged callers - Remove bad Inworld voices (Pixie, Dominus), reduce repeat caller frequency - Add audio monitor device routing, uvicorn --reload-dir fix - Publish episode 13 Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
3.3 KiB
3.3 KiB
AI Podcast - Project Instructions
Git Remote (Gitea)
- Repo:
git@gitea-nas:luke/ai-podcast.git - Web: http://mmgnas:3000/luke/ai-podcast
- SSH Host:
gitea-nas(configured in ~/.ssh/config)- HostName:
mmgnas(usemmgnas-10gif wired connection issues) - Port:
2222 - User:
git - IdentityFile:
~/.ssh/gitea_mmgnas
- HostName:
NAS Access
- Hostname:
mmgnas(wireless) ormmgnas-10g(wired/10G) - SSH Port: 8001
- User: luke
- Docker path:
/share/CACHEDEV1_DATA/.qpkg/container-station/bin/docker
Castopod (Podcast Publishing)
- URL: https://podcast.macneilmediagroup.com
- Podcast handle:
@LukeAtTheRoost - API Auth: Basic auth (admin/podcast2026api)
- Container:
castopod-castopod-1 - Database:
castopod-mariadb-1(user: castopod, db: castopod)
Running the App
# Start backend — ALWAYS use --reload-dir to avoid CPU thrashing from file watchers
python -m uvicorn backend.main:app --reload --reload-dir backend --host 0.0.0.0 --port 8000
# Or use run.sh
./run.sh
Publishing Episodes
python publish_episode.py ~/Desktop/episode.mp3
Environment Variables
Required in .env:
- OPENROUTER_API_KEY
- ELEVENLABS_API_KEY (optional)
- INWORLD_API_KEY (for Inworld TTS)
Post-Production Pipeline (added Feb 2026)
- Branch:
feature/real-callers— all current work is here, pushed to gitea - Stem Recorder (
backend/services/stem_recorder.py): Records 5 WAV stems (host, caller, music, sfx, ads) during live shows. Uses lock-free deque architecture — audio callbacks just append to deques, a background writer thread drains to disk.write()for continuous streams (host mic, music, ads),write_sporadic()for burst sources (caller TTS, SFX) with time-aligned silence padding. - Audio hooks in
backend/services/audio.py: 7 tap points guarded byif self.stem_recorder:. Persistent mic stream (start_stem_mic/stop_stem_mic) runs during recording to capture host voice continuously, not just during push-to-talk. - API endpoints:
POST /api/recording/start,POST /api/recording/stop(auto-runs postprod in background thread),POST /api/recording/process - Frontend: REC button in header with red pulse animation when recording
- Post-prod script (
postprod.py): 6-step pipeline — load stems → gap removal → voice compression (ffmpeg acompressor) → music ducking → stereo mix → EBU R128 loudness normalization to -16 LUFS. All steps skippable via CLI flags. - Known issues resolved: Lock-free recorder (old version used threading.Lock in audio callbacks causing crashes), scipy.signal.resample replaced with nearest-neighbor (was producing artifacts on small chunks), sys import bug in auto-postprod, host mic not captured without persistent stream
LLM Settings
_pick_response_budget()in main.py controls caller dialog token limits (150-450 tokens). MiniMax respects limits strictly — if responses seem short, check these values.- Default max_tokens in llm.py is 300 (for non-caller uses)
- Grok (
x-ai/grok-4-fast) works well for natural dialog; MiniMax tends toward terse responses
Website
- Domain: lukeattheroost.com (behind Cloudflare)
- Analytics: Cloudflare Web Analytics (enable in Cloudflare dashboard, no code changes needed)
Episodes Published
- Episode 6 published 2026-02-08 (podcast6.mp3, ~31 min)