2026-04-05 - 2026-04-06
Overview
There has been no commit activity in this period.
8 Issues closed from 1 user
Closed
#24 [P2-4] Run full quality comparison: turbo4 vs f16 on 10 test prompts
Closed
#1 TurboQuant — KV Cache Compression for Local Inference on M4 Max
Closed
#25 [P2-5] Download qwen3.5:27b and benchmark turbo4 at 64K/128K context
Closed
#23 [P2-3] Fix Ollama install and build custom Ollama with TurboQuant fork
Closed
#26 [P2-6] Production cutover: swap Timmy's llama-server to TurboQuant
Closed
#27 [TQ-2] Build TheTom/llama-cpp-turboquant for M3 Max Metal
Closed
#30 [EPIC] TurboQuant + Gemma 4 Local Mac Deployment
Closed
#31 [TQ-1] Download Gemma 4 via Ollama on Mac
10 Unresolved Conversations
Open
#29
[TQ-5] Benchmark: latency, memory, quality comparison
Open
#28
[TQ-4] Create Hermes profile for local Gemma 4 + TurboQuant
Open
#32
[TQ-3] Perplexity quality gate: turbo4 vs f16
Open
#21
[P2-1] Download wikitext-2-raw and run perplexity quality gate
Open
#22
[P2-2] Write 10 test prompts for quality comparison
Open
#17
TurboQuant Initiative Review & Contributor Feedback
Open
#15
[P4] Upstream llama.cpp / Ollama TurboQuant watch
Open
#16
[P1-PREP] Write 10 predefined test prompts
Open
#11
[P2] Full test matrix — 10 prompts + quality + performance
Open
#12
[P2] Long-session quality test — 50-turn conversation