2026-04-03 - 2026-04-06
Overview
There has been no commit activity in this period.
1 Release published by 1 user
Published
GoldenRockachopa
8 Issues closed from 1 user
Closed
#24 [P2-4] Run full quality comparison: turbo4 vs f16 on 10 test prompts
Closed
#1 TurboQuant — KV Cache Compression for Local Inference on M4 Max
Closed
#25 [P2-5] Download qwen3.5:27b and benchmark turbo4 at 64K/128K context
Closed
#23 [P2-3] Fix Ollama install and build custom Ollama with TurboQuant fork
Closed
#26 [P2-6] Production cutover: swap Timmy's llama-server to TurboQuant
Closed
#27 [TQ-2] Build TheTom/llama-cpp-turboquant for M3 Max Metal
Closed
#30 [EPIC] TurboQuant + Gemma 4 Local Mac Deployment
Closed
#31 [TQ-1] Download Gemma 4 via Ollama on Mac
6 Issues created by 1 user
Opened
#27 [TQ-2] Build TheTom/llama-cpp-turboquant for M3 Max Metal
Opened
#28 [TQ-4] Create Hermes profile for local Gemma 4 + TurboQuant
Opened
#29 [TQ-5] Benchmark: latency, memory, quality comparison
Opened
#30 [EPIC] TurboQuant + Gemma 4 Local Mac Deployment
Opened
#31 [TQ-1] Download Gemma 4 via Ollama on Mac
Opened
#32 [TQ-3] Perplexity quality gate: turbo4 vs f16
7 Unresolved Conversations
Open
#21
[P2-1] Download wikitext-2-raw and run perplexity quality gate
Open
#22
[P2-2] Write 10 test prompts for quality comparison
Open
#17
TurboQuant Initiative Review & Contributor Feedback
Open
#16
[P1-PREP] Write 10 predefined test prompts
Open
#15
[P4] Upstream llama.cpp / Ollama TurboQuant watch
Open
#11
[P2] Full test matrix — 10 prompts + quality + performance
Open
#12
[P2] Long-session quality test — 50-turn conversation