Compare commits

...

1 Commits

Author SHA1 Message Date
Alexander Whitestone
d2ef914edd feat: Comprehensive review and improvements for TurboQuant (#17)
All checks were successful
Smoke Test / smoke (pull_request) Successful in 24s
This commit addresses issue #17 by providing a comprehensive review
of the TurboQuant initiative and implementing key improvements.

## Changes

### 1. Initiative Review (docs/INITIATIVE_REVIEW.md)
- Comprehensive assessment of current state
- Code quality findings and recommendations
- Contributor feedback for @manus, @Timmy, @Rockachopa
- Implementation plan with clear milestones

### 2. Code Improvements

#### llama-turbo.cpp
- Added input validation with assertions
- Optimized Lloyd-Max search with binary search (O(log n) vs O(n))
- Added stack allocation for d=128 (avoids heap allocation in hot path)
- Added error handling for edge cases
- Added decision boundaries for efficient quantization

#### ggml-metal-turbo.metal
- Added bounds checking to all kernels
- Added NaN/Inf handling for numerical stability
- Completed fused attention kernel (was stub)
- Added fused attention with softmax kernel
- Added Metal encoding kernel for completeness
- Added binary search for quantization

### 3. Testing (tests/test_turbo.cpp)
- Unit tests for encode/decode round-trip
- Tests for known values (zeros, ones)
- Tests for edge cases (large/small values)
- Error handling tests

### 4. Build System (CMakeLists.txt)
- Added CMake configuration for building library
- Added test executable
- Added install targets

### 5. Documentation (README.md)
- Added build instructions
- Added API documentation
- Added contributing guidelines
- Added code style guide

## Key Improvements

1. **Performance**: Binary search instead of linear search for Lloyd-Max quantization
2. **Memory**: Stack allocation for common case (d=128)
3. **Reliability**: Input validation and error handling
4. **Metal Integration**: Complete fused attention implementation
5. **Testing**: Unit tests for correctness verification
6. **Documentation**: Contributor guidelines and API docs

## Next Steps

1. Run benchmarks to verify performance improvements
2. Test with actual models (qwen3.5:27b)
3. Integrate with llama.cpp fork
4. Deploy to production

Closes #17
2026-04-14 22:07:21 -04:00
6 changed files with 732 additions and 21 deletions

31
CMakeLists.txt Normal file
View File

@@ -0,0 +1,31 @@
cmake_minimum_required(VERSION 3.10)
project(turboquant)
set(CMAKE_CXX_STANDARD 11)
set(CMAKE_CXX_STANDARD_REQUIRED ON)
# Source files
set(SOURCES
llama-turbo.cpp
)
# Header files
set(HEADERS
llama-turbo.h
)
# Create library
add_library(turboquant STATIC ${SOURCES} ${HEADERS})
target_include_directories(turboquant PUBLIC ${CMAKE_CURRENT_SOURCE_DIR})
# Test executable
add_executable(test_turbo tests/test_turbo.cpp)
target_link_libraries(test_turbo turboquant)
# Install
install(TARGETS turboquant ARCHIVE DESTINATION lib)
install(FILES ${HEADERS} DESTINATION include)
# Tests
enable_testing()
add_test(NAME turboquant_tests COMMAND test_turbo)

View File

@@ -15,6 +15,93 @@ A 27B model at 128K context with TurboQuant beats a 72B at Q2 with 8K context.
## Status
See [issues](http://143.198.27.163:3000/Timmy_Foundation/turboquant/issues) for current progress.
## Building
### Prerequisites
- CMake 3.10+
- C++11 compiler
- Xcode Command Line Tools (for Metal on macOS)
### Build Instructions
```bash
# Clone the repository
git clone https://forge.alexanderwhitestone.com/Timmy_Foundation/turboquant.git
cd turboquant
# Build with CMake
cmake -B build -DCMAKE_BUILD_TYPE=Release
cmake --build build
# Run tests
cd build && ctest
```
### Integration with llama.cpp
See [PR-IMPLEMENTATION-PLAN.md](PR-IMPLEMENTATION-PLAN.md) for integration steps.
## API
### CPU Reference Implementation
```c
// Encode: Compress float vector to 4-bit packed representation
void polar_quant_encode_turbo4(
const float* src, // Input: float array [d]
uint8_t* dst, // Output: packed 4-bit indices [d/2]
float* norm, // Output: L2 norm (radius)
int d // Dimension (must be power of 2, e.g., 128)
);
// Decode: Decompress 4-bit packed representation to float vector
void polar_quant_decode_turbo4(
const uint8_t* src, // Input: packed 4-bit indices [d/2]
float* dst, // Output: float array [d]
float norm, // Input: L2 norm (radius)
int d // Dimension (must be power of 2, e.g., 128)
);
```
### Metal Shaders
See `ggml-metal-turbo.metal` for GPU-accelerated kernels:
- `kernel_fwht_128`: Fast Walsh-Hadamard Transform
- `kernel_turbo4_dequant`: Dequantization for attention
- `kernel_attention_turbo4`: Fused attention computation
- `kernel_attention_turbo4_softmax`: Fused attention with softmax
- `kernel_turbo4_encode`: Encoding on GPU
## Contributing
### Getting Started
1. Fork the repository
2. Create a feature branch: `git checkout -b feature/your-feature`
3. Make your changes
4. Add tests for new functionality
5. Run the test suite: `cd build && ctest`
6. Submit a pull request
### Code Style
- C++11 standard
- 4-space indentation
- Snake_case for functions and variables
- UPPER_CASE for constants
- Add comments for complex algorithms
### Testing
- All new code must have unit tests
- Run tests before submitting PR: `cd build && ctest`
- Test on both CPU and Metal (if applicable)
### Pull Request Process
1. Update documentation if needed
2. Add tests for new functionality
3. Ensure all tests pass
4. Request review from maintainers
### Issues
- Use issue templates when available
- Tag issues appropriately (`bug`, `enhancement`, `documentation`)
- Include reproduction steps for bugs
- For performance issues, include benchmark results
## Roles
- **Strago:** Build spec author
- **Cid:** Implementation, benchmarks, deployment
@@ -30,3 +117,5 @@ See [issues](http://143.198.27.163:3000/Timmy_Foundation/turboquant/issues) for
## Docs
- [BUILD-SPEC.md](BUILD-SPEC.md) — Full build specification (Strago, v2.2)
- [docs/PROJECT_STATUS.md](docs/PROJECT_STATUS.md) — Current project status
- [docs/INITIATIVE_REVIEW.md](docs/INITIATIVE_REVIEW.md) — Initiative review and feedback

167
docs/INITIATIVE_REVIEW.md Normal file
View File

@@ -0,0 +1,167 @@
# TurboQuant Initiative Review & Contributor Feedback
## Executive Summary
The TurboQuant initiative shows promising results with 73% KV memory savings and minimal performance overhead. However, the transition from 'Build Spec' to 'Code Implementation' needs acceleration. This review provides actionable feedback for contributors.
## Current State Assessment
### ✅ What's Working
1. **Phase 1 Results**: 73% KV memory savings with 1% prompt overhead
2. **Algorithm Correctness**: PolarQuant implementation matches paper specifications
3. **Metal Shaders**: Basic dequantization and WHT kernels exist
4. **Documentation**: Comprehensive build spec and status reports
### ⚠️ What Needs Improvement
1. **Repository Activity**: Only 3 commits — implementation needs acceleration
2. **Code Quality**: Several issues in current implementation
3. **Metal Integration**: Fused attention kernel is incomplete (stub only)
4. **Testing**: No unit tests or integration tests
5. **Documentation**: Missing contributor guidelines and API docs
## Code Review Findings
### 1. llama-turbo.cpp Issues
#### Issue 1.1: Inefficient Lloyd-Max Search
```cpp
// Current: O(n) linear search through 16 centroids
int best_idx = 0;
float min_dist = fabsf(val - turbo4_centroids[0]);
for (int j = 1; j < 16; j++) {
float dist = fabsf(val - turbo4_centroids[j]);
if (dist < min_dist) {
min_dist = dist;
best_idx = j;
}
}
```
**Problem**: Linear search is inefficient. With 128 dimensions per vector, this runs 128 × 16 = 2048 comparisons per vector.
**Solution**: Use binary search or precomputed decision boundaries.
#### Issue 1.2: Missing Error Handling
```cpp
void polar_quant_encode_turbo4(const float* src, uint8_t* dst, float* norm, int d) {
// No validation of inputs
// No check for d being power of 2
// No check for null pointers
}
```
**Solution**: Add input validation.
#### Issue 1.3: Memory Allocation
```cpp
std::vector<float> rotated(src, src + d); // Heap allocation per call
```
**Problem**: Heap allocation in hot path. For 1000 vectors, this is 1000 allocations.
**Solution**: Use stack allocation for small d (d=128) or preallocated buffer.
### 2. ggml-metal-turbo.metal Issues
#### Issue 2.1: Incomplete Fused Attention Kernel
```metal
kernel void kernel_attention_turbo4(...) {
// 1. Dequantize K on the fly
// 2. Compute dot product with Q
// 3. Store score
}
```
**Problem**: This is a stub. The real performance win comes from fusing dequantization with attention computation.
**Solution**: Implement the fused kernel.
#### Issue 2.2: Missing Error Checking
```metal
kernel void kernel_fwht_128(...) {
// No bounds checking
// No NaN/Inf handling
}
```
**Solution**: Add bounds checking and numerical stability.
### 3. Integration Issues
#### Issue 3.1: Missing CMake Integration
The PR-IMPLEMENTATION-PLAN.md mentions updating CMake, but there's no CMakeLists.txt in the repo.
#### Issue 3.2: No Test Suite
No unit tests for the CPU implementation, no integration tests for Metal.
## Contributor Feedback
### For @manus (Implementation)
1. **Priority 1**: Complete the fused attention kernel in Metal
2. **Priority 2**: Add input validation to all functions
3. **Priority 3**: Optimize Lloyd-Max search with binary search
4. **Priority 4**: Add unit tests for encode/decode round-trip
### For @Timmy (Spec Alignment)
1. **Action**: Review Metal shader performance against spec benchmarks
2. **Action**: Verify that WHT rotation is correctly implemented in Metal
3. **Action**: Ensure codebook boundaries match the paper's specifications
### For @Rockachopa (Quality Oversight)
1. **Risk**: CPU turbo4 reference path is incompatible with Metal dequant
2. **Action**: Add integration tests that verify CPU and Metal produce same results
3. **Action**: Implement PPL testing with wikitext-2-raw corpus
## Implementation Plan
### Phase 1: Code Quality (Week 1)
1. Add input validation to all functions
2. Fix memory allocation issues
3. Add error handling
4. Create unit tests
### Phase 2: Metal Integration (Week 2)
1. Complete fused attention kernel
2. Add bounds checking to all kernels
3. Optimize memory access patterns
4. Add integration tests
### Phase 3: Documentation (Week 3)
1. Create API documentation
2. Write contributor guidelines
3. Add code examples
4. Create performance benchmarks
### Phase 4: Production Readiness (Week 4)
1. Run full test suite
2. Performance optimization
3. Memory leak detection
4. Production deployment guide
## Action Items
### Immediate (This Week)
- [ ] Fix input validation in llama-turbo.cpp
- [ ] Add error handling to Metal shaders
- [ ] Create unit test framework
- [ ] Document API surface
### Short-term (Next 2 Weeks)
- [ ] Complete fused attention kernel
- [ ] Optimize Lloyd-Max search
- [ ] Add integration tests
- [ ] Create contributor guidelines
### Long-term (Next Month)
- [ ] Performance benchmarking
- [ ] Memory optimization
- [ ] Production deployment
- [ ] Upstream integration
## Conclusion
TurboQuant has strong technical foundations but needs focused implementation effort. The biggest risk is the incomplete Metal fused attention kernel — this is where the real performance win lives. Contributors should prioritize completing this work to accelerate the transition from 'Build Spec' to 'Code Implementation'.
**Rating**: 7/10 — Strong algorithm, needs implementation polish
**Next Steps**: Focus on Metal integration and testing to achieve production readiness.

View File

@@ -10,6 +10,14 @@ constant float turbo4_centroids[16] = {
0.1523, 0.2154, 0.2800, 0.3500
};
// Decision boundaries for binary search (precomputed)
constant float turbo4_boundaries[15] = {
-0.18385, -0.1322, -0.09665, -0.0683,
-0.04375, -0.0213, 0.0, 0.0213,
0.04375, 0.0683, 0.09665, 0.1322,
0.18385, 0.2477, 0.315
};
// Fast Walsh-Hadamard Transform (In-place, SIMD-optimized)
// Assumes d=128 (standard head dimension)
kernel void kernel_fwht_128(
@@ -19,6 +27,11 @@ kernel void kernel_fwht_128(
const uint d = 128;
uint base = tid * d;
// Bounds check
if (base + d > 128 * 1024) { // Reasonable upper bound
return;
}
// Stage 1-7 (128 = 2^7)
for (uint h = 1; h < d; h <<= 1) {
for (uint i = 0; i < d; i += (h << 1)) {
@@ -38,6 +51,20 @@ kernel void kernel_fwht_128(
}
}
// Binary search for Lloyd-Max quantization (Metal version)
inline uint quantize_turbo4_metal(float val) {
uint left = 0, right = 14;
while (left < right) {
uint mid = (left + right) / 2;
if (val < turbo4_boundaries[mid]) {
right = mid;
} else {
left = mid + 1;
}
}
return left;
}
// PolarQuant Turbo4 Dequantization (Attention Hot Path)
// Unpacks 4-bit indices, looks up centroids, scales by radius
kernel void kernel_turbo4_dequant(
@@ -49,28 +76,218 @@ kernel void kernel_turbo4_dequant(
const uint d = 128;
uint base_src = tid * (d / 2);
uint base_dst = tid * d;
// Bounds check
if (base_dst + d > 128 * 1024) {
return;
}
float norm = norms[tid];
// Check for NaN/Inf in norm
if (isnan(norm) || isinf(norm) || norm < 0) {
// Fill with zeros for invalid norm
for (uint i = 0; i < d; i++) {
dst[base_dst + i] = 0.0;
}
return;
}
for (uint i = 0; i < d; i++) {
uchar packed = src[base_src + (i / 2)];
uint idx = (i % 2 == 0) ? (packed & 0x0F) : (packed >> 4);
dst[base_dst + i] = turbo4_centroids[idx] * norm;
// Bounds check for index
if (idx >= 16) {
dst[base_dst + i] = 0.0;
} else {
dst[base_dst + i] = turbo4_centroids[idx] * norm;
}
}
// Note: FWHT is applied separately or fused into attention
}
// Fused Attention with TurboQuant (Conceptual)
// Fused Attention with TurboQuant (Complete Implementation)
// This is where the real speed win happens
kernel void kernel_attention_turbo4(
device const float* q [[buffer(0)]],
device const uchar* k_packed [[buffer(1)]],
device const float* k_norms [[buffer(2)]],
device float* scores [[buffer(3)]],
constant uint& d [[buffer(4)]],
constant uint& seq_len [[buffer(4)]],
constant uint& head_dim [[buffer(5)]],
uint tid [[thread_position_in_grid]]
) {
// 1. Dequantize K on the fly
// 2. Compute dot product with Q
// 3. Store score
// Each thread computes one attention score
uint query_idx = tid / seq_len;
uint key_idx = tid % seq_len;
// Bounds check
if (query_idx >= seq_len || key_idx >= seq_len) {
return;
}
// Dequantize key on the fly
uint key_base = key_idx * (head_dim / 2);
float key_norm = k_norms[key_idx];
// Check for invalid norm
if (isnan(key_norm) || isinf(key_norm) || key_norm < 0) {
scores[tid] = -INFINITY;
return;
}
// Compute dot product: Q · K
float dot_product = 0.0;
for (uint i = 0; i < head_dim; i++) {
uchar packed = k_packed[key_base + (i / 2)];
uint idx = (i % 2 == 0) ? (packed & 0x0F) : (packed >> 4);
if (idx < 16) {
float k_val = turbo4_centroids[idx] * key_norm;
float q_val = q[query_idx * head_dim + i];
dot_product += q_val * k_val;
}
}
// Scale by sqrt(head_dim) for attention stability
float scale = 1.0 / sqrt(float(head_dim));
scores[tid] = dot_product * scale;
}
// Fused Attention with TurboQuant and Softmax (Complete)
// Computes attention scores and applies softmax in one kernel
kernel void kernel_attention_turbo4_softmax(
device const float* q [[buffer(0)]],
device const uchar* k_packed [[buffer(1)]],
device const float* k_norms [[buffer(2)]],
device float* attention_weights [[buffer(3)]],
constant uint& seq_len [[buffer(4)]],
constant uint& head_dim [[buffer(5)]],
uint tid [[thread_position_in_grid]]
) {
// Each thread computes attention for one query position
uint query_idx = tid;
if (query_idx >= seq_len) {
return;
}
// Compute all attention scores for this query
threadgroup float scores[1024]; // Assuming max seq_len = 1024
float max_score = -INFINITY;
for (uint key_idx = 0; key_idx < seq_len; key_idx++) {
uint key_base = key_idx * (head_dim / 2);
float key_norm = k_norms[key_idx];
if (isnan(key_norm) || isinf(key_norm) || key_norm < 0) {
scores[key_idx] = -INFINITY;
continue;
}
// Compute dot product
float dot_product = 0.0;
for (uint i = 0; i < head_dim; i++) {
uchar packed = k_packed[key_base + (i / 2)];
uint idx = (i % 2 == 0) ? (packed & 0x0F) : (packed >> 4);
if (idx < 16) {
float k_val = turbo4_centroids[idx] * key_norm;
float q_val = q[query_idx * head_dim + i];
dot_product += q_val * k_val;
}
}
// Scale by sqrt(head_dim)
float scale = 1.0 / sqrt(float(head_dim));
scores[key_idx] = dot_product * scale;
// Track max for numerical stability
if (scores[key_idx] > max_score) {
max_score = scores[key_idx];
}
}
// Compute softmax
float sum_exp = 0.0;
for (uint key_idx = 0; key_idx < seq_len; key_idx++) {
if (scores[key_idx] == -INFINITY) {
attention_weights[query_idx * seq_len + key_idx] = 0.0;
} else {
float exp_val = exp(scores[key_idx] - max_score);
attention_weights[query_idx * seq_len + key_idx] = exp_val;
sum_exp += exp_val;
}
}
// Normalize
if (sum_exp > 0.0) {
for (uint key_idx = 0; key_idx < seq_len; key_idx++) {
attention_weights[query_idx * seq_len + key_idx] /= sum_exp;
}
}
}
// PolarQuant Turbo4 Encoding (Metal version for completeness)
kernel void kernel_turbo4_encode(
device const float* src [[buffer(0)]],
device uchar* dst [[buffer(1)]],
device float* norms [[buffer(2)]],
constant uint& head_dim [[buffer(3)]],
uint tid [[thread_position_in_grid]]
) {
uint base_src = tid * head_dim;
uint base_dst = tid * (head_dim / 2);
// Bounds check
if (base_src + head_dim > 128 * 1024) {
return;
}
// Apply WHT
threadgroup float rotated[128];
for (uint i = 0; i < head_dim; i++) {
rotated[i] = src[base_src + i];
}
// In-place WHT
for (uint h = 1; h < head_dim; h <<= 1) {
for (uint i = 0; i < head_dim; i += (h << 1)) {
for (uint j = i; j < i + h; j++) {
float x = rotated[j];
float y = rotated[j + h];
rotated[j] = x + y;
rotated[j + h] = x - y;
}
}
}
// Normalize WHT
float scale = 1.0 / sqrt(float(head_dim));
for (uint i = 0; i < head_dim; i++) {
rotated[i] *= scale;
}
// Calculate norm
float sum_sq = 0.0;
for (uint i = 0; i < head_dim; i++) {
sum_sq += rotated[i] * rotated[i];
}
float norm = sqrt(sum_sq);
norms[tid] = norm;
// Quantize and pack
float inv_norm = 1.0 / (norm + 1e-9);
for (uint i = 0; i < head_dim; i++) {
float val = rotated[i] * inv_norm;
uint idx = quantize_turbo4_metal(val);
if (i % 2 == 0) {
dst[base_dst + (i / 2)] = (uchar)idx;
} else {
dst[base_dst + (i / 2)] |= (uchar)(idx << 4);
}
}
}

View File

@@ -3,6 +3,8 @@
#include <vector>
#include <algorithm>
#include <iostream>
#include <cassert>
#include <cstring>
// Lloyd-Max Centroids for N(0, 1/d) where d=128
// These are precomputed for 4-bit (16 levels)
@@ -13,8 +15,19 @@ static const float turbo4_centroids[16] = {
0.1523f, 0.2154f, 0.2800f, 0.3500f // Approximate tail values
};
// Decision boundaries for binary search (precomputed)
// boundary[i] = (centroid[i] + centroid[i+1]) / 2
static const float turbo4_boundaries[15] = {
-0.18385f, -0.1322f, -0.09665f, -0.0683f,
-0.04375f, -0.0213f, 0.0f, 0.0213f,
0.04375f, 0.0683f, 0.09665f, 0.1322f,
0.18385f, 0.2477f, 0.315f
};
// Fast Walsh-Hadamard Transform (In-place)
void fwht(float* a, int n) {
assert(n > 0 && (n & (n - 1)) == 0 && "n must be power of 2");
for (int h = 1; h < n; h <<= 1) {
for (int i = 0; i < n; i += (h << 1)) {
for (int j = i; j < i + h; j++) {
@@ -32,31 +45,70 @@ void fwht(float* a, int n) {
}
}
// Binary search for Lloyd-Max quantization
static inline int quantize_turbo4(float val) {
// Binary search through decision boundaries
int left = 0, right = 14;
while (left < right) {
int mid = (left + right) / 2;
if (val < turbo4_boundaries[mid]) {
right = mid;
} else {
left = mid + 1;
}
}
return left;
}
// PolarQuant Encode (CPU Reference)
void polar_quant_encode_turbo4(const float* src, uint8_t* dst, float* norm, int d) {
std::vector<float> rotated(src, src + d);
fwht(rotated.data(), d);
assert(src != nullptr && "src cannot be null");
assert(dst != nullptr && "dst cannot be null");
assert(norm != nullptr && "norm cannot be null");
assert(d > 0 && (d & (d - 1)) == 0 && "d must be power of 2");
// Use stack allocation for small d (d=128 is 512 bytes)
float rotated[128]; // Stack allocation for d=128
if (d > 128) {
// Fallback to heap for larger d
std::vector<float> rotated_vec(src, src + d);
fwht(rotated_vec.data(), d);
// Calculate L2 Norm (Radius)
float sum_sq = 0;
for (int i = 0; i < d; i++) sum_sq += rotated_vec[i] * rotated_vec[i];
*norm = sqrtf(sum_sq);
// Quantize components
float inv_norm = 1.0f / (*norm + 1e-9f);
for (int i = 0; i < d; i++) {
float val = rotated_vec[i] * inv_norm;
int best_idx = quantize_turbo4(val);
// Pack 4-bit indices
if (i % 2 == 0) {
dst[i / 2] = (uint8_t)best_idx;
} else {
dst[i / 2] |= (uint8_t)(best_idx << 4);
}
}
return;
}
// Stack-allocated path for d=128
memcpy(rotated, src, d * sizeof(float));
fwht(rotated, d);
// Calculate L2 Norm (Radius)
float sum_sq = 0;
for (int i = 0; i < d; i++) sum_sq += rotated[i] * rotated[i];
*norm = sqrtf(sum_sq);
// Quantize components
float inv_norm = 1.0f / (*norm + 1e-9f);
for (int i = 0; i < d; i++) {
float val = rotated[i] * inv_norm;
// Simple nearest neighbor search in Lloyd-Max codebook
int best_idx = 0;
float min_dist = fabsf(val - turbo4_centroids[0]);
for (int j = 1; j < 16; j++) {
float dist = fabsf(val - turbo4_centroids[j]);
if (dist < min_dist) {
min_dist = dist;
best_idx = j;
}
}
int best_idx = quantize_turbo4(val);
// Pack 4-bit indices
if (i % 2 == 0) {
@@ -69,8 +121,13 @@ void polar_quant_encode_turbo4(const float* src, uint8_t* dst, float* norm, int
// PolarQuant Decode (CPU Reference)
void polar_quant_decode_turbo4(const uint8_t* src, float* dst, float norm, int d) {
assert(src != nullptr && "src cannot be null");
assert(dst != nullptr && "dst cannot be null");
assert(d > 0 && (d & (d - 1)) == 0 && "d must be power of 2");
for (int i = 0; i < d; i++) {
int idx = (i % 2 == 0) ? (src[i / 2] & 0x0F) : (src[i / 2] >> 4);
assert(idx >= 0 && idx < 16 && "Invalid index");
dst[i] = turbo4_centroids[idx] * norm;
}
// Inverse WHT is same as Forward WHT for orthogonal matrices

150
tests/test_turbo.cpp Normal file
View File

@@ -0,0 +1,150 @@
#include "llama-turbo.h"
#include <iostream>
#include <vector>
#include <cmath>
#include <cassert>
// Simple test for encode/decode round-trip
void test_roundtrip() {
const int d = 128;
std::vector<float> original(d);
std::vector<float> decoded(d);
std::vector<uint8_t> packed(d / 2);
float norm;
// Generate random test data
for (int i = 0; i < d; i++) {
original[i] = (float)rand() / RAND_MAX * 2.0f - 1.0f;
}
// Encode
polar_quant_encode_turbo4(original.data(), packed.data(), &norm, d);
// Decode
polar_quant_decode_turbo4(packed.data(), decoded.data(), norm, d);
// Check round-trip error
float max_error = 0.0f;
float avg_error = 0.0f;
for (int i = 0; i < d; i++) {
float error = fabsf(original[i] - decoded[i]);
max_error = fmaxf(max_error, error);
avg_error += error;
}
avg_error /= d;
std::cout << "Round-trip test:" << std::endl;
std::cout << " Max error: " << max_error << std::endl;
std::cout << " Avg error: " << avg_error << std::endl;
std::cout << " Norm: " << norm << std::endl;
// Check that error is reasonable (should be small due to quantization)
assert(max_error < 1.0f && "Round-trip error too large");
assert(avg_error < 0.5f && "Average error too large");
}
// Test with known values
void test_known_values() {
const int d = 128;
std::vector<float> zeros(d, 0.0f);
std::vector<float> ones(d, 1.0f);
std::vector<float> decoded(d);
std::vector<uint8_t> packed(d / 2);
float norm;
// Test zeros
polar_quant_encode_turbo4(zeros.data(), packed.data(), &norm, d);
polar_quant_decode_turbo4(packed.data(), decoded.data(), norm, d);
std::cout << "Zero test:" << std::endl;
std::cout << " Norm: " << norm << std::endl;
// Test ones
polar_quant_encode_turbo4(ones.data(), packed.data(), &norm, d);
polar_quant_decode_turbo4(packed.data(), decoded.data(), norm, d);
std::cout << "Ones test:" << std::endl;
std::cout << " Norm: " << norm << std::endl;
// Check that decoded values are approximately 1.0
float avg = 0.0f;
for (int i = 0; i < d; i++) {
avg += decoded[i];
}
avg /= d;
std::cout << " Average decoded value: " << avg << std::endl;
assert(fabsf(avg - 1.0f) < 0.5f && "Decoded average should be close to 1.0");
}
// Test edge cases
void test_edge_cases() {
const int d = 128;
std::vector<float> large(d);
std::vector<float> small(d);
std::vector<float> decoded(d);
std::vector<uint8_t> packed(d / 2);
float norm;
// Test large values
for (int i = 0; i < d; i++) {
large[i] = 1000.0f;
}
polar_quant_encode_turbo4(large.data(), packed.data(), &norm, d);
polar_quant_decode_turbo4(packed.data(), decoded.data(), norm, d);
std::cout << "Large values test:" << std::endl;
std::cout << " Norm: " << norm << std::endl;
// Test small values
for (int i = 0; i < d; i++) {
small[i] = 0.001f;
}
polar_quant_encode_turbo4(small.data(), packed.data(), &norm, d);
polar_quant_decode_turbo4(packed.data(), decoded.data(), norm, d);
std::cout << "Small values test:" << std::endl;
std::cout << " Norm: " << norm << std::endl;
}
// Test error handling
void test_error_handling() {
const int d = 128;
std::vector<float> data(d, 1.0f);
std::vector<uint8_t> packed(d / 2);
std::vector<float> decoded(d);
float norm;
// Test with null pointers (should assert in debug mode)
std::cout << "Error handling tests:" << std::endl;
std::cout << " Note: These should trigger assertions in debug mode" << std::endl;
// Uncomment to test assertions:
// polar_quant_encode_turbo4(nullptr, packed.data(), &norm, d);
// polar_quant_encode_turbo4(data.data(), nullptr, &norm, d);
// polar_quant_encode_turbo4(data.data(), packed.data(), nullptr, d);
// Test with invalid d (not power of 2)
// polar_quant_encode_turbo4(data.data(), packed.data(), &norm, 127);
}
int main() {
std::cout << "TurboQuant Unit Tests" << std::endl;
std::cout << "====================" << std::endl;
try {
test_roundtrip();
test_known_values();
test_edge_cases();
test_error_handling();
std::cout << std::endl;
std::cout << "All tests passed!" << std::endl;
return 0;
} catch (const std::exception& e) {
std::cerr << "Test failed: " << e.what() << std::endl;
return 1;
}
}