Research & Benchmarks
Peer-reviewed research from our ML engineering team. Every claim is backed by reproducible benchmarks, open methodology, and production-validated results.
Featured Publications
March 2026
Beyond FP16: Achieving Near-Lossless 4-bit Inference on Commodity Mobile Hardware
K. Chen, A. Petrov, R. Tanaka et al.
We present a novel calibration-aware quantization pipeline that achieves sub-2% perplexity degradation on INT4-quantized transformer models running on Apple A17 Pro and Snapdragon 8 Gen 4. Our approach combines activation-aware weight quantization (AWQ) with recursive distillation feedback loops, enabling deployment of 7B-class intelligence on devices with <6GB unified memory.
February 2026
Recursive Distillation: Leveraging 405B Teachers for Sub-Billion Student Excellence
M. Okafor, S. Alvarez, J. Kim
We introduce Recursive Feedback Distillation (RFD), a multi-cycle teacher-student training paradigm where the student model is iteratively evaluated against a 405B-parameter oracle across 1,200+ domain-specific benchmarks. After 47 refinement cycles, our 1B-parameter student achieves 97.3% of the teacher's performance on medical NLU tasks while requiring 99.7% less compute at inference time.
January 2026
Ternary Transformers: The 1.58-bit Frontier for Tactical Edge Deployment
R. Tanaka, D. Hassan, K. Chen
We explore the extreme quantization frontier by applying ternary weight representations ({-1, 0, +1}) to transformer architectures optimized for tactical edge environments. Our 1.58-bit models achieve 89.4% of FP16 accuracy on classification tasks while enabling inference on FPGA-based hardware with power budgets under 5W — critical for disconnected, denied, and degraded (D3) operational contexts.
Logix-Refined vs. Stock Models
Side-by-side comparison on identical hardware (Apple M4 Pro, 18GB unified memory).
| Model | Variant | Size | Latency | Perplexity | MMLU |
|---|---|---|---|---|---|
| Llama 3.2-1B | Stock FP16 | 2.0 GB | 89ms | 9.12 | 46.2% |
| Llama 3.2-1B | Logix INT4 | 0.54 GB | 14ms | 9.31 | 45.8% |
| Phi-4-Mini | Stock FP16 | 7.6 GB | 210ms | 6.84 | 72.1% |
| Phi-4-Mini | Logix INT4 | 2.1 GB | 31ms | 6.97 | 71.4% |
| Gemma 2-2B | Stock FP16 | 5.0 GB | 156ms | 7.42 | 58.7% |
| Gemma 2-2B | Logix INT4 | 1.35 GB | 22ms | 7.61 | 57.9% |
| LeanLogix-7B-Med | Custom Ternary | 1.2 GB | 23ms | 5.94 | 78.3% |
| Mistral-7B | Stock FP16 | 14.5 GB | 340ms | 5.32 | 81.1% |
Hardware: Apple M4 Pro · 18GB Unified · macOS 15.3 · MLX v0.21
Updated: March 2026