Building on HF
126.0 TFLOPS OzTianlu
28
followers ·
30 following AI & ML interests My research focuses on deep reasoning with small language models, Transformer architecture innovation, and knowledge distillation for efficient alignment and transfer.
Recent Activity reacted to reaperdoesntknow 's post with 👍 8 days ago We present a methodology for training small language models on CPU at FP32 precision
that achieves capability-per-dollar efficiency orders of magnitude beyond GPU-based training.
Across15modelsspanningfournovelarchitecturefamilies—MixtureofAttentions(MoA),cross-
architecture fusion (Qemma), swarm intelligence (SAGI), and metric-space causal language
models (DiscoverLM)—total compute cost was $24 on a single AMD EPYC 9454P proces-
sor. We introduce seven methodological pillars: (1) FP32 precision preservation, with exper-
iments demonstrating 5,810×single-operation error and 23,225×compounding error ratio for
FP16 at network depth; (2) sparse cognitive architectures where 0.02–7% of parameters activate
per token, matching CPU branching rather than GPU SIMD; (3) developmental curriculum
training progressing from language to logic to transfer to depth; (4) continuous belt-fed data
ingestion eliminating truncation waste; (5) hardware-native optimization for AMD Zen 4 via
AOCL/OpenMP/NUMA-aware allocation; (6) self-regulating thermodynamic governance with
emergent temperature measurement grounded in L2-star discrepancy; and (7) open-standard
compute (AVX2 SIMD at FP32) free of proprietary vendor dependency. We argue that transformers were designed for GPU hardware rather than mathematical optimality, and that architecture designed for geometric correctness—metric-space attention, triangle inequality enforcement, sparse expert routing—naturally favor CPU execution. For sub-2B parameter models, CPU training produces more capable models at a fraction of the cost. View all activity Organizations OzTianlu/A_Reasoning_Critique_of_Diffusion_Models Viewer
• Updated Dec 15, 2025 • 6 • 20
• 1
OzTianlu/Semigroup_Reasoning_Model_A_Scalpel Viewer
• Updated Dec 14, 2025 • 5 • 23
• 1
OzTianlu/Abstract_of_Structural_Critique_of_Reasoning Viewer
• Updated Dec 10, 2025 • 1 • 15
• 1
OzTianlu/Reasoning_and_Jacobian_Collapse Preview
• Updated Dec 9, 2025 • 53
• 1
OzTianlu/From_Reasoning_Structure_to_the_Ancient_Problem_of_Primes Viewer
• Updated Dec 3, 2025 • 5 • 44
• 1
OzTianlu/When_Euler_Meets_Stack Viewer
• Updated Nov 27, 2025 • 1 • 19
OzTianlu/Reasoning_as_Fluid Viewer
• Updated Nov 25, 2025 • 4 • 104
• 1
OzTianlu/The_Geometric_Incompleteness_of_Reasoning Viewer
• Updated Nov 24, 2025 • 3 • 56
OzTianlu/The_Incompleteness_of_Reasoning Viewer
• Updated Nov 23, 2025 • 1 • 20
• 1
OzTianlu/Why_Reasoning_Models_Collapse_Themselves_in_Reasoning Viewer
• Updated Nov 23, 2025 • 1 • 19
OzTianlu/Computational_Phase_Transitions_in_Reasoning Viewer
• Updated Nov 23, 2025 • 4 • 4
OzTianlu/Quantitative_Mapping_of_Computational_Boundaries Viewer
• Updated Nov 23, 2025 • 6 • 34