Looking for an arXiv endorser in cs.LG: Endorsement link: https://arxiv.org/auth/endorse?x=PWEZJ7 Endorsement link 2: http://arxiv.org/auth/endorse.php Endorsement code: PWEZJ7 Paper: https://zenodo.org/records/18805367 Code: https://github.com/panindratg/RD-Sphota RD-SPHOTA is a character-level language model using reaction-diffusion dynamics instead of attention or gating, with architecture derived from Bhartrhari’s sphota theory and Dharmakirti’s epistemology, mapped to computational operations and validated through ablation, not used as metaphor. The dual-channel architecture independently resembles the U/V decomposition in Turing’s unpublished 1953-1954 manuscripts. A 7th century Indian epistemologist and a 20th century British mathematician arriving at the same multi-scale structure through completely different routes. Results on Penn Treebank (215K parameters): 1.493 BPC vs LSTM 1.647 (9.3% improvement) 1.493 BPC vs GRU 1.681 (11.2% improvement) Worst RD-SPHOTA seed beats best baseline seed across all initialisations Three philosophical components failed ablation and were removed. The methodology is falsifiabl submitted by /u/panindratg276
Originally posted by u/panindratg276 on r/ArtificialInteligence
