“In the beginning there was neither existence nor non-existence.” — Ṛgveda
You can Call me Cash. Stanford gave me a PhD in AI. For 13 years I have been working on Deep Learning. Language modeling has always been a big draw for me, be it Optical Character Recognition or text prediction for Sanskrit.
Current focus is on more advanced stuff, specialising in Transformer based Diffusion Models. An independent research building permutation-invariant DDPM/DDIM models for aperiodic chaotic tilings on fractal manifolds called Penrose Diffusion
Also training nano-GPT style models on Sanskrit text. Repo Sanskrit GPT
Implemented first-order methods from my PhD thesis on L₁-regularized infinite-dimensional convex neural networks. AnyBoost.jl – Julia package for boosting any loss, activation, and constraint.
Worked extensively on Universal script-agnostic OCR (CNN + LSTM + CTC). Just with a basic set of fonts and sample corpus data, we can build an complete segmentation free OCR system
- Chamantt OCR – Universal OCR for any language/script (Hindi, Arabic, Telugu, etc.)
- Banti OCR – End-to-end Telugu OCR framework. CNN-based, font-free system with n-gram language modelling, glyph stitching & erasure recovery. Extensible to other Indic scripts. (arXiv:1509.05962)
- Rnn CTC – One of the earliest GitHub RNN+CTC sequence learning libraries
- Theanet – Keras-style CNN library in Theano with full augmentations & regularizations. Built this before Keras was a thing!
- PhD Statistics, Stanford (advisor: Prof. Trevor Hastie)
- Thesis: Path algorithms for ℓ₁-regularized infinite-dimensional convex neural networks
- 1750+ hours classical yoga teacher training
Languages & Frameworks
Python PyTorch TensorFlow JAX/XLA Julia C++ CUDA
Languages
Spanish • Sanskit • Hindi • English • Telugu • Kannada • Tamil
Currently open to collaboration or roles at the frontier of generative AI research.
Reach me at my github id@gmail.com