Tanul Singh

Tanul Singh

ML Engineer · 5+ years in NLP & LLMs

# trained from scratch. no pre-trained weights.

Initialized with a Mechanical Engineering degree, pre-trained on curiosity, and fine-tuned by an unreasonable number of late nights — 5+ years of gradient descent through NLP, LLMs, and generative AI. Currently inference-serving at Apple. I write about ML here so you don't have to train from scratch too.

Senior ML Engineer

Apple — multi-agentic systems & LLM research

Kaggle Grandmaster

Notebooks GM, Competitions Master

US Patent Holder

Dynamic intent detection system

Self-Taught

ME degree → ML through sheer will

model.fit(life, epochs=∞, lr=persistence)

My Training Curve

Each milestone is a neuron. Experiences flow forward. Lessons backpropagate. The loss is still decreasing.

The SparkThe Hardest YearFound KaggleNotebooks GMMLE at LevelAISenior MLEPatent & PaperLead MLEAppleforward pass →← backprop (gradients)
forward pass (life moving forward)backprop (lessons learned)turning points

training_loss.plot()

epochs (time)1.00.0loss'17'18'19'20'21'22'23'24'25?
“You don’t need a low initial loss. You need a good learning rate and the patience to keep training.”

— the philosophy that took me from Mechanical Engineering to Apple

Want to know more?

Watch my conversations with people who shaped the Indian ML community: