Understanding Transformers and Attention Mechanisms
A deep dive into the transformer architecture and self-attention mechanism that powers modern NLP models.
Master's student in AI and Machine Learning at Drexel University (GPA: 4.00). Currently working as a Research Assistant developing novel neural network initialization algorithms and optimizing large language models.
Developing novel initialization algorithms achieving 81% on MNIST with 20% data
Yearly continual pretraining on longitudinal PubMed corpus with embedding-space and concept-proximity analyses to quantify evolving biomedical knowledge
I'm an AI/ML Researcher and Engineer pursuing a Master's in AI and Machine Learning at Drexel University. My research focuses on neural network initialization and semantic shift analysis in biomedical NLP. Previous experience includes building enterprise RAG pipelines and deepfake systems at Ernst & Young, and backend infrastructure at Neuflo Solutions.
Drexel University, Philadelphia, PA
Expected May 2026 | Drexel CCI Merit Scholarship Recipient
Mahatma Gandhi University, India
Graduated 2022 | Best Outgoing Student
Drexel University
Drexel University
Neuflo Solutions Pvt Ltd
Ernst & Young
Zoho Technologies
Novel initialization algorithm using co-occurrence statistics and product quantization. Achieved 81% accuracy on MNIST with 20% data, 65% on CIFAR-10 without gradient descent.
Yearly continual pretraining on longitudinal PubMed corpus with embedding-space, neighborhood, and concept-proximity analyses to quantify evolving biomedical knowledge.
Chain-of-thought LLM agent leveraging OpenAI's function-calling to orchestrate Wikipedia API tools.
White-box (PGD) and black-box attacks on ResNet-18 and YOLOv5 to quantify vulnerabilities.
Fine-tuning BioBERT to detect semantic shift in medical text from PubMed (1975-2025).
COVID-19 news video sentiment analysis using LSTM, Bi-LSTM, CNN, GRU. Published at ICITA-2021.
A deep dive into the transformer architecture and self-attention mechanism that powers modern NLP models.
A beginner-friendly introduction to PyTorch, covering tensors, autograd, and building your first neural network.
Interested in AI/ML research collaboration, consulting, or just want to chat about machine learning? I'd love to hear from you.
Built with Astro, TailwindCSS, and DaisyUI.
•
© 2026 Milan Varghese. All rights reserved.