Back to Home

All Projects

A complete archive of my open source work, research implementations, and system designs.

Production350k+ Downloads

Language Identification Model

83.5M-parameter model optimized for source-code classification. Fine-tuned across 25+ programming languages to deliver high-accuracy language detection at scale. Adopted by 350k+ developers worldwide.

PyTorchTransformersHugging FaceONNX
Research

Vision Transformer (ViT) From Scratch

A from-scratch PyTorch implementation of Vision Transformer with configurable hyperparameters and intuitive explanations of self-attention, patch embeddings, and training logic, focused on learning fundamentals and clean engineering. Documented end-to-end on Medium.

PyTorchResearch PaperhydraMLflow
Research

Transformer From Scratch

A clean PyTorch implementation of Transformer encoder and decoder from first principles, deeply exploring attention mechanics and causal masking. Guided code and narrative help understand how Transformer blocks actually work.

PyTorchPaper ReproductionSelf-AttentionTransformer
Systems

Distributed Training with PyTorch DDP

A practical deep dive into multi-GPU training using PyTorch DistributedDataParallel (DDP). Explains how gradient synchronization works under the hood, how to scale batch sizes correctly, and how to avoid common pitfalls when moving from single-GPU to distributed training.

PyTorch DDPCUDANCCLDistributed Systems
Education

Self-Attention Explained (From Intuition to Math)

A deep yet intuitive breakdown of self-attention, explaining how tokens interact, why attention works, and how queries, keys, and values emerge, using simple language, diagrams, and minimal math to build real understanding.

Self-AttentionTransformersMathPaper Understanding
Engineering

Knowledge Graph + LLM Chatbot

An RAG-powered conversational system that combines Neo4j knowledge graphs with large language models to answer natural language queries with structured and unstructured data. Includes Cypher integration and vector search strategies.

LLMsNeo4jVector SearchRAG
ML Systems

Efficient LLM Fine-Tuning

A step-by-step practical demonstrations on fine-tuning large language models using parameter-efficient techniques (e.g., LoRA), reducing training cost while maintaining strong performance, backed by hands-on notebooks.

LoRALLMsHugging FaceML OptimizationParameter Efficient Fine-Tuning
Research

Understanding LoRA: Theory Behind Efficient LLM Fine-Tuning

A theory-first breakdown of Low-Rank Adaptation (LoRA), derived from the original research paper. Explains the mathematical intuition behind low-rank updates, why full fine-tuning is inefficient, and how LoRA enables scalable adaptation of large models in modern ML systems.

LoRALLMsTransformersLinear AlgebraPEFTPaper Understanding
Foundations

Understanding Categorical Correlations

Explored how to measure relationships between categorical variables where standard correlation metrics fail. The project explains when and why the Chi-Square test is used to detect statistical dependence, and how Cramér’s V quantifies the strength of that relationship.

StatisticsChi-Square TestCramér’s VData Analysis
Foundations

Principal Component Analysis (PCA) Explained

A clear, intuition-driven breakdown of Principal Component Analysis (PCA), connecting the underlying linear algebra with practical data insights. Explains variance, eigenvectors, dimensionality reduction, and when PCA helps or hurtsreal-world machine learning pipelines.

PCALinear AlgebraStatisticsData Analysis
Applied ML

Video Frame Deduplication for Efficient Processing

Designed a lightweight computer vision pipeline to reduce long screen-recording and coding videos by removing redundant frames while preserving meaningful content. The approach targets scenarios where most frames show incremental changes, but only a few frames carry actual informational value.

Computer VisionFrame SimilarityVideo ProcessingComputer Vision AlgorithmsOpenCV
Foundations

Mathematics of Convolution & Deconvolution in Vision

A first-principles breakdown of convolution and deconvolution operations in computer vision. Explains the underlying mathematics and how these operations shape feature extraction and reconstruction in modern CNN-based systems.

Linear AlgebraComputer VisionCNNsDeep LearningMathematics