I am an ML researcher and software engineer with 8+ years of experience building production AI systems and researching trustworthy, auditable AI. Previously a Graduate Research Assistant at the Institute of Artificial Intelligence at the University of Georgia, where I completed an MS thesis on Interpretable AI for Image Classification under Dr. Jaewoo Lee.
Most recently, I was a Software Development Engineer at Amazon (Austin, TX), where I worked on payment and identity systems and prototyped LLM-based root cause analysis pipelines for incident remediation — work that directly motivated my current research.
I am now an ML Engineer on Omdena innovation challenges, building multi-agent AI systems and evaluation infrastructure. My active research focuses on a question that matters in production: how do you evaluate whether an LLM-generated remediation playbook is operationally safe — not just fluent — in environments where hallucinated recommendations cause real incidents?
My research interests sit at the intersection of LLM evaluation, agentic systems, and explainable AI — with a consistent focus on making AI decisions auditable, attributable, and trustworthy.
LLM Evaluation for AI-Driven Incident Remediation (Active)
Developing a multi-dimensional evaluation framework using RAG with must-cite decoding constraints, measuring grounding fidelity and source attributability beyond BLEU/ROUGE. The core question: can we evaluate operational safety, not just fluency, in LLM-generated playbooks?
TPNet — Interpretable Image Classification (MS Thesis, UGA 2021)
Designed TPNet, a transformer-based image classifier using learned prototypes that generate human-readable explanations for every prediction — enabling auditable AI decisions in medical imaging. The interpretability mechanism is the point: every classification is traceable to a human-verifiable prototype.
AI-Powered STEM Autograding
Built ML text classifiers automating 100% of STEM assignment grading, deployed as a Django web application serving an educational research lab. (Adviser: Dr. Xiaoming Zhai)
A. Joshi. Interpretable Image Classification. University of Georgia ProQuest Dissertations & Theses, 2021. [UGA Repository]
A. Joshi, M. Kulkarni. Analysis of Fractal Image Compression Using Quadtree Partitioning. Bulletin of Marine Science & Technology, Vol. 10, 2015.
Technical Paper Reviewer — ACL Workshop on Computational Methods for Endangered Languages (ComputEL), 2024.
Designed a LangGraph-based multi-agent system automating carbon emissions tracking and EU CSRD/US SEC-aligned compliance reporting for SMEs. Led 3 sub-teams coordinating agents for data extraction, calculation, and regulatory strategy.
Designed a two-phase schema validation harness integrated into CI that reduced noise alerts by 28% and triage time by 19% across LLM, agent, and CV pipelines. Established versioned JSON logging contracts for reproducible evaluation.
Transformer-based prototype network providing human-readable explanations for every classification decision. Designed for auditability in high-stakes medical imaging.
Django application automating 100% of assignment grading for an educational research lab, reducing manual effort by 90%.
A few of my paintings.
I have a passion for sports and fitness. I hold a Black Belt (Dan) in Muay Thai (2018) and have been strength training since 2017, currently training CrossFit. Staying active helps me build the discipline and focus that carry into research.