3.5+ years engineering end-to-end LLM & ML systems. From RAG architectures to GPU inference optimization — I bridge the gap between AI research and production-ready infrastructure.
AI-driven anomaly detection engine for time-series system metrics using PCA, statistical analysis, and hypothesis testing. LLM-powered incident summarizer that automates ServiceNow workflows for real-time triage.
RAG system centered on Indian ancient texts (Bhagavad Gita) interpreted through Shankaracharya's Advaita Vedanta. Indian-only model stack — MuRIL embeddings + Sarvam-M LLM. Fine-tuned retrieval on scripture structure for philosophical fidelity.
MCP-driven AI/RAG recommendation system that maps user workloads to optimal AMD EPYC processor choices. End-to-end AI advisory flow combining LLM reasoning with production-ready deployment for cost-efficient resource utilization.
Complete audio pipeline — upload → transcription → live transcription → TTS feedback loop designed specifically for users with dysarthria. Cloud-native deployment with real-time processing.
Enterprise co-pilot integrated with Salesforce and Jira. Enhanced LangChain/LangGraph agentic workflows with serverless deployments via GitHub Actions and Azure, enabling automated business process orchestration.
Systematic benchmarking of Llama models on NVIDIA A100 and L40S GPU clusters using TensorRT and Triton Inference Server. Achieved significant throughput improvements and latency reduction at production scale.
Kalasalingam University
Great Lakes Institute of Management
Vivekananda College
Career Essentials in Generative AI — Microsoft/LinkedIn
Introduction to AI/ML Toolkits with Kubeflow (LFS147)
Open to AI/ML engineering roles, consulting engagements, and collaborative research. I specialize in turning complex ML research into production-grade systems.