ONLINE LOC: NEW DELHI ROLE: AI_ENGINEER VIEWS: traffic

AMAN
MISHRA

Hands-on with LLM systems, RAG pipelines, and production AI infrastructure. Strong foundation in ML, DL, and GenAI — engineering first, always.

EXPERIENCE
AUG 2025 — JAN 2026 // REMOTE
Observo Tech Studio
AI Engineer — Evntro

• Contributed to a production-grade agentic B2B outreach system integrating email sending, receiving, replying, and campaign-level tracking across multiple mailboxes.
• Designed and implemented the LLM tool layer (send/track/reply) with persistent conversation state in PostgreSQL, enabling autonomous, multi-threaded communication workflows.
• Engineered a real-time feedback pipeline capturing 16+ delivery and engagement signals, enabling adaptive, data-driven campaign optimization.
• Developed a context-aware RAG pipeline leveraging prospect data for personalized email generation, improving relevance over template-based outreach.
• Refactored the FastAPI backend to resolve critical issues and deployed services on Azure Function Apps, enabling a scalable, event-driven system.

APR 2025 — JUL 2025 // REMOTE
Confedo AI
GenAI Engineer (R&D)

• Built multi-stage RAG pipelines and agentic workflows to simulate real-world LLM usage and enable systematic evaluation of faithfulness, relevance, and failure modes in complex pipelines.
• Designed advanced retrieval pipelines (query routing, HyDE, reranking) to generate challenging, multistep queries, improving robustness and realism of evaluation benchmarks.
• Reverse engineered 3 LLM observability frameworks to understand tracing architectures and extract design insights for building in-house evaluation and observability tooling.
• Built a PoC for an in-house LLM tracing system to validate design decisions and inform the product team in developing a production-grade tracing SDK.

JAN 2025 — JUN 2025 // ONSITE
USG Lab @ IIIT Delhi
Research Intern

• Designed and validated a data-free, zero-shot machine unlearning framework to enable class-level data removal without requiring access to original training data, supporting privacy compliance.
• Leveraged contrastive model inversion to generate 5K synthetic samples, enabling effective targeted unlearning while preserving model performance on unaffected classes.
• Achieved 90%+ unlearning efficacy across 8 classes on the SVHN dataset, while maintaining 95%+ accuracy on retained classes, minimizing unintended knowledge loss.
• Improved unlearning efficiency with a 77% reduction in runtime compared to the GKT baseline, enhancing feasibility for real-world deployment.
• Conducted comparative evaluation against 4 baseline methods (EMMN, GKT, zMuGAN, retraining) and explored subset/coreset strategies to optimize scalability and effectiveness of unlearning.

STACK

AI / ML / DL

PyTorch Transformers PEFT scikit-learn

LLM Systems

LangChain LangGraph LangSmith RAG

Backend & Data

FastAPI PostgreSQL MongoDB Supabase Pinecone FAISS ChromaDB

Cloud & Infra

Azure Functions Azure Logic Apps Blob Storage Docker Git

Tracking

Weights & Biases
EDUCATION
2023 — 2025

MSc Data Science

DIAT, Pune
2018 — 2021

BSc Mathematics

IGNOU
PROJECTS VIEW ALL →

Enterprise-grade Multi-Tiered Retrieval-Augmented Generation System. Integrates dynamic fallback strategies and diverse retrieval architectures to ensure high-fidelity context extraction across complex datasets.

RAG ENTERPRISE SYSTEM ARCHITECTURE

Custom-built, local-only LLM tracing and observability tooling. Engineered for async agentic workflows using contextvars to provide granular telemetry, span tracking, and metric evaluation with minimal overhead.

LLMOPS OBSERVABILITY TRACING

Multi-path autonomous RAG pipeline featuring intelligent query routing, self-corrective retrieval loops, and iterative web search fallbacks to actively detect and eliminate hallucinations in real time.

AGENTIC RAG ORCHESTRATION LLM AGENTS

On-device RAG engineered around LEANN's graph-based search. Indexes millions of documents locally in gigabytes without significant accuracy loss, achieving zero cloud dependency and strict privacy.

ON-DEVICE RAG GRAPH SEARCH PRIVACY

Structure-aware documentation RAG pipeline built from scratch. Crawls 104 FastAPI pages, preserves section hierarchy, generates 971 metadata-rich chunks, and delivers citation-grounded answers via Pinecone + Groq.

RAG NLP LLM

RAG system with SHA-256 document deduplication, namespace isolation, semantic reranking, and inline-cited answers. Upload a 200-page manual, ask a question, get a precise cited response.

RAG DEDUPLICATION VECTOR SEARCH