Skip to content
Anmol Kumar
Open to AI Engineer / GenAI roles - anywhere in India or worldwide

Anmol Kumar

AI Engineer & GenAI Specialist

I build AI systems that have to work on day one - RAG frameworks, multi-agent workflows, and conversational copilots built for global enterprises.

Bengaluru, IndiaAsia/Kolkata (IST, UTC+5:30)
Years shipping production AI & backend
4+

Enterprise products used by Fortune 500 customers.

Product teams using the RAG framework
4+

New AI POCs went from ~3 weeks of scaffolding to under 5 days.

Cut in SME investigation time
60-70%

For incident classes handled by the multi-agent RCA system.

Product roadmap decisions influenced
3

Driven by AI analytics POCs on IMS, Db2, and z/OS telemetry.

01 · About

Building AI systems that hold up in enterprise

Where applied GenAI meets the messy realities of enterprise-scale operations.

I'm a software engineer with 4+ years across backend and AI engineering. These days I spend most of my time on retrieval systems, multi-agent workflows, and LLM copilots that have to actually work in front of customers - right now at BMC Software.

Most of what I do lives in the gap between what GenAI can demo and what survives inside an air-gapped, observability-heavy enterprise. In practice that looks like RAG frameworks with confidence scoring and hallucination checks, agents that genuinely tool-call into live IMS / Db2 / z/OS, and fine-tuned sub-3B transformers for the customers where frontier APIs simply aren't allowed in the room.

Open to opportunities anywhere in India or anywhere in the world where I can keep building AI systems that earn their place in production.

RAG Systems

Confidence-scored, hallucination-aware retrieval - built for the kind of enterprise support workflow where a wrong answer has consequences.

Multi-Agent Workflows

Supervisor and specialist agents that actually call tools against live enterprise data, instead of roleplaying it from the context window.

Enterprise Copilots

Natural-language access to observability and performance data, so engineers ask questions instead of hunting through dashboards.

On-Prem LLMs

Fine-tuning sub-3B models for air-gapped customers whose data can't leave the building - the cases where frontier APIs aren't allowed in the room.

02 · Experience

Where I've been shipping

Four years across backend and AI engineering - with the last two focused entirely on production GenAI for enterprise-scale customers.

  1. Associate Product Developer · BMC Software

    Bengaluru, India

    Jul 2022 - Apr 2026

    I work on the AI layer inside BMC's mainframe products - the RAG framework, the multi-agent copilots, and the on-prem LLM experiments that ultimately land in front of Fortune 500 IMS, Db2, and z/OS customers.

    • Built a reusable RAG framework in Python with confidence scoring, hallucination detection, and context management that other teams could drop in. 4+ product teams picked it up, which took new AI POCs from ~3 weeks of scaffolding down to under 5 days and quietly became the internal pattern for retrieval-augmented features.
    • Designed multi-agent systems with tool-calling and live data access and wired them into existing enterprise web platforms. They automate root-cause analysis workflows that used to eat 2-4 hours of SME time per incident, cutting investigation time by an estimated 60-70%.
    • Shipped AI-driven analytics POCs on real-time data streams coming out of IMS, Db2, and z/OS. They fed directly into 3 product roadmap decisions and contributed to work recognized with the BMC IMS Impact Makers Award (Q1 FY26).
    • Built conversational copilots that pull from live enterprise data so support engineers can ask questions in English instead of hunting across dashboards - natural-language access to IMS and database performance observability.
    • Fine-tuned (SFT) small transformer models on modest hardware to show on-prem AI was real for air-gapped mainframe shops - and that sub-3B parameter models could handle the everyday domain-specific tasks without paying frontier-model inference costs.
    • Wrote internal research on agentic AI and RAG architectures that other mainframe, IMS, and database-performance teams picked up and started building on.
    • Python
    • FastAPI
    • LangGraph
    • LangChain
    • PyTorch
    • AWS SageMaker
    • Angular
    • TypeScript
    • IMS
    • Db2
    • z/OS
  2. Product Development Intern · BMC Software

    Bengaluru, India

    Jan 2022 - Jul 2022

    Automated the mainframe dataset lifecycle for the development team with Jenkins and Git so people stopped waiting on manual provisioning.

    • Built Jenkins CI/CD pipelines hooked into Git to handle creation, modification, and deletion of VSAM and GDG mainframe datasets. It replaced a manual, error-prone flow and took provisioning from ~30 minutes per request down to under 2 minutes.
    • Jenkins
    • Git
    • JCL
    • VSAM
    • GDG
    • z/OS
  3. Back End Developer (Intern) · Celebal Technologies

    Remote

    Jul 2021

    Shipped an NLP chatbot end-to-end during a 4-week internship - my first taste of delivering something real instead of a classroom exercise.

    • Built an NLP-powered conversational chatbot in Python and TensorFlow, covering the full loop from intent classification to response generation - and learning on the job what it means to ship something, not just prototype it.
    • Python
    • TensorFlow
    • NLP
03 · Projects

Selected work - production, open source, and research

The systems I've built that other teams now rely on. Enterprise client work is described in the detail the NDA allows; open-source work links to the repo.

Enterprise RAG Framework

Enterprise
2024 - 2026 · BMC Software

A reusable Python framework for retrieval-augmented AI features - confidence-scored, hallucination-aware, and context-managed so teams don't have to reinvent the plumbing.

  • Designed a pluggable retrieval core with ranked multi-source fusion, conversational context management, and per-claim confidence scoring - so the framework can say 'I don't actually know' instead of guessing.
  • A hallucination detection layer that flags weakly-grounded responses before they reach a user, which matters a lot when the user is a support engineer looking at a customer ticket.
  • Adopted across 4+ product teams inside BMC - which pulled new AI POCs from ~3 weeks of scaffolding down to under 5 days.
  • Quietly became the standard pattern for how retrieval-augmented features get built across the IMS, Db2, and z/OS product lines.
  • Python
  • FastAPI
  • LangChain
  • Vector DBs
  • OpenAI
  • Azure OpenAI

IMS AI Agent - Multi-Agent RCA

Enterprise
2024 - 2026 · BMC Software

A tool-calling multi-agent system that takes the first pass at root-cause analysis on IMS mainframe telemetry, so SMEs walk into a review, not a blank page.

  • A set of specialist agents - log triage, metric correlation, knowledge retrieval, narrative summarization - with a supervisor agent that routes work based on the shape of the incident.
  • The tool layer wraps live IMS, Db2, and z/OS data, so agents pull fresh telemetry at decision time instead of guessing from whatever was in the context window.
  • On supported incident classes, SME investigation time dropped by an estimated 60-70% - from 2-4 hours of manual digging to a few minutes of reviewing what the agents already surfaced.
  • Part of the work recognized with the BMC IMS Impact Makers Award (Q1 FY26).
  • Python
  • LangGraph
  • Tool-Calling
  • FastAPI
  • IMS
  • Observability

AI Log Analyzer

Enterprise
2023 - 2024 · BMC Software

Real-time log-stream analysis that turns mainframe chatter into anomalies and early-warning signals you can actually act on.

  • Pulls in real-time log streams from IMS, Db2, and z/OS and classifies events against a domain-specific taxonomy we keep inside the RAG framework.
  • Surfaces predictive signals - capacity cliffs, recurring subsystem issues, subtle regressions - rather than only after-the-fact alerts.
  • Fed directly into 3 product roadmap decisions by making previously invisible patterns legible to product managers and SMEs.
  • Python
  • Streaming
  • LLM Classification
  • Angular
  • TypeScript

Mainframe Support Copilot

Enterprise
2024 - 2025 · BMC Software

A conversational copilot that replaces manual dashboard-hopping with natural-language questions against live enterprise data.

  • Natural-language access to system observability data across IMS and database performance products - support engineers stop hunting across dashboards and just ask the copilot.
  • Live data retrieval through the agent tool layer, so answers reflect the system as it is right now rather than a stale snapshot.
  • Built on the internal RAG framework, with confidence scoring so the copilot can gracefully say 'I'm not sure' when grounding is weak - better than a confident wrong answer.
  • Python
  • FastAPI
  • Angular
  • LangChain
  • RAG

On-Prem SFT for Air-Gapped AI

Research
2024 · BMC Software

Supervised fine-tuning of sub-3B transformer models to show on-prem AI is realistic for air-gapped mainframe shops that can't touch hosted APIs.

  • Fine-tuned small transformer models on modest hardware, with domain-specific corpora drawn from IMS and Db2 documentation and real telemetry patterns.
  • Showed that sub-3B parameter models can handle the most common domain-specific tasks at a small fraction of what frontier models cost per inference.
  • Unblocked AI conversations with air-gapped enterprise customers who simply can't send data to hosted LLM APIs - suddenly there was a believable story for them.
  • PyTorch
  • Transformers
  • LoRA/SFT
  • AWS SageMaker

HridaySetu - AI Healthcare Platform

Open Source
2025 · Open Source

A unified, AI-powered healthcare platform I built to pull hospitals, clinics, labs, and patients into a single secure place - so medical records stop living in silos.

  • A document extraction pipeline that takes medical report PDFs and images through OCR and turns them into structured parameter tables.
  • A clinical analysis layer on AWS SageMaker-hosted Med42 (Llama3-8B) that generates severity-aware summaries and patient-friendly explanations on top of those parameters.
  • A conversational health assistant with role-isolated sessions for patient, doctor, and admin dashboards - and a context strategy that prioritizes structured parameters over raw OCR text inside the 2048-token window.
  • Trend tracking across reports with Recharts visualisations, so patients and doctors can see parameters move over time instead of reading isolated one-off results.
  • React 18
  • TypeScript
  • Vite
  • Tailwind
  • shadcn/ui
  • AWS SageMaker
  • AWS API Gateway
  • Med42

Also worth a look

  • Archestra - OSS Agent OrchestrationOpen Source
    Ongoing · Open Source
  • Early NLP ChatbotPersonal
    2021 · Celebal Technologies
04 · Tech Stack

The tools I reach for

Across the AI stack - from agent frameworks and retrieval pipelines down to the cloud and backend plumbing that makes it all ship.

AI & Agentic Frameworks

How I put agents together - orchestration, tool-calling, and copilot design.

  • LangGraph
  • LangChain
  • Agentic Workflows
  • Multi-Agent Orchestration
  • Tool-Calling
  • Copilot Design

LLM Ops & Engineering

Everything from retrieval pipelines to fine-tuning on hardware that isn't a data center.

  • RAG Pipelines
  • Self-RAG
  • Prompt Engineering
  • Fine-Tuning (SFT)
  • Hallucination Detection
  • Confidence Scoring
  • Chain-of-Thought
  • Transformer Architecture
  • Vector Databases
  • PyTorch
  • TensorFlow

Enterprise Systems

Where the AI work meets actual enterprise customers.

  • Mainframe / z/OS
  • BMC AMI
  • IBM IMS
  • IBM Db2
  • Log Analysis
  • System Observability

Cloud, Backend & MLOps

The platform and MLOps layer underneath the models in production.

  • Python
  • FastAPI
  • AWS Bedrock
  • AWS Bedrock AgentCore
  • AWS SageMaker
  • AWS API Gateway
  • AWS EC2
  • Docker
  • Kubernetes
  • Helm
  • Tilt
  • PostgreSQL
  • Terraform
  • GitHub Actions
  • Git

Frontend

The UI layer where copilots actually meet the user.

  • TypeScript
  • React
  • Angular
  • Next.js
  • Vite
  • Tailwind CSS
  • shadcn/ui
  • Recharts

CS Fundamentals

The parts that aren't exciting but quietly keep everything honest.

  • Algorithms
  • Data Structures
  • Software Engineering Fundamentals
05 · Open Source

Contributions beyond the day job

Open-source work where I bring lessons from production back out into the ecosystem.

Archestra

Contributor
View repo

Contributing patterns around multi-agent orchestration, tool-calling, and retrieval - shaped by the production lessons the BMC RAG framework taught me. The focus is on the unglamorous parts: deterministic tool contracts, grounded responses, and graceful degradation when retrieval is weak.

  • Agentic AI
  • Orchestration
  • Tool-Calling

HridaySetu

Author & Maintainer
View repo

A unified AI-powered healthcare platform for India that I designed and built - OCR-driven document extraction, SageMaker-hosted Med42 clinical reasoning, and a role-isolated conversational assistant for patients, doctors, and admins.

  • AWS SageMaker
  • Med42 / Llama3
  • Healthcare
  • OCR
06 · Awards & Education

Recognition and the paper trail

Awards

  • BMC IMS Impact Makers Award

    Q1 FY26

    BMC Software

    Recognized for AI work inside the IMS product line - multi-agent RCA, RAG-backed copilots, and AI-driven analytics on top of mainframe telemetry.

  • IBM ICE 2022 - Gold Medalist

    2022

    IBM

    Took first position in IBM ICE 2022 - a competition focused on cross-platform innovation and working knowledge of the IBM ecosystem.

Research

  • Agentic AI & RAG Architectures

    2024 - 2026

    BMC Software - Internal Research

    Authored internal research on agentic AI and RAG architectures that drove broader adoption across teams working on mainframe systems, IMS, and database performance analytics - established foundational patterns later reused in multiple product initiatives.

Education

  • UPES

    2018 - 2022

    B.Tech, Computer Science Engineering

    Dehradun, India

07 · Contact

Let's build something

Open to AI Engineer / GenAI roles anywhere in India or anywhere in the world. The fastest way to reach me is email - I reply within a day.

Whether you're shipping RAG into an enterprise product, standing up agentic workflows that actually work, or bringing GenAI into an air-gapped environment where frontier APIs aren't an option - I'd love to hear what you're working on.

Location
Bengaluru, India
Timezone
Asia/Kolkata (IST, UTC+5:30)
Status
Open to AI Engineer / GenAI roles - anywhere in India or worldwide