// welcome to my portfolio
Work Experience Education Research Projects Skills Certifications Volunteering Contact
Open to Opportunities

Karthik
Mulugu

AI Engineer · Machine Learning · Data Scientist

karthik@ai-engineer ~
cat about.txt
Karthik Mulugu
scroll

Work Experience

AI Engineer Intern
Pronix Inc
● Present Remote · Oct 2025
Remote · Oct 2025 – Present
  • Boosted retrieval accuracy by 35% by building RAG pipelines with Azure/OpenAI LLMs using embedding-based and hybrid retrieval techniques, enabling more context-aware chatbot responses.
  • Reduced HR support workload by 40% by designing and deploying chatbot workflows on Kore.ai Agentic Platform, integrating Workday APIs for real-time employee profile and leave management.
  • Increased chatbot task success rate by 25% and reduced fallback responses by 30% by optimizing intent classification, dialog state management, and workflow orchestration.
  • Eliminated critical edge-case failures by designing and executing scenario-based validation strategies within Salesforce for JBL, improving chatbot response consistency across high-priority workflows.
RAGNLPOpenAIAzureWorkday APIKore.aiGenerative AI
AI Automation Extern
Extern
Top Performer Remote · Jun–Sep 2025
Remote · June 2025 – September 2025
  • Increased mortgage document-processing efficiency by 70% by building scalable Python automation pipelines to parse, chunk, and structure 1,000+ financial documents, reducing manual review time from hours to minutes.
  • Improved OCR accuracy by 35% and retrieval precision by 40% by optimizing Tesseract/EasyOCR and architecting a RAG pipeline using LangChain, LlamaIndex, and FAISS for high-precision document search and QA.
  • Boosted answer accuracy by 28% and reduced fallback responses by 35% by benchmarking and optimizing Gemini vs. Mistral models — earning recognition as Top Performer for technical impact.
PythonOCRRAGLlamaIndexFAISSGradioLLMs

Beyond work

ISS Volunteer
University at Buffalo · Aug 2024
Welcomed international grad students during Fall 2024 orientation, guiding them on campus resources and community integration.
Hackathon Coordinator
HackFiesta — TechnoMist 2K23 · Hyderabad
Organised and coordinated a large-scale hackathon with multiple teams, mentors, and event logistics end-to-end.

Education

UB
M.S. Computer Science — AI/ML 3.7 GPA
University at Buffalo, SUNY
Jan 2024 – Jun 2025 · Buffalo, NY

Specialised in ML, Deep Learning, Computer Vision, and Data Visualization — maintaining a 3.7 GPA while concurrently building production AI systems across NLP, RAG, and forecasting domains.

JNTU
B.Tech. Information Technology
Jawaharlal Nehru Technological University, Hyderabad
Aug 2019 – Jul 2023 · Hyderabad, India

Built solid foundations in DSA, DBMS, OOP, and systems programming. Ranked in the top 1% of the IT department for academic performance across core engineering coursework.

Research

Submitted · IEEE Access · Apr 2026
DocuQuery: Hybrid Lexical-Dense Retrieval with LangGraph Orchestration for Robust PDF Question Answering

An open-source system for natural-language question answering over PDF documents. Combines hybrid retrieval (BM25, FAISS, TF-IDF) with LangGraph multi-intent orchestration — routing queries across summarisation, comparison, refinement, and QA modes — backed by Gemini generation. A key finding reveals that fixed-weight hybrid fusion can degrade below pure lexical retrieval when dense similarity is misleading, motivating query-aware fusion gating.

RAG Hybrid Retrieval LangGraph BM25 + FAISS Gemini NLP PDF QA

Projects

↑ 7 projects — click a tab to explore

Customer Churn Prediction System
End-to-end ML pipeline that predicts at-risk telecom customers, explains the drivers behind each prediction, and surfaces actionable retention strategies — designed for direct business use.
  • Trained XGBoost on 7,000+ telecom records with 12 SQL-engineered features; achieved 88% recall on the minority (churn) class — the metric that matters most for retention campaigns
  • Cohort SQL analysis pinpointed that high-charge month-to-month customers drive 63% of churn, directly informing which segment to target first with discount offers
  • Built a 3-view Tableau dashboard (risk heatmap, churn drivers, segment breakdown) that converts raw model output into one-page executive summaries
  • Deployed Streamlit app with a live decision-threshold slider and per-customer SHAP waterfall charts — lets non-technical stakeholders explore predictions without touching code
XGBoostSHAPSQLStreamlitTableau
Live app may take ~30s to wake
DocuQuery AI Assistant
Production-ready RAG system for natural-language QA over arbitrary PDF documents. Subject of peer-reviewed research submitted to IEEE Access.
  • Reduced document analysis time by 75% by building and deploying a production RAG system using LangGraph, Gemini, and FAISS/Pinecone with low-latency inference of 1.5–4.0s
  • Improved retrieval precision by 30% by architecting a hybrid search pipeline combining semantic search, BM25, TF-IDF, and cross-encoder reranking — the core contribution of the IEEE Access paper

→ full benchmark results in the Research section

LangGraphGeminiFAISSPineconeRAG
Live app may take ~30s to wake
Amazon Stock Forecasting Dashboard
Multi-model forecasting dashboard comparing Linear Regression, LSTM, and Seq2Seq architectures — enriched with technical indicators, sentiment signals, and deployed on AWS EC2 for production serving.
  • Reduced stock price forecasting error by 49% (MAPE: 17.45% → 8.86%) by training and benchmarking Seq2Seq LSTM, LSTM, and Linear Regression models enhanced with RSI, MACD, and Bollinger Bands
  • Delivered actionable insights across 6,300+ Amazon historical records (2000–2025) with an interactive dashboard integrating sentiment analysis and backtesting for data-driven buy/sell evaluation
  • Deployed a production-grade pipeline on AWS EC2 — optimizing CPU-based PyTorch inference, enforcing secure network policies, and automating execution with systemd for reliable model serving
LSTMSeq2SeqStreamlitPython
Live app may take ~30s to wake
Human Action Recognition — CNN Models
Systematic benchmark comparing four CNN architectures — two built from scratch, two via transfer learning — for classifying 40 distinct human actions across 9,500+ images.
  • Implemented VGG16 and ResNet50 from scratch in TensorFlow/Keras; then applied transfer learning with GoogLeNet and DenseNet to measure the accuracy-vs-compute tradeoff across all four
  • ResNet50 with transfer learning achieved the best accuracy, outperforming the from-scratch VGG baseline by a significant margin on Stanford 40 Actions dataset
  • Reached 80% real-time inference accuracy, with Streamlit app allowing image upload and live prediction — demonstrating viability for surveillance and fitness-tracking applications
TensorFlowResNet50Transfer LearningStreamlit
Text Summarization — BART Transformer
Fine-tuned BART transformer for abstractive summarization — produces fluent, condensed summaries of long articles rather than just extracting sentences verbatim.
  • Fine-tuned facebook/bart-large-cnn on the CNN/DailyMail benchmark (300K+ article-summary pairs); achieved ROUGE-L of 0.41 — competitive with published abstractive baselines
  • Abstractive output generates novel phrasings rather than copying source sentences — verified via n-gram novelty analysis against the reference summaries
  • Streamlit interface accepts any URL or pasted text and returns a summary in under 3 seconds, validated across news, research abstracts, and legal documents
PyTorchBARTHugging FaceStreamlit
Hospital Inpatient Cost Prediction
End-to-end ML system predicting inpatient costs from 2.5M NY SPARCS discharge records — built to surface key clinical cost drivers with explainability, uncertainty quantification, and full production deployment.
  • Engineered a scikit-learn ColumnTransformer pipeline (imputation, scaling, encoding) with strict train/val/test isolation — eliminated 100% of null values and prevented data leakage across all 5 model families.
  • Benchmarked Ridge, Random Forest, XGBoost, LightGBM, and PyTorch MLP with Optuna Bayesian tuning (30 trials/model); XGBoost achieved R2 = 0.969, RMSE = $2,035, MAE = $1,622 — 58% RMSE reduction over baseline.
  • SHAP TreeExplainer identified Length of Stay, APR Severity Code, and Medical/Surgical pathway as the top three cost drivers; quantile regression (Q10/Q90) produced calibrated 80% prediction intervals.
  • Deployed FastAPI REST API + 6-tab Streamlit dashboard (Predict, What-If, Batch, EDA, History) via Docker Compose; experiments tracked in MLflow.
PythonXGBoostSHAPFastAPIStreamlitMLflow
Live app may take ~30s to wake
Sentiment Analysis — DistilBERT + BiLSTM
Comparative study of recurrent and transformer architectures for 3-class news headline sentiment — fine-tuning DistilBERT against a custom BiLSTM+Attention baseline on 204K HuffPost articles, measuring what each architectural decision actually buys.
  • Fine-tuned distilbert-base-uncased on 204K HuffPost headlines with AdamW + linear warmup; converged at epoch 2 reaching 80.4% accuracy and 0.919 ROC-AUC — outperforming the BiLSTM baseline by 4.4 points
  • Traced the BiLSTM's 76% ceiling to noisy category-based pseudo-labels: Business, Science, and Tech span the full sentiment spectrum, creating a data quality ceiling no architecture alone could overcome
  • Built a live RSS pipeline fetching real-time headlines from 6 outlets (BBC, CNN, NPR), running batch sentiment prediction, and rendering hour-by-hour trend charts per source
  • Deployed end-to-end: FastAPI inference backend, 4-tab Streamlit dashboard on Community Cloud, Docker for local deployment, MLflow experiment tracking, and GitHub Actions CI
PyTorchNLPDeep LearningFastAPIStreamlitDocker
Live app may take ~30s to wake

Technical Skills

Programming
PythonSQLJavaC
Machine Learning
Supervised LearningUnsupervised LearningFeature EngineeringModel OptimizationCross-ValidationHyperparameter TuningEDAModel Evaluation
Deep Learning
CNNsLSTMsTransformersRNNsTransfer LearningFine-TuningHugging Face
Generative AI & LLMs
RAG PipelinesLangChainLlamaIndexAgentic AIPrompt EngineeringVector DatabasesSemantic SearchLLM Fine-tuningFAISSPineconeEmbedding Models
Libraries
PyTorchScikit-learnPandasNumPyTensorFlowKerasSciPyStreamlitGradio
Data & Platforms
ETL/ELT PipelinesSnowflakePostgreSQLMySQLVertex AIAzure ML StudioSageMakerDatabricksBigQuery
MLOps & Deployment
MLflowDockerFastAPICI/CDExperiment TrackingInference Optimization
Cloud
AWS (EC2, S3, SageMaker)GCP
Visualization
TableauPower BIMatplotlibSeaborn
Developer Tools
GitGitHubVS CodeJupyterCursor
currently learning
Terraform
Infrastructure as Code
Databricks
Data Analytics Platform
BigQuery
Data Warehouse

Let's Build Something

Have a project in mind, a role to fill, or just want to talk AI?
My inbox is always open — I'll get back to you within 24 hours.

karthikmulugu14@gmail.com
send_message.sh