Home

Stephen Gurram - Senior AI/ML Engineer
[email protected]
Location: Overland Park, Kansas, USA
Relocation:
Visa: Green card
Resume file: Stephen Gurram_1776095081169.pdf
Please check the file(s) for viruses. Files are checked manually and then made available for download.
Client: KeyBank Pittsburgh, PA
Role: Senior AI/ML Engineer
Duration: May 2024 Present | USA
Designed highly scalable machine learning and deep learning models using PyTorch (2.x) and TensorFlow
(2.11+) for critical banking use cases including fraud detection, credit risk scoring, and transaction anomaly
detection, improving detection accuracy by over 30% and strengthening enterprise risk mitigation.
Architected and implemented end-to-end machine learning pipelines using Azure Machine Learning, Azure
Data Lake Storage (ADLS Gen2), Azure Functions, and Azure Data Factory, enabling automated
workflows for model training, validation, deployment, and orchestration, reducing deployment cycle time by
approximately 40%.
Built and deployed high-performance, real-time inference systems using FastAPI, Docker, and Azure
Kubernetes Service (AKS), supporting millions of financial transactions with low latency, high throughput,
and high availability.
Developed advanced NLP solutions using BERT, RoBERTa, and Hugging Face Transformers (v4.x) for
financial document classification, KYC/AML text analysis, and customer sentiment modeling, improving
operational efficiency and decision-making.
Designed and implemented Retrieval-Augmented Generation (RAG) pipelines using Azure OpenAI
Service, LangChain, and Azure AI Search, enabling secure, context-aware responses over enterprise
financial data and internal knowledge systems.
Built scalable vector search solutions using Azure AI Search and FAISS, enabling semantic document
retrieval for applications such as regulatory Q&A, policy search, and internal knowledge assistants.
Developed LLM-powered intelligent assistants for banking use cases including customer support
automation, document summarization, and compliance reporting, significantly reducing manual effort and
improving response accuracy.
Engineered end-to-end RAG workflows, including document ingestion, embedding generation, indexing,
retrieval, and prompt orchestration, ensuring accurate and grounded LLM outputs.
Implemented prompt engineering and response optimization techniques, reducing hallucinations and
improving factual accuracy of LLM-generated outputs in regulated financial environments.
Integrated security, governance, and access control mechanisms within AI systems, ensuring compliance
with PCI-DSS, GDPR, and internal banking regulations.
Implemented enterprise-grade MLOps pipelines using MLflow (2.x), Azure Machine Learning,
Kubernetes (AKS), and Jenkins/Azure DevOps, enabling automated CI/CD, model versioning,
monitoring, and lifecycle management.
Engineered scalable data processing pipelines using Apache Spark (3.x) on Azure Databricks and
Snowflake, optimizing feature engineering workflows and reducing data processing time by approximately
35%.
Designed and deployed model monitoring and drift detection frameworks using Azure Monitor, Application
Insights, Prometheus, and custom Python-based solutions to ensure model reliability and performance.
Built and optimized recommendation systems using collaborative filtering and deep learning techniques,
increasing customer engagement and product adoption rates by approximately 25%.
Applied advanced statistical modeling, hypothesis testing, and A/B testing methodologies, ensuring
measurable, data-driven business impact across banking use cases.
Performed extensive hyperparameter optimization using Grid Search, Random Search, and Bayesian
Optimization, improving model precision, recall, and overall predictive performance.
Designed and implemented feature stores (Azure ML / Databricks Feature Store) and reusable ML
components, improving model reusability, consistency, and accelerating development cycles.
Collaborated with cross-functional teams including data engineering, risk analytics, product, and business
stakeholders to deliver scalable AI/ML solutions aligned with enterprise goals.
Integrated CI/CD pipelines using Git, Docker, Jenkins, and Azure DevOps, enabling faster, reliable, and
auditable deployment of machine learning models.
Supported and maintained mission-critical 24/7 production ML systems, ensuring high availability, rapid
incident response, and minimal downtime in high-stakes financial environments.
Client: CVS Health- Atlanta, GA
Role: AI/ML Engineer
Duration: Jan 2022 Apr 2024 | USA
Developed predictive machine learning models using Scikit-learn (1.x) and TensorFlow (2.10+) for
healthcare use cases such as patient risk stratification, hospital readmission prediction, and claims
analytics, improving prediction accuracy by approximately 28% and enabling proactive clinical decision
making.
Designed and implemented scalable, production-grade data pipelines using Azure Data Factory, Azure
Databricks, and Azure Data Lake Storage (ADLS), enabling efficient ingestion, validation, transformation,
and orchestration of large-scale healthcare datasets for machine learning workflows.
Built and optimized recommendation systems for personalized healthcare interventions and treatment
suggestions using collaborative filtering and deep learning techniques, improving patient engagement and
care outcomes.
Engineered and deployed end-to-end machine learning systems using Docker, Azure Kubernetes Service
(AKS), and Azure Machine Learning, ensuring high availability, fault tolerance, and scalability for real
time healthcare applications.
Performed extensive feature engineering and dataset construction using Pandas, NumPy, and SQL,
including handling missing clinical data, normalization, encoding, and feature selection to improve model
robustness and predictive performance.
Developed and exposed RESTful APIs using Flask and FastAPI, deployed via Azure App Service and
containerized environments, enabling low-latency, real-time inference for healthcare analytics systems.
Utilized MLflow (2.x) integrated with Azure Machine Learning for experiment tracking, model versioning,
and lifecycle management, ensuring reproducibility, auditability, and governance across ML systems.
Leveraged Apache Spark (3.x) on Azure Databricks for distributed data processing and large-scale model
training, reducing training time by approximately 35% and improving system efficiency.
Applied advanced hyperparameter optimization techniques, including Grid Search, Random Search, and
Bayesian Optimization, improving model generalization and performance across diverse healthcare datasets.
Developed NLP-based solutions using spaCy (3.x) and Hugging Face Transformers, enabling automated
clinical text classification, medical document summarization, and patient feedback analysis.
Implemented robust model monitoring, logging, and alerting systems using Azure Monitor, Application
Insights, and custom Python-based solutions, detecting performance degradation, data drift, and anomalies
in production environments.
Collaborated with cross-functional teams including data engineering, clinical analysts, and product
stakeholders, translating healthcare requirements into scalable AI/ML solutions aligned with organizational
goals.
Built interactive dashboards and reporting tools using Power BI and Tableau, enabling visualization of model
outputs, patient risk scores, and key healthcare KPIs.
Ensured compliance with healthcare data regulations such as HIPAA, implementing secure data handling,
access control, and audit-ready ML systems.
Automated end-to-end machine learning workflows using Azure Machine Learning Pipelines, Azure
DevOps, Jenkins, and Git, improving deployment efficiency, reliability, and reducing manual intervention.
Client: GEICO Lakeland, FL
Role: Machine Learning Engineer
Duration: Sep 2019 Dec 2021 | USA
Developed machine learning models for insurance claim fraud detection using Python (3.7/3.8) and Scikit
learn, improving fraud detection accuracy by over 25% and reducing financial losses across high-volume
claims processing systems.
Designed and implemented scalable data processing pipelines using Apache Spark (2.x/early 3.x) and SQL,
integrated with Azure Data Lake Storage (ADLS) for efficient storage and processing of large-scale
insurance datasets, reducing data preparation time by approximately 30%.
Built predictive models for claims severity estimation and policy risk scoring using TensorFlow (2.x) and
Keras, supporting underwriting strategies and improving loss ratio performance.
Developed NLP-based solutions using NLTK and early spaCy, enabling automated extraction of key
information from claim reports, adjuster notes, and customer communications, improving operational
efficiency.
Engineered and deployed production-ready ML services using Flask-based REST APIs and Docker, hosted
on Azure Virtual Machines and Azure App Service, enabling near real-time inference and integration with
internal systems.
Utilized Azure cloud services (Azure Blob Storage, Virtual Machines, and Azure Functions) for data
storage, batch processing, and lightweight model deployment workflows in a scalable cloud environment.
Performed comprehensive feature engineering and data preprocessing using Pandas, NumPy, and SQL,
including handling missing values, encoding categorical variables, normalization, and outlier detection to
improve model performance.
Applied hyperparameter tuning techniques such as Grid Search and Random Search, improving model
precision, recall, and F1-score for fraud detection and risk modeling tasks.
Implemented basic model monitoring and logging mechanisms using Azure Monitor and custom logging
solutions, tracking model performance and detecting early signs of data drift.
Collaborated closely with data engineering teams to integrate ML workflows into existing ETL pipelines,
leveraging Azure Data Factory for data ingestion and transformation.
Built dashboards and reporting solutions using Power BI and Tableau, enabling stakeholders to monitor
fraud trends, claims insights, and key performance metrics.
Ensured compliance with insurance regulations and data privacy standards, maintaining secure handling
of sensitive policyholder and claims data.
Partnered with business analysts and underwriting teams to translate insurance requirements into scalable ML
solutions aligned with business goals.
Automated ML workflows including data ingestion, model training, validation, and deployment using
Python scripting and Jenkins-based CI/CD pipelines, improving operational efficiency.
Supported and maintained production ML systems in a 24/7 environment, ensuring high availability, rapid
issue resolution, and minimal disruption to insurance operations.
Client: Verizon Sandy, UT
Role: Data Scientist / ML Engineer
Duration: Jul 2017 Aug 2019 | USA
Developed machine learning models using Python (3.6/3.7), Scikit-learn, and TensorFlow (1.x / early 2.0)
for telecom use cases such as network anomaly detection, customer churn prediction, and service quality
optimization, improving prediction accuracy and reducing customer attrition.
Processed and analyzed large-scale telecom datasets (network logs, call detail records, customer usage data)
using Apache Spark (2.x) on AWS EMR, enabling efficient distributed data processing and reducing data
latency for analytics workflows.
Built anomaly detection models using statistical methods and Scikit-learn, identifying network performance
issues and service disruptions, leading to improved network reliability and reduced downtime.
Designed and orchestrated data pipelines using Python, SQL, and Apache Airflow (1.x), ensuring reliable
ingestion, transformation, and scheduling of high-volume telecom data for downstream analytics and ML
applications.
Developed time-series forecasting models using LSTM (TensorFlow/Keras) to predict network traffic
patterns and demand fluctuations, improving capacity planning and resource allocation.
Engineered and deployed machine learning services using Flask-based REST APIs and Docker, enabling
scalable inference for internal telecom analytics platforms.
Leveraged AWS services including EC2, S3, Lambda, and EMR for distributed data storage, batch
processing, and model deployment, supporting scalable ML workflows in a cloud environment.
Performed feature engineering and preprocessing using Pandas and NumPy, including handling missing
data, normalization, and time-based feature extraction to improve model performance.
Applied statistical analysis and hypothesis testing to validate models and support data-driven decision
making for telecom operations and customer analytics.
Built dashboards and visualization tools using Tableau and Power BI, enabling stakeholders to monitor
network performance, customer trends, and key KPIs.
Implemented model performance tracking and logging mechanisms, ensuring consistent accuracy and
reliability of ML models in production.
Collaborated with cross-functional teams including network engineers, operations teams, and business
stakeholders, aligning ML solutions with telecom infrastructure and business objectives.
Automated data workflows and model retraining pipelines using Python scripts and Jenkins-based CI/CD,
improving efficiency and reducing manual effort.
Ensured adherence to data governance, security, and telecom compliance standards, maintaining integrity
and confidentiality of customer and network data.
Supported and maintained production ML systems in a 24/7 environment, ensuring high availability, rapid
issue resolution, and minimal disruption to critical telecom services.
Client: Ruksun Software Technologies Ltd Pune
Role: Junior Data Scientist / ML Engineer
Duration: Jun 2016 May 2017 | India
Developed machine learning models using Python (Scikit-learn) for client-focused analytics use cases such
as customer segmentation and sales forecasting, improving prediction accuracy by 15 20% across multiple
projects.
Performed end-to-end data preprocessing, cleaning, and transformation using Pandas and NumPy,
improving dataset quality and enabling more efficient model development.
Designed and executed SQL-based data extraction and transformation workflows, preparing structured
datasets for analytics and reporting across client engagements.
Built and evaluated classification and regression models, applying appropriate algorithms and validating
performance using metrics such as accuracy, precision, recall, and F1-score.
Implemented foundational NLP solutions using NLTK and early spaCy, enabling text classification and
keyword extraction from unstructured data sources.
Created data visualizations and dashboards using Tableau and Matplotlib, delivering actionable insights to
business stakeholders and supporting data-driven decision-making.
Developed ETL workflows using Python, automating data ingestion from structured and semi-structured
sources (CSV, JSON, relational databases).
Collaborated with senior data scientists and cross-functional teams to improve model performance through
feature engineering and basic hyperparameter tuning.
Gained hands-on exposure to AWS services (EC2, S3) for data storage and experimentation, supporting
scalable data processing and model execution.
Assisted in deploying machine learning models via Flask-based APIs, gaining initial experience in
integrating models into application workflows.
Participated in Agile development processes, contributing to sprint planning, task execution, and timely
delivery of analytics and ML solutions.
Keywords: continuous integration continuous deployment artificial intelligence machine learning business intelligence sthree Florida Georgia Pennsylvania Utah

To remove this resume please click here or send an email from [email protected] to [email protected] with subject as "delete" (without inverted commas)
[email protected];7165
Enter the captcha code and we will send and email at [email protected]
with a link to edit / delete this resume
Captcha Image: