Home

Nanda Kumar - Data Analyst/Engineer
[email protected]
Location: Dallas, Texas, USA
Relocation: YES
Visa: H1B
Resume file: Nanda Kumar Data Analyst Engineer_1775844011517.docx
Please check the file(s) for viruses. Files are checked manually and then made available for download.
Nanda Kumar
Data Analyst/Engineer
[email protected]
469-581-9864
__________________________________________________________________________________
Professional Summary
Data Analyst/Engineer with 5+ years of experience in ETL & ELT processes and data warehousing across multiple platforms, including Hadoop, Snowflake, and Talend, with expertise in analyzing both structured and unstructured data.
Analyzed and managed large-scale banking/healthcare data using Hadoop Hue, SQL Server handling millions of records in both production and development environments.
Experience in designing SQL-driven ETL pipelines, Snowflake data warehouses, and Power BI dashboards for hospital data, supply chain, and clinical quality reporting.
Proven expertise in Epic EMR integration, HEDIS measures, ICD/CPT code analysis, and data quality automation to deliver scalable, compliant, and actionable insights for clinical and operational teams.
Designed and developed scalable ETL pipelines in Azure Databricks integrating multi-source data (SQL Server, S3, APIs) into Delta Lake with optimized transformation and validation logic
Experienced working in secure and regulated environments (HIPAA, financial audits); comfortable aligning with public sector compliance standards such as CJIS.
Analyzed and Migrated data from enterprise data lake to Snowflake for advanced analytics and reporting.
Proficient in SQL and Python for data querying and transformation, with a strong understanding of machine learning techniques including regression, classification, clustering and collaborative filtering.
Integrated REST APIs, SFTP, and JSON/XML message queues into analytical models.
Experienced in designing and automating data pipelines using Python, Shell, and Snow SQL to extract, validate, and monitor Snowflake data, with seamless integration into BI tools and cloud platforms such as AWS S3 and Tableau.
Extracted data from enterprise data lake by developing complex ETL/ELT jobs in Talend and build stored procedures to support report building.
Performed advanced data analysis in Hive, Snowflake, and optimized query performance in ETL/ELT jobs, creating partitioned tables and Views in snowflake on top of migrated tables for reporting.
Experienced in normalizing and de-normalizing data to optimize database design and support various analytical and reporting requirements.
Developed and executed Linux shell scripts to automate routine data tasks including file ingestion, transformation, archiving, and job scheduling.
Built data observability frameworks including audit tables, row-level logging, and status monitoring.
Experience in implementing and designing database views to stakeholder requirement for reporting efficiency in snowflake.
Experience in creating Ad hoc Reports , Complex Crystal Reports, Drill-Down Reports,
Created interactive dashboards and visualizations using Tableau to provide actionable insights and support data-driven decision-making.
Implemented Tableau for visualizations and views including scatter plots, box plots, heatmaps, tree maps, donut charts, highlight tables, word clouds, reference lines, etc.
Applied data mining techniques to uncover fraud patterns and customer behavior insights across large-scale banking datasets using SQL, Snowflake, and Python.
Collaborated with project managers and business analysts to align data analysis processes with business requirements and project goals.
Worked on RELTIO for MDM(Meta Data Management).
Developed data definitions and standardization rules to maintain consistency and improve data quality across different environments.

Category Tools & Technologies
Data Analysis Hadoop Hue (EDL), Hive, Toad, Snowflake, Tableau
Data Visualization Tableau, Power BI, Microsoft Excel, PowerPoint
ETL/ELT Tools Talend, Snowpipe, Snowpark, Shell Scripting (CRON, AWS CLI), QuerySurge (ETL Testing)
Data Modeling Erwin, Snowflake Star Schema, OLTP & OLAP Models
Data Governance & MDM Ataccama ONE, Reltio, Alation (Data Catalog, Lineage, Data Quality, PII/HRSDE)
Databases & Programming SQL, T-SQL, Python, VBScript, Windows PowerShell
Machine Learning Techniques Regression, Classification, Clustering, Collaborative Filtering
Database Systems Snowflake, SQL Server, MySQL, Hadoop EDL, Salesforce
Version Control
Office & Collaboration Tools
Git
Microsoft Word, Excel, PowerPoint, Jira, Confluence, SharePoint

Cloud Platforms & Services AWS (S3, EC2, IAM)

Professional Experience:
Truist, Coppell, TX Feb 2023 - Present
Data Analyst/Engineer.
Key Responsibilities:
Analyzed and managed banking data using Hadoop Hue, handling both production and development data environments involving millions of records.
Analyzed trends and patterns in data related to COMPLAINTS, TSYS, DISPUTES, FRAUD (CFO, CFC) and Enterprise Data.
Worked on OLAP data models in Snowflake and transforming OLTP banking data (fraud, disputes, complaints, TSYS) into Star Schema and reporting layers.
Designed ETL pipelines using Talend and SnowSQL to move data from Hadoop/transaction systems into OLAP structures for high-performance reporting and dashboards
Built data definitions and standardization rules to maintain consistency and improve data quality based on production data.
Good hands on in preparing Mappings from source to target with writing necessary business transformation rules.
Great understanding towards core data modeling concepts like FACTS, DIMENSIONS and worked closely with data modelers and have very good understanding in conceptual, logical and physical data models.
Worked on Golden source Data Model environment and have deep understanding in GS Model.
Experience in writing and performing complex SQL Queries for analysis and developmen in the snowflake environment.
Implemented Snowpipe for real-time data ingestion from AWS S3 buckets into Snowflake to support continuous data flow.
Developed ETL workflows using Talend to transform data from source and to load data into snowflake target by performing transformations according to the model from EDL.
Designed and implemented Talend transformation jobs to cleanse, standardize, and integrate data, handling Slowly Changing Dimensions (SCD) Types 1, 2, and 3 for account, customer and product dimensions to ensure historical tracking and accurate reporting.
Build ELT workflows in Snowflake using Snowpark to data for auditing and future uses.
Implemented Star Schema data models in Snowflake to support analytical reporting.
Automated daily data extraction from Snowflake to CSV using Snow SQL and Python, streamlining recurring report delivery tasks.
Automated Snowflake data ingestion processes using shell scripts that integrate copy into commands and trigger downstream workflows.
Scheduled ETL jobs using shell scripting and CRON to export query results and securely upload files to AWS S3 via the AWS CLI, ensuring automated and reliable data movement
Implemented email-based alert system to distribute daily extracts to stakeholders using MAILX with dynamic attachments.
Experience in preparing DDL s, performing data profiling and have expertise in writing data quality rules.
Experience in identifying High Risk Data Elements(HRSDE) or Critical data elements.
Collaborated with cross-functional teams to gather data requirements and implement data integration solutions.
Performed data quality checks and hash-based comparisons (like MD5) to ensure reliable and accurate data movement between source and target systems.
Successfully identified and resolved data mismatches between source systems (RDBMS/flat files) and Snowflake, which improved data pipeline stability and overall data quality
Worked closely with QA teams to define testing strategies, write and maintain test cases, and automate validation processes for both incremental and full data loads using tools like Snowflake COPY INTO, Snow pipe, and Talend.
Experience in defining KPI s on top of the business consumption layer.
Utilized Snowflake for data warehousing, including writing complex SQL queries for data extraction, transformation, and analysis.
Delivered Git-based pipelines, integrated with GitFlow and version control systems.
Conducted segmentation analysis on customer complaints and disputes data to group users by product type, transaction channel, fraud indicators, and region, improving reporting accuracy and fraud detection response.
Built interactive Tableau dashboards with geo-spatial heatmaps to identify fraud patterns across banking products, highlighting transaction hotspots, vulnerable regions, and abnormal activity.
Integrated dynamic KPIs, custom fields, and filters (amount thresholds, banking channel, transaction type, and location) to trigger alerts and provide real-time fraud monitoring insights.
DATA QUALITY AND GOVERNANCE
Integrated Data Quality (DQ) rules in Ataccama ONE to automatically profile data, detect anomalies, and generate scorecards for completeness, duplication, and consistency checks.
Implemented Master Data Management (MDM) using Reltio, building golden customer records, applying survivorship rules, and ensuring sensitive data attributes were governed properly.
Designed and automated ETL pipelines in Talend and Snowflake, embedding governance controls (PII masking, deduplication, change data capture (CDC)logic) directly into workflows for compliance.
Automated ETL validation and reconciliation with QuerySurge, reducing manual data quality checks by 40% and improving accuracy of reporting datasets.
Developed Tableau dashboards to monitor Data Governance KPIs, including null percentages, classification coverage, lineage documentation, and rule compliance, enabling leadership to track adoption and maturity.
Acted as a Data Steward by investigating records that failed multiple Data Quality (DQ) rules in Ataccama, working with business teams to determine whether issues were due to manual data entry errors, ETL transformation logic, or source system problems, and driving resolution to improve trust in data
Consulted with business teams to align data definitions and policies, standardizing attributes such as complaints, counterparties, and sensitive identifiers across regions.
Led migration of governance workflows from Hadoop to Snowflake, integrating Ataccama for data quality and Reltio for MDM, ensuring alignment with enterprise governance standards.
Defined business rules, access policies, and metadata standards in Erwin and Ataccama to align data management practices with enterprise governance policies.
Implemented data cataloging and data lineage tracking in Ataccama, enabling end users to discover data assets, understand lineage from source to reporting, and ensure compliance with enterprise taxonomy.
Applied segmentation and classification rules in Ataccama to organize datasets by domain (customer, account, complaints) and to enforce enterprise-level tagging of PII and sensitive attributes.
Collaborated with stakeholders to resolve conflicts between business needs and governance requirements, recommending safe alternatives such as surrogate IDs in place of sensitive attributes.
Built stored procedures in Snowflake for customer hierarchy resolution, leveraging recursive CTEs and graph-style traversal to identify Master and Link customers from match pairs while ensuring accuracy, preventing loops with ARRAY_CONTAINS, and supporting production stability
KIMS, India May 2019 Jul 2021
Data Analyst
Responsibilities:
Collaborated with nursing leaders, supply chain managers, finance analysts, and business analysts to gather requirements for clinical quality, utilization, and cost performance tracking.
Defined and operationalized KPIs for HEDIS quality measures, ICD-10/CPT/HCPCS code analysis, pharmacy claims, authorizations, and eligibility data, aligning with payer regulations and operational KPIs.
Designed SQL- and Alteryx-based ETL pipelines to process and validate high-volume healthcare data, including claims, drug prescriptions, encounter data, and provider records, enabling rapid turnaround for ad-hoc and scheduled reporting.
Used Alteryx Designer to build reusable data workflows for data cleansing, merging payer claims with EMR and Workday extracts, improving data prep time by 30%.
Integrated Epic EMR, Workday Supply Chain, and external clinical data into Snowflake, Databricks SQL, and Power BI pipelines, delivering full-funnel analytics on cost, quality, and utilization metrics.
Built claims-based cost utilization models using Alteryx, SQL, and Excel, identifying high-cost diagnosis categories and outlier procedures through ICD-10 and CPT mapping logic.
Created ICD-10/CPT/HCPCS based dashboards to support care coordination, provider benchmarking, and coverage trend monitoring in Power BI and Tableau.
Developed Snowflake star-schema models for claims and eligibility data marts, optimizing partitioning and surrogate key strategies for performance across millions of rows.
Automated recurring HEDIS quality and supply chain reports using Power BI (DAX, Power Query), SSRS, and Excel VBA, improving data refresh reliability and reducing manual work.
Implemented data quality checks in Snowflake, Workday, and Alteryx to identify issues like missing vendor IDs, duplicate purchase orders, mismatched NDCs, and incomplete eligibility records, ensuring HIPAA, FDA, and Joint Commission compliance.
Certifications
Snowflake: Certified by Snowflake.
Ataccama: Certified by Ataccama.
Keywords: quality analyst business intelligence sthree active directory Delaware Texas

To remove this resume please click here or send an email from [email protected] to [email protected] with subject as "delete" (without inverted commas)
[email protected];7157
Enter the captcha code and we will send and email at [email protected]
with a link to edit / delete this resume
Captcha Image: