| Siddartha prasad - Data Engineer (Datastage, Dbt, Snowflake) |
| [email protected] |
| Location: Plano, Texas, USA |
| Relocation: yes |
| Visa: H1B |
| Resume file: Siddartha Prasad_Resume_1771281890757.docx Please check the file(s) for viruses. Files are checked manually and then made available for download. |
|
Siddartha prasad B S
Email: [email protected] Ph: +1 346-375-0837 Ext: 225 https://www.linkedin.com/in/siddarthaprasadbs/ PROFESSIONAL SUMMARY Results-oriented DataStage Developer leveraging a decade of experience in Data Warehousing to design, implement, and optimize robust ETL solutions, design and development. Involved in complete software Development lifecycle (SDLC) of various projects, including Requirements gathering, system design, Data modeling, and ETL design, development, Production Enhancements, support, and Maintenance. Excellent Interpersonal and communication skills with an ability to remain highly focused and self-assured in fast-paced and high-pressure environments. Extensive ETL tool experience using IBM info sphere/ Web Sphere DataStage. Worked on DataStage tools like DataStage Designer, DataStage Director and DataStage Administrator. Strong understanding of the principles of Data Warehousing using fact tables, dimension tables and star/snowflake schema modeling. Worked extensively with Dimensional modeling, Data migration, Data cleansing, ETL Processes for data warehouses. Developed parallel jobs using different processing stages like Transformer, XML files, Aggregator, Lookup, Join, Sort, copy, merge, Funnel, CDC, Change Apply and Filter. User Enterprise Edition/Parallel stages like Datasets, change Data capture, Row Generator, and many other stages in accomplishing the ETL coding. Familiar in using highly scalable parallel processing infrastructure using parallel jobs and multiple node configuration files. Experience in data migration from various data sources like excel, SQL Server, Oracle, XML files, Flat Files using Data Stage and SQL commands. Proficient in designing, developing, and optimizing scalable ETL solutions, with expertise in managing and implementing Data Stage/SSIS projects from conception through deployment. Experienced in scheduling sequence and parallel jobs using DataStage Director and Unix scripts and scheduling tools. Assisted in development efforts for Data marts and reporting. Experience in troubleshooting of jobs and addressing production issues like data issues, ENV issues, performance tuning and enhancements. Technical and analytical skills with clear understanding of design goals of ER Modeling for OLTP and dimension modeling for OLAP. Extensive experience in Unit Testing, Functional Testing, System Testing, Integration, Regression Testing, User Acceptance Testing (UAT) and Performance Testing. Worked with various databases like IBM Netezza, Oracle, SQL Server, snowflake. TECHNICAL SKILLS ETL Tool IBM DataStage -11.7, Data Build Tool (DBT) Databases Oracle, Netezza, SQL Server, Snowflake Hadoop Components HDFS, Hive Operating Systems Windows, Unix Languages SQL, Unix Shell Scripting Testing/Defect Tracking qtest, JIRA PROFESSIONAL EXPEREINCE Client: USAA Bank, Plano TX 09/21/2024 Present Technical Lead (Senior project Engineer) Responsibilities Application Development Using DBT cloud to migrate stored procedures into DBT jobs. Implemented Snowflake role based access control (RBAC) to secure the data layer for unstructured datasets and flattened structured tables. Worked on Anti Money Laundering banking project to create SAR reports, 314a, subpoena, PEP Negative report using ETL process and tools. Design and build data warehouses and all related processes, including data loading, processing, and extraction. Create High level and Low-Level Design documents which will translate the business requirements into specific systems, applications, and process designs for custom technical solutions. Analyze and interpret the business goals to develop optimal warehouse and ETL model. Create a source to target mapping documents with data translation rules for each data movement process from heterogeneous sources to staging and then to Data Warehouse. Working on all facets of data management, including normalization, cleansing and standardization, merging, change data capture, data integration, reconciliation, and transformations. Designed and developed SQL-based data transformation models using DBT, ensuring scalable and efficient data pipelines for analytics and reporting. Implement Change Data Capture (CDC) techniques for Incremental or SCD Type 2 / Type 1 data loads. Develop SQL scripts including complex query building and query performance tuning. Implement performance tuning techniques like table partitioning, Bulk Load at DB/ETL level. Automate the deployment process for ETL and DB components using CI/CD tools Gitlab and UCD. Develop shell scripts to orchestrate, invoke and execute the ETL workflows. Migrate existing on-prem oracle based regulatory reporting platform to snowflake-based cloud architecture. Develop Data Ingestion flows using Snow Pipe to load the data files from external cloud services (AWS/Azure) to Snowflake Cloud Warehouse. Lead the technical planning & requirements gathering phases including estimate, develop, test, and deliver. Technologies : Extract, Transform, Load (ETL), Data Stage, Data Build Tool (DBT), Oracle/Snowflake Database, SQL, Linux, Shell Scripting. HCL Technologies, Bangalore, India 10/15/2018 09/20/2024 Client: USAA Bank Senior Software Engineer Responsibilities Collaborated with EDW team in High Level Design documents for extract, transform, validate and load ETL process data dictionaries, Metadata descriptions, file layouts and flow diagram. Orchestrated the designing, testing, support, and debugging of new and existing ETL & reporting processes. Used to read and write to Mainframe machine by using complex flat fie stage. Used different stages of DataStage Designer like lookup, join, transformer, change capture, aggregator etc Collaborated with EDW team in, Low level design document for mapping the files from source to target and implementing business logic. Created shell script to automate data stage jobs from Unix and then schedule this script to run data stage through scheduling tool. Extensively used IBM DataStage Designer for creating DataStage Jobs and created Shared Containers for reusability. Involved in creating Table definitions, indexes, views, sequences, Materialized view creation. Involved in writing test plans, test scenarios, test cases and test scripts and performed unit, integration, system testing and user acceptance testing. Converted complex job designs to different job segments and executed through job sequencer for better performance and easy maintenance. Performed performance tuning of jobs by interpreting performance statistics of the jobs developed. Liaised with the Development team of 10 to design a data warehousing system that meets specific business needs. Extracted the necessary data and transferred it to the new system after the development of warehousing system. Tested the designs to ensure the proper functioning of the system and fixed any issues that may arise. Supervised impact analysis of changing upstream processes to Data Warehouse and Reporting systems. Technologies: IBM DataStage, Informatica, SQL, Unix, Unix Shell Scripting, JIRA, Control M Scheduler. SLK Software Services, Bangalore, India 02/15/2016 10/12/2018 Client: Fifth / Third Bank Software Engineer Responsibilities Extensively used DataStage for extracting, transforming, and loading databases from sources including Oracle, Netezza and Flat files. Generation of surrogate keys for the dimensions and fact tables for indexing and faster access of data in Data warehouse. Tuned transformations and jobs for performance enhancement. Extracted data from flat files and then transformed according to the requirement and loaded into target using various stages like sequential file, look up, Aggregator, Transformer, Join, Remove Duplicates, Change Capture data, sort, column generators, Funnel and Oracle Enterprise. Created batches (DS Job controls) and sequences to Control set of Job. Extensively used DataStage change data capture for DB2 and Oracle files and employed change capture stage in parallel jobs. Execute pre and post session commands on source and target database using shell scripting. Extensively worked on Job sequences to Control the execution of the job flow using various Activities & Triggers (Conditional and unconditional) like Job Activity, wait for file, email notification, sequencer, exception handler activity and execute command. Utilized parallelism through different partition methods to optimize performance in a large database environment. Developed DS Jobs to populate the data into staging and data mart. Executed jobs through sequencer for better performance and easy maintenance. Performed the unit testing for jobs developed to ensure that it meets the requirements. Developed UNIX shell scripts to automate file manipulation and data loading procedures. Responsible for daily verification that all scripts, downloads, and file copies were executed as planned, troubleshooting any steps that failed, and providing both immediate and long-term problem resolutions. Provided technical assistance and support to IT analysts and the business community. Technologies: IBM DataStage, SQL, Unix, Unix Shell Scripting, JIRA/RTC. Societe Generale , Bangalore, India 03/27/2014 03/12/2015 Support Engineer Responsibilities L1 support of Monitoring DataStage jobs in control-m scheduling tool on daily, weekly and monthly Basis on Production Server L1, L2 support to the application and handling the tickets and resolving in time Generating performance report on a daily, weekly and monthly basis on production jobs Finding the root cause of the job failure and check if the issue can be resolved without L3 support. Working the L3 team to get the issue resolved. Implemented and enforced best practices and standard operating procedures. Reviews code to identify SQL changes to improve system performance. Handled performance tuning and break/fix of existing SQL code. Technologies: IBM DataStage, SQL, Unix, Unix Shell Scripting, JIRA/RTC. Education B.E (Telecommunication) Atria Institute of Technology, Bangalore (VTU) , 2013. Keywords: continuous integration continuous deployment database information technology Texas Keywords: continuous integration continuous deployment database information technology Texas Keywords: continuous integration continuous deployment database information technology Texas |