| VENKATESWARA REDDY DURGEMPUDI - ETL INFORMATICA/TERADATA/GCP DEVELOPER |
| [email protected] |
| Location: Dallas, Texas, USA |
| Relocation: |
| Visa: |
| Resume file: Venkateswara Reddy Durgempudi Teradata Developer_1769199039855.docx Please check the file(s) for viruses. Files are checked manually and then made available for download. |
|
VENKATESWARA REDDY DURGEMPUDI
ETL INFORMATICA/TERADATA/GCP DEVELOPER LOCATION: Dallas, TX PHONE: +14014412682 EMAIL: [email protected] PROFESSIONAL SUMMARY: Over 10+ years of experience in Information Technology with building & supporting Data Warehouse/ DataMart using Informatica PowerCenter 10.4.0/10.2/10.1.1/10.0.1/9.6.1/9.5.1/9.1.1/8.6.1. Strong work experience in Data Warehouse lifecycle process. Involved in understanding of Business Processes, grain identification, identification of dimensions and measures (Facts). Extensive knowledge on Understanding Data Modeling(ER & Dimensional Modeling), doing Data Integration and Data Migration. Extensive experience in working with different RDBMS Oracle, Teradata, My Sql, SQL Server, Azure SQL Server, DB2, Sybase and also with File base system Flat Files & XML Files. Extensive experience in designing and developing complex mappings using transformations lookups(Connected & Un Connected), Normalizer, Source Qualifier, Router, Filter, Expression, Aggregator, Joiner, java and Update Strategy. Expert on implementing Slowly Changing Dimensions Type 1, Type 2 and Type 3 for inserting and updating Target tables for maintaining the history. Expert on Implementing Change Data Capture (CDC) for handling Incremental loads. Experience on Mapping Debugger to validate the mappings and gain troubleshooting information about data and error conditions. Experience in using Automation Scheduling tools like Autosys, Control-M, Tivoli, Maestro scripts. Experience on post-session and pre-session shell scripts for tasks like merging flat files after Creating, deleting temporary files, changing the file name to reflect the file generated date etc. Extensively used Informatica Mapping Parameters and variables. Extensively worked with Informatica performance tuning involving Identifying and eliminating bottlenecks. Experience with Integrating Informatica with Teradata and using Teradata features. Extensive experience with Teradata utilities like BTEQ, Fast Export, Fast Load, Multi Load&Tpump also with TPT. Proficient in Teradata EXPLAIN plans, Collect Stats option, Primary Indexes (PI, NUPI), Secondary Indexes (USI, NUSI), Partition Primary Index (PPI), Join Indexes (JI), Volatile, global temporary, derived tables etc. Expertise in performance tuning and query optimization of the Teradata SQLs. Experience with Unit Testing, working with QA team on System testing also involved in UAT. Experience with ETL Migrations & Code deployments also involved with Post production validations. Solid experience in writing SQL queries, Stored Procedures. Experience in working with UNIX Shell Scripts for automatically running sessions, aborting sessions and creating adhoc parameter files. Written number of shell scripts to run various batch jobs. Implemented Data Warehouse projects both in Agile and Water Fall Methodologies and good understanding on Scrum process. Good Experience with snowflake cloud data warehouse, AWS S3. Good experience with Snowflake utility SnowSQL. Strong understanding of the principles of Data Warehousing concepts using fact tables, dimension tables and Star/ Snowflake schema modeling. Excellent interpersonal and communication skills, capable of driving the DWH projects independently. Very keen in knowing the newer techno stack that Google Cloud Platform (GCP) adds. Set up GCP Firewall rules in order to allow or deny traffic to and from the VM's instances based on specified configuration and used GCP cloud CDN (content delivery network) to deliver content from GCP cache locations drastically improving user experience and latency. TECHNICAL SKILLS: Operating Systems: Windows, UNIX, LINUX, MS-DOS Modeling: Dimensional data modeling, star schema modeling, snowflake Schema modeling, E-R modeling, visio RDBMS: Oracle 12c/11g/10g/9i, Teradata 15/14/13/12, DB2, SQL Server 2008/2012/2014/2016, DB2, MySQL, Sybase, Azure SQL Server ETL Tools: Informatica PowerCenter 10.x,9.x/8.6.1, IICS. Reporting Tools: Cognos, Business Objects, Tableau, Denodo Scheduling Tools: Autosys, Control-M Languages: XML, UNIX Shell Scripting, SQL, PL/SQL, Python, Powershell Miscellaneous: GIThub, SVN Google Cloud Platform: GCP Cloud Storage, Big Query, Composer, Cloud Dataproc, Cloud SQL, Cloud Functions, Cloud Pub/Sub, Dataflow etc. Cloud technologies: GCP, Snowflake, SnowSQL EXPERIENCE: JAN 2022 PRESENT TERADATA DEVELOPER, VERIZON, IRVING, TX Involved in requirements gathering and data gathering to support developers in handling the design specification. Extracted the data from various source systems like Oracle, SQL Server and flat files as per the requirement. Extensive experience in writing and executing BTEQ scripts for validation and testing of the sessions, data integrity between source and target database and for report generation. Involved in loading of data into Teradata from legacy systems and flat files using complex MLOAD scripts and FASTLOAD scripts. Created Teradata External loader connections such as Mload, Upsert and Update, Fastload while loading data into the target tables in Teradata Database. Created proper Primary Index (PI) talking into consideration of both planned access of data and even distribution of data across all the available AMPS. Write numerous BTEQ scripts to run complex queries on the teradata database. Loading data by using the Teradata loader connection, writing Teradata utilities scripts (Fastload, Multiload) and working with loader logs. Reduced Teradata space used by optimizing tables adding compression where appropriate and ensuring optimum column definitions. Responsible for migration and production support. Wrote hundreds of DDL scripts to create tables, views and indexes in the company Data Warehouse. Prepared Unit Test specification requirements. Migrated on-premise environment in GCP (Google Cloud Platform) Involved in porting the existing on-premise code migration to GCP (Google Cloud Platform) Bigquery. Experience in GCP Dataproc, Dataflow, PubSub, GCS, Cloud functions, BigQuery, Stackdriver, Cloud logging, IAM, Data studio for reporting etc. Build data pipelines in airflow in GCP for ETL related jobs using different airflow operators. Used cloud shell SDK in GCP to configure the services data proc, storage, BigQuery. Coordinated with team and Developed framework to generate Daily adhoc reports and Extracts from enterprise data from BigQuery. Experience in working with product teams to create various store level metrics and supporting data pipelines written in GCP s big data stack. Vast experience in identifying production bugs in the data using stack driver logs in GCP. Experience in GCP Dataproc, GCS, Cloud functions, Cloud SQL & BigQuery. Used Cloud shell SDK in GCP to configure the services Data Proc, Storage, BigQuery. Extensive use of cloud shell SDK in GCP to configure/deploy the services like Cloud Dataproc (Managed Hadoop), Google Cloud Storage and Cloud Bigquery. Created BigQuery jobs for loading the data into BigQuery tables from data files stored in Google Cloud storage daily. JUNE 2021 JAN 2022 ETL INFORMATICA / TERADATA DEVELOPER, CHARLES SCHWAB, WESTLAKE, TX Worked with Business Analyst & Data Modelers in understanding the BRD - Business Requirement Document, Mapping Document and Data Model. Extracted data from different source systems - Oracle, DB2, My Sql, Flat Files and XML Files. Developed ETL programs using Informatica Power center 10.4.0 to implement the business requirements. Involved in enhancements and maintenance activities of the Enterprise Data Warehouse. Communicated with business customers to discuss the issues and requirements. Used most of the Transformations available in Informatica - Source Qualifier, Filter, Router, Lookup (Connected & Un Connected), Expression, Update Strategy, Transaction Control and Sequence Generator. Implemented Slowly Changing Dimensions - Type 1 & Type 2 to maintain the history in Dimension tables. Worked with Huge data sets to load Fact Tables. Implemented Change Data Capture (CDC) for handling delta loads. Involved in Informatica upgrade process and testing the whole existing Informatica flow in new upgrade environment. Experienced in doing performance Tuning of Informatica objects- Finding the bottle necks at source, Target and mapping level and eliminating the with tuning methods. Used Informatica file watch events to pole the FTP sites for the external files. Involved in Enhancing existing Production informatica objects for change or additional requirements and pushing it back to production after successful QA testing. Expertise in using Teradata Utilities BTEQ, M-Load, F-Load, TPT and F-Export in combination with Informatica for better Load in to Teradata Ware House. Built several BTEQ to load data from Stage to Base after considering several performance techniques in Teradata sql. Involved in Teradata upgrade process from TD 12/TD 14. Production Support has been done to resolve the ongoing issues and troubleshoot the problems. Effectively used Informatica parameter files for defining mapping variables, workflow variables, FTP connections and relational connections. Effectively worked in Informatica version based environment and used deployment groups to migrate the objects. Used debugger in identifying bugs in existing mappings by analyzing data flow, evaluating transformations. Pre and post session assignment variables were used to pass the variable values from one session to other. Designed workflows with many sessions with decision, assignment task, event wait, and event raise tasks, Used Control-M to schedule jobs. Performed unit testing at various levels of the ETL and actively involved in team code reviews. Implemented & followed organization level Informatica best Standards & Procedures. Involved in Solving Trouble tickets raised by Business users as part of application support team. Created tables, views, secure views user defined functions in Snowflake cloud data warehouse. Extracted and loaded CSV files, Json files data from AWS S3 to Snowflake cloud data warehouse. Migrated Oracle database tables data into Snowflake cloud data warehouse. JULY 2019 - JUNE 2021 ETL/ TERADATA DEVELOPER, WELLS FARGO BANK, SAN LEANDRO, CA Developed internal and external Interfaces to send the data in regular intervals to Data warehouse systems. Extensively used Power Center to design multiple Mappings with embedded business logic. Involved in discussion of user and business requirements with business team. Performed Data Migration in different sites on regular basis. Created complex mappings using Unconnected Lookup, Sorter, and Aggregator and Router transformations for populating target tables in efficient manner. Attended the meetings with business integrators to discuss in-depth analysis of design level issues. Involved in data design and modeling by specifying the physical infrastructure, system study, design, and development. Extensively involved in performance tuning of the Informatica ETL mappings by using the caches and overriding the SQL queries and also by using Parameter files. Developed complex SQL queries to develop the Interfaces to extract the data in regular intervals to meet the business requirements and extensively used Teradata Utilities like M - load, F- Load, TPT, BTEQ and Fast Export. Analyzed Session Log files in session failures to resolve errors in mapping or session configuration. Written various UNIX shell Scripts for scheduling various data cleansing scripts, loading process and automating the execution of maps. Created transformations like Expression, Lookup, Joiner, Rank, Update Strategy and Source Qualifier Transformation using the Informatica designer. Created Mapplets and used them in different mappings. Worked on Flat Files and XML, DB2, Oracle as sources. Written PL/SQL Procedures and functions and involved in Change Data Capture (CDC) ETL process. Implemented Slowly Changing Dimension Type II for different Dimensions. Involved in Informatica, Teradata and oracle upgrade process and testing the environment while up gradation. Worked with Informatica version Control excessively. Experience in using SVN as version control for migration. Written Unit test scripts to test the developed interfaces. Experience in Autosys scheduling tool to automate the jobs. Managed enhancements and coordinated with every release with in Informatica objects. Provided support for the production department in handling the data warehouse. Worked under Agile methodology and used Rally tool one to track the tasks. Written thorough design docs, unit test documentation, Installation and configuration guide documents. Performed bulk data imports and created Stored procedures, Functions, Views and Queries. JUNE 2018 - JULY 2019 ETL DEVELOPER, UHG(OPTUM), EDEN PRAIRIE, MN Collaborate with Lead Developers, System Analysts, Business Users, Architects, Test Analysts, Project Managers and peer developers to analyze system requirements. Collaborate with Lead Developers, System Analysts, Business Users, Architects, Test Analysts, Project Managers and peer developers to analyze system requirements. Worked withSQL, PL/SQL procedures and functions, stored procedures and packages within the mappings. Involved in all activities related to the development, implementation, and support of ETL Process using Informatica Power Center 10.x Worked with most of the transformations such as the Source Qualifier, Expression, Aggregator and Connected & Unconnected lookups, Filter, Router, Sequence Generator, Sorter, Joiner, SQL and Update Strategy. Developed complex store procedures using input/output parameters, cursors, views, triggers and complex queries using temp tables and joins. Develop of scripts for loading the data into the tables using FastLoad, MultiLoad and BTEQ utilities of Teradata. Used control-M to schedule Jobs. Used Snowflake schema to be joined with the fact table Involved in requirement analysis, ETL design and development for extracting data from the source systems like sales force, Mainframe, DB2, sybase, Oracle, flat files. Responsible for determining the bottlenecks and fixing the bottlenecks with performance tuning. Extensively involved in the Analysis, design and Modeling. Worked on Snowflake Schema, Data Modeling, Data Elements, Issue/Question Resolution Logs, and Source to Target Mappings, Interface Matrix and Design elements. Design and develop logical and physical data models that utilize concepts such as Star Schema, Snowflake Schema and Slowly Changing Dimensions Work on Test Driven Development and conduct unit testing, system testing and user acceptance testing. Create deployment packages to deploy the developed Informatica mappings, Mapplets, Worklets and workflows into test and then to production. Trouble shoots any deployment issues and coordinates to deploy the code into production on the target date. Created dashboards for analyzing POC data and applied Filter Actions between different worksheets and dashboards. Designed and developed various analytical reports from multiple data sources by blending data on a single worksheet in Tableau Desktop. Participate in agile daily stand-up, sprint planning, sprint review, retrospective, Backlog refinement and Feature Overview meetings. JULY 2016- JUNE 2018 INFORMATICA/ TERADATA DEVELOPER, FREDDIE MAC, FAIRFAX, VA Extensively worked with various Active transformations like Filter, Sorter, Aggregator, Router and Joiner transformations. Extensively worked with various Passive transformations like Expression, Lookup, Sequence Generator, Mapplet Input and Mapplet Output transformations. Worked with source databases like Oracle, SQL Server and Flat Files. Worked with extracting data from SFDC. Extensively worked with Teradata utilities BTEQ, F-Load, M-load & TPT to load data in to Teradata ware house. Created complex mappings using Unconnected and Connected lookup Transformations. Responsible for the performance tuning of the ETL process at source level, target level, mapping level and session level. Tuned performance of Informatica session for large data files by increasing block size, data cache size and target based commit interval. Responsible for Performance Tuning of Teradata scripts using explain plans, indexing and Statistics. Implemented slowly changing dimension Type 1 and Type 2. Worked with various look up cache like Dynamic Cache, Static Cache, Persistent Cache, Re Cache from database and Shared Cache. Worked extensively with update strategy transformation for implementing inserts and updates. Worked with various Informatica Power Center objects like Mappings, transformations, Mapplet, Workflows and Session Tasks. Auditing is captured in the audit table and EOD snapshot of daily entry is sent to the distributed list to analyze if there are any abnormalities. As per business we implemented Auditing and Balancing on the transactional sources so that every record read is either captured in the maintenance tables or wrote to Target tables. Extensively used the tasks like email task to deliver the generated reports to the mailboxes and command tasks to write post session and pre session commands. Extensively used debugger to test the logic implemented in the mappings. Performed error handing using session logs. Involved in production support when required. Monitored workflows and session using Power Center workflows monitor. Used Informatica Scheduler for scheduling the workflows in dev for testing. Provided 24*7 support for Production environment jobs. Monitoring the Extraction and loading processes of data and Involved in writing UNIX shell scripting for automating the jobs. APR 2014 - FEB 2015 JR.ETL DEVELOPER, ADAPTIVE SOFTWARE SOLUTIONS PVT LTD, HYDERABAD, INDIA Involved in the requirement gathering and Business Analysis of the specifications provided by the business analysts. Designed the mappings between sources (external files and databases) to operational staging targets. Experience with high volume datasets from sources like DB2, Oracle and Flat Files. Loaded data from various sources using different transformations like Source Qualifier, Joiner, Aggregators, Connected & Unconnected lookups, Filters, Router, Expression, Rank Union, and Update Strategy & Sequence Generator. Experience in writing PL/SQL scripts, Stored Procedures and functions and debugging them. Responsible for Migration of Stored Procedures into Informatica Mappings for improvingPerformance issue. Involved in Performance Tuning of application by identifying bottlenecks in SQL, thus providinginputs to the application programmer, thereby correcting and implementing the right components. Created Session Task, Email and Workflow to execute the mappings. Used Workflow Monitor tomonitor the jobs, reviewed error logs that were generated for each session, and rectified any cause of failure. Experience in ETL testing, Created Unit test plans and Integration test plans to check whether the data has been loaded into the target is accurate, which was extracted from different source systems according to the user requirements. Setting up of the local Informatica environment on the client machines which included the connectivity and access to the data sources, taking the necessary steps to set up the RelationalConnectivity variables in the Workflow manager etc. Used SQL override to perform certain tasks essential for the business. Mapplets and Reusable Transformations were used to prevent redundancy of transformation usage and modularity. Defined Target Load Order Plan for loading data into Target Tables. Involved in the documentation of the ETL process with information about the various mappings, the order of execution for them and the dependencies. EDUCATION: B.E-JNTUH/Electronics & communication Engineering/Hyderabad/2014 Master's-Trine University/Indiana/Engineering Management/2015-2016 Keywords: quality analyst sthree rlang information technology microsoft mississippi procedural language California Minnesota Texas Virginia |