Rohit - Sr. ETL/Sowflake/Informatica/Data engineer |
venkata.dasari@kanerika.com |
Location: Atlanta, Georgia, USA |
Relocation: Remote |
Visa: H1B |
Resume file: RohitReddyBojja_1746716977803.docx Please check the file(s) for viruses. Files are checked manually and then made available for download. |
Professional Summary:
Over 14 Years of IT experience in SSIS and Informatica PowerCenter, Informatica Cloud (Power Center and b2b), IDQ, informatica Developer b2b, Matillion, Data Architect, Data Stage, Talend, Teradata developer, PostgreSQL (Greenplum), Tableau, QlikView, OBIEE, Denodo, Spotfire, data warehousing, Data Integration Hub, Cloud interfaces like IICS (Uses Rest & Soap API), Cloud Data Integration Hub, Oracle, Control M Unix Scripts, Perl scripting, Python, Python Spark(PySpark),PL/SQL, Approx., Denodo, Knowledgeable in ETL tools - Informatica (7.1, 8.1, 9.1, 10), Matillion, ADF, Databricks, Talend Studio V5.6, SSIS Having good knowledge on data warehousing concepts Snowflake, Sql Server, PostgreSQL (Greenplum), Oracle 9i/10g, Denodo & Spotfire, Dynamo DB(No-Sql), Data Integration, Data Architect Having good experience in Informatica IDQ and Data Stage. Having profound skills in writing Perl scripting, Teradata developer, Microsoft Visual Studio, HTML, PHP. Maintained large databases and configured services to reduce software maintenance expens-es and improved data protection and security. Led the successful implementation of a Big Data project using ETL - Informatica PowerCenter, IICS cloud and using tableau, a key component in optimizing data processing and analysis. Dynamic Partitioning, Parallel Load, Cache processing in Cloud, EDI split and merge are few big data technologies for handling Big data in Informatica within given performance. Uses computer and analysis knowledge to analyze, design, and write new application systems. Involves color perception to design color-enhanced computer screens. Creates and modifies complex applications programs, tests, debugs, documents, and main-tains those programs. Created business requirement docs and ETL architecture, technical specifications and pro-cessed documents for business practice and support. Worked on testing, preparing use cases, creating unit test cases and system integrated testing. Having Good Experience on Azure Data Factory (ADF), Databricks for Big Data Processing Having sound knowledge and basic coding in C, C++, Java, Db2, Rest & Soap API. Familiar with SDLC and worked on Agile and waterfall methodologies in multiple projects. Experience in working on all phases of software Development life cycle (SDLC) - requirement gathering, test planning, analysis, design, implementation, and execution and closure phases. Having vast experience on UNIX/BASH by creating Unix/Bash Scripts. Experience in creating new Unix Scripts for end-end flow in SSIS and Informatica jobs. Worked on creating Scripts using Python and PySpark to modify few source files for unit test-ing. Uses Control M to Create Informatica Workflows as jobs and Schedule and Monitor Experience Oracle Queries (SQL Developer, Toad), data warehousing, and Dimension data modeling (physical and logical), policies, rules or standards and optimizing SQL/oracle queries to increase performance. Having Knowledge in TDD (technical Design), HLD (high Level Design) and LLD (Low Level De-sign Having knowledge in Proficient in error handling, performance analysis and performance tun-ing of SSIS and Informatica ETL components and Teradata utilities, UNIX scripts and SQL scripts etc. Having good experience in SCD (type1 to 4) and CDC (Change data Capture). Having knowledge in tuning Sql queries, writing complex queries using nested, correlated, Ana-lytic functions, Using with clause. Experience in analyzing and creating existing Pl/SQL, PostgreSQL queries (in Greenplum) to load or Update data in new tables using Views or existing tables. A team player and used to work on multiple tasks at a time to meet deadlines and having good communication and interpersonal skills. Having knowledge in JIRA, Rally and confluence where tasks and sub tasks can be up-dated and which in turn stores all versions. Expertise in multiple programming languages, API development, and micro-services architec-ture. Optimized ETL processes by leveraging Pandas for data manipulation, SQLAlchemy for data-base interactions, and pyodbc for ODBC connections, resulting in significant improvements in data processing efficiency and reduced processing times. Excellent problem-solving skills. Knowledge in project planning, management, and production deployment Technical Skills: ETL Tools: Informatica Power Center, Informatica Power Exchange, Informatica Cloud, Informatica Da-ta Quality (IDQ), Informatica Data Transformation B2B, Matillion, Big Data, Informatica Data Explorer (IDE), SSIS, Talend, DataStage, etc, Cloud interfaces like Cloud Integration Hub, Azure Data factory Reporting Tools: Business Objects, QlikView, OBIEE, Tableau, MicroStrategy, Oracle Analytics, etc. Scheduling tools: Informatica Scheduler, CA Scheduler (Autosys), Approx, DAC, Tidal, Control-M. Data Modeling: Relational Modeling, Dimensional Modeling (Star Schema, Snowflake, FACT, Dimen-sions), Physical, Logical Data Modeling, and ER Diagrams. DB Tools: SQL Server Management Studio (2008), Oracle SQL Developer (3.0), snowflake, Toad 11.6, SSIS, Microsoft SSRS (Oracle), DB2, Teradata, AQT v9 (Advanced query tool) (Oracle/Netezza), DB Arti-san 9.0 (Sybase), SQL Browser (Oracle Sybase), Visio, Languages: C, C++, Java, .Net, Python, PySpark, Perl Scripting, UNIX/BASH Shell Scripting, XSLT, PL/SQL, T-SQL. Python Frameworks: Django, Flask Operating Systems: UNIX, Linux, Windows PROFESSIONAL EXPERIENCE Project: Anthem Healthcare Project Employer: Marquee Technology Apr 2021 till date Title: ETL Senior Developer Place: Virginia Beach, VA Project description: Medicaid markets (i.e., states) and Medicare publish revised fee schedules pro-cess the claims on the websites or FTP locations over internet. Currently, the data is extracted, stored, and processed separately for each market. FSP retrieve files from market websites or FTP locations, create ACCR and send notification to Market Analysts, Extraction of data from market files to raw data tables, apply business rules, decide the action based on comparison with existing data. Responsibilities: Worked on Informatica Power Center and Informatica Developer tool to handle big data un-structured pdf and excel files of around 300MB into a centralized Big Data platform. Working as an analyst for several markets while performing peer review to provide approvals for further action. Designed and developed ETL pipelines using Python libraries like Pandas, SQLAlchemy, and py-odbc, extracting data from multiple sources, transforming and loading it into target databases, and ensuring data quality and integrity. Oracle to Snowflake Migration, migration of data infrastructure from Oracle databases to Snowflake Data Warehouse, improving data processing performance and scalability. Designed and implemented ETL pipelines using Python, SQL, and Snowflake s native tools to ensure seamless data transfer and integration across multiple business applications and data sources. Led the successful implementation of a Big Data project using Informatica, a key component in optimizing data processing and analysis. Developed and implemented data integration strategies to seamlessly transfer and transform large datasets from various sources, ensuring data accuracy and consistency. Working on Bitbucket, GitHub and SVN for files versioning. Worked on ADF (Azure Data factory) to connect to Salesforce DB and create Pipelines and gen-erate Incremental files. Started working on Data Bricks for Sql load - On SQL Load Migration from SSIS to Data Bricks. Optimized data processing pipelines for scalability and performance, achieving a [percent-age]% improvement in processing times. Utilizing self-defined migration tool to migrate data from one environment to another. Working on design and develop Informatica mappings using expression, Filter, Joiner, Lookups, Aggregator, Update strategy, Normalizer, Transaction control, Union, and SQL transformations. Worked on design and development of Informatica Data Transformations B2b using Informatica developer which converts Excel sheets with multiple tabs, pdf files into XML files by applying Tool scripts like Parser, Serializer, Mapper, Transformer and these DT codes is developed using Infor-matica developer. Having end- end SDLC phase from Requirement gathering, Development, QA Testing, Produc-tion Deployment and Support in Production. Proposed IDTP and ODTP Job concept using Python Spark (PySpark) along Informatica for effi-cient and scalable data processing to ingest, transform, and analyze large datasets, leveraging computing for improved performance. Also used PySpark for Data Quality Reconciliation checks between multiple sources and used for Optimal Performance and reducing processing time by 25%. Conducted thorough testing and debugging of Informatica workflows to ensure the reliability and integrity of data transformations in a high-volume environment. Tools: Matillion, Informatica Power Center, Data Transformation B2B, Jira, Snowflake, Oracle, Toad, Python, PySpark, Control M, Informatica Developer, ADF, Azure, Data bricks, project specific migration tools. Project: Wonderful Citrus BI Analytics Employer: Kanerika Jan 2019 Apr 2021 Title: ETL Cloud team Place: Hyderabad, India Client: The Wonderful Company Project description: wonderful citrus has big Fusion oracle apps DWH from which we need to handle multiple modules like P2PY, GL, OTC, ADP etc and load data into Snowflake DWH using Informatica cloud and built Power Bi Dashboards on 60 KPI s and also uses DT and DX to use X12 EDI and EDIFACT Files Responsibilities: Developed and optimized ETL pipelines using Matillion for efficient data extraction, transfor-mation, and loading (ETL) into Snowflake Data Warehouse, improving data processing speed and accuracy. Designed and implemented scalable data models within Snowflake, leveraging Snowflake s na-tive features like Snowpipe, Streams, and Tasks for efficient data loading and management. Optimized SQL queries and Snowflake schema to ensure maximum efficiency, reducing query processing time by 30%. Worked on Python for Api call using endpoint URL and generate the incremental files. Worked on Informatica Power Center and Informatica Developer tool to handle Big data un-structured pdf, EDI files of around 500MB to 1GB into a centralized Big Data platform by using Big data technologies like splitting the files by Invoice#, Dynamic partitioning, parallel load and merge them while Pushing to target DB Web services in IICS are used by Rest & Soap API. Had worked on Cloud Integration Hub which is a combination of multiple cloud applications. Designed and developed multiple Informatica mappings using different transformations. Developed and implemented ETL processes using Informatica Cloud to efficiently move and transform data between on-premises and Snowflake cloud environments. Designed and developed tasks and task flows and loaded data into Snowflake DWH. Designed and optimized data pipelines to ensure high-performance data integration, contributing to a 30% reduction in data processing time. Proposed IDTP and ODTP Job concept using Python Spark (PySpark) along Informatica for effi-cient and scalable data processing to ingest, transform, and analyze large datasets, leveraging computing for improved performance. Also used PySpark for Data Quality Reconciliation checks between multiple sources. Explored more features like cloning, security in snowflake DB and implemented in existing sys-tems. Compared to other tools, snowflake data load and DWH analysis are faster. Designed data modelling of the database and Power BI Dashboards. Worked on Informatica Power Center and Informatica Developer DT and DX console for X12 EDI and EDIFACT Files Design and develop Informatica mappings using different transformations. Project: Fee Schedule Automation Employer: Rapid IT Aug 2018 Dec 2018 Title: ETL Team Lead and Developer Client: Anthem Place: Virginia Beach, VA Project description: Amerigroup Medicaid markets (i.e. states) and Medicare publish revised fee schedules on the websites or FTP locations over internet. These fee schedules rates are used by Amer-igroup systems to process the claims. Currently, the data is extracted, stored and processed separately for each market. FSP retrieve files from market websites or FTP locations, create ACCR and send notifi-cation to Market Analysts, Extraction of data from market files to raw data tables, Apply business rules, decide the action based on comparison with existing data in FACETS and send the reports to Gate Keeper and Once GA approves, Move the Fee Schedule data to FACETS Responsibilities: Worked on Informatica Power Center and Informatica Developer. Design, develop Informatica mappings using expression, Filter, Joiner, Lookups, Aggregator, Up-date strategy, Normalizer, Transaction control, Union, and SQL transformations. Design, Develop Informatica Data Transformation B2b using Informatica developer which con-verts Excel sheets with multiple tabs, pdf files into XML files by applying Tool scripts like Parser, Serializer Project: CMG Analytics implementation Employer: Ascendum May 2018 Aug 2018 Title: ETL Team Lead and Developer Client: Cox Place: Atlanta, GA Project description: CMG Analytics a reporting solution based on the Oracle Business Intelligence applications, Oracle E-Business, WideOrbit, News Cycle, Darwin & Hyperion Planning feed data to the CMG Analytics system. Informatica loads data into staging tables which are transformed and loaded into various dimensions and fact tables in the Data Warehouse. Oracle BI Applications supports Third Party Applications like Microsoft Great Plains and ReqLogic through Universal Adapters. Universal Adapters Extract and Load the Data from CSV files into the Out of the Box Dimension & Fact staging tables Responsibilities: Worked on Informatica Power Center, SSIS, Data Stage, Cognos. Web services in the project are used by Rest & Soap API. Had data Integration Hub which is combination of multiple applications. Designed ETL Mapping, workflows and monitor jobs in Informatica Repository, Packages in SSIS, Jobs in Data Stage using many transformations which includes data Extraction, Transform Project: SPS data profiling Employer: Ascendum Nov 2017 May 2018 Title: ETL Team Lead and Developer Client: Anthem Place: Atlanta, GA Project description: The SPS project comprises of an enterprise modeling engine owned by the An-them group. This enables simulations for provider contract negotiations with existing or expected payment calculations, analytics for contract performance monitoring, ad hoc and standard reporting. The enterprise modeling solution allows it. Responsibilities: Worked on data profiling using SQL Queries in Db2 and Idq. Created Data Models by using policies, rules or standards. Worked on how Data is stored, arranged, integrated, and put to use in data systems. Project: GE Treasury Employer: Tech Mahindra Americas Mar 2016 Nov 2017 Title: ETL Team Lead and Developer Client: GE Place: Stamford, CT Project Description General Electric Company is an American multinational conglomerate corporation in US. I work for Treasury app support project under GE Capital. In Treasury, I work as L2 Team member where we de-velop ETL mappings for Treasury Data HUB and load data into TDH DB and generate reports and dash-boards using Tableau. GE treasury is Finance project. Responsibilities: Analyzed specifications provided by the clients. Responsible for solutions design and development Designed data modelling of the target database. Created Data Models by using policies, rules or standards. Worked on how Data is stored, arranged, integrated, and put to use in data systems. Project: GE Industrial PLR project Employer: Tech Mahindra Jan 2013 Mar 2016 Title: ETL Senior software Engineering Client: GE Place: Hyderabad, India Project Description General Electric Company is an American multinational conglomerate corporation in US. I worked Product line reporting (PLR) under Energy Management. In PLR project, there are different modules like PLR Orders, PLR Sales, PLR Backlog, PLR SFDC, and PLR Power Quality. The PLR project strives com-mercial excellence, and provides data analysis on sales & commercial performance, orders forecast, Backlog information etc. To provide central global database for all source systems where users can directly check orders & sales amounts, cancel orders and backlog orders based on Product Hierarchy. Responsibilities: I am one of key resources of Project at ETL side and Reporting Side. Developed ETL mappings, workflows in both SSIS and Informatica and monitored the jobs in production. Created Data Models by using policies, rules or standards. Worked on how Data is stored, arranged, integrated, and put to use in data systems. Project: SAP DLE (Deposit Liquidity Engine) Employer: Infosys May2012-Dec 2012 Title: ETL software Engineering Client: RBS Place: Chennai, India Project Description: To establish a central repository (DLE package) to support Corporate Bank Division launch attract new deposits by providing propositions tailored to segments and sectors and their associated cash flow pat-terns with associated pricing. RBS Is Banking Project. Responsibilities: Communicated with Business regarding the requirements. As a crucial team member worked on the Requirement Analysis, test planning, design, exe-cution. Always help support team on any issues. Uses SCD type 1 and Type 2 and CDC logic in SSIS and informatica. Project: Lvts Employer: Infosys Jun 2010- May2012 Title: ETL software Engineer Client: Ameriprise Place: Chennai, India Project Description To develop Accounts Interface for Banking Accounts, calculating prices and perform Securities by Securities Interface. Responsibilities: Communicated with Business regarding the requirements & Key resource of Project at ETL and Reporting Side. Keywords: cprogramm cplusplus quality analyst access management business intelligence database active directory information technology procedural language California Connecticut Delaware Georgia Virginia |