Mohammed Irshad - Data Engineer |
mi.data198@gmail.com |
Location: Chicago, Illinois, USA |
Relocation: No |
Visa: USC |
Resume file: Ghori_Mohammed_Snr StaffEng_V1_1747865537787.docx Please check the file(s) for viruses. Files are checked manually and then made available for download. |
Ghori, Mohammed
Snr Data Engineer | Phone: 248-805-9706 | Email: mi.data198@gmail.com| US Citizen Professional Summary Results driven Snr Data Engineer with 13+ years of experience in Data Integration, Data Warehousing, and ETL development. Expertise in using Informatica PowerCenter and Data Management Cloud (IDMC). Proficient in migrating ETL pipelines from SSIS to Informatica Cloud, and experienced with cloud data platforms including Azure Synapse, AWS Redshift, and Snowflake. Adept at developing complex ETL mappings, data quality frameworks, and real-time integrations using REST/SOAP APIs. Skilled in performance tuning, data validation, and end-to-end project delivery within Agile and Waterfall methodologies. TECHNICAL SUMMARY: ETL Tools: IDMC, Informatica B2B- DX, Informatica DQ, Informatica PowerCenter (10.2/9.x/8.6), Cloud Platforms: Azure (Data Factory, Synapse, ADLS), AWS (Redshift, S3, Glue, Lambda), Snowflake Databases: Oracle, SQL Server, DB2, Netezza, Cassandra, Teradata, Azure SQL DB Programming/Scripting: PL/SQL, T-SQL, Python, Unix Shell Scripting, Batch Scripting Integration & APIs: REST/SOAP APIs, API Development, Salesforce, SharePoint ServiceNow Data Modeling: Star Schema, Snowflake Schema Version Control & CI/CD: Git, Control-M, AWS Code Pipeline, Azure DevOps Methodologies: Agile, Waterfall PROFESSIONAL EXPERIENCE: Wise IT Inc, Remote Aug 2022 Current Snr Staff Software Developer Role Highlights: Developed data integration components using Informatica Cloud (IDMC), Azure Databricks, Azure Synapse, Data Lake Store, and Blob Storage. Implemented IDMC Mass Ingestion to efficiently process large volumes of structured and unstructured data from multiple sources into AWS Redshift and Snowflake, reducing ingestion time by 40%. Utilized input/output parameters extensively in IDMC workflows. Experienced with cloud-based data platforms including Azure Synapse, Data Lake Store, AWS Redshift, and Snowflake. Hands-on with core data integration components: parsers, UDFs, fixed-width files, mapplets, mappings, mapping tasks, and task flows. Managed real-time and file-based database ingestion within IDMC's Mass Ingestion domain. Created external tables with partitions using Hive, AWS Athena, and Redshift. Loaded data into AWS S3 as flat files. Converted legacy Informatica PowerCenter mapping into Informatica Cloud mappings in collaboration with the Data Migration Factory team. Conducted unit and functional testing of migrated assets in IDMC, deploying to higher environments. Loaded data into Snowflake staging layers via Snowflake connector in IICS/IDMC for analytics and insights within the Education domain. Extensively used Snowflake SQL and notebooks for data transformation and loading between databases. Developed data processing workflows in Azure Databricks, creating delta tables for curated layers. Published data from Databricks into semantic layers for reporting and AI/ML use cases. Hands-on experience with Databricks components: clusters, Azure Key Vaults, notebooks, workflows, and Unity Catalog. Involved in development, unit testing, UAT, and deployment activities. Environment: Informatica Intelligent Cloud Services (IICS), Informatica Data Quality, PowerCenter 10.x, Salesforce Cloud, Snowflake, Azure SQL Data Warehouse, Azure Data Lake Storage (ADLS), Python, SQL Server 2016, CI/CD pipelines, Azure Data Factory (ADF), Azure Data Lake, Azure Blob Storage. ACV, Remote June 2021 July 2022 Senior Data Integration Engineer Role Highlights: Designed and deployed end-to-end data integration solutions across cloud and on-premises environments. Extracted flat files from Amazon S3 into Snowflake using IICS Data Integration services. Loaded customer, teller, and account data into Salesforce, performing Upserts with External ID (Hash Key) to maintain SCD Type 1 via Task Flows, Bulk API, and Standard API. Integrated data sources including Salesforce, SharePoint, ServiceNow, SQL Server, Vertica, and flat files using IICS CIH pub/subtopics. Built AWS data pipelines and data lakes utilizing EC2, Glue, and Lambda for scalable processing. Developed bulk ingestion workflows from FTP/SFTP/local folders and on-premises databases into Salesforce and cloud data warehouses. Configured Secure Agents for data loads from S3 and heterogeneous databases into cloud platforms. Created web services and APIs for third-party application integrations, including migration of legacy Informatica workflows into IICS. Developed real-time data pipelines with Informatica Cloud, REST, and SOAP web services supporting cloud-to-cloud and cloud-to-on-premises flows. Managed large-scale transformations into Redshift, S3, Hive, and HDFS; utilized Snowflake features like cloning and Time Travel for dataset management. Migrated SQL Server data into multi-cluster Snowflake environment, establishing data sharing and virtual warehouses. Experienced with AWS services (EC2, S3, Redshift, EMR, RDS) for scalable cloud data solutions. Performed data profiling, standardization, address validation, and fuzzy matching using identity, Hamming distance, and bigram algorithms via WSDL web services. Developed data governance rules and real-time batch web services for data quality, enabling immediate correction and re-validation. Optimized web service performance, reducing response times from 900ms to 34ms. Proficient in Informatica transformations, T-SQL (stored procedures, triggers, views), and MDM architecture. Environment: Informatica Intelligent Cloud Services (IICS), Informatica Data Quality, 10.x, Redshift, RDS, CI/CD, Amazon Elastic MapReduce, Python, SQL Server 2016. Beacon Health Option, Remote Dec 2019 May 2021 Senior Data Integration Engineer Role Highlights: Architected and developed scalable Data Integration workflows utilizing Informatica Cloud (IICS), Azure SQL Data Warehouse (Synapse Analytics), Azure Data Lake Store Gen2, and Azure Blob Storage to support enterprise data pipelines. Conducted comprehensive analysis of existing ETL processes within Data Warehouse environments and ERP/NON-ERP application interfaces; authored detailed design specifications aligned with migration to cloud-native platforms such as Azure Synapse Analytics and Data Lake Store. Created and maintained ETL standards including naming conventions, ETL methodologies, data validation/cleansing strategies, and data preprocessing protocols to ensure consistency, scalability, and maintainability. Developed detailed source-to-target mapping documents covering transformation logic, source data column attributes, target schema definitions, and data lineage tracing. Designed, developed, and deployed ETL workflows using Informatica Intelligent Cloud Services (IICS), leveraging cloud-native transformations and connectors for optimal performance. Configured and managed IICS cloud connections utilizing connectors for Azure Synapse, Azure Data Lake Store V3, Azure Blob Storage, Oracle, Oracle CDC, and SQL Server environments. Applied advanced performance tuning techniques include push-down optimization, data partitioning, and parallel processing to enhance load efficiency into Azure Synapse Analytics. Extensively utilized cloud transformations such as Aggregator, Expression, Filter, Joiner, Lookup (connected and unconnected), Rank, Router, Sequence Generator, Sorter, Update Strategy, and Union to implement complex transformation logic. Developed reusable, parameterized mapping templates supporting scalable incremental loads, SCD Type 1 & 2, CDC, and fact table processing workflows. Leveraged Push Down Optimization to maximize processing capacity and exploit Azure Synapse s compute power for large data loads. Extracted data from Snowflake using Snowflake connectors within IICS, orchestrating data loads into Azure Data Warehouse to support reporting and analytics needs. Implemented CDC (Change Data Capture) workflows for real-time data replication from PeopleSoft to Azure Data Warehouse using Informatica Cloud CDC for Oracle. Designed and orchestrated complex, parallel task flows involving multiple mapping and control flow tasks to optimize data processing pipelines. Built high-volume Mass Ingestion workflows to efficiently load large datasets from on-premises sources into Azure Data Lake Store Gen2. Developed a Data Integration Audit framework within Azure SQL Data Warehouse to automate load tracking, workload management, and generate compliance reports supporting SOX requirements. Collaborated with a distributed team of four onshore and six offshore developers, prioritizing tasks and ensuring delivery of high-quality solutions within project timelines. Environment: Informatica Intelligent Cloud Services (IICS), PowerCenter 10.2, PowerExchange 10.2, SSIS, Windows Secure Agent, Teradata v1310, Azure Synapse (Azure SQL Data Warehouse), Azure Data Lake Store, Azure Blob Storage, SQL Database, Power BI reporting. Daimler-Mercedes Benz Financial Services Nov 2018 Dec 2019 Snr SQL/Informatica Developer Role Highlights: Developed high-level, detailed, and source-to-target mapping documents for end-to-end ETL processes. Extracted and loaded data across multiple sources (Oracle, SQL Server, Salesforce, XML, Flat Files), pushing data into AWS Redshift via Informatica PowerCenter for analytics. Implemented complex transformations including Transaction Control, Lookup (Active/Passive), Router, Filter, Expression, Aggregator, Joiner, Update Strategy, Stored Procedures, and SQL transformations. Designed incremental/Delta load mechanisms utilizing mapping variables, parameters, and parameter tables for full and incremental loads. Built layered mappings for landing, staging, and publish zones, implementing SCD Type I & II using MD5 hashing for change detection. Profiled data with Informatica Data Quality (IDQ), and integrated data loads directly into Salesforce objects via PowerCenter using API and Bulk API. Enhanced performance by optimizing session properties, tuning SQL/stored procedures, and applying Partitioning and Push-Down Optimization techniques. Developed reusable transformations, Mapplets, and Worklets with Transformation Developer and Mapplet Designer. Loaded unstructured data into Oracle using Unstructured Data Transformation. Universal McCann, Birmingham MI Mar 2016 Nov 2018 Data Developer Role Highlights: Developed a reusable ETL framework and detailed dataflow documentation. Designed complex mappings, sessions, and workflows in PowerCenter for interaction with MDM and EDW systems. Implemented full/incremental loads, Type 1 & 2 SCDs, and integrated IDQ for data quality validation, including address validation rules. Managed ETL requirement gathering, environment setup, code migration, and production support. Monitored performance, optimized mappings via transformations such as Router, Lookup, Joiner, Expression, Sorter, and Aggregator. Developed reusable Mapplets and performed unit testing. Huntington Bank, Cleveland, OH July 2012 Feb2016 Informatica Developer Role Highlights: Gathered requirements, translated into technical designs, and developed ETL mappings aligned with business goals. Designed and optimized data models including star and snowflake schemas for OLTP/OLAP systems. Developed data integration workflows using PowerExchange for mainframe data access, and created batch scripts for file management. Conducted performance tuning by optimizing cache, buffer, and partitioning settings in mappings and sessions. Reviewed ETL mappings, SQL scripts, and data warehouse constructs to ensure accuracy and adherence to business requirements. Managed source/target definitions, shared folders, and user access for efficient project delivery. Education: Master s in management information systems - USA - Year 2013 Professional Reference: Available upon request. Keywords: continuous integration continuous deployment artificial intelligence machine learning business intelligence sthree database information technology procedural language Idaho Michigan Ohio |