Rohith Dhane - Senior Data Engineer |
rohith.dhane69@gmail.com |
Location: Wilmington, Delaware, USA |
Relocation: YES |
Visa: GC |
Resume file: Rohith Dhane Data Engineer_1749138021615.docx Please check the file(s) for viruses. Files are checked manually and then made available for download. |
Profile Summary: Rohith Dhane Senior Data EngineerOverview:
Rohith Dhane is a highly experienced and technically proficient Senior Data Engineer with over 12 years of hands-on experience in building and managing end-to-end data engineering solutions. His expertise spans across Big Data technologies, cloud platforms (AWS, Azure, GCP), and modern data pipelines for large-scale, real-time, and batch processing systems.Core Competencies:Cloud Platforms: Proficient in AWS (S3, EMR, Lambda, Glue, Redshift), Azure (Data Factory, Synapse, Databricks), and GCP (BigQuery, Dataflow, Dataproc).Big Data Technologies: Deep expertise in Hadoop ecosystem (Hive, HDFS, MapReduce, Pig, Spark, Sqoop), Spark (PySpark, Scala), Kafka, HBase, MongoDB, Cassandra.ETL & Data Pipelines: Skilled in designing scalable ETL pipelines using Apache Airflow, Spark, Python, Talend, Informatica, and Data Fusion.Programming Languages: Proficient in Python, Scala, Shell scripting, SQL, and HiveQL.DevOps & Automation: Experience with CI/CD pipelines using Jenkins, Terraform, AWS SAM, and infrastructure automation.BI & Visualization: Developed reports and dashboards using Tableau and Power BI.Monitoring & Performance: Worked with tools like AppDynamics, Datadog, ELK Stack, Dynatrace, and IDERA.Professional Highlights:CareFirst (2023 Present): Led cloud-native ETL workflows using PySpark and AWS services, built robust data ingestion and transformation pipelines, and implemented real-time analytics using Kafka and Spark Streaming.CBRE (2021 2023): Migrated legacy systems to AWS EMR, built real-time data processing apps using Spark and Kafka, and managed infrastructure using CloudFormation and Jenkins.Merck Pharma (2019 2021): Developed and deployed data pipelines on GCP using Airflow, BigQuery, and Dataflow. Built REST APIs for data ingestion and managed real-time streaming data solutions.Dollar Tree (2016 2019): Played a key role in Azure cloud migration, data warehousing using Synapse Analytics, developed ETL pipelines, and built dashboards using Tableau.Concentrix (2013 2016): Designed and executed ETL processes using Talend and Informatica, optimized database performance, and implemented cloud-based Redshift solutions via AWS Data Pipeline.Technical Summary:Big Data Tools: Hadoop, Spark, Hive, Kafka, Flume, Sqoop, HBase, PigLanguages: Python, Scala, SQL, Shell ScriptingDatabases: Snowflake, Teradata, Oracle, SQL Server, MongoDB, DynamoDBCloud: AWS, Azure, GCPETL Tools: Talend, Informatica, Airflow, Data FactoryDevOps: Jenkins, Git, Terraform, AWS SAMBI Tools: Tableau, Power BISoft Skills & Methodologies:Strong communication and problem-solving skillsAgile/Scrum project delivery experienceAdept at stakeholder collaboration and requirement gatheringProven ability to mentor and guide junior engineers Keywords: continuous integration continuous deployment business intelligence sthree |