Home

Google Cloud Data Engineer || F2F Interview at Charlotte, North Carolina, USA
Email: [email protected]
https://rb.gy/r1ud0k
https://jobs.nvoids.com/job_details.jsp?id=2495182&uid=c96793d7239d4eb9b303b00b0385e94f
From:

Shubham sharma,

VyzeInc

[email protected]

Reply to:   [email protected]

Google Cloud Data Engineer || F2F Interview

Location: Charlotte, NC (Need local to Charlotte, NC)
Contract Duration: 1 year

Notes:

Must have GCP Certified

Must have experience with Python libraries and Airflow

Experience in the finance industry is required

Candidates that does not need sponsorship now or in the future is required

Job Description

Build the future of data engineering with us, were building a set of native Python libraries that will leverage cloud native technologies on GCP. We are building a foundational set of libraries that allow users to author data and enforce data governance standards.

What youll do

Develop and enhance Python frameworks and libraries to support data processing, quality, lineage, governance, analysis, and machine learning operations.

Design, build, and maintain scalable and efficient data pipelines on GCP.

Implement robust monitoring, logging, and alerting systems to ensure the reliability and stability of data infrastructure.

Build scalable batch pipelines leveraging Bigquery, Dataflow and Airflow/Composer scheduler/executor framework on Google Cloud Platform

Building data pipelines, leveraging Scala, PubSub, Akka, Dataflow on Google Cloud Platform

Design our data models for optimal storage and retrieval and to meet machine learning modeling using technologies like Bigtable and Vertex Feature Store

Contribute to shared Data Engineering tooling & standards to improve the productivity and quality of output for Data Engineers across the company

Minimum Basic Requirements

Python Expertise: Write and maintain Python frameworks and libraries to support data processing and integration tasks.

Code Management: Use Git and GitHub for source control, code reviews, and version management.

GCP Proficiency: Extensive experience working with GCP services (e.g., BigQuery, Cloud Dataflow, Pub/Sub, Cloud Storage).

Python Mastery: Proficient in Python with experience in writing, maintaining, and optimizing data processing frameworks and libraries.

Software Engineering: Strong understanding of software engineering best practices, including version control (Git), collaborative development (GitHub), code reviews, and CI/CD.

Data Management: Deep knowledge of data modeling, ETL/ELT, and data warehousing concepts.

Problem-Solving: Excellent problem-solving skills with the ability to tackle complex data engineering challenges.

Communication: Strong communication skills, including the ability to explain complex technical details to non-technical stakeholders.

Data Science Stack: Proficiency in data analysis and familiarity with tools such as Jupyter Notebook, pandas, NumPy, and other Python data analysis libraries.

Frameworks/Tools: Familiarity with machine learning and data processing tools and frameworks such as TensorFlow, Apache Spark, and scikit-learn.

Bachelors or Masters degree in Computer Science, Engineering, Computer Information Systems, Mathematics, Physics, or a related field or software development training program

Preferred Qualifications

Experience in Scala, Java, and/or any functional language. We code primarily in Scala, so youll be excited to either ramp or continue with such

Experience in microservices architecture, messaging patterns, and deployment models

Experience in API design and building robust and extendable client/server contracts

Education

Bachelor's degree (or foreign equivalent) in Computer Science, Engineering, Computer Information Systems, Mathematics, Physics, or a related field & 5 years of experience involving the following

Special Requirements

Dynamic server-side OOP languages; Scala, Java, C++, Python, or similar languages; design patterns, algorithms, statistics, programming languages, networking and operating systems; web application internals and common technologies; deployment strategies

Production infrastructure; Kafka, BigQuery, Dataflow, Spark, Akka-Http, GRPC, BigTable, JavaScript frameworks; application scalability at any application tier; SQL, relational database schema design and ORM technologies; and Agile/Scrum practices.

Keywords: cplusplus continuous integration continuous deployment North Carolina
Google Cloud Data Engineer || F2F Interview
[email protected]
https://rb.gy/r1ud0k
https://jobs.nvoids.com/job_details.jsp?id=2495182&uid=c96793d7239d4eb9b303b00b0385e94f
[email protected]
View All
10:03 AM 10-Jun-25


To remove this job post send "job_kill 2495182" as subject from [email protected] to [email protected]. Do not write anything extra in the subject line as this is a automatic system which will not work otherwise.


Your reply to [email protected] -
To       

Subject   
Message -

Your email id:

Captcha Image:
Captcha Code:


Pages not loading, taking too much time to load, server timeout or unavailable, or any other issues please contact admin at [email protected]


Time Taken: 8

Location: Charlotte, North Carolina