Home

C2C Role : DataLake Architect || Remote in the USA at Remote, Remote, USA
Email: [email protected]
http://bit.ly/4ey8w48
https://jobs.nvoids.com/job_details.jsp?id=2300837&uid=

Dear,

Hope you are doing great.

Im Yogesh from the
Recruitment team at Elgebra LLC, an IT consulting firm that helps clients meet their immediate and long-term staffing needs. Id like to discuss the following opportunity with you. If youre interested, kindly share your updated
resume in Word format, ensuring it includes your visa status and LinkedIn profile URL for reference.

Job Title: 
DataLake
Architect

Location:  
Remote
in the USA

Contract (12 Months)

Need LI URL and PP number for submission.

No OPT/H1B-T profiles for this specific role.

Rate: $75/hr. C2C

seeking a
DataLake Administrator/Architect to join us. 

Duties and Responsibilities:

Manage and maintain Data Lake clusters infrastructure on premise and in cloud: installation, configuration, performance tuning and monitoring of Hadoop clusters

Should demonstrate strong concepts in Unix/Linux, Windows OS, cloud platforms (AWS, GCP), Kubernetes, Open Shift & Docker

Must have good exposure to Cloudera manager, Cloudera Navigator or similar cluster management tool

Collaborate and assist developers in successful implementation of their code, monitor and fine tune their process for optimum resource utilization on cluster, ability to automate run time process  

Must have good knowledge of HDFS, Ranger/Sentry, Hive, Impala, Spark, HBase, Kudu, Kafka, Kafka Connect,  Schema Registry, Ni-Fi, Sqoop and other Hadoop related services

Exposure to Data Science collaborative tools such as data science workbench, CML, anaconda, etc.  

Strong Networking concepts: topology, proxy, F5, firewall

Strong security concepts: Active directory ,Kerberos, LDAP, SAML, SSL, data encryption @rest

Programming language concepts: Java, Perl, python, PySpark and Unix scripting

Possess experience in cluster management, perform cluster upgrade, migration, and testing

Perform periodic updates to cluster and keeping the stack current

Ability to expand clusters by adding new nodes and rebalance cluster storage systems

Manage application databases, application integration, users, roles, permissions within cluster

Collaborate with OpenShift, Unix, network, database and security teams on cluster related matters

Must monitor cluster for maximum uptime, ability to research on cluster issues via logs and collaborate with support in a proactive way

Education and Years of Experience: 

Minimum 10 years experience in advanced technologies including a minimum of 5+ years as data lake admin/architect

BS degree, preferably in Computer science or equivalent

Good communication skills with right attitude to blend in with team

Minimum 5 years work experience in Hadoop ecosystems (Horton HDP or Clouderas CDP)

Required and Desired Skills/Certifications:

Solid experience in Cloudera data lake environments both on prem and cloud

Solid experience in administration and set up including security topics related to a data lake

Strong experience architecting and designing solutions for new business needs

Thorough understanding and hands-on experience with implementing robust logging and tracing implementation for end-to-end systems traceability

Familiarity with Clouderas BDR tool to perform and monitor backups of critical data and able to restore data when in need

Willing and ready to get hands on code development with dev team for developing and troubleshooting, doing quick proof of concepts for exploring new solutions, products etc.

Experienced in working with technical teams to discuss, analyze, understand and negotiate business requirements, being able to explain to architects about the technical considerations and associated implications on the user journey/experience/requirements.

Experience in tuning and optimizing Hadoop environment in keeping clusters healthy and available for end users and applications with maximum cluster uptime as defined in SLA

Deep knowledge and related experience with Hadoop and its ecosystem components i.e. HDFS, Yarn, Hive, MapReduce, Pig, Sqoop, Oozie, Kafka, Spark, Presto and other Hadoop components

Warm regards,

Yogesh Pratap Singh( Yogi)

Direct (205)-775-0773

244 Fifth Avenue, Suite R295,

New York, NY 10001

Email: 
[email protected]

Web: 
http://elgebra.com/

https://www.linkedin.com/in/yogesh-pratap-singh-282a744a/

Donate Red
 || 
Save
Blue
 || 
Spread Green

Keywords: information technology ffive New York
C2C Role : DataLake Architect || Remote in the USA
[email protected]
http://bit.ly/4ey8w48
https://jobs.nvoids.com/job_details.jsp?id=2300837&uid=
[email protected]
View All
09:01 PM 31-Mar-25


To remove this job post send "job_kill 2300837" as subject from [email protected] to [email protected]. Do not write anything extra in the subject line as this is a automatic system which will not work otherwise.


Your reply to [email protected] -
To       

Subject   
Message -

Your email id:

Captcha Image:
Captcha Code:


Pages not loading, taking too much time to load, server timeout or unavailable, or any other issues please contact admin at [email protected]


Time Taken: 0

Location: ,