C2C Role : DataLake/ Hadoop Admin || Remote in the USA at Remote, Remote, USA |
Email: [email protected] |
http://bit.ly/4ey8w48 https://jobs.nvoids.com/job_details.jsp?id=2305729&uid= Dear, Hope you are doing great. Im Yogesh from the Recruitment team at Elgebra LLC, an IT consulting firm that helps clients meet their immediate and long-term staffing needs. Id like to discuss the following opportunity with you. If youre interested, kindly share your updated resume in Word format, ensuring it includes your visa status and LinkedIn profile URL for reference. Job Title: DataLake/ Hadoop Administrator Location: Remote in the USA Contract (12 Months) Need LI URL and PP number for submission. No OPT/H1B-T profiles for this specific role. Rate: $65-$68/hr. C2C S eeking a DataLake/Hadoop Administrator to join us. Duties and Responsibilities: Manage and maintain Data Lake clusters infrastructure on premise and in cloud: installation, configuration, performance tuning and monitoring of Hadoop clusters Should demonstrate strong concepts in Unix/Linux, Windows OS, cloud platforms (AWS, GCP), Kubernetes, Open Shift & Docker Must have good exposure to Cloudera manager, Cloudera Navigator or similar cluster management tool Collaborate and assist developers in successful implementation of their code, monitor and fine tune their process for optimum resource utilization on cluster, ability to automate run time process Must have good knowledge of HDFS, Ranger/Sentry, Hive, Impala, Spark, HBase, Kudu, Kafka, Kafka Connect, Schema Registry, Ni-Fi, Sqoop and other Hadoop related services Exposure to Data Science collaborative tools such as data science workbench, CML, anaconda, etc. Strong Networking concepts: topology, proxy, F5, firewall Strong security concepts: Active directory ,Kerberos, LDAP, SAML, SSL, data encryption @rest Programming language concepts: Java, Perl, python, PySpark and Unix scripting Possess experience in cluster management, perform cluster upgrade, migration, and testing Perform periodic updates to cluster and keeping the stack current Ability to expand clusters by adding new nodes and rebalance cluster storage systems Manage application databases, application integration, users, roles, permissions within cluster Collaborate with OpenShift, Unix, network, database and security teams on cluster related matters Must monitor cluster for maximum uptime, ability to research on cluster issues via logs and collaborate with support in a proactive way Education and Years of Experience: Minimum 10 years experience in advanced technologies including a minimum of 5+ years as data lake admin/architect BS degree, preferably in Computer science or equivalent Good communication skills with right attitude to blend in with team Minimum 5 years work experience in Hadoop ecosystems (Horton HDP or Clouderas CDP) Required and Desired Skills/Certifications: Solid experience in Cloudera data lake environments both on prem and cloud Solid experience in administration and set up including security topics related to a data lake Strong experience architecting and designing solutions for new business needs Thorough understanding and hands-on experience with implementing robust logging and tracing implementation for end-to-end systems traceability Familiarity with Clouderas BDR tool to perform and monitor backups of critical data and able to restore data when in need Willing and ready to get hands on code development with dev team for developing and troubleshooting, doing quick proof of concepts for exploring new solutions, products etc. Experienced in working with technical teams to discuss, analyze, understand and negotiate business requirements, being able to explain to architects about the technical considerations and associated implications on the user journey/experience/requirements. Experience in tuning and optimizing Hadoop environment in keeping clusters healthy and available for end users and applications with maximum cluster uptime as defined in SLA Deep knowledge and related experience with Hadoop and its ecosystem components i.e. HDFS, Yarn, Hive, MapReduce, Pig, Sqoop, Oozie, Kafka, Spark, Presto and other Hadoop components Warm regards, Yogesh Pratap Singh( Yogi) Direct (205)-775-0773 244 Fifth Avenue, Suite R295, New York, NY 10001 Email: [email protected] Web: http://elgebra.com/ https://www.linkedin.com/in/yogesh-pratap-singh-282a744a/ Donate Red || Save Blue || Spread Green Keywords: information technology ffive New York C2C Role : DataLake/ Hadoop Admin || Remote in the USA [email protected] http://bit.ly/4ey8w48 https://jobs.nvoids.com/job_details.jsp?id=2305729&uid= |
[email protected] View All |
11:50 PM 01-Apr-25 |