*No C2C, please. H1B and OPT visas are NOT accepted for this role.*
“Job Title: Hadoop Developer
Contract term: 18-24 months with possible conversion to FTE
Work Model – Hybrid
LOCATIONS:
– Kennesaw, GA
– Chandler, AZ
MUST-HAVE SKILLS:
– Cloudera
– Hadoop
– Hive
-MySQL
– Python
– Pyspark
*No C2C, please. H1B and OPT visas are NOT accepted for this role.*
POSITION SUMMARY:
As a Cloudera Developer, you will be responsible for developing and maintaining data solutions using the Cloudera platform. You will work closely with cross-functional teams, including data engineers, data scientists, and business stakeholders, to understand data requirements and deliver robust and scalable data solutions. Your primary focus will be on designing, developing, and implementing data processing pipelines and data ingestion frameworks. The ideal candidate is adept at using pyspark, ezflow, and other methods for moving data within a Hadoop environment.
Key Responsibilities:
• Utilize multiple architectural components in design and development of client requirements.
• Maintain, improve, clean and manipulate data for the operational and/or analytics data systems.
• Constantly looking for better ways of solving technical problems and designing the solution without being afraid of challenging the status quo.
• Document and communicate required information for deployment, maintenance, support, and business functionality.
• Adhere to team delivery/release process and cadence pertaining to code deployment and release
· They must have a proven ability to drive business results with their data-based insights
· The right candidate will have a passion for discovering solutions hidden in large data sets and working with stakeholders to improve business outcomes.
• Design, develop, and maintain data processing pipelines using Cloudera technologies such as Apache Hadoop, Apache Spark, Apache Hive, and Python.
• Collaborate with data engineers and data scientists to understand data requirements and translate them into technical specifications.
• Develop and maintain data ingestion frameworks for efficiently extracting, transforming, and loading data from various sources into the Cloudera platform.
• Optimize and tune data processing jobs to ensure high performance and scalability.
• Implement data governance and security policies to ensure data integrity and compliance.
• Monitor and troubleshoot data processing jobs to identify and resolve issues in a timely manner.
• Perform unit testing and debugging of data solutions to ensure high quality and reliability.
• Document technical specifications, data flows, and data architecture diagrams.
• Stay updated with the latest advancements and best practices in Cloudera technologies and big data analytics.
Qualifications:
• Experience with Agile Methodologies
• Add experience with Python scripting for data engineering to qualifications.
• Ability to pick new products quickly to understand their data
• Strong problem-solving skills with an emphasis on product development.
• Experience working with and creating data architectures.
• Excellent written and verbal communication skills for coordinating across teams.
• A drive to learn and master new technologies and techniques.
• Experience working with AutoSys
• 5-7 years of experience with distributed data/computing tools: Hadoop, Hive, MySQL, etc.”
To apply for this job please visit itjobpro.com.