Overview
We are seeking a Cloudera Developer to build and optimize scalable Big Data solutions using Cloudera Data Platform (CDP), NiFi, Spark, and related technologies. This role focuses on hands-on development, data pipeline engineering, and performance optimization within distributed environments.
Responsibilities
- Develop and maintain data pipelines and workflows using CDP, Apache NiFi, and Apache Spark
- Implement data transformation and processing solutions for batch and real-time use cases
- Write and optimize queries using Apache Hive and HDFS-based data systems
- Collaborate with architects and stakeholders to translate requirements into technical implementations
- Troubleshoot and resolve issues in distributed data processing environments
- Support platform deployment, upgrades, and configurations where required
- Ensure data quality, integrity, and performance across systems
- Contribute to technical documentation and knowledge sharing
Requirements
- 3–6 years of experience in Big Data development
- Hands-on experience with Cloudera (CDP), Apache NiFi, Apache Spark
- Strong knowledge of data processing, ETL, and data integration concepts
- Experience working with HDFS and Apache Hive
- Familiarity with Linux environments and distributed systems
- Understanding of Java / Python / Scala ecosystems
- Basic knowledge of security configurations (LDAP, Kerberos) is a plus
- Good problem-solving and debugging skills
Nice to Have
- Exposure to streaming technologies (Kafka, Spark Streaming)
- Experience with CI/CD pipelines
- Cloudera or Hadoop-related certifications
