Job Type
Work Type
Location
Experience
We are seeking a highly skilled Senior Cloudera Engineer to design, build, optimize, and operate enterprise-scale data platforms based on the Cloudera Data Platform (CDP). The role requires deep hands-on expertise in Cloudera administration, big data engineering, security, performance tuning, and hybrid/on-prem environments, supporting mission-critical analytics, AI, and data-driven workloads for large enterprises and government customers. This is a key technical role with ownership across design, implementation, operations, security, and optimization of Cloudera platforms. Key Responsibilities 1. Cloudera Platform Design & Implementation Design, deploy, and configure Cloudera Data Platform (CDP) environments (on-prem and hybrid) Implement and manage core Cloudera components including: HDFS, YARN, Hive, Impala, HBase Spark, Kafka Ozone, Kudu (where applicable) Plan and execute cluster sizing, capacity planning, and scalability Support platform upgrades, migrations, and patch management 2. Administration, Operations & Support Perform day-to-day administration and monitoring of Cloudera clusters Ensure high availability, reliability, and performance of data services Troubleshoot complex production issues across compute, storage, and network layers Define and maintain operational SOPs, runbooks, and documentation Provide Level 3 / expert support and mentor junior engineers 3. Security, Governance & Compliance Implement and manage Kerberos, Ranger, Atlas, and Knox Enforce data access controls, encryption, auditing, and compliance Integrate with enterprise LDAP / Active Directory Support regulatory, sovereign, and data residency requirements 4. Performance Tuning & Optimization Tune Hive, Impala, Spark, and HDFS for optimal performance Optimize resource management using YARN and workload management Analyze bottlenecks and improve query performance and cluster efficiency Reduce operational costs through smart capacity and workload planning 5. Data Engineering & Integration Support ingestion pipelines using Kafka, NiFi, Sqoop, Flume Work closely with data engineers, data scientists, and application teams Enable analytics, AI/ML, and reporting workloads Integrate Cloudera with BI tools, AI platforms, and external systems 6. Automation & DevOps Automate cluster deployment and operations using: Ansible, Terraform (preferred) Shell scripting, Python Implement CI/CD best practices for data platforms Support containerized and Kubernetes-based integrations where required Key Success Metrics Platform availability and stability Performance and query response improvements Security and compliance adherence Reduction in operational incidents Successful upgrades and migrations Mandatory Requirements Bachelor’s degree in Computer Science, Engineering, or related field 8–12+ years of experience in Big Data platforms 5+ years of hands-on experience with Cloudera Strong experience with: HDFS, Hive, Impala, Spark Kafka, NiFi Linux system administration Hands-on experience with Cloudera Security (Kerberos, Ranger, Atlas) Strong troubleshooting and performance tuning expertise Experience supporting large-scale, production-grade environments CDP Private Cloud or CDP Public Cloud exposure Experience with AI/ML workloads on Cloudera Knowledge of Ozone, Kudu Exposure to cloud platforms (AWS / Azure / GCP) Cloudera Certification (CCA, CCP) Experience in government, oil & gas, or regulated environments Soft Skills & Attributes Strong analytical and problem-solving abilities Excellent communication and documentation skills Ability to work in high-pressure, mission-critical environments Ownership mindset with attention to detail Team player with mentoring capability