Overview
Skills
Job Details
We're looking for a Data Engineer - Scala/Spark with deep expertise in big data processing and distributed systems to join our growing team! If you're passionate about building scalable data solutions and thrive in fast-paced environments, this could be the perfect opportunity.
Key Responsibilities:
Develop, test, and deploy data processing applications using Apache Spark and Scala.
Optimize and tune Spark applications for better performance on large-scale data sets.
Work with the Cloudera Hadoop ecosystem (e.g., HDFS, Hive, Impala, HBase, Kafka) to build data pipelines and storage solutions.
Collaborate with data scientists, business analysts, and other developers to understand data requirements and deliver solutions.
Design and implement high-performance data processing and analytics solutions.
Ensure data integrity, accuracy, and security across all processing tasks.
Troubleshoot and resolve performance issues in Spark, Cloudera, and related technologies.
Implement version control and CI/CD pipelines for Spark applications.