We are looking for a skilled Spark Scala Developer to join our dynamic team. The ideal candidate will have a deep understanding of big data technologies, particularly Apache Spark, and experience in Scala programming. You will be responsible for developing, optimizing, and maintaining our data processing pipelines, ensuring data integrity, performance, and scalability.
- Design and Develop: Implement high-performance data processing pipelines using Apache Spark and Scala.
- Optimize: Enhance the performance of existing data processing jobs by tuning Spark jobs, improving data storage, and optimizing the overall data flow.
- Data Integration: Work with various data sources, including structured, semi-structured, and unstructured data, to integrate them into the data processing framework.
- Collaboration: Collaborate with data engineers, data scientists, and other stakeholders to define requirements, refine processes, and ensure successful project delivery.
- Testing: Develop unit tests and integration tests to ensure the reliability and accuracy of the data processing pipelines.
Requirements
- 5+ Years as Spark Scala Developer
- Min 5 Years experience in Scala Programming
- Proficient in Scala, with experience in functional programming paradigms.
- Strong experience in Apache Spark and Scala programming.
- Familiarity with Hadoop, HDFS, Hive, or similar big data technologies.
- Experience with ETL processes, data integration, and handling large-scale datasets.
- Understanding of data warehousing concepts and principles.
- Proficiency in SQL and data query optimization.
- Experience with real-time data processing using Kafka or similar technologies is a plus.
- Experience with version control systems like Git.
- Familiarity with CI/CD pipelines and tools like Jenkins or similar.
- Mandatory Skills : Min 5+ Years in Scala
•
Last updated on Aug 11, 2024