We are looking for an experienced Senior Data Engineer to design and build scalable data pipelines and data platforms. The role involves working on high-volume data processing systems, ensuring performance, reliability, and efficiency across both batch and real-time workflows. The position is fully remote (based in Israel) and part of a global engineering team.
Key Responsibilities
- Design, develop, and maintain scalable data pipelines (ETL/ELT)
- Build systems for data ingestion, transformation, and processing
- Work with event-driven architectures using Kafka
- Deploy and manage data workloads on Kubernetes-based environments
- Optimize pipelines for performance, scalability, and cost efficiency
- Ensure data quality, validation, and reliability
- Collaborate with DevOps teams on CI/CD and infrastructure scaling
- Mentor junior engineers and promote best engineering practices
Required Skills & Experience (Must Have)
- 5–8+ years of experience in Data Engineering or Backend Engineering
- Strong programming skills in Python, Java, or Scala
- Hands-on experience with Apache Kafka (event streaming and processing)
- Strong experience with Kubernetes (container orchestration and scaling)
- Experience building distributed data processing systems
- Strong knowledge of ETL/ELT pipelines and data workflows
- Proficiency in SQL and data modeling
- Experience with cloud platforms (AWS, Azure, or GCP)
Nice to Have
- Experience with real-time streaming pipelines
- Exposure to data warehousing and analytics platforms
- Familiarity with CI/CD and infrastructure automation
- Working knowledge of observability tools such as Prometheus and Grafana for pipeline monitoring and troubleshooting
- Understanding of data governance and security practices
Success Metrics
- Pipeline performance and throughput
- System reliability and uptime
- Data accuracy and quality
- Efficiency and cost optimization