Description:
Design, develop, and maintain scalable ETL/ELT data pipelines for batch and real-time processing. Build and manage real-time streaming solutions using Apache Kafka and similar technologies. Develop high-performance data workflows using Apache Spark (PySpark/Scala) and Structured Streaming. Architect and maintain modern data platforms including data lakes and data warehouses. Optimize Spark jobs, SQL queries, and streaming workloads for performance and cost efficiency. Implement data quality chec
Feb 25, 2026;
from:
dice.com