Description:
Data Pipeline Development Design develop and implement efficient and scalable ETLELT processes using Apache Spark with Java integrating data from diverse sources into data lakes (e.g. S3) and data warehouses (e.g. Redshift). AWS Service Utilization Leverage a range of AWS services for data storage processing and analytics including but not limited to S3 Redshift Glue EMR Lambda Kinesis and DynamoDB. Optimise Spark applications and data pipelines for performance, cost-efficiency, and reliability,
Feb 6, 2026;
from:
dice.com