Key Skills:
Hands-on experience with Python, SQL, Spark
Expertise in ETL/ELT pipeline design and data warehousing (Snowflake / Redshift / BigQuery)
Cloud experience (AWS / Azure / GCP)
Strong knowledge of data modeling, partitioning, performance tuning
Exposure to CI/CD, Docker, Kubernetes is a plus
Responsibilities:
Build and maintain high-performance data pipelines and architectures
Collaborate with data scientists, analysts, and business teams to ensure data availability and quality
Implement best practices for data governance, quality, and security
Optimize workflows for scalability and cost-efficiency
Support real-time and batch processing requirements