This job post is closed and the position is probably filled. Please do not apply.
🤖 Automatically closed by a robot after apply link
was detected as broken.
Description:
Oversee the operations of the data warehouse on Databricks within AWS, ensuring optimal performance and reliability.
Continuously optimize data warehouse performance, including query performance, storage, and data processing efficiency.
Design, develop, and maintain ETL (Extract, Transform, Load) processes to integrate data from various sources into the data warehouse.
Manage and optimize data pipelines to ensure timely and accurate data processing and delivery.
Implement and monitor data quality checks to ensure accuracy, completeness, and consistency of data within the warehouse.
Set up and manage monitoring systems to detect and address data issues, performance bottlenecks, and system anomalies.
Collaborate with cross-functional teams to design and architect scalable, high-performance data solutions aligned with business objectives.
Provide technical support and expertise on data-related issues and queries from other teams.
Create and maintain comprehensive documentation for data processes, ETL jobs, and data warehouse configurations.
Implement and promote best practices for data engineering, including coding standards, data governance, and security measures.
Requirements:
Proven experience as a Data Engineer, with significant experience working with data warehouses, ETL processes, and data quality management.
Strong proficiency in Databricks, AWS services, SQL, Python, Spark, Fivetran, Segment event tracking, and ETL tools.
Excellent analytical and problem-solving skills with a strong ability to derive insights from complex data sets.
Effective communication skills with the ability to convey technical information to non-technical stakeholders.
Bachelor’s degree in Computer Science, Engineering, or a related field, or equivalent working experience.