This job post is closed and the position is probably filled. Please do not apply.
π€ Automatically closed by a robot after apply link
was detected as broken.
Description:
Design, build, and maintain scalable and efficient data pipelines for ingesting, processing, and transforming large volumes of structured and unstructured data.
Implement data models and schemas to support analytical and operational requirements, ensuring data integrity, consistency, and performance.
Build and optimize data warehouses and data lakes to store and manage structured and unstructured data for analytics and reporting purposes.
Integrate data from disparate sources and systems, ensuring data consistency, quality, and completeness.
Establish and enforce data governance policies and best practices to ensure data quality, security, and compliance.
Optimize data pipelines and queries for performance and efficiency, improving system scalability and reliability.
Implement monitoring and alerting systems to track data pipeline performance and health, minimizing downtime and data loss.
Document data infrastructure and pipelines to facilitate understanding and collaboration among team members.
Collaborate with cross-functional teams to understand requirements and deliver data solutions that meet business needs.
Mentor junior engineers, providing guidance, support, and technical leadership in data engineering best practices and technologies.
Requirements:
Bachelor's degree or higher in Computer Science, Engineering, Mathematics, or related field.
5+ years of experience in data engineering, focusing on designing, building, and maintaining data infrastructure and pipelines.
Proficiency in programming languages like Python, Java, or Scala, and experience with data engineering frameworks and tools such as Apache Spark, Apache Kafka, Apache Airflow, or similar.
Strong understanding of data modeling concepts and techniques, with experience designing and implementing data models and schemas for relational and non-relational databases.
Experience with cloud platforms like AWS, Azure, or Google Cloud Platform, and familiarity with cloud-based data services such as Amazon Redshift, Google BigQuery, or Azure Synapse Analytics.
Familiarity with SQL and NoSQL databases, data warehousing, and ETL/ELT processes.
Strong problem-solving skills and analytical thinking to troubleshoot complex data issues and optimize system performance.
Excellent communication and collaboration skills to work effectively in cross-functional teams and communicate technical concepts to non-technical stakeholders.
Benefits:
Competitive salary ranging from $170,000 to $230,000 per year, depending on experience and qualifications.
Comprehensive health, dental, and vision insurance plans.
Flexible work hours and remote work options.
Generous vacation and paid time off.
Professional development opportunities, including access to training programs, conferences, and workshops.
State-of-the-art technology environment with access to cutting-edge tools and resources.
Vibrant and inclusive company culture with opportunities for growth and advancement.
Exciting projects with real-world impact at the forefront of data-driven innovation.