This position is for a Data Engineer at Harvard Business Publishing, located in the United States, and is offered as a remote role.
The role involves designing, building, and maintaining robust data infrastructure to support large-scale data processing and analysis across various business functions.
As a key member of the data engineering team, the Data Engineer will collaborate with product engineers, data scientists, and AI developers to create scalable data pipelines and models that provide critical business insights and enhance user-facing applications.
Responsibilities include leading the entire data engineering lifecycle, which encompasses requirements gathering, technical design, development, deployment, and production support.
The Data Engineer will design, develop, and maintain scalable, secure, and high-performance ELT/ETL pipelines and data warehouses.
Collaboration with analytics, data science, and business teams is essential to create and evolve data models for dashboards, AI/ML projects, and SaaS products.
The role requires producing clear technical documentation, including system diagrams, data flow charts, and runbooks to aid in decision-making and maintenance.
The Data Engineer will evaluate design trade-offs to balance cost, performance, flexibility, and supportability, ensuring alignment with IT strategy and scalability goals.
Ensuring smooth project transitions to operations with comprehensive documentation and knowledge transfer is also a key responsibility.
Requirements:
A minimum of 5 years of experience in data engineering or related fields, specifically in delivering production-grade data pipelines and platforms, is required.
Expertise in designing, building, and operating high-volume ELT/ETL data pipelines from diverse sources is essential.
Strong SQL skills, particularly in Snowflake SQL, with experience in complex analytics and DBT standards are necessary.
Proficiency in developing modular dbt models, implementing tests, documentation, and enforcing coding standards via CI/CD is required.
Hands-on experience with data modeling and maintaining data warehouses on cloud platforms such as Snowflake is expected.
Experience with data quality frameworks to ensure accuracy, consistency, and lineage is important.
Excellent communication and stakeholder management skills are required to translate business needs into technical solutions.
An analytical mindset with the ability to assess trade-offs and recommend optimal data architectures is necessary.
Proven ability to monitor pipelines, diagnose issues, and provide production support with root-cause analysis is essential.
Proficiency in Python programming is required.
Benefits:
The position offers a competitive salary range of $140,000 - $150,000, with eligibility for performance-based variable pay.
The company promotes an inclusive culture that fosters trust, engagement, and a sense of belonging among employees.
There are programs supporting career development and employee wellness, including education reimbursement and early-release Summer Fridays.
A comprehensive benefits package is provided, with a strong focus on work-life balance.