Renew Home is seeking a Senior Engineer, Data to architect and deploy secure, scalable, and highly available batch and real-time data pipelines.
The role involves implementing and optimizing data lake architectures for structured and unstructured data from millions of thermostats.
The engineer will work closely with development teams to integrate data engineering services into the broader system architecture.
Collaboration with cross-functional teams consisting of engineers, data scientists, and analysts is essential to deliver clean, reliable data.
Responsibilities include analyzing and enhancing the performance of PostgreSQL Aurora and MySQL databases through query tuning, indexing strategies, and efficient resource allocation.
The engineer will strive for a 99.999% uptime SLA for the systems and participate in on-call rotations, responding to application and data infrastructure incidents.
Ensuring data quality, integrity, and compliance with best practices and governance standards is a key responsibility.
The role also involves contributing to the design and evolution of the data architecture to support growing business needs.
The engineer will work on various aspects of the tech stack, including Python, MySQL, Postgres, AWS/GCP, Redis, Git, and Jira.
Implementing application monitoring tools and proactively monitoring application performance is required.
Participation in the agile development process, including regular team updates, stand-up meetings, and one-on-ones, is expected.
Requirements:
Candidates should have 5-10 years of industry experience.
A Bachelor's or Master's degree in computer science or equivalent experience in the software industry is required.
The ideal candidate is a self-starter who takes initiative to identify improvement areas and proposes actionable enhancements.
Hands-on experience building scalable batch and real-time data pipelines using structured and unstructured data is necessary.
Experience with orchestration tools like Prefect, Airflow, or Dagster is required.
Familiarity with streaming technologies such as Apache Kafka, AWS Kinesis, Apache Flink, or GCP Pub/Sub is essential.
Strong knowledge of data lake architectures and technologies (e.g., AWS S3, AWS Glue, Delta Lake) is required.
Proven ability to analyze and optimize database performance, including query tuning and indexing strategies, preferably with Redshift and Postgres.
Proficiency in using CDK and Terraform for automating infrastructure deployment and management is necessary.
A strong software engineering background and proficiency in one or more programming languages such as Python, Java, PHP, or Ruby is required.
The candidate should have the ability to work collaboratively with development teams, providing guidance and mentorship on data infrastructure-related issues.
A commitment to staying up-to-date with the latest advancements in cloud infrastructure and database technologies is essential.
Bonuses include knowledge of containerization and orchestration tools like Docker and Kubernetes, experience with Prefect, familiarity with data warehousing best practices, exposure to machine learning pipelines, and contributions to open-source data projects or relevant certifications.
Benefits:
This is a full-time position with a competitive salary based on experience, ranging from $145k to $205k.
The role offers a fully remote work environment with a home office set-up allowance.
The company promotes a real work-life balance with no pre-set vacation limits and a culture of taking meaningful PTO.
Parental leave benefits and a corporate value of working sustainably and putting families first are included.
A competitive benefits package that includes numerous health and wellness benefits is provided.
The position includes a 401(k) plan with employer contributions.
Employees will have the opportunity to work with passionate individuals in a fully-remote environment, fostering learning and growth every day.