Remote Senior Data Engineer

Posted

Apply now
Please, let Wave HQ know you found this job on RemoteYeah. This helps us grow 🌱.

Description:

  • The Senior Data Engineer position is located in Toronto, Ontario, and is a full-time remote role within the Data & AI team.
  • The role involves building tools and infrastructure to support the Data Products and Insights & Innovation teams, as well as the business as a whole.
  • The ideal candidate is a talented, curious self-starter who is driven to solve complex problems and can manage multiple domains and stakeholders.
  • Responsibilities include designing, building, and deploying components of a modern data stack, including CDC ingestion using Debezium, a centralized Hudi data lake, and various data pipelines.
  • The engineer will help manage a fault-tolerant data platform, maintain legacy Python ELT scripts, and transition to dbt models in Redshift.
  • Collaboration with cross-functional teams is essential to plan and roll out data infrastructure and processing pipelines for analytics, machine learning, and GenAI services.
  • The role requires self-motivation and the ability to work autonomously, identifying opportunities to optimize pipelines and improve data workflows.
  • The engineer will respond to PagerDuty alerts, implement monitoring solutions, and ensure high availability and reliability of data systems.
  • Strong communication skills are necessary to assist technical and non-technical audiences and to help internal teams surface actionable insights for customer satisfaction.

Requirements:

  • Candidates must have 3+ years of experience in building data pipelines and managing a secure, modern data stack, including CDC streaming ingestion using tools like Debezium into a Hudi data lake.
  • At least 3 years of experience with AWS cloud infrastructure, including Kafka (MSK), Spark/AWS Glue, and infrastructure as code (IaC) using Terraform is required.
  • Strong coding skills in Python, SQL, and dbt are necessary, with the ability to write and review high-quality, maintainable code.
  • Prior experience in building data lakes on S3 using Apache Hudi with various file formats (Parquet, Avro, JSON, CSV) is essential.
  • Candidates should have experience in building and managing multi-stage workflows using serverless Lambdas and AWS Step Functions.
  • Familiarity with data governance practices, including data quality, lineage, and privacy, is required, along with experience using cataloging tools.
  • Experience with CI/CD best practices for developing and deploying data pipeline solutions is necessary.
  • Working knowledge of data integration tools such as Stitch and Segment CDP is required.
  • Knowledge and practical experience with analytical and ML tools like Athena, Redshift, or Sagemaker Feature Store is a bonus.

Benefits:

  • Employees have the flexibility to work from the office in downtown Toronto or remotely, depending on their preference.
  • The company invests in employee health and wellness, offering comprehensive benefits that address body, mind, and soul.
  • There are diverse learning experiences, educational allowances, and mentorship opportunities to support personal and professional growth.
  • Fair compensation and various office perks are provided, along with the expected benefits of a growing tech company.
  • Wave fosters an inclusive culture that values diverse perspectives and encourages open feedback, ensuring every team member has the opportunity to grow.
  • The company has been recognized as one of Canada's Top Ten Most Admired Corporate Cultures and a great place to work in various categories.
Apply now
Please, let Wave HQ know you found this job on RemoteYeah . This helps us grow 🌱.
About the job
Posted on
Job type
Salary
-
Report this job

Job expired or something else is wrong with this job?

Report this job
Leave a feedback