Remote Staff Data Engineer

Posted

This job is closed

This job post is closed and the position is probably filled. Please do not apply.  Automatically closed by a robot after apply link was detected as broken.

Description:

  • As a Staff Data Engineer, you will act as the data lake subject matter expert and develop the vision for the data engineering architecture.
  • You will be responsible for developing data products and platforms that create flexible, scalable, and reliable data pipelines with high accuracy and quality.
  • You will own and develop the long-term technical vision and blueprint for the team.
  • Your role includes designing a technical strategy to develop a well-architected data lakehouse.
  • You will collaborate with internal architects to design and build the ELT process from data ingestion to analytics marts.
  • You will create reliable, reusable frameworks and abstractions to standardize software development.
  • You will provide support to development teams while mentoring engineering team members on modern database architecture principles and best practices.
  • Your responsibilities include monitoring and analyzing database performance and following best practices of data engineering such as code reviews, scrum, and SDLC.
  • You will develop schema design for reports and analytics.
  • You will engage in hands-on development from microservices and sub-systems to the entire technical stack.
  • You will deal with ambiguity and figure things out with minimal guidance.
  • You will mentor junior engineers on the team and create standards for engineering and operational excellence.
  • You will identify and educate on industry-relevant technical trends.

Requirements:

  • You must have 7+ years of experience in a Data Engineering role, with experience in data warehouses, fetching data from APIs, and developing and maintaining ETL/ELT processes.
  • You should have experience creating data products and internal platforms that accelerate the development of data pipelines.
  • Advanced SQL experience and database design skills are required.
  • You must have experience building and maintaining pipeline monitoring for latency, traffic, saturation, and errors.
  • You should demonstrate advanced computer and analytical skills, particularly with GCP or equivalent cloud-based data lake/OLAP/OLTP environments.
  • Familiarity with APIs, Airflow, Cron, dbt, and Git is necessary.
  • Experience creating microservices for data passing is required.
  • You should have experience working with CDC (PostGres) and technologies like Kafka.
  • Working experience with Software Engineering development and deployment practices is essential.
  • You must be comfortable with agile methodologies such as Scrum or Kanban and work in iterative product-driven cycles.
  • Proficiency in object-oriented/object function scripting languages such as Python or Scala is required.
  • Experience with CI/CD processes and source control tools like Github is necessary.
  • Additional preferred experience includes handling unstructured text and data, infrastructure as code (ideally Terraform), non-tabular database management systems (e.g., MongoDB, NeptuneDB), BI dashboards (e.g., Tableau or PowerBI), implementing machine learning systems, and utilizing logging/observability software such as DataDog.

Benefits:

  • You will receive insurance or subsidies based on your country of residence.
  • The position offers unlimited time off, with a minimum time off recommendation.
  • You will enjoy 10 company-paid holidays.
  • There is an official company-wide holiday for the last week of the calendar year.
  • You will have access to free data skills training through the company's programs.
  • The company culture empowers individuals and embraces diversity through its core mission.
Leave a feedback