This job post is closed and the position is probably filled. Please do not apply.
🤖 Automatically closed by a robot after apply link
was detected as broken.
Description:
As a Staff Data Engineer, you will act as the data lake subject matter expert and develop the vision for the data engineering architecture.
You will be responsible for developing data products and platforms that create flexible, scalable, and reliable data pipelines with high accuracy and quality.
You will own and develop the long-term technical vision and blueprint for the team.
Your role includes designing a technical strategy to develop a well-architected data lakehouse.
You will collaborate with internal architects to design and build the ELT process from data ingestion to analytics marts.
You will create reliable, reusable frameworks and abstractions to standardize software development.
You will provide support to development teams while mentoring engineering team members on modern database architecture principles and best practices.
Your responsibilities include monitoring and analyzing database performance and following best practices of data engineering such as code reviews, scrum, and SDLC.
You will develop schema design for reports and analytics.
You will engage in hands-on development from microservices and sub-systems to the entire technical stack.
You will deal with ambiguity and figure things out with minimal guidance.
You will mentor junior engineers on the team and create standards for engineering and operational excellence.
You will identify and educate on industry-relevant technical trends.
Requirements:
You must have 7+ years of experience in a Data Engineering role, with experience in data warehouses, fetching data from APIs, and developing and maintaining ETL/ELT processes.
You should have experience creating data products and internal platforms that accelerate the development of data pipelines.
Advanced SQL experience and database design skills are required.
You must have experience building and maintaining pipeline monitoring for latency, traffic, saturation, and errors.
You should demonstrate advanced computer and analytical skills, particularly with GCP or equivalent cloud-based data lake/OLAP/OLTP environments.
Familiarity with APIs, Airflow, Cron, dbt, and Git is necessary.
Experience creating microservices for data passing is required.
You should have experience working with CDC (PostGres) and technologies like Kafka.
Working experience with Software Engineering development and deployment practices is essential.
You must be comfortable with agile methodologies such as Scrum or Kanban and work in iterative product-driven cycles.
Proficiency in object-oriented/object function scripting languages such as Python or Scala is required.
Experience with CI/CD processes and source control tools like Github is necessary.
Additional preferred experience includes handling unstructured text and data, infrastructure as code (ideally Terraform), non-tabular database management systems (e.g., MongoDB, NeptuneDB), BI dashboards (e.g., Tableau or PowerBI), implementing machine learning systems, and utilizing logging/observability software such as DataDog.
Benefits:
You will receive insurance or subsidies based on your country of residence.
The position offers unlimited time off, with a minimum time off recommendation.
You will enjoy 10 company-paid holidays.
There is an official company-wide holiday for the last week of the calendar year.
You will have access to free data skills training through the company's programs.
The company culture empowers individuals and embraces diversity through its core mission.