This job post is closed and the position is probably filled. Please do not apply.
π€ Automatically closed by a robot after apply link
was detected as broken.
Description:
We are seeking a Staff Data Engineer to build the next generation of data pipelines and applications using emerging technologies, including generative AI and large language models.
The role involves operational management of large language models in production environments.
The Staff Data Engineer will design data integration frameworks and pipelines while being hands-on when necessary.
Responsibilities include designing, building, and scaling data pipelines across various source systems and streams, including internal, third-party, and cloud-based systems.
A deep understanding of Machine Learning best practices and algorithms is required.
The candidate should have a solid understanding of data modeling, warehousing, and architecture principles.
The role requires collaboration with cross-functional teams to understand data requirements and develop efficient data acquisition and integration strategies.
The Staff Data Engineer will coach and guide others in building scalable pipelines based on foundational data engineering principles.
The position involves developing automated tests for code and creating documentation for developers and business users.
The candidate will deploy to cloud-based platforms and troubleshoot application, cloud, and configuration issues as necessary.
Requirements:
A minimum of 8 years of experience in a data engineering role, with a strong emphasis on leading data engineering teams is required.
The candidate must be able to think critically about problems and effectively use resources to find solutions.
Proven ability to thrive in an agile development environment and adapt to changing priorities is essential.
Understanding of REST-based APIs, vectorized embeddings, and other Retrieval Augmented Generation AI workload components is necessary.
Direct experience with data modeling, ETL/ELT development principles, cloud development, and data warehousing concepts is required.
Knowledge of cloud technologies such as AWS, Azure, and GCP is essential.
Experience in building data pipelines using Python/SQL or similar programming languages is required.
A Bachelor's degree in computer science, Data Science, Statistics, Informatics, Information Systems, or a related field is necessary.
Benefits:
This position is fully remote, allowing employees to work from their residence.
The role is eligible for company-sponsored benefits, including medical, dental, and vision insurance.
Employees will have access to a 401(k) plan, paid leave, and tuition reimbursement.
A variety of discounts and perks are available to employees.
The salary range for this position is $130,000 - $170,000 and is bonus eligible.