The Senior Data Engineer will design, build, and maintain reliable data pipelines to support internal analytics and customer-facing product features.
This role involves architecting, developing, and maintaining robust data infrastructure using structured (Postgres) and non-structured databases (MongoDB).
The engineer will write, optimize, and maintain extensive SQL queries and scripts to support complex data retrieval and analysis.
Responsibilities include designing, developing, and managing ETL processes using APIs to integrate various data sources seamlessly into the analytics environment.
The position requires performing database optimization, performance tuning, and capacity planning to ensure maximum efficiency and reliability.
The engineer will own data integrations across systems, ensuring data is clean, validated, and well-modeled.
Collaboration with data scientists, analysts, and software engineers is essential to ensure the right data is available in the right format at the right time.
The role involves leveraging modern data tools and cloud infrastructure to build scalable solutions that meet business and technical requirements.
Troubleshooting data issues and continuously improving pipeline performance and observability are key tasks.
Requirements:
Candidates must have advanced proficiency in Python and SQL in a production environment, including query optimization and schema design.
Proven experience working programmatically with structured (Postgres) and unstructured (MongoDB) databases is required.
A strong background in database performance optimization and capacity planning is essential.
Demonstrated expertise in developing and maintaining ETLs using APIs and automation frameworks is necessary.
Experience with cloud platforms such as AWS, particularly services like Glue, Redshift, or S3, is required.
A solid understanding of data modeling and best practices for data integrity and security is needed.
Familiarity with big data tools such as Spark, Hive, Databricks, or similar is preferred.
Working knowledge of Terraform, Kubernetes, and CI/CD workflows is necessary.
Excellent communication and collaboration skills are required, with the ability to work cross-functionally and translate technical concepts to non-technical audiences.
Experience with orchestration tools like Prefect or Airflow is preferred.
Exposure to modern data workflows, notebooks, and warehouse modeling is beneficial.
Candidates must ensure data quality and integrity through proactive monitoring, testing, and validation processes.
Benefits:
Polly offers competitive salaries designed to reward expertise and direct impact/contributions.
Employees enjoy 100% company-paid medical, vision, dental, disability, and life insurance, providing peace of mind.
The company has a flexible, non-accrued vacation policy, allowing employees to take time off to recharge when needed.
Polly employees work on-site three days a week (Tues./Wed./Thurs.) at Innovation Hubs located in Dallas-Fort Worth and San Francisco, promoting a hybrid work environment.