Remote Senior Data Engineer

at Polly

Posted 1 day ago 2 applied

Description:

  • The Senior Data Engineer will design, build, and maintain reliable data pipelines to support internal analytics and customer-facing product features.
  • This role involves architecting, developing, and maintaining robust data infrastructure using structured (Postgres) and non-structured databases (MongoDB).
  • The engineer will write, optimize, and maintain extensive SQL queries and scripts to support complex data retrieval and analysis.
  • Responsibilities include designing, developing, and managing ETL processes using APIs to integrate various data sources seamlessly into the analytics environment.
  • The position requires performing database optimization, performance tuning, and capacity planning to ensure maximum efficiency and reliability.
  • The engineer will own data integrations across systems, ensuring data is clean, validated, and well-modeled.
  • Collaboration with data scientists, analysts, and software engineers is essential to ensure the right data is available in the right format at the right time.
  • The role involves leveraging modern data tools and cloud infrastructure to build scalable solutions that meet business and technical requirements.
  • Troubleshooting data issues and continuously improving pipeline performance and observability are key tasks.

Requirements:

  • Candidates must have advanced proficiency in Python and SQL in a production environment, including query optimization and schema design.
  • Proven experience working programmatically with structured (Postgres) and unstructured (MongoDB) databases is required.
  • A strong background in database performance optimization and capacity planning is essential.
  • Demonstrated expertise in developing and maintaining ETLs using APIs and automation frameworks is necessary.
  • Experience with cloud platforms such as AWS, particularly services like Glue, Redshift, or S3, is required.
  • A solid understanding of data modeling and best practices for data integrity and security is needed.
  • Familiarity with big data tools such as Spark, Hive, Databricks, or similar is preferred.
  • Working knowledge of Terraform, Kubernetes, and CI/CD workflows is necessary.
  • Excellent communication and collaboration skills are required, with the ability to work cross-functionally and translate technical concepts to non-technical audiences.
  • Experience with orchestration tools like Prefect or Airflow is preferred.
  • Exposure to modern data workflows, notebooks, and warehouse modeling is beneficial.
  • Candidates must ensure data quality and integrity through proactive monitoring, testing, and validation processes.

Benefits:

  • Polly offers competitive salaries designed to reward expertise and direct impact/contributions.
  • Employees enjoy 100% company-paid medical, vision, dental, disability, and life insurance, providing peace of mind.
  • The company has a flexible, non-accrued vacation policy, allowing employees to take time off to recharge when needed.
  • Polly employees work on-site three days a week (Tues./Wed./Thurs.) at Innovation Hubs located in Dallas-Fort Worth and San Francisco, promoting a hybrid work environment.