Remote Data Engineer

Posted

Apply now
Please, let PadSplit know you found this job on RemoteYeah. This helps us grow 🌱.

Description:

  • PadSplit is hiring for a Data Engineer to build and maintain scalable data infrastructure that drives analytics, reporting, and decision-making across the organization.
  • This role is critical to optimizing data pipelines, ensuring data reliability, and enabling cross-functional teams to unlock valuable insights in a remote, high-growth environment.
  • The Data Engineer will design, build, and optimize scalable ETL/ELT pipelines to facilitate seamless data ingestion and transformation processes.
  • They will develop and maintain data models to enable self-service analytics and reporting across the organization.
  • The role involves optimizing database performance in PostgreSQL, ensuring efficient data storage, retrieval, and query execution.
  • The Data Engineer will implement and enhance search capabilities using NoSQL technologies like ElasticSearch or Solr to improve data discovery.
  • Collaboration with data analysts to create insightful dashboards that support data-driven decision-making is essential.
  • The position requires ensuring data quality, governance, and security by adhering to best practices in cloud-based data environments.
  • Monitoring and troubleshooting issues within data pipelines, focusing on optimizing efficiency and reliability, is a key responsibility.
  • The Data Engineer will work closely with software engineers and product teams to integrate data solutions into operational workflows and product development.

Requirements:

  • Candidates must have 5+ years of experience in data engineering or a similar role, with a proven track record of designing scalable data solutions.
  • Expertise in PostgreSQL, including database management, query optimization, and performance tuning, is required.
  • Hands-on experience with AWS cloud services such as S3, Lambda, Glue, Redshift, and IAM is necessary.
  • Proficiency in data warehousing technologies like Snowflake, Redshift, or BigQuery for cloud-based data storage and analysis is essential.
  • Strong skills in data transformation, modeling, and building efficient ETL/ELT pipelines are required.
  • Experience with data visualization tools like Mode, Looker, Tableau, or Hex to support analytics and reporting is needed.
  • Knowledge of ElasticSearch or Solr for implementing search indexing and query capabilities is important.
  • Proficiency in SQL and Python, with experience in automation, scripting, and workflow orchestration (e.g., Airflow) is required.
  • Understanding of CI/CD pipelines, infrastructure-as-code principles, and cloud-based deployment practices is necessary.
  • Strong analytical and problem-solving abilities, with a passion for leveraging data-driven insights to inform decisions, are essential.
  • Nice-to-Have: Experience with streaming data solutions like Kafka or Kinesis, knowledge of machine learning pipelines, and familiarity with data privacy regulations such as GDPR or CCPA.

Benefits:

  • This is a fully remote position.
  • The company offers a competitive compensation package including an equity incentive plan.
  • National medical, dental, and vision healthcare plans are provided.
  • A company-provided life insurance policy is included.
  • Optional accidental insurances, FSA, and DCFSA benefits are available.
  • The position includes an unlimited paid-time off (PTO) policy with eleven company-observed holidays.
  • A 401(k) plan is offered.
  • Twelve weeks of paid time off for both birth and non-birth parents is provided.
  • Employees have the opportunity to do what they love at a company that is at the forefront of solving the affordable housing crisis.
Apply now
Please, let PadSplit know you found this job on RemoteYeah . This helps us grow 🌱.
About the job
Posted on
Job type
Salary
$ 135,000 - 175,000 USD / year
Report this job

Job expired or something else is wrong with this job?

Report this job
Leave a feedback