Remote Senior Data Engineer / Architect

at Enable Data Incorporated

Posted 4 days ago 2 applied

Description:

  • The position is for a Senior Data Engineer / Architect with over 8 years of experience.
  • The role is remote, allowing for flexible work arrangements.
  • Responsibilities include designing, developing, and maintaining scalable and robust data solutions in the cloud using Apache Spark and Databricks.
  • The candidate will gather and analyze data requirements from business stakeholders and identify opportunities for data-driven insights.
  • Building and optimizing data pipelines for data ingestion, processing, and integration using Spark and Databricks is a key task.
  • Ensuring data quality, integrity, and security throughout all stages of the data lifecycle is essential.
  • The role involves collaborating with cross-functional teams to design and implement data models, schemas, and storage solutions.
  • The candidate will optimize data processing and analytics performance by tuning Spark jobs and leveraging Databricks features.
  • Providing technical guidance and expertise to junior data engineers and developers is expected.
  • Staying up to date with emerging trends and technologies in cloud computing, big data, and data engineering is important.
  • The candidate will contribute to the continuous improvement of data engineering processes, tools, and best practices.

Requirements:

  • A Bachelor’s or master’s degree in computer science, engineering, or a related field is required.
  • The candidate must have 10+ years of experience as a Data Engineer, Software Engineer, or in a similar role, focusing on building cloud-based data solutions.
  • Strong knowledge and experience with the Azure cloud platform, Databricks, EventHub, Architecture, Spark, Kafka, ETL Pipeline, Python/Pyspark, and SQL are necessary.
  • The candidate should have strong experience with cloud platforms, particularly Azure.
  • Experience with big data systems, including Apache Spark and Kafka, is required.
  • The candidate must have experience contributing to the architecture and design of large-scale distributed systems.
  • Expertise in the Databricks Lakehouse Platform, its architecture, and its capabilities is essential.
  • Experience building production pipelines using Databricks and Azure services is required.
  • Proficiency in multiple coding languages such as Python or SQL is necessary.

Benefits:

  • The position offers the flexibility of remote work.
  • Candidates can join immediately or within a specified notice period, with options for permanent or contract roles.
  • The role provides opportunities for professional growth and development in a cutting-edge field.
  • The candidate will have the chance to work with advanced technologies and contribute to innovative data solutions.

Get realtime job alerts

Be the first to know about new jobs