Remote Senior Data Engineer - Datalake (Remote)

Posted

This job is closed

This job post is closed and the position is probably filled. Please do not apply.  Automatically closed by a robot after apply link was detected as broken.

Description:

  • The Senior Data Engineer - Datalake is responsible for analyzing models, designing, creating, modifying, and supporting complex systems, processes, or operations to enable optimal business capabilities.
  • This role involves creating Application and System Design Documents and developing Applications, Reports, Systems, and Enterprise Solutions.
  • The engineer will estimate component/application system level and Enterprise solution work efforts.
  • Responsibilities include creating RFI/RFP requests/responses for vendor product evaluations and designing, developing, and implementing complex business rules.
  • The position requires fulfilling end-user requests and providing on-call support as needed, while also guiding and assisting in training less experienced individuals.
  • The engineer will deliver personal tasks on time and lead the delivery of tasks for natural or cross-functional workgroups.
  • Participation in initiatives with deliverables is expected, ensuring quality standards are met on time.
  • The role also involves leading cross-functional initiatives with deliverables and meeting quality standards on time.

Requirements:

  • A Bachelor’s degree or equivalent work experience is required.
  • Candidates must have 5+ years of experience in a Data Engineering or ETL Development role.
  • Strong experience with PySpark and Python for building solid data pipelines is essential.
  • Experience with Iceberg, Hive, S3, and Trino is required.
  • Hands-on experience with Hadoop ecosystems, relational databases, and SQL queries is necessary.
  • Familiarity with Apache Ranger and Rancher/Kubernetes is preferred.
  • Experience with Talend, Red Point, or other ETL technologies is an advantage.
  • Candidates should have experience with Agile Software Development methodologies.
  • Proficiency in GitLab, CI/CD processes, and ServiceNow is required.
  • Solid programming skills in object-oriented/functional scripting languages like Python and PySpark for building data pipelines are required, along with experience in testing and logging for quality assurance.
  • Experience in distributed systems and parallel data processing using big data tools such as Spark, PySpark, Hadoop, Kafka, and Hive is required.
  • Strong knowledge of Linux/Unix-based computer systems is necessary.
  • Experience in building Data Processing pipelines using ETL tools like Talend or SSIS is required.
  • Understanding of Machine Learning models and algorithms is beneficial.
  • Proficiency in data visualization tools such as Tableau or matplotlib is a plus.
  • AWS cloud experience in Redshift, Lambda, Sage Maker, and Glue is nice to have.
  • Experience with building Rest APIs is preferred.
  • Excellent data analytical, conceptual, and problem-solving skills are required.
  • Strong communication skills to promote cross-team collaboration are essential.

Benefits:

  • The base salary for this position ranges from $100,000 to $130,000, depending on skill level, cost of living, experience, and responsibilities.
  • Vericast offers a generous total rewards benefits package that includes medical, dental, and vision coverage.
  • A 401K plan with company match is provided, along with a generous PTO allowance.
  • Additional benefits include life insurance, employee assistance programs, and pet insurance.
  • Employees can enjoy a supportive work environment with smart and friendly coworkers.
About the job
Posted on
Job type
Salary
$ 100,000 - 130,000 USD / year
Leave a feedback