This job post is closed and the position is probably filled. Please do not apply.
🤖 Automatically closed by a robot after apply link
was detected as broken.
Description:
The Senior Data Engineer - Datalake is responsible for analyzing models, designing, creating, modifying, and supporting complex systems, processes, or operations to enable optimal business capabilities.
This role involves creating Application and System Design Documents and developing Applications, Reports, Systems, and Enterprise Solutions.
The engineer will estimate component/application system level and Enterprise solution work efforts.
Responsibilities include creating RFI/RFP requests/responses for vendor product evaluations and designing, developing, and implementing complex business rules.
The position requires fulfilling end-user requests and providing on-call support as needed, while also guiding and assisting in training less experienced individuals.
The engineer will deliver personal tasks on time and lead the delivery of tasks for natural or cross-functional workgroups.
Participation in initiatives with deliverables is expected, ensuring quality standards are met on time.
The role also involves leading cross-functional initiatives with deliverables and meeting quality standards on time.
Requirements:
A Bachelor’s degree or equivalent work experience is required.
Candidates must have 5+ years of experience in a Data Engineering or ETL Development role.
Strong experience with PySpark and Python for building solid data pipelines is essential.
Experience with Iceberg, Hive, S3, and Trino is required.
Hands-on experience with Hadoop ecosystems, relational databases, and SQL queries is necessary.
Familiarity with Apache Ranger and Rancher/Kubernetes is preferred.
Experience with Talend, Red Point, or other ETL technologies is an advantage.
Candidates should have experience with Agile Software Development methodologies.
Proficiency in GitLab, CI/CD processes, and ServiceNow is required.
Solid programming skills in object-oriented/functional scripting languages like Python and PySpark for building data pipelines are required, along with experience in testing and logging for quality assurance.
Experience in distributed systems and parallel data processing using big data tools such as Spark, PySpark, Hadoop, Kafka, and Hive is required.
Strong knowledge of Linux/Unix-based computer systems is necessary.
Experience in building Data Processing pipelines using ETL tools like Talend or SSIS is required.
Understanding of Machine Learning models and algorithms is beneficial.
Proficiency in data visualization tools such as Tableau or matplotlib is a plus.
AWS cloud experience in Redshift, Lambda, Sage Maker, and Glue is nice to have.
Experience with building Rest APIs is preferred.
Excellent data analytical, conceptual, and problem-solving skills are required.
Strong communication skills to promote cross-team collaboration are essential.
Benefits:
The base salary for this position ranges from $100,000 to $130,000, depending on skill level, cost of living, experience, and responsibilities.
Vericast offers a generous total rewards benefits package that includes medical, dental, and vision coverage.
A 401K plan with company match is provided, along with a generous PTO allowance.
Additional benefits include life insurance, employee assistance programs, and pet insurance.
Employees can enjoy a supportive work environment with smart and friendly coworkers.