This job post is closed and the position is probably filled. Please do not apply.
🤖 Automatically closed by a robot after apply link
was detected as broken.
Description:
The Senior Data Engineer - Datalake is responsible for analyzing models, designing, creating, modifying, and supporting complex systems, processes, or operations to enable optimal business capabilities.
Key duties include creating Application and System Design Documents, developing applications, reports, systems, and enterprise solutions.
The role involves estimating component/application system level and enterprise solution work efforts, creating RFI/RFP requests/responses for vendor product evaluations, and designing, developing, and implementing complex business rules.
The engineer will fulfill end-user requests, provide on-call support as required, and assist in training less experienced individuals.
The position requires delivering personal tasks on time and leading the delivery of tasks for natural or cross-functional workgroups.
Participation in initiatives with deliverables and meeting quality standards on time is essential, as well as leading cross-functional initiatives.
Requirements:
A Bachelor’s degree or equivalent work experience is required.
Candidates must have 5+ years of experience in Data Engineering or ETL Development roles.
Strong experience with PySpark and Python for building solid data pipelines is necessary.
Experience with Iceberg, Hive, S3, and Trino is required.
Hands-on experience with Hadoop ecosystems, relational databases, and SQL queries is essential.
Familiarity with Apache Ranger, Rancher/Kubernetes is preferred.
Experience with Talend, Red Point, or other ETL technologies is an advantage.
Knowledge of Agile Software Development methodologies is required.
Proficiency in GitLab, CI/CD processes, and ServiceNow is necessary.
Solid programming skills in object-oriented/functional scripting languages like Python and PySpark are required.
Experience in distributed systems and parallel data processing using big data tools such as Spark, PySpark, Hadoop, Kafka, and Hive is essential.
Strong knowledge of Linux/Unix-based computer systems is required.
Experience in building Data Processing pipelines using ETL tools like Talend and SSIS is necessary.
Understanding of Machine Learning models and algorithms is preferred.
Proficiency in data visualization tools such as Tableau and matplotlib is a plus.
AWS cloud experience in Redshift, Lambda, Sage Maker, and Glue is nice to have.
Experience with building Rest APIs is beneficial.
Excellent data analytical, conceptual, and problem-solving skills are required.
Strong communication skills to promote cross-team collaboration are essential.
Benefits:
The base salary for this position ranges from $100,000 to $130,000, depending on skill level, cost of living, experience, and responsibilities.
Vericast offers a generous total rewards benefits package that includes medical, dental, and vision coverage.
A 401K plan with company match and a generous PTO allowance are provided.
Additional benefits include life insurance, employee assistance programs, and pet insurance.
Employees can enjoy a supportive work environment with smart and friendly coworkers.