Please, let Rackspace know you found this job
on RemoteYeah.
This helps us grow 🌱.
Description:
The Cloud Data Services Team is seeking a skilled Data Architect (Abinitio) for a customer-facing role.
This position involves working with cutting-edge technologies and tools to build, optimize, and maintain large-scale data processing frameworks on the cloud.
The role requires collaboration with customers to design and deliver efficient, scalable data solutions.
Responsibilities include leveraging AWS services (S3, EMR, Athena, Glue) or equivalent cloud platforms for data storage and processing.
Proficiency with the AbInitio product suite is essential, including experience in PDL, Meta-programming, Conduct>It, Express>It, and GDE-based development.
The candidate will implement and maintain CI/CD pipelines using tools like Jenkins and GitHub.
Working with schedulers such as Apache Airflow to automate workflows and data pipelines is a key responsibility.
The role involves automating tasks and processes using Python, shell scripting, and other relevant programming languages.
The candidate will develop, troubleshoot, and optimize PL/SQL queries for complex data workflows.
Extensive work on Big Data platforms like Hadoop and Hive to process large datasets is required.
The role includes driving the development of large-scale, self-service enterprise data frameworks.
Exposure to open table formats like Iceberg is highly desirable.
Requirements:
Candidates must have experience with AWS services (S3, EMR, Athena, Glue) or equivalent cloud platforms.
Proficiency with the AbInitio product suite is required, including PDL, Meta-programming, Conduct>It, Express>It, and GDE-based development.
Experience in implementing and maintaining CI/CD pipelines using tools like Jenkins and GitHub is necessary.
Familiarity with schedulers such as Apache Airflow for automating workflows and data pipelines is essential.
Candidates should have skills in automating tasks and processes using Python, shell scripting, and other relevant programming languages.
Experience in developing, troubleshooting, and optimizing PL/SQL queries for complex data workflows is required.
Extensive experience with Big Data platforms like Hadoop and Hive is necessary for processing large datasets.
Candidates should have a proven track record in driving the development of large-scale, self-service enterprise data frameworks.
Exposure to open table formats like Iceberg is highly desirable.
Benefits:
The position offers the flexibility of working from home or remotely.
Employees will have the opportunity to work with cutting-edge technologies and tools in the cloud data services domain.
The role provides a chance to collaborate with customers and contribute to the design and delivery of scalable data solutions.
There is potential for professional growth and development in a rapidly evolving field.
The company promotes a supportive work environment that values innovation and teamwork.
Apply now
Please, let Rackspace know you found this job
on RemoteYeah
.
This helps us grow 🌱.