Remote Data Architect (Abinitio)

Posted

Apply now
Please, let Rackspace know you found this job on RemoteYeah. This helps us grow 🌱.

Description:

  • The Cloud Data Services Team is seeking a skilled Data Architect (Abinitio) for a customer-facing role.
  • This position involves working with cutting-edge technologies and tools to build, optimize, and maintain large-scale data processing frameworks on the cloud.
  • The role requires collaboration with customers to design and deliver efficient, scalable data solutions.
  • Responsibilities include leveraging AWS services (S3, EMR, Athena, Glue) or equivalent cloud platforms for data storage and processing.
  • Proficiency with the AbInitio product suite is essential, including experience in PDL, Meta-programming, Conduct>It, Express>It, and GDE-based development.
  • The candidate will implement and maintain CI/CD pipelines using tools like Jenkins and GitHub.
  • Working with schedulers such as Apache Airflow to automate workflows and data pipelines is a key responsibility.
  • The role involves automating tasks and processes using Python, shell scripting, and other relevant programming languages.
  • The candidate will develop, troubleshoot, and optimize PL/SQL queries for complex data workflows.
  • Extensive work on Big Data platforms like Hadoop and Hive to process large datasets is required.
  • The role includes driving the development of large-scale, self-service enterprise data frameworks.
  • Exposure to open table formats like Iceberg is highly desirable.

Requirements:

  • Candidates must have experience with AWS services (S3, EMR, Athena, Glue) or equivalent cloud platforms.
  • Proficiency with the AbInitio product suite is required, including PDL, Meta-programming, Conduct>It, Express>It, and GDE-based development.
  • Experience in implementing and maintaining CI/CD pipelines using tools like Jenkins and GitHub is necessary.
  • Familiarity with schedulers such as Apache Airflow for automating workflows and data pipelines is essential.
  • Candidates should have skills in automating tasks and processes using Python, shell scripting, and other relevant programming languages.
  • Experience in developing, troubleshooting, and optimizing PL/SQL queries for complex data workflows is required.
  • Extensive experience with Big Data platforms like Hadoop and Hive is necessary for processing large datasets.
  • Candidates should have a proven track record in driving the development of large-scale, self-service enterprise data frameworks.
  • Exposure to open table formats like Iceberg is highly desirable.

Benefits:

  • The position offers the flexibility of working from home or remotely.
  • Employees will have the opportunity to work with cutting-edge technologies and tools in the cloud data services domain.
  • The role provides a chance to collaborate with customers and contribute to the design and delivery of scalable data solutions.
  • There is potential for professional growth and development in a rapidly evolving field.
  • The company promotes a supportive work environment that values innovation and teamwork.
Apply now
Please, let Rackspace know you found this job on RemoteYeah . This helps us grow 🌱.
About the job
Posted on
Job type
Salary
-
Position
Experience level
Report this job

Job expired or something else is wrong with this job?

Report this job
Leave a feedback