Incedo is seeking a skilled and experienced Java + PySpark Developer to join its data engineering team.
The ideal candidate will work on scalable data pipeline development, data integration, and backend microservices, contributing to analytics and business insights platforms.
Key responsibilities include designing, developing, and maintaining scalable data processing pipelines using PySpark on big data platforms like Hadoop or Databricks.
The candidate will build and enhance Java-based microservices/APIs to support data ingestion and transformation.
Collaboration with data scientists, data analysts, and architects is essential to deliver high-quality, secure, and performant solutions.
The role involves optimizing data workflows and troubleshooting performance bottlenecks.
Participation in code reviews, design discussions, and providing technical leadership when needed is expected.
The candidate will implement data validation, quality checks, and error-handling mechanisms.
Ensuring adherence to data security and compliance standards within financial environments is crucial.
Requirements:
A minimum of 6 years of hands-on experience in Java, with Spring Boot preferred.
At least 3 years of experience with PySpark in distributed data processing.
Experience with the Hadoop ecosystem, Hive, HDFS, and Spark SQL is required.
A solid understanding of RESTful APIs, JSON/XML, and integration patterns is necessary.
Strong knowledge of SQL and data modeling is essential.
Familiarity with CI/CD tools, Git, and Agile methodology is required.
Benefits:
The position offers a full-time employment opportunity.
The role is based in Dallas TX, Pittsburgh PA, or Cleveland OH.
Candidates will have the opportunity to work on innovative data engineering projects.
The position provides a chance to collaborate with a talented team of data professionals.
Opportunities for professional growth and development within the company are available.