We are seeking a highly skilled, execution-focused Senior Data Scientist with a minimum of 5 years of experience.
This role demands hands-on expertise in building, deploying, and optimizing machine learning models at scale, while working with big data technologies and modern cloud platforms.
You will be responsible for driving data-driven solutions from experimentation to production, leveraging advanced tools and frameworks across Python, SQL, Spark, and AWS.
The role requires strong technical depth, problem-solving ability, and ownership in delivering business impact through data science.
Responsibilities include designing, building, and deploying scalable machine learning models into production systems.
You will develop advanced analytics and predictive models using Python, SQL, and popular ML/DL frameworks such as Pandas, Scikit-learn, TensorFlow, and PyTorch.
You will leverage Databricks, Apache Spark, and Hadoop for large-scale data processing and model training.
Implementing workflows and pipelines using Airflow and AWS EMR for automation and orchestration is also part of the role.
You will collaborate with engineering teams to integrate models into cloud-based applications on AWS.
Optimizing query performance, storage usage, and data pipelines for efficiency is essential.
Conducting end-to-end experiments, including data preprocessing, feature engineering, model training, validation, and deployment is required.
You will drive initiatives independently with high ownership and accountability.
Staying up to date with industry best practices in machine learning, big data, and cloud-native deployments is necessary.
Requirements:
A minimum of 5 years of experience in Data Science or Applied Machine Learning is required.
Strong proficiency in Python, SQL, and ML libraries such as Pandas, Scikit-learn, TensorFlow, and PyTorch is essential.
Proven expertise in deploying ML models into production systems is necessary.
Experience with big data platforms such as Hadoop and Spark, and distributed data processing is required.
Hands-on experience with Databricks, Airflow, and AWS EMR is essential.
Strong knowledge of AWS cloud services including S3, Lambda, SageMaker, and EC2 is required.
A solid understanding of query optimization, storage systems, and data pipelines is necessary.
Excellent problem-solving skills, with the ability to design scalable solutions, are required.
Strong communication and collaboration skills to work in cross-functional teams are essential.
Benefits:
We offer a best-in-class salary, as we hire only the best and pay accordingly.
Proximity Talks provide opportunities to meet other designers, engineers, and product geeks, allowing you to learn from experts in the field.
You will have the chance to keep on learning with a world-class team, working with the best in the field, challenging yourself constantly, and learning something new every day.