Remote Senior Data Scientist

at Proximity Works

Posted 20 hours ago 2 applied

Description:

  • We are seeking a highly skilled, execution-focused Senior Data Scientist with a minimum of 5 years of experience.
  • This role demands hands-on expertise in building, deploying, and optimizing machine learning models at scale, while working with big data technologies and modern cloud platforms.
  • You will be responsible for driving data-driven solutions from experimentation to production, leveraging advanced tools and frameworks across Python, SQL, Spark, and AWS.
  • The role requires strong technical depth, problem-solving ability, and ownership in delivering business impact through data science.
  • Responsibilities include designing, building, and deploying scalable machine learning models into production systems.
  • You will develop advanced analytics and predictive models using Python, SQL, and popular ML/DL frameworks such as Pandas, Scikit-learn, TensorFlow, and PyTorch.
  • You will leverage Databricks, Apache Spark, and Hadoop for large-scale data processing and model training.
  • Implementing workflows and pipelines using Airflow and AWS EMR for automation and orchestration is also part of the role.
  • You will collaborate with engineering teams to integrate models into cloud-based applications on AWS.
  • Optimizing query performance, storage usage, and data pipelines for efficiency is essential.
  • Conducting end-to-end experiments, including data preprocessing, feature engineering, model training, validation, and deployment is required.
  • You will drive initiatives independently with high ownership and accountability.
  • Staying up to date with industry best practices in machine learning, big data, and cloud-native deployments is necessary.

Requirements:

  • A minimum of 5 years of experience in Data Science or Applied Machine Learning is required.
  • Strong proficiency in Python, SQL, and ML libraries such as Pandas, Scikit-learn, TensorFlow, and PyTorch is essential.
  • Proven expertise in deploying ML models into production systems is necessary.
  • Experience with big data platforms such as Hadoop and Spark, and distributed data processing is required.
  • Hands-on experience with Databricks, Airflow, and AWS EMR is essential.
  • Strong knowledge of AWS cloud services including S3, Lambda, SageMaker, and EC2 is required.
  • A solid understanding of query optimization, storage systems, and data pipelines is necessary.
  • Excellent problem-solving skills, with the ability to design scalable solutions, are required.
  • Strong communication and collaboration skills to work in cross-functional teams are essential.

Benefits:

  • We offer a best-in-class salary, as we hire only the best and pay accordingly.
  • Proximity Talks provide opportunities to meet other designers, engineers, and product geeks, allowing you to learn from experts in the field.
  • You will have the chance to keep on learning with a world-class team, working with the best in the field, challenging yourself constantly, and learning something new every day.

Get realtime job alerts

Be the first to know about new jobs