Senior Data Scientist
Role details
Job location
Tech stack
Job description
As a Senior Data Scientist, you will play a key role in transforming raw sensor data into valuable insights. You will design heuristics and algorithms that enrich this data in real time, empowering customers to better understand and optimize their processes. On top of that, you will develop models that generate actionable and personalized advice. You'll be involved in every step of the development process: from data quality checks for new sensor types to deploying and maintaining prediction pipelines. With data at the core of our platform, you will be a vital member of a cross-disciplinary team driving meaningful impact.
-
You will be part of our Asset Management Product Team, where you will work on data quality improvements and algorithms tailored to circuit breakers. Besides that you will partner up with our VP of Product Innovation on data science heavy R&D projects.
-
You will be responsible for designing algorithms that turn our raw and enriched data into actionable advice in a scalable way, such as flagging early signs of machine degradation or estimating efficiency gains from machine upgrades or process changes.
-
You will work closely with our Data Engineers to create deployable artifacts which will continuously generate insights across our customer base in a mix of streaming and batch applications.
-
You will interface directly with customers to gain a deep understanding of their processes and translate these into pragmatic and scalable algorithms that deliver real-world impact.
-
We have many different skills and personalities in the team, and promote an open and collaborative working environment. You are encouraged to work together on complex tasks, to give feedback when you can and actively think about your personal development.
-
Being part of a fast-moving environment, you can work autonomously and proactively prioritize and solve the needs of our growing group of customers.
The key technologies you will be working with:
Our current AWS stack is focused on ingesting and processing raw sensor data in real time using Kafka and Flink. Processed sensor data is stored in Clickhouse, other types of data live in Postgres. Batch processing is done using Prefect and Fargate, on-demand services are deployed using Lambda. We have a powerful internal GraphQL API to expose data to end users, managed by Hasura. As a data scientist you will use Poetry for package management and Gitlab CI/CD to test, build and deploy our code. You can spin up a fresh Fargate cluster at any time and parallelize your workload using Dask.
How we do it
We do Scrum with 2-week sprints, sprint planning and retrospective sessions. We have daily stand ups over Google Meet. The course is determined by quarterly goals, set collaboratively by business, data, development and product teams. We know how important it is to get in the zone and write beautiful code so we try to keep meeting pressure low. We work from home about 70% of the time, but we enjoy meeting each other in the office regularly.
Requirements
Do you have experience in Python?, Do you have a Master's degree?, * Have an MSc in Business Analytics, Data Science, Computer Science, or a comparable analytical / technical field
- Are a Data Scientist with 3+ years of experience who is fluent in data science and analysis with Python using Pandas and Numpy
- Have moved 'beyond Jupyter notebooks' and are comfortable with building Python modules and packages
- Have experience with or are eager to learn more about programming best practices such as version control, testing, code quality and CI/CD.
- Are a strong communicator, effective in both close team collaboration and independent, self-directed work.
- (Bonus) Have experience with working in AWS
- (Bonus) Have experience with use case discovery and running PoCs, for example from working in data science consultancy.
- Business level fluency in English is mandatory. Given the customer-facing aspect of this role, Dutch proficiency is a strong plus.