Data Scientist - Python (Mid-senior, Senior)
Paris, Île-de-France, France - Remote
Applications have closed
Pathway
Pathway is the data processing framework which handles streaming data updates for you.About Pathway
Deeptech start-up, founded in March 2020.
- Our primary developer offering is an ultra-performant Data Processing Framework (unified streaming + batch) with a Python API, distributed Rust engine, and capabilities for data source integration & transformation at scale (Kafka, S3, databases/CDC,...).
- The single-machine version is provided on a free-to-use license (`pip install pathway`).
- Major data use cases are around event-stream data (including real-world data such as IoT), and graph data that changes over time.
- Our enterprise offering is currently used by leaders of the logistics industry, such as DB Schenker or La Poste, and tested across multiple industries. Pathway has been featured in Gartner's market guide for Event Stream Processing.
- Learn more at http://pathway.com/ and https://github.com/pathwaycom/.
Pathway is VC-funded, with amazing BAs from the AI space and industry. We have operations across Europe and in the US. We are headquartered in Paris, with significant support from the French ecosystem (BPI, Agoranov, WILCO,...).
The Team
Pathway is built by and for overachievers. Its co-founders and employees have worked in the best AI labs in the world (Microsoft Research, Google Brain, ETH Zurich), worked at Google, and graduated from top universities (Polytechnique, ENSAE, Sciences Po, HEC Paris, PhD obtained at the age of 20, etc…). Pathway’s CTO is a co-author with Goeff Hinton and Yoshua Bengio. The management team also includes the co-founder of Spoj.com (1M+ developer users) and NK.pl (13.5M+ users) and experienced growth leader who has scaled companies with multiple exits.
The opportunity
We are currently searching for Data Scientists with experience in the Python stack, to help explore and discover the most pertinent insights in datasets on spatio-temporal event streams. In this job, statistical rigor and beauty of visualization meet on equal footing.
You Will
- be working with spatiotemporal data with advanced schemas (time-changing graph models)/
- be designing data cross-sections, proposing analytics metrics and KPI’s in line with clients’ objectives, selecting clustering algorithms, and preparing visualizations, to enable fast data exploration and insight discovery – all within our product.
- be designing dashboards in SQL with some Python elements/extensions.
- be directly helping us with Customer Conversion and Adoption within Customer organizations, by contributing to both deployment instances and “demonstrators” of our product, performed on client data sets.
- work directly with our Product Owner and CTO to propose and implement extensions to our product, based on repetitive client needs.
- depending on your seniority, implement machine learning algorithms on spatiotemporal event streams and other geospatial data.
The results of your work will play a crucial role in proving how our technology can help with compelling industry use cases.
Requirements
You Are
- Ready for hands-on contribution to the product, helping to ensure the success of demonstrators for clients, and contribution to product codebase.
- Intuitive, with good visual taste, and good common sense judgment.
- Committed to beautiful user-centered design: you know that stories are made for people, and you are willing to listen to what they have to say.
- Curious at heart and thrilled to work with real-world data, especially spatio-temporal data.
- Like trains, trucks, cranes, pythons, pandas, and other things that move.
- Not afraid to switch between the roles of data scientist, data-vis magician, statistician, engineer, and detective, at a moment’s notice.
- Have 2 years+ experience in positions related to Data Science.
- Have a very good working knowledge of Python.
- Know SQL. Are able to work with tables and other data types (arrays, json,…).
- Would be able to implement the Transit Node Routing algorithm in Python just based on reading its Wikipedia article.
- Have experience with git, build systems, and CI/CD.
- Have at least basic undergrad textbook familiarity with graph algorithms, finite automata, and text (string) search algorithms.
- Understand statistical concepts, such as correlated random variables, significance, and non-Gaussian noise.
- Prepared to be quizzed & grilled by the datasets you encounter, everyday. Here are some questions you should be able to answer off the top of your head: what can “-273.15” signify; why “65535” is a suspicious integer value; how many months does it take a containership to go around the world; and, roughly what order of g-force is attained by an astronaut in a space rocket at liftoff?
- Respectful of others
- Fluent in English
Bonus Points
- Showing a portfolio: code on github, visualization works, a research paper or a PhD thesis with an original statistical / probabilistic analysis or experiment design,…
- Successful track-record in Data Science or algorithms contests (Kaggle, Codeforces,…)
- Experience in topics linked to logistics/moving assets.
- Familiarity with some form of GIS software.
- Familiarity with Pandas, SciPy, NetworkX, and similar tools from the Python stack.
- Experience in Data Visualization and UX.
- Some knowledge of French, Polish, or German.
Why You Should Apply
- Join an intellectually stimulating work environment.
- Be a pioneer: you get to work with a new type of data processing.
- Work in one of the hottest data/AI startups in France.
- Uncover exciting career prospects.
- Make significant contribution to our success.
- Join & co-create an inclusive workplace culture.
Benefits
- Type of contract: Permanent employment contract
- Preferable joining date: February 2023. The positions (at least 2) are open until filled.
- Compensation: annual salary of €50K-€70K (mid) up to €60K-€90K (senior, upper band negotiable) + Employee stock option plan.
- Location: Remote work from home. Possibility to work or meet with other team members in one of our offices:
- Paris Area – Drahi X-Novation Center, Ecole Polytechnique, Palaiseau.
- Paris – Agoranov (where Doctolib, Alan, and Criteo were born) near Saint-Placide Metro (75006).
- Wroclaw – University area.
Permanent residence will be required in France or Poland, exceptional candidates will be considered anywhere in the EU.
If you meet our broad requirements but are missing some experience, don’t hesitate to reach out to us.
Note: CS & engineering school students with exceptional profiles and/or strong motivation to join Pathway are invited to apply for Data Science internships. (Minimum duration: 5-6 months, remuneration level: €1500 / month.)
Tags: APIs CI/CD Clustering Data visualization Engineering Git GitHub JSON Kafka KPIs Machine Learning Pandas PhD Python Research Rust SciPy SQL Statistics Streaming UX
Perks/benefits: Career development Equity Salary bonus Startup environment
More jobs like this
Explore more AI, ML, Data Science career opportunities
Find even more open roles in Artificial Intelligence (AI), Machine Learning (ML), Natural Language Processing (NLP), Computer Vision (CV), Data Engineering, Data Analytics, Big Data, and Data Science in general - ordered by popularity of job title or skills, toolset and products used - below.
- Open Data Manager jobs
- Open Lead Data Analyst jobs
- Open MLOps Engineer jobs
- Open Senior Business Intelligence Analyst jobs
- Open Data Science Manager jobs
- Open Principal Data Engineer jobs
- Open Data Engineer II jobs
- Open Power BI Developer jobs
- Open Sr Data Engineer jobs
- Open Data Scientist II jobs
- Open Product Data Analyst jobs
- Open Business Intelligence Developer jobs
- Open Data Analytics Engineer jobs
- Open Junior Data Scientist jobs
- Open Sr. Data Scientist jobs
- Open Senior Data Architect jobs
- Open Business Data Analyst jobs
- Open Data Analyst Intern jobs
- Open Big Data Engineer jobs
- Open Principal Data Scientist jobs
- Open Junior Data Engineer jobs
- Open Manager, Data Engineering jobs
- Open Data Product Manager jobs
- Open Data Quality Analyst jobs
- Open Research Scientist jobs
- Open GCP-related jobs
- Open Data quality-related jobs
- Open Business Intelligence-related jobs
- Open Java-related jobs
- Open ML models-related jobs
- Open Data management-related jobs
- Open Privacy-related jobs
- Open Data visualization-related jobs
- Open Finance-related jobs
- Open Deep Learning-related jobs
- Open PhD-related jobs
- Open APIs-related jobs
- Open TensorFlow-related jobs
- Open PyTorch-related jobs
- Open NLP-related jobs
- Open Consulting-related jobs
- Open Snowflake-related jobs
- Open CI/CD-related jobs
- Open LLMs-related jobs
- Open Generative AI-related jobs
- Open Kubernetes-related jobs
- Open Data governance-related jobs
- Open Hadoop-related jobs
- Open Airflow-related jobs
- Open Docker-related jobs