We are hiring: Data Scientist for New Product Development
Join GEOLYTIX! Build the next gen of location intelligence. Lead deep learning, spatial R&D, and production-ready models. Innovate with us today.
GEOLYTIX is seeking a Data Scientist to help us build the next generation of location intelligence products.
You will join the New Product Development (NPD) team, responsible for turning ideas into products. The NPD team builds anything from customer favourites, like our Busyness suite of mobility data sets, to internal tooling and automation. With our eyes firmly set on model innovation, we will develop and evolve the solutions that the rest of the business, and our clients, will use for years to come.
Your challenge is to combine our rich proprietary geodata and open data with high-volume human activity signals from banking and mobile device data to build deep learning models, which are not only accurate during training, but robust enough to generalise effectively to the real world.
Responsibilities
- Designing & Training Models: You will spend significant time hands-on with code, leading the design of model architectures that interpret complex interactions between people and places.
- Wrangling & Engineering Data: You won’t just wait for clean datasets. You will be building the rigorous pipelines required to clean, normalise, and restructure massive signals (like banking or mobility data), ensuring model inputs are stable and grounded in reality.
- Guiding R&D Direction: You will act as a technical compass for the team. Instead of just executing tickets, you will help steer our research efforts, distinguishing between “AI hype” and commercial utility to save us from costly dead-ends.
- Bridging Teams and Mentoring: You will actively collaborate with our Data and Software Engineering teams to ensure solutions align with client needs and technical friction is managed proactively. You will also get the chance to upskill junior colleagues and help them to elevate their understanding of modern ML architectures.
- Building for Production: You will think about the destination from the start. You’ll make early architectural decisions, regarding data structures and containerisation, to ensure what you are building can scale cost-effectively in our cloud environment.
About You
- You are a highly skilled technical practitioner: You possess the deep expertise required to build sophisticated models.
- You are driven by a desire to solve real world problems: You are excited by the prospect of your outputs directly benefitting our clients.
- You are pragmatic: You can pair your technical skills with the pragmatism; understanding that a model is only valuable if it is robust, scalable, and solves the problem at hand.
- You are excited by spatial analytics: Your background does not have to be related to retail modelling or location planning; we encourage applications from Data Scientists who bring a fresh perspective, having used spatial analytics in other domains.
Skills & Experience
- Deep Learning: You have strong conceptual understanding and practical experience with modern architectures (Transformers, Attention Mechanisms, Embeddings, GNNs) and frameworks like PyTorch or Tensorflow. Ideally, you already know the common pitfalls of applying these to spatial/vector data. You know the shortcuts that save R&D time, and you have the intuition to select the right architecture without wasting compute cycles on blind experimentation.
- Spatial Data Science: A deep understanding of the maths behind the maps. You are familiar with algorithmic geometry (e.g. Spatial Indexing, Topology, Voronoi) and know your way around spatial extensions in various coding and query languages (e.g. PostGIS, Geopandas, h3).
- Statistical Foundations: A solid grasp of probability, distributions, and hypotheses testing. You know how to rigorously quantify uncertainty in your predictions and use established metrics and evaluation techniques to identify weak spots in your models.
- Big Data & Pipelines: Experience handling large-scale datasets (e.g. Banking or Mobility Signals). You are comfortable processing data in BigQuery and may be familiar with modern orchestration frameworks (like dbt, Airflow or Dagster) to build robust data flows.
- Coding & Version Control: Proficiency in Python (and/or other general-purpose programming languages) and git.
- Advanced SQL: High proficiency in one or more SQL dialects (ideally BigQuery and PostgreSQL). You know how to utilise advanced features like window functions, stored procedures, and clustering/partitioning. You understand the query planner and how to optimise for parallelisation and performance.
- AI-Aided Workflows: You know how to utilise AI coding assistants effectively to accelerate your development speed.
- Cloud First: You have experience deploying models and data pipelines to cloud environments (GCP preferred). You are comfortable building scalable, cloud-native applications from day one.
- AGILE & Ways of Working: You are comfortable working in an iterative, agile environment. You value rapid prototyping and feedback loops over rigid, long-term specifications, and you are happy to “ship and iterate”.
Other stuff
Our offices are in Leeds and Clerkenwell, London. The role can be located in either. Competitive salary. We’re a young and growing company who embrace flexible working; full time or part time, family friendly hours and/or working from home days considered.
Interested?
If you are interested in this role and are looking for your next challenge, please send your CV and cover letter and include DS/NPD001 - NPD in the subject to careers@geolytix.co.uk. No agencies please.