Senior Software Engineer - Machine Learning

Reposted 18 Days Ago
6 Locations
Remote
Senior level
Artificial Intelligence • Cloud • Analytics • Automation
The Role
Design, build, and operate large-scale geospatial data pipelines and production ML/agent systems. Own data ingestion, distributed processing, model training, deployment, monitoring, and observability. Develop backend services and tooling, integrate agent frameworks, collaborate with product and customer teams to drive revenue and reliability.
Summary Generated by Built In

About SafeGraph

SafeGraph is a Data as a Service (DaaS) company with one focus: curating the most accurate, precise, and fresh points of interest (POI) database on the planet. We provide product builders, data scientists, and analytics teams with the location data they need to power site selection, transaction enrichment, advertising audiences, competitive intelligence, and more.

Our customers include companies like Plaid, Mapbox, Clear Channel — spanning fintech, retail, real estate, adtech, logistics, and government. We’re fully remote, lean by design, and serious about data quality.

The Role

You’ll be a generalist responsible for building and running large-scale data, machine learning, and agentic systems. The focus is operational ML/AI, including agentic systems and geospatial data pipelines.  You should be comfortable owning the full lifecycle: from data ingestion and distributed processing to model development, deployment, and monitoring. This role requires the ability to iterate quickly from initial concept to a robust, production-ready solution.


Key Responsibilities

  • Take ownership of the end-to-end AI/ML lifecycle, with a strong focus on dealing with complex and messy data, thorough evaluation of different approaches, and successfully deploying robust models, and handling cost vs performance tradeoffs.
  • Implement and integrate large-scale, agent-based systems with access to external systems, building these solutions from the ground up and integrating them with our existing infrastructure.
  • Establish observability for pipelines, models, and agents (metrics, tracing, alerting).
  • Collaborate with product and customer teams to drive revenue.

Requirements
  • Strong experience with distributed data processing, particularly Spark and SQL.
  • Proven expertise in building production machine learning systems, including working with large, wide datasets, effective training, deployment, and monitoring.
  • Experience designing and deploying task-oriented AI agents and working with coding agents.
  • Experience working with cloud services across data, compute, and ML.
  • Strong communication abilities, including code architecture and documentation, at a level where any technical team member can troubleshoot and contribute easily.

Languages: Scala, Python

Tools / Frameworks: Spark, AWS Sagemaker / Bedrock, Kubernetes

Nice to Haves

  • Startup experience or growing projects from 0 to production in a larger org.
  • Experience with large geospatial datasets, formats, and indexing strategies.
  • Experience building operational AI agents that work at scale (millions of separate, complex tasks including web research)
  • Experience with fine-tuning, distilling, and self-hosting LLM models.
  • Experience in traditional ML, with a focus on working with messy data and robust evaluation of model approaches.
  • Proficiency with CI/CD, infrastructure as code, and containerization.

What Success Looks Like

  • ML/AI models deployed with robust monitoring and significant customer impact.
  • Agentic workflows improving internal/external operations.
  • Infrastructure that is stable, observable, and automated.
  • Successful iteration and delivery of new ML/AI products from concept to production.
  • Ability to contribute to existing geospatial pipelines directly or through the use of AI

Top Skills

Alerting)
AWS
Ci/Cd
Containerization
Embeddings
Graph Databases
Infrastructure As Code
Langchain
Llm Fine-Tuning
Observability (Metrics
Openai Assistants
Python
Scala
Spark
Spatial Databases
Tracing
Vector Databases
Am I A Good Fit?
beta
Get Personalized Job Insights.
Our AI-powered fit analysis compares your resume with a job listing so you know if your skills & experience align.

The Company
Montreal, Quebec
62 Employees

What We Do

PartnerOne is an enterprise software company that manages the world’s largest data environments through virtualized cloud storage, hyper-automation, artificial intelligence, and metadata analytics. Contrary to other software companies, we play a mission-critical role in not just one, but many aspects of the enterprise Big Data cycle. Over 1250 of the world’s largest data environments rely on our software for their most critical needs and to safeguard their most valuable data.

Similar Jobs

Capco Logo Capco

Data & AI Warsaw Tech Summit 2026: TRINITY GAME

Fintech • Professional Services • Consulting • Energy • Financial Services • Cybersecurity • Generative AI
Remote or Hybrid
Poland
6000 Employees

Mondelēz International Logo Mondelēz International

Sr Manager Global Engineering Digital Rollout

Big Data • Food • Hardware • Machine Learning • Retail • Automation • Manufacturing
Remote or Hybrid
4 Locations
90000 Employees
140K-193K Annually

Affirm Logo Affirm

Senior Software Engineer

Big Data • Fintech • Mobile • Payments • Financial Services
Easy Apply
Remote
Poland
2200 Employees
285K-385K Annually

DuckDuckGo Logo DuckDuckGo

Senior Data Scientist

Information Technology
Remote
14 Locations
393 Employees
179K-179K Annually

Similar Companies Hiring

GC AI Thumbnail
Artificial Intelligence • Legal Tech
San Mateo, California
80 Employees
Idler Thumbnail
Artificial Intelligence
San Francisco, California
6 Employees
Bellagent Thumbnail
Artificial Intelligence • Machine Learning • Business Intelligence • Generative AI
Chicago, IL
20 Employees

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account