Senior Software Engineer, Data Backend(Data Platform)

Reposted 3 Days Ago
Be an Early Applicant
2 Locations
In-Office
Senior level
Artificial Intelligence
The Role
Design, build, and operate large-scale distributed data pipelines and platforms (Spark, Kafka, data warehouses/lakes). Optimize performance, establish architecture and tooling, and support ETL, deployment, monitoring, and scalability.
Summary Generated by Built In

About Appier 

Appier is a software-as-a-service (SaaS) company that uses artificial intelligence (AI) to power business decision-making. Founded in 2012 with a vision of democratizing AI, Appier’s mission is turning AI into ROI by making software intelligent. Appier now has 17 offices across APAC, Europe and U.S., and is listed on the Tokyo Stock Exchange (Ticker number: 4180). Visit www.appier.com for more information.


About the role

We're looking for an experienced data engineer to build and scale our data infrastructure. You'll work on mission-critical systems that power our data-driven decision making, designing robust pipelines and optimizing distributed platforms that process massive datasets.


Responsibilities

  • Design, develop, and maintain production data pipelines processing hundreds of terabytes of data daily across distributed systems
  • Build, manage, and optimize large-scale data platforms including Spark clusters, Kafka clusters, and related big data infrastructure
  • Develop and operate data warehouse systems and ETL pipelines for data collection, processing, storage, and analytics
  • Profile and tune performance of critical components to optimize throughput, reduce latency, and improve resource efficiency
  • Establish architectural foundations and engineering best practices for the platform
  • Design frameworks and tooling to streamline software development, deployment, monitoring, and observability
  • Provide technical expertise and guidance on big data system design, performance optimization, and scalability challenges

About you

[Minimum qualifications]

  • BS/MS degree in Computer Science or related technical field
  • 2+ years of hands-on experience building and operating large-scale distributed systems
  • Proficiency in Java and/or Scala for production-level development
  • Proven experience building ETL pipelines and data processing workflows using Apache Spark
  • Experience managing and optimizing data lakes or data warehouses at scale
  • Ability to work independently and adapt quickly in fast-paced, dynamic environments
  • Self-motivated learner who takes ownership, manages multiple priorities effectively, and thrives in collaborative team settings
  • Experience deploying and managing infrastructure on public cloud platforms (AWS, GCP, or Azure)

[Preferred qualifications]

  • Proficiency in Python and/or Golang for building production systems
  • Experience profiling and optimizing JVM performance for high-throughput applications
  • Hands-on experience managing and operating data platforms such as Hadoop, Kafka, Flink, Hive, Trino, Presto, or ClickHouse
  • Knowledge of modern open table formats including Apache Iceberg, Delta Lake, or Hudi 
  • Active contributions to open source projects (please include your GitHub profile if applicable)

#LI-AK1


Skills Required

  • BS/MS degree in Computer Science or related technical field
  • 2+ years of hands-on experience building and operating large-scale distributed systems
  • Proficiency in Java and/or Scala for production-level development
  • Proven experience building ETL pipelines and data processing workflows using Apache Spark
  • Experience managing and optimizing data lakes or data warehouses at scale
  • Ability to work independently and adapt quickly in fast-paced, dynamic environments
  • Self-motivated learner who takes ownership and manages multiple priorities effectively
  • Experience deploying and managing infrastructure on public cloud platforms (AWS, GCP, or Azure)
  • Proficiency in Python and/or Golang for building production systems
  • Experience profiling and optimizing JVM performance for high-throughput applications
  • Hands-on experience managing and operating data platforms such as Hadoop, Kafka, Flink, Hive, Trino, Presto, or ClickHouse
  • Knowledge of modern open table formats including Apache Iceberg, Delta Lake, or Hudi
  • Active contributions to open source projects (please include your GitHub profile if applicable)
Am I A Good Fit?
beta
Get Personalized Job Insights.
Our AI-powered fit analysis compares your resume with a job listing so you know if your skills & experience align.

The Company
HQ: San Francisco, CA
642 Employees
Year Founded: 2012

What We Do

Appier is a software-as-a-service (SaaS) company that uses artificial intelligence (AI) to power business decision-making. Founded in 2012 with a vision of democratizing AI, Appier now has 17 offices across APAC, Europe and U.S., and is listed on the Tokyo Stock Exchange (Ticker number: 4180). Visit www.appier.com for more information.

Similar Jobs

Appier Logo Appier

Senior Software Engineer

Artificial Intelligence
In-Office
2 Locations
642 Employees

Morningstar Logo Morningstar

Account Manager

Artificial Intelligence • Big Data • Enterprise Web • Fintech • Software • Financial Services
Hybrid
Tokyo, JPN
11500 Employees

Morningstar Logo Morningstar

Customer Success Manager

Artificial Intelligence • Big Data • Enterprise Web • Fintech • Software • Financial Services
Hybrid
Tokyo, JPN
11500 Employees
Hybrid
Tokyo, JPN
897 Employees

Similar Companies Hiring

GC AI Thumbnail
Artificial Intelligence • Legal Tech
San Mateo, California
100 Employees
Idler Thumbnail
Artificial Intelligence
San Francisco, California
6 Employees
Bellagent Thumbnail
Artificial Intelligence • Machine Learning • Business Intelligence • Generative AI
Chicago, IL
20 Employees

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account