Overview

Build high-performance Go services that power AI/ML workflows and data pipelines on Tower Cloud. You will work with researchers and backend engineers to ship reliable, observable APIs and batch jobs.

What you'll do

  • Design and implement Go microservices for inference orchestration, feature stores, and integrations.
  • Work with ML engineers and data teams on model deployment, versioning, and GPU-adjacent workloads where applicable.
  • Optimize latency, throughput, and resource usage across distributed systems.
  • Operate what you build: metrics, tracing, alerts, and runbooks.
  • Contribute to internal tooling and SDKs used by AI/ML customers.

What we're looking for

  • About 2 years of experience shipping production Go (or strong backend experience with a path to Go).
  • Familiarity with AI/ML concepts: training/serving boundaries, batch vs real-time, and common frameworks.
  • Experience with Kubernetes and public cloud environments.
  • Solid grasp of distributed systems, REST/gRPC, and infrastructure-as-code.
  • Clear communication and ability to collaborate across research, product, and engineering teams.