Overview
Build high-performance Go services that power AI/ML workflows and data pipelines on Tower Cloud. You will work with researchers and backend engineers to ship reliable, observable APIs and batch jobs.
What you'll do
- Design and implement Go microservices for inference orchestration, feature stores, and integrations.
- Work with ML engineers and data teams on model deployment, versioning, and GPU-adjacent workloads where applicable.
- Optimize latency, throughput, and resource usage across distributed systems.
- Operate what you build: metrics, tracing, alerts, and runbooks.
- Contribute to internal tooling and SDKs used by AI/ML customers.
What we're looking for
- About 2 years of experience shipping production Go (or strong backend experience with a path to Go).
- Familiarity with AI/ML concepts: training/serving boundaries, batch vs real-time, and common frameworks.
- Experience with Kubernetes and public cloud environments.
- Solid grasp of distributed systems, REST/gRPC, and infrastructure-as-code.
- Clear communication and ability to collaborate across research, product, and engineering teams.