We are seeking a Staff Engineer to support the design and delivery of next-generation AI and Generative AI platforms within Sia’s AI Factory. This is a hands-on engineering role focused on building scalable, production-grade systems where LLMs, agentic workflows, and machine learning models integrate seamlessly with cloud-native backend services.
You will work closely with product, data science, and platform teams to translate business and client requirements into robust technical solutions. The role requires strong Python expertise, experience with distributed systems and microservices, and a pragmatic approach to delivering reliable, scalable systems in enterprise environments. In addition to technical delivery, you will contribute to architectural decisions, raise engineering standards, and support client-facing engagements through technical leadership and clear communication.
Key Responsibilities
Define and own architecture for backend systems (primarily Python) that integrate AI/ML into production services.
Deploy, optimize, and scale ML models in collaboration with Data Science and Data Engineering.
Design and maintain cloud infrastructure using Terraform and Helm; support deployments on GCP/AWS.
Lead containerization and orchestration best practices (Docker, Kubernetes) for development and production.
Ensure data integrity and performance across SQL and NoSQL systems (Postgres, MongoDB, etc.).
Establish and maintain monitoring and observability: Prometheus, Grafana, logging (Loki/ELK) and alerting.
Ship backend microservices and platform tooling: APIs, auth, data pipelines, batch/streaming components.
Contribute to product architecture decisions for both SaaS products and client implementations.
Act as a technical contact on client-facing projects—translate requirements into technical designs and guide delivery.