OPEN POSITION

Senior AI Platform Engineer

We are now expanding our team and are looking for skilled, goal-oriented backend developers to join our platform and custom solution teams.

Apply now

Our stack

Python 3.12 and Go as our core languages - Python for rapid development, Go for performance-critical services and internal toolsCloud-native infrastructure: Kubernetes orchestration with Helm, multi-cloud deployment across AWS, GCP, and AzuregRPC, REST APIs, WebSockets and Kafka for service communicationGitHub Actions for CI/CD, Docker for containerization, with Grafana, Prometheus, and OpenTelemetry for observabilityPostgreSQL, ClickHouse for analytics, and specialized vector databases for AI workloads

Responsibilities

Architect high-performance real-time systems: Design micro-services and streaming architectures that handle millions of concurrent voice/text AI interactions with ultra-low latency

Create foundational platform components: Build reusable infrastructure blocks (rate limiters, service mesh, circuit breakers), developer tools (APIs, SDKs), and observability solutions that enable teams to ship AI features rapidly

Optimize AI infrastructure end-to-end: Manage GPU clusters, design model serving pipelines, implement vector similarity search, and collaborate with ML engineers to ensure efficient model deployment and inference

Enable engineering excellence: Establish platform standards, deployment patterns, and self-service automation that empower product teams to move fast while maintaining reliability

Requirements

Pragmatic engineering approach: Balance speed and quality by shipping working solutions quickly when needed, while keeping code clear and maintainable over clever abstractions

Strong programming skills in Python (expert level with type systems and modern ecosystem) and Go (for performance-critical components)

Production-scale distributed systems experience: Building and operating micro-services architectures with service discovery, API gateways, event-driven patterns, and proven ability to handle large-scale production workloads

Real-time systems and observability expertise: Experience with streaming architectures, message queues, low-latency optimization, plus deep knowledge of distributed tracing, metrics aggregation, and log analysis at scale

Strategic thinking: For critical architectural decisions, you invest time to evaluate multiple approaches, present top candidates with trade-offs, and document why alternatives were rejected

Outstanding technical communication: Ability to clearly explain complex architectural decisions, write comprehensive documentation, facilitate productive technical discussions, and effectively communicate with both engineering teams and stakeholders

What we offer?

  • High-impact projects with prominent clients and our proprietary AI platform development

  • Cutting-edge technology stack with the latest tools in AI infrastructure and platform engineering

  • Budgets for the latest AI models (GPT-4, Claude, etc.) to ensure our team has the best tools available

  • Rapid career progression working alongside senior professionals from leading tech companies

  • Remote work flexibility from anywhere globally

  • Direct influence on platform architecture decisions that shape our AI products

Our stack

Python 3.12 and Go as our core languages - Python for rapid development, Go for performance-critical services and internal toolsCloud-native infrastructure: Kubernetes orchestration with Helm, multi-cloud deployment across AWS, GCP, and AzuregRPC, REST APIs, WebSockets and Kafka for service communicationGitHub Actions for CI/CD, Docker for containerization, with Grafana, Prometheus, and OpenTelemetry for observabilityPostgreSQL, ClickHouse for analytics, and specialized vector databases for AI workloads

Responsibilities

Architect high-performance real-time systems: Design micro-services and streaming architectures that handle millions of concurrent voice/text AI interactions with ultra-low latency

Create foundational platform components: Build reusable infrastructure blocks (rate limiters, service mesh, circuit breakers), developer tools (APIs, SDKs), and observability solutions that enable teams to ship AI features rapidly

Optimize AI infrastructure end-to-end: Manage GPU clusters, design model serving pipelines, implement vector similarity search, and collaborate with ML engineers to ensure efficient model deployment and inference

Enable engineering excellence: Establish platform standards, deployment patterns, and self-service automation that empower product teams to move fast while maintaining reliability

Requirements

Pragmatic engineering approach: Balance speed and quality by shipping working solutions quickly when needed, while keeping code clear and maintainable over clever abstractions

Strong programming skills in Python (expert level with type systems and modern ecosystem) and Go (for performance-critical components)

Production-scale distributed systems experience: Building and operating micro-services architectures with service discovery, API gateways, event-driven patterns, and proven ability to handle large-scale production workloads

Real-time systems and observability expertise: Experience with streaming architectures, message queues, low-latency optimization, plus deep knowledge of distributed tracing, metrics aggregation, and log analysis at scale

Strategic thinking: For critical architectural decisions, you invest time to evaluate multiple approaches, present top candidates with trade-offs, and document why alternatives were rejected

Outstanding technical communication: Ability to clearly explain complex architectural decisions, write comprehensive documentation, facilitate productive technical discussions, and effectively communicate with both engineering teams and stakeholders

What we offer?

  • High-impact projects with prominent clients and our proprietary AI platform development

  • Cutting-edge technology stack with the latest tools in AI infrastructure and platform engineering

  • Budgets for the latest AI models (GPT-4, Claude, etc.) to ensure our team has the best tools available

  • Rapid career progression working alongside senior professionals from leading tech companies

  • Remote work flexibility from anywhere globally

  • Direct influence on platform architecture decisions that shape our AI products

© 2022—2024 Aiphoria

UK, Cyprus, Portugal, UAE