SR. DEVOPS ENGINEER

XTN-E810728

City
N/A, N/A
Schedule
Office Location
KMC site or other business locations as may be instructed by KMC Solutions from time to time

Make your next big career move by applying as KMC Solutions' next SR. DEVOPS ENGINEER

.About the Role
We are looking for a DevOps Engineer to own and evolve the infrastructure that powers Anervea's AI platform and supporting services. You will be responsible for designing, deploying, and maintaining a reliable, secure, and cost-efficient cloud environment on AWS, with Docker-based workloads at its core.
This is a hands-on role for someone who enjoys the full lifecycle of infrastructure work — from provisioning and automation to monitoring, incident response, and continuous improvement. You will partner closely with engineering, AI/ML, and product teams to ship reliably and scale gracefully.

On top of your salary, here are the exciting benefits you can look forward to:

•  Health Insurance/HMO 
•  Enjoy unlimited MadMax Coffee
•  Diverse learning & growth opportunities
•  Accessible Cloud HR platform (Sprout)
•  Above standard leaves

The main responsibilities of a SR. DEVOPS ENGINEER include:

Tech Stack You’ll Support
You will be deploying, scaling, and maintaining infrastructure for the following stack:

  • Mobile applications built with Flutter.
  • Web frontend built with React.
  • Backend services built with Python (FastAPI).
  • AI/ML workloads including LLM inference, RAG pipelines, and supporting services.
  • Cloud platform: AWS (primary).
  • Containerization with Docker; orchestration via ECS or EKS.
  • Databases and storage: PostgreSQL/RDS, S3, vector stores, and caching layers (Redis).

    Key Responsibilities
  • AWS Infrastructure
  • Design, provision, and maintain AWS infrastructure across services including EC2, ECS/EKS, S3, RDS, VPC, IAM, CloudFront, Route 53, ELB/ALB, Lambda, and CloudWatch.
  • Implement Infrastructure as Code (IaC) using Terraform, CloudFormation, or AWS CDK to ensure repeatable and version-controlled environments.
  • Manage networking (VPCs, subnets, security groups, NAT gateways, VPN/peering) with a strong focus on security and least-privilege access.
  • Optimize AWS spend through right-sizing, reserved instances/savings plans, autoscaling, and continuous cost monitoring.
  • Docker & Containerization
  • Build, optimize, and maintain Docker images for backend services, AI/ML workloads, and supporting tooling.
  • Manage container orchestration on Amazon ECS (Fargate/EC2) or EKS (Kubernetes), including service definitions, task scaling, and rolling deployments.
  • Maintain private container registries (ECR) with tagging, scanning, and lifecycle policies.
  • Troubleshoot container-level issues across networking, storage, and runtime performance.
  • CI/CD & Automation
  • Build and maintain CI/CD pipelines using GitHub Actions, GitLab CI, Jenkins, or AWS CodePipeline / CodeBuild.
  • Automate testing, container builds, image scanning, and zero-downtime deployments to staging and production environments.
  • Standardize deployment workflows across services so engineers can ship safely and quickly.
  • Monitoring, Reliability & Incident Response
  • Implement and maintain observability across the stack using CloudWatch, Prometheus, Grafana, ELK/OpenSearch, Datadog, or similar.
  • Define and track SLOs/SLAs, set up actionable alerting, and reduce noise in on-call workflows.
  • Lead incident response, conduct root-cause analyses, and drive postmortems with clear follow-up actions.
  • Plan and test backup, disaster recovery, and high-availability strategies.
  • Security & Compliance
  • Apply security best practices across IAM, secrets management (AWS Secrets Manager / Parameter Store), encryption at rest and in transit, and network security.
  • Manage SSL/TLS certificates, WAF rules, and DDoS protection.
  • Support compliance, audit, and data-protection requirements relevant to healthcare/AI workloads (e.g., data residency, access logging).
  • Generative AI Infrastructure
  • Support and maintain infrastructure for LLM and generative AI workloads on AWS, including Amazon Bedrock, SageMaker, and self-hosted model endpoints.
  • Help manage GPU-based compute (EC2 P/G instances, Inferentia) for model inference and training workloads, with a focus on cost and throughput optimization.
  • Set up and operate vector databases (OpenSearch, pgvector, Pinecone) and supporting RAG infrastructure.
  • Build secure, observable pipelines for model deployment, versioning, and rollback.
  • Monitor token usage, inference latency, and model-serving costs; implement guardrails and rate limiting where needed.
  • Manage API gateways and authentication for AI endpoints exposed to internal and external consumers.
  • Collaboration
  • Work closely with backend, frontend, and AI/ML engineers to understand workload requirements and provide reliable infrastructure.
  • Document infrastructure, runbooks, and operational procedures so the team can operate confidently.
  • Mentor engineers on DevOps practices, deployment hygiene, and cloud-cost awareness.

To apply, you must be an expert on the following requirements:

Required Qualifications

  • 3-6 years of professional experience in DevOps, SRE, Cloud, or Platform Engineering roles.
  • Strong hands-on experience with AWS across compute, networking, storage, and managed services.
  • Solid expertise in Docker — image building, multi-stage builds, and container runtime troubleshooting.
  • Production experience with at least one container orchestrator (ECS, EKS, or Kubernetes).
  • Proficiency in at least one IaC tool (Terraform preferred; CloudFormation or AWS CDK acceptable).
  • Strong scripting skills in Bash and Python (or Go) for automation and tooling.
  • Working knowledge of Linux administration, networking fundamentals (DNS, TCP/IP, HTTP/S, load balancing), and Git-based workflows.
  • Comfort deploying and operating applications across mobile (Flutter), web (React), and Python backend (FastAPI) stacks — you don’t need to write the code, but you should understand how to build, ship, and troubleshoot it.
  • Hands-on experience setting up and maintaining CI/CD pipelines.
  • Experience with monitoring, logging, and alerting tools in production environments.
  • Exposure to or genuine interest in operating generative AI / LLM workloads in production (Bedrock, SageMaker, self-hosted inference, or similar) — prior experience preferred but not required.


Nice to Have

  • AWS certifications (Solutions Architect, DevOps Engineer, SysOps Administrator).
  • Experience operating Kubernetes (EKS) at scale, including Helm and GitOps tools (ArgoCD, Flux).
  • Hands-on experience with AI/ML infrastructure at scale: GPU instance optimization, model serving frameworks (vLLM, Triton, TGI), fine-tuning pipelines, or production RAG systems.
  • Experience with service meshes (Istio, App Mesh) and API gateways.
  • Familiarity with multi-account AWS architectures (AWS Organizations, Control Tower, Landing Zone).
  • Experience with healthcare or regulated-data environments (HIPAA, GDPR, SOC 2).
  • Familiarity with NGINX, reverse proxies, and domain/SSL management workflows.

It will also be favorable if you are knowledgeable in:

Additional relevant knowledge or experience related to the above requirements will be considered an advantage.

The successful candidate must submit the following pre-employment requirements

  • Scanned copy of valid NBI Clearance
  • Accomplished Medical or PEME Slip (covered by KMC)
  • 2x2 & Half body picture with white background
  • Proof of government numbers (TIN, SSS, Pag-ibig, & Philhealth)
  • Photocopy of 2 valid IDs – front & back (government-issued)
  • Clear copy of your Birth Certificate (PSA or NSO)
  • Accomplished HR Forms & Promissory Note (will be provided by KMC’s Onboarding Team

Click here to view the complete list of KMC’s pre-employment requirements.

KMC Careers

If you're a rockstar at what you do and looking to be a part of our amazing story, we want to hear from you!

We offer attractive salaries and benefits plus you get to work in some of the Philippines' best flexible workspaces. Our employees also get to enjoy exclusive discounts, rewards and freebies, and invites to our monthly events. We are always recruiting for roles in IT & Development, Marketing, Business Administration, HR & Recruitment and Legal & Finance Roles.

KMC provides quality employment opportunities for job-seekers looking for a career that is both challenging and fulfilling. We are also committed to providing equal opportunities at every selection stage. We do not discriminate due to age, gender, sexual orientation, ethnicity, nationality, and religion.

Work with Us. Grow with Us.

KMC Solutions offers a variety of career opportunities in Metro Manila, Cebu and Clark & Iloilo. We are always looking for talented and enthusiastic individuals who are ready to make their next big career move.

Our Culture

At KMC, we foster an inclusive and positive workplace for all. We push our members to succeed in everything they do through our collaborative work environment. We encourage our community to work hard and reach their full potential while delivering results that matter for our members and you as professionals.

We host amazing and quality events and implement people-centric policies to work flexibly. We ensure that everyone in our expansive network is engaged, from our internal employees and those who work on behalf our offshore partners.

Life within KMC: Work Hard Party Harder

At KMC, we work hard and we are committed to putting our best foot forward in everything we do. Everyone is encouraged to be an individual while also working for the collective good of the KMC Community. We believe mistakes are opportunities and that you should not present a solution without a problem.

We also know when hard work deserves to be recognized so we reward our employees with monthly parties, free trips and much much more!

No account yet

Sign up to view exciting career opportunities!