Mexico
21 days ago
Remote Senior Devops Engineer

Key Responsibilities

Design, implement, and automate ML lifecycle workflows using tools like MLflow, Kubeflow, Airflow and OCI Data Science Pipelines. Build and maintain CI/CD pipelines for model training, validation, and deployment using GitHub Actions, Jenkins, or Argo Workflows. Collaborate with data engineers to deploy models within modern data lakehouse architectures (e.g., Apache Iceberg, Delta Lake, Apache Hudi). Integrate machine learning frameworks such as TensorFlow, PyTorch, and Scikit-learn into distributed environments like Apache Spark, Ray, or Dask. Operationalize model tracking, versioning, and drift detection using DVC, model registries, and ML metadata stores. Manage infrastructure as code (IaC) using tools like Terraform, Helm, or Ansible to support dynamic GPU/CPU training clusters. Configure real-time and batch data ingestion and feature transformation pipelines using Kafka, Goldengate and OCI Streaming. Collaborate with DevOps and platform teams to implement robust monitoring, observability, and alerting with tools like Prometheus, Grafana, and the ELK Stack. Support AI governance by enabling model explainability, audit logging, and compliance mechanisms aligned with enterprise data and security policies.  

Required Qualifications

Bachelor’s or Master’s degree in Computer Science, Data Science, or a related technical discipline. 5–8 years of experience in ML engineering, DevOps, or data platform engineering, with at least 2 years in MLOps or model operations. Proficiency in Python, particularly for automation, data processing, and ML model development. Solid experience with SQL and distributed query engines (e.g., Trino, Spark SQL). Deep expertise in Docker, Kubernetes, and cloud-native container orchestration tools (e.g., OCI Container Engine, EKS, GKE). Working knowledge of open-source data lakehouse frameworks and data versioning tools (e.g., Delta Lake, Apache Iceberg, DVC). Familiarity with model deployment strategies, including batch, real-time inference, and edge deployments. Experience with CI/CD pipelines (GitHub Actions, GitLab CI, Jenkins) and MLOps frameworks (Kubeflow, MLflow, Seldon Core). Competence in implementing monitoring and logging systems (e.g., Prometheus, ELK Stack, Datadog) for ML applications. Strong understanding of cloud platforms (OCI, AWS, GCP) and IaC tools (Terraform, CloudFormation).  

Preferred Qualifications

Experience integrating AI workflows with Oracle Data Lakehouse, Databricks, or Snowflake. Hands-on experience with orchestration tools like Apache Airflow, Prefect, or Dagster. Exposure to real-time ML systems using Kafka or Oracle Stream Analytics. Understanding of vector databases (e.g., Oracle 23ai Vector Search). Knowledge of AI governance, including model explainability, auditability, and reproducibility frameworks.  

Soft Skills

Strong problem-solving skills and an automation-first mindset. Excellent cross-functional communication, especially when collaborating with data scientists, DevOps, and platform engineering teams. A collaborative and knowledge-sharing attitude, with good documentation habits. Passion for continuous learning, especially in the areas of AI/ML tooling, open-source platforms, and data engineering innovation.  

 

Confirm your E-mail: Send Email
All Jobs from Oracle