Oracle logo

Remote Senior Devops Engineer

Oracle
Full-time
Remote
Mexico
Remote Developer and Engineering
Description

Key Responsibilities

  • Design, implement, and automate ML lifecycle workflows using tools like MLflow, Kubeflow, Airflow and OCI Data Science Pipelines.
  • Build and maintain CI/CD pipelines for model training, validation, and deployment using GitHub Actions, Jenkins, or Argo Workflows.
  • Collaborate with data engineers to deploy models within modern data lakehouse architectures (e.g., Apache Iceberg, Delta Lake, Apache Hudi).
  • Integrate machine learning frameworks such as TensorFlow, PyTorch, and Scikit-learn into distributed environments like Apache Spark, Ray, or Dask.
  • Operationalize model tracking, versioning, and drift detection using DVC, model registries, and ML metadata stores.
  • Manage infrastructure as code (IaC) using tools like Terraform, Helm, or Ansible to support dynamic GPU/CPU training clusters.
  • Configure real-time and batch data ingestion and feature transformation pipelines using Kafka, Goldengate and OCI Streaming.
  • Collaborate with DevOps and platform teams to implement robust monitoring, observability, and alerting with tools like Prometheus, Grafana, and the ELK Stack.
  • Support AI governance by enabling model explainability, audit logging, and compliance mechanisms aligned with enterprise data and security policies.
 

Required Qualifications

  • Bachelor’s or Master’s degree in Computer Science, Data Science, or a related technical discipline.
  • 5–8 years of experience in ML engineering, DevOps, or data platform engineering, with at least 2 years in MLOps or model operations.
  • Proficiency in Python, particularly for automation, data processing, and ML model development.
  • Solid experience with SQL and distributed query engines (e.g., Trino, Spark SQL).
  • Deep expertise in Docker, Kubernetes, and cloud-native container orchestration tools (e.g., OCI Container Engine, EKS, GKE).
  • Working knowledge of open-source data lakehouse frameworks and data versioning tools (e.g., Delta Lake, Apache Iceberg, DVC).
  • Familiarity with model deployment strategies, including batch, real-time inference, and edge deployments.
  • Experience with CI/CD pipelines (GitHub Actions, GitLab CI, Jenkins) and MLOps frameworks (Kubeflow, MLflow, Seldon Core).
  • Competence in implementing monitoring and logging systems (e.g., Prometheus, ELK Stack, Datadog) for ML applications.
  • Strong understanding of cloud platforms (OCI, AWS, GCP) and IaC tools (Terraform, CloudFormation).
 

Preferred Qualifications

  • Experience integrating AI workflows with Oracle Data Lakehouse, Databricks, or Snowflake.
  • Hands-on experience with orchestration tools like Apache Airflow, Prefect, or Dagster.
  • Exposure to real-time ML systems using Kafka or Oracle Stream Analytics.
  • Understanding of vector databases (e.g., Oracle 23ai Vector Search).
  • Knowledge of AI governance, including model explainability, auditability, and reproducibility frameworks.
 

Soft Skills

  • Strong problem-solving skills and an automation-first mindset.
  • Excellent cross-functional communication, especially when collaborating with data scientists, DevOps, and platform engineering teams.
  • A collaborative and knowledge-sharing attitude, with good documentation habits.
  • Passion for continuous learning, especially in the areas of AI/ML tooling, open-source platforms, and data engineering innovation.
 

 



Qualifications

Career Level - IC4