Apply for Job
DataOps Engineer
SG
Job Description
As a DataOps Engineer, you will operate large-scale big data platforms across hybrid (on-premises and cloud) environments, enabling reliable analytics and data-driven use cases. You will work closely with data engineers, data scientists, infrastructure, security, and business stakeholders to ensure data quality, platform stability, and operational excellence.This role focuses on building, running, and optimizing production-grade data platforms and pipelines, with strong ownership of infrastructure, automation, reliability, and operations.
Key Responsibilities
1. Manage On-Prem and Cloud Data Platforms
- Maintain and support on-premises clusters, including compute, storage, networking, and system configurations.
- Provision, configure, and manage cloud infrastructure, including AWS S3, EMR, Redshift, and RDS.
- Monitor platform performance, capacity, and availability; implement operational alerts, monitoring, and runbooks.
2. Build, Operate & Support Data Pipelines
- Develop and support ETL/ELT pipelines using PySpark and Airflow, ensuring reliable ingestion, transformation, and data loading.
- Operate pipelines across hybrid environments, handling job execution, retries, basic performance tuning, and failure recovery.
- Validate, clean, and standardize datasets; monitor pipeline health, data freshness SLAs, and failure patterns, and perform root cause analysis and remediation.
- Support data storage platforms such as S3, PostgreSQL, Redshift, and MongoDB from an operational and platform perspective.
- Automate DataOps and MLOps workflows using CI/CD pipelines (e.g., GitLab CI/CD, Jenkins).
3. Security, Compliance & Governance
- Implement secure access controls (IAM, VPC, security groups), encryption, backups, and disaster recovery mechanisms.
- Partner with infrastructure and security teams to ensure compliance with PDPA, GDPR, and internal governance policies.
4. Cross-Functional Collaboration & Documentation
- Work with data engineers, data scientists, infrastructure, security, and business stakeholders to clarify requirements and operational expectations.
- Document schemas, pipelines, configurations, and operational runbooks to support maintainability and knowledge sharing.
Requirements
- Degree in Computer Science, Software Engineering, Data Science, or equivalent experience
- 2–5+ years of hands-on experience in DevOps, Data Platform Engineering, or Infrastructure Engineering, supporting big data or analytics platforms.
- Experience managing on-premises and cloud infrastructure, including compute, storage, and network configuration.
- Hands-on experience with cloud and/or hybrid environments supporting big data pipelines, with Spark / PySpark and Airflow at an operational level (job execution, basic tuning, failure handling).
- Knowledge of containerization and orchestration, including Docker fundamentals and Kubernetes (deployments, services, ingress, scaling, resource limits).
- Strong understanding of networking fundamentals, including VPC design, routing, DNS, firewalls, and load balancing.
- Knowledge of security concepts, such as IAM, access control, and compliance basics.
- Experience with CI/CD pipelines and automation using tools such as GitLab CI/CD, Jenkins, or similar.
- Basic understanding of MLOps concepts from a platform and infrastructure support perspective.
- Strong troubleshooting and problem-solving skills in production environments.
- Comfortable working with cross-functional teams (data engineers, ML engineers, security, product).