Back to search:Data Engineer / Jakarta

Requirements:

  • Bachelor's degree in Computer Science, Information Systems, or related field.
  • 3+ years of experience in data engineering or data operations roles.
  • Hands-on experience with GCP data services (BigQuery, Dataflow, Pub/Sub, Composer, etc.).
  • Strong experience managing and optimizing IBM Db2 databases.
  • Proficiency in SQL (advanced query tuning and performance analysis).
  • Experience with Python or other scripting languages for automation.
  • Familiarity with ETL/ELT frameworks and data workflow orchestration tools (Airflow, Cloud Composer).
  • Strong understanding of data modeling, data warehousing, and performance optimization.
  • Knowledge of cloud networking, IAM, and security best practices.
  • Good analytical, troubleshooting, and communication skills.

Responsibilities:

  • Manage and maintain data infrastructure across GCP (BigQuery, Cloud Storage, Dataflow, Composer, Pub/Sub) and Db2 databases (on-prem or cloud).
  • Develop, deploy, and monitor ETL/ELT data pipelines using GCP tools or open-source frameworks (e.g., Apache Beam, Airflow, Dataform).
  • Implement data ingestion, transformation, and integration processes between Db2, GCP, and other enterprise systems.
  • Perform data quality checks, error handling, and pipeline optimization.
  • Administer and tune Db2 databases, including query optimization, backup/recovery, and replication.
  • Monitor and troubleshoot data jobs, ensuring SLAs and data freshness are met.
  • Automate routine operational tasks using scripting languages such as Python, Bash, or SQL.
  • Collaborate with DevOps and platform teams to implement CI/CD pipelines for data workloads.
  • Support data governance, compliance, and security policies within GCP and Db2 environments.
  • Create and maintain technical documentation, dashboards, and operational runbooks.