Back to search:Data Engineer / Jakarta (central)

Requirements:

  • At least 1 year of experience in designing and building data pipelines and ETL processes.
  • Bachelor's degree in Engineering or a related field; a Master's degree is a plus.
  • Strong proficiency in Google Cloud Platform (GCP) services such as BigQuery, Dataflow, Pub/Sub, Composer/Airflow, and Cloud Storage.
  • Solid programming skills in languages such as Python, Java, or Scala.
  • Experience with data modeling, data warehousing concepts, and relational databases.
  • Proficiency in SQL and database design principles.
  • Familiarity with data integration and ETL tools.
  • Familiarity with cloud platforms (e.g., AWS, Azure, Google Cloud).
  • Strong understanding of data governance, data security, and data quality best practices.

Job desc:

  • Design, develop, and maintain data pipelines, ETL processes, and data warehouses.
  • Ensure data quality, integrity, and reliability throughout the data pipeline.
  • Work with GCP services such as Google Cloud Storage, BigQuery, Dataflow, and Pub/Sub to architect and implement scalable and efficient data processing solutions.
  • Monitor and troubleshoot data pipeline issues.
  • Implement security and access controls to protect sensitive data and ensure compliance with data privacy regulations.
  • Document data processes, data flows, and architectural decisions.