Back to search:Data Engineer / Jakarta

The Data Engineer is responsible for designing, building, and maintaining scalable data pipelines and analytics-ready datasets on modern cloud data platforms. This role enables business teams to generate insights and supports AI-driven analytics delivery, sitting at the intersection of data engineering, BI enablement, and advanced analytics.

Key Responsibilities

Data Engineering & Platform Development

o      Design, build, and maintain scalable ETL/ELT pipelines for structured and semi-structured data.

o      Develop and optimize cloud data warehouses or lakehouses (e.g., BigQuery, Snowflake).

o      Build reusable data transformation models (SQL, Python, Dataform/dbt).

o      Implement data quality, validation, and lineage controls.

o      Integrate data from multiple systems (APIs, SaaS, databases, streaming).

o      Support ingestion & streaming workloads (e.g., Pub/Sub, Kafka).

o      Collaborate with analytics and AI teams to ensure reliable and accessible data for downstream use.

Business Intelligence & Data Enablement

o      Build and automate dashboards and reporting pipelines, focusing on Looker Studio.

o      Design and maintain semantic layers/data marts optimized for BI workloads.

o      Translate business questions into analytical datasets and dashboards.

o      Work closely with business stakeholders to define metrics, KPIs, and data models.

o      Ensure data accuracy, usability, and performance for reporting and self-service analytics.

Collaboration & Delivery

o      Partner with architects, analysts, and AI teams to deliver end-to-end solutions.

o      Document pipelines, data models, and processes.

o      Participate in code reviews and engineering best practices.

o      Provide knowledge sharing and mentorship to junior team members.

[Optional Bonus] Generative AI & Advanced Analytics

o      Exposure to GenAI use cases (chatbots, RAG, document Q&A).

o      Familiarity with vector DBs (FAISS, Pinecone) and frameworks (LangChain/LangGraph).

o      Experience integrating analytics models into Vertex AI/OpenAI pipelines.

Requirements

Experience

o      ~5 years of hands-on data engineering or backend data platform experience.

o years working with cloud ecosystems (GCP preferred).

o      Demonstrated experience delivering production-grade data pipelines & BI reports.

Technical Skills

o      Programming: Proficient in Python and SQL; experience with data transformation and automation scripts.

o      Data Platforms: Experience with BigQuery, Snowflake, Redshift, or Databricks.

o      Orchestration Tools: Familiar with Airflow, Composer, or dbt/Dataform.

o      BI & Visualization: Looker Studio (preferred), Looker, Power BI, Tableau.

o      Big Data Frameworks: Exposure to Spark, Beam, or Kafka is an advantage.

o      Version Control & CI/CD: Familiarity with Git, Cloud Build, or similar tools.

o      Data Governance: Understanding of metadata, lineage, and access control best practices (bonus).

Soft Skills

o      Strong analytical and problem-solving skills.

o      Good communication and collaboration skills to work with cross-functional teams.

o      Proactive, collaborative, and eager to learn—especially in AI & analytics.

Preferred Qualifications

o      Bachelor's degree in Computer Science, Engineering, or related field.

o      Google Cloud Professional Data Engineer or equivalent cloud certification (preferred).

o      Experience contributing to AI/ML or Gen AI projects (nice to have).

o      Fluent in English.