Requirements:
- Bachelor's degree in Computer Science, Information Systems, or related field.
- 3+ years of experience in data engineering or data operations roles.
- Hands-on experience with GCP data services (BigQuery, Dataflow, Pub/Sub, Composer, etc.).
- Strong experience managing and optimizing IBM Db2 databases.
- Proficiency in SQL (advanced query tuning and performance analysis).
- Experience with Python or other scripting languages for automation.
- Familiarity with ETL/ELT frameworks and data workflow orchestration tools (Airflow, Cloud Composer).
- Strong understanding of data modeling, data warehousing, and performance optimization.
- Knowledge of cloud networking, IAM, and security best practices.
- Good analytical, troubleshooting, and communication skills.
Responsibilities:
- Manage and maintain data infrastructure across GCP (BigQuery, Cloud Storage, Dataflow, Composer, Pub/Sub) and Db2 databases (on-prem or cloud).
- Develop, deploy, and monitor ETL/ELT data pipelines using GCP tools or open-source frameworks (e.g., Apache Beam, Airflow, Dataform).
- Implement data ingestion, transformation, and integration processes between Db2, GCP, and other enterprise systems.
- Perform data quality checks, error handling, and pipeline optimization.
- Administer and tune Db2 databases, including query optimization, backup/recovery, and replication.
- Monitor and troubleshoot data jobs, ensuring SLAs and data freshness are met.
- Automate routine operational tasks using scripting languages such as Python, Bash, or SQL.
- Collaborate with DevOps and platform teams to implement CI/CD pipelines for data workloads.
- Support data governance, compliance, and security policies within GCP and Db2 environments.
- Create and maintain technical documentation, dashboards, and operational runbooks.