• Proven experience as a Data Engineer or similar role with a strong track record of designing and implementing data solutions.
• Proficiency in programming languages such as Python (primary), Java, or Scala
• Hands-on experience with stream processing and CDC pipelines, particularly using Apache Kafka.
• Experience with analytical databases and real-time data warehouses, especially RisingWave and ClickHouse.
• Solid understanding of database concepts (SQL & NoSQL), including query optimization for high-performance analytics.
• Familiarity with data ingestion pipelines, ETL/ELT processes, and system integrations.
• Strong knowledge of data modeling, partitioning, and schema evolution in streaming/analytical environments.
• Experience with cloud platforms (AWS, GCP, or Azure) and services such as S3, BigQuery, Redshift, EMR, or EC2.
• Familiarity with observability and monitoring tools (e.g., Grafana, Prometheus, Kibana) for data pipeline monitoring.
• Experience in performance tuning for highly concurrent workloads.
• Basic understanding of data governance, security, and compliance practices.