A company located in Seoul, South Korea, is seeking a Data Pipeline Engineer to join its team.
The company's organization serves as a data hub, reliably managing large-scale data and building and operating flexible and scalable systems tailored to business needs. It ensures data storage and processing with stability and performance in mind, while quickly analyzing and processing data to contribute to business outcomes across various services and applications.
Role involves:
- Developing and operating batch/real-time pipelines for data from various sources.
- Developing and operating a governance system that provides services such as data quality and discovery for highly reliable data provision.
- Performing table processing/performance improvement tasks for data analysis, machine learning, and data product development of internal consumption organizations.
Requirements:
- Minimum of 5 years of data engineering experience.
- Sufficient knowledge and operational experience with data lakehouses such as Apache Iceberg and Apache Hudi.
- Extensive knowledge and experience with large-volume data processing and distributed systems such as Apache Spark and Apache Flink.
- Experience in building and operating data pipelines in AWS cloud and Kubernetes environments.
- Experience in directly building a service monitoring environment and proactively monitoring for failures.
Role offers:
- The opportunity to develop and manage large-scale distributed processing systems proactively.
- The opportunity to designate and manage company-wide central data management standards.
- High level of compliance and governance experience.
- Sharing new technologies and know-how through regular team and data weekly technical seminars.