Data Engineer (Control M, Hadoop, Spark, Hive, Big Data)
NEPTUNEZ SINGAPORE PTE. LTD.
Date: 2 days ago
Area: Singapore, Singapore
Salary:
SGD 8,000
-
SGD 9,000
per month
Contract type: Full time

Responsibilities:
- Design, implement, and maintain data pipelines for the migration of data from on-premises systems to Hadoop-based platforms.
- Develop and optimize data processing jobs using Spark, Hive, and Python.
- Manage job orchestration and scheduling using Control-M, ensuring timely and accurate data delivery.
- Collaborate with cross-functional teams to understand data requirements and deliver efficient solutions.
- Perform code quality checks and peer reviews to ensure best practices, maintainability, and adherence to coding standards.
- Ensure end-to-end operational stability of data pipelines by proactively identifying and resolving bottlenecks, failures, and data quality issues.
- Ensure data quality through rigorous cleaning and validation processes.
- Documented data flow processes, transformation logic, and framework usage to support onboarding and troubleshooting.
Requirements:
- Proficiency in Python and SQL.
- Strong experience with Hadoop ecosystem tools (Hive, Spark).
- Worked extensively with transformation components, mapping development, and workflow orchestration in Informatica/DataStage.
- Experience with job scheduling and monitoring using Control-M.
- Familiar with pipeline-as-code concepts and using Jenkins files for automation of build and deployment processes.
- Solid understanding of database systems including Teradata, Oracle, and SQL Server.
- Ability to analyze and troubleshoot large-scale data processing systems.
- Experience in the banking or financial services industry.
- Knowledge of data warehousing concepts and star/snowflake schema design.
See more jobs in Singapore