Job Scope:
- Support the development of data-pipeline (spark, scala, pyspark)
- Support the conversion of developed ML model to be part of batch / streaming data pipelines (python to pySpark)
- Support the deployment of the of these ML models + data pipelines to production
- Support the operational running of these data pipelines within existing data lakes
Job Qualifications:
- Bachelor's Degree in Information Technology or equivalent
- 4 years of experience in Spark data pipeline development, possess PySpark (python + spark) development skills
- Experience in Big data environment development (HDFS, Shell Scripts)
- Familiar with Singapore Government data centre environment
- Familiar with Data Science / Machine Learning mode development & operationalisation concepts
- Ability to operate & develop within existing Big Data environment
- Ability to develop, test, deploy and manage data pipelines
- Ability to convert existing python code to pySpark or Spark/Scala code to run on distributed manner
- Ability to understand & operationalize ML / Analytics model
- Familiar with DataOps, MLOps and the data / model governance
- Good to have Apache Airflow, Spark, Python, ETL experience
- Proactive to learn, communicate, team players, consulting background will be good
Bernice Mae Nocum Rallonza EA License No. 02C3423 Personnel Registration No. R1442141
Please note that your response to this advertisement and communications with us pursuant to this advertisement will constitute informed consent to the collection, use and/or disclosure of personal data by ManpowerGroup Singapore for the purpose of carrying out its business, in compliance with the relevant provisions of the Personal Data Protection Act 2012. To learn more about ManpowerGroup's Global Privacy Policy, please visit https://www.manpower.com.sg/privacy-policy