Job Openings Data Executive/Specialist (Newton)

About the job Data Executive/Specialist (Newton)


You will work closely with the DCR Data Manager to:

  1. handle V3 data queries and requests from various government agencies, Institutes of Higher Learning (IHLs), Research Institutes (RIs), other non-government/private entities, and the public.
  1. help manage DCR data collections on the National Super-Computing Centre (NSCC) platform, which includes V3 data, and other observational datasets.

Your responsibilities include:

  1. monitoring the Organisation email inbox to ensure timely response to queries about V3 data.
  2. ensuring that the MSS Data Sharing Form is filled accordingly by the Data Requestor.
  3. processing the data requests within 7 days following the MSS Data Sharing Procedure for Standard/Time-sensitive Datasets.
  4. working in a Linux environment and accessing the National Supercomputer Centre (NSCC) to retrieve/extract the required data.
  5. maintaining the log of V3 data requests.
  6. maintaining the external V3 Data Catalogue used for public/agency reference (Excel-based).
  7. Creating and maintaining the overall V3 Data Collections Inventory for internal reference.
  8. Helping to create a digital V3 data catalog using Python Intake package and exploring cloud-friendly formats for efficient data storage.
  9. Undertaking any other data-related work delegated by Supervisor.

Position Requirements:

  • Familiarity in working in a High-Performance Computing environment (i.e., with Linux)
  • Proficiency in handling/processing/serving big data requests (in the hundreds of TB)
  • Proficiency in Bash scripting
  • Proficiency in Python (Xarray library).
  • Ability to work independently and learn on-the-job quickly.
  • Good written and oral communication skills.
  • Meticulous with an eye for detail.
  • Proficiency in Microsoft Office applications.

Position Desirables:

  • Prior 1 to 2-year experience working in Data/Information Management
  • Familiarity with the NetCDF file format (commonly used for storing gridded/array-oriented geospatial data).
  • Familiarity with online wiki-like documentation software tools (e.g. Confluence).
  • Familiarity with the Python package Intake.
  • Familiarity with the Conda/Mamba package manager in Python.
  • Familiarity with cloud-friendly data formats (e.g. zarr)

Duration/ Working hours:

7.5 months contract (Immediate)

42 hours per week

** We regret to inform only shortlisted candidates will be notified.