Skip to content

aodn/rimrep-documentation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

97 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Great Barrier Reef Data Management System (GBR DMS)

IMOS is leading a project to deliver a fit-for-purpose Data Management System (DMS) for the Reef 2050 Integrated Monitoring and Reporting Program (RIMReP).

The DMS will be critical infrastructure that underpins the Program to create a better understanding of what’s happening on the Great Barrier Reef. Also, the DMS will be a critical service in support of RIMReP’s Reef Knowledge System (https://reefknowledgesystem.gbrmpa.gov.au/).

The DMS architecture is conceived as a scalable data-agnostic system based on a cloud-native environment and built using open-source tools based on standards. The system will collect and update data and metadata from data providers in an automated way, transform the data collections into analysis-ready cloud-optimised (ARCO) formats, and provide the end user with a delivery mechanism through dedicated Application Programming Interface (API) services or direct connection to the cloud storage.

For more details go to the project webpage or contact us at info-dms at utas dot edu dot au.

Access the DMS catalogue and data services

The DMS catalogue is available at http://stac.reefdata.io/browser. If your organisation is an AAF partner you can use your email credentials to log in. If not, you can write to info-dms at utas dot edu dot au and we will provide your access credentials.

Docs

Public Repositories

Table of contents

More information

What is GBR DMS?

GBR DMS is an Open Geospatial Consortium (OGC) API service and analysis-ready, cloud-optimised (ARCO) repository for data and metadata relevant to the management of the Great Barrier Reef. GBR DMS offers services to allow the discovery of the data and the interaction with external systems.

In simple terms, GBR DMS is a data system that aims to be a "one-stop-shop" for all data related to the Great Barrier Reef World Heritage Area, which can be easily accessed by the Great Barrier Reef Marine Park Authority (GBRMPA) to support evidence-based management strategies. All datasets have a standard format regardless of their origin, which not only facilitates access to data, but also their analysis as it removes the need to understand and parse different data formats.

Datasets that have an open licence are publicly available in the GBR DMS, while datasets that have a restricted licence are only available to users that have been granted access to them by the data provider.

Table of contents

Discovering datasets in GBR DMS

To browse datasets available in the GBR DMS you can use the SpatioTemporal Asset Catalogs (STAC). STAC provides a common language to describe a range of geospatial information, so that data can be indexed and easily discovered. Our STAC catalogue is available at https://stac.reefdata.io/browser/.

Alternatively, we provide a link to the original source of the dataset to give users the option of accessing the data directly from the data provider.

Discovering datasets via STAC

The STAC catalogue is a web-based interface that allows users to search for datasets using a range of filters, such as dataset name, data provider, keywords and date range. To search for datasets, you have the option of clicking on the Search button on the top right corner of the page, or you can use the search bar on the top left corner of the page. These two options are highlighted in red boxes in the image below.

Screenshot of STAC catalogue home page showing the two search options mentioned in the previous paragraph

Datasets available via STAC are organised by collections, each containing one or multiple datasets or items that are related to each other. To illustrate this, we will use the AIMS Sea Water Temperature Observing System as an example.

Screenshot of AIMS Sea Water Temperature collection page showing a single item is available in this collection

The collection level page includes the following information:

  • A description of the collection, which is a brief summary of the datasets available in the collection.
  • The items or datasets available in the collection. In this case, we can see that there is a single item available in the collection.
  • The license under which the datasets are available.
  • The temporal coverage of the datasets.
  • A map showing the spatial coverage of the datasets.
  • Information about the data provider .

If you click on the item name (in this case, AIMS Sea Water Temperature Observing System), you will be taken to the item level page.

Screenshot of AIMS Sea Surface Temperature Monitoring Program item page showing the item level information. Links to S3 bucket and API are highlighted in red boxes

The item level page includes the following information:

  • A map showing the spatial coverage of the dataset.
  • A description of the dataset.
  • A link to the collection level page.
  • A link to the dataset available in a GBR DMS S3 bucket under the Assets section.
  • Under the Additional Resources section, there will be a link to the data API under and to the original source of the dataset.
  • Metadata about the dataset, including the projection system, preferred citation and the names of the columns in the dataset.

The API and S3 links are highlighted in red boxes in the image above because these are the two methods shown in this repository to access datasets available in the GBR DMS.

Table of contents

Code snippets

In this section, we are including code snippets that will help you get started with the GBR DMS. These snippets are available in R and Python, simply select the language you want to use from the tabs below.

Connecting to tabular dataset in S3 bucket

To run this code in R or Python, you will need to have the S3 URL address for the dataset of your interest. For this example, we are using the AIMS Sea Surface Temperature Monitoring Program dataset, but you can simply replace the S3 URL address with the one for the dataset you want to access.

You can get this URL following the instructions in the Searching for datasets via STAC section above.

Instructions for R users
# Loading arrow library to connect to S3 bucket
library(arrow)
# Providing S3 URL address for dataset of interest
dataset_s3 <- "s3://gbr-dms-data-public/aims-temp-loggers/data.parquet"
# Connecting to S3 bucket
s3_conn <- s3_bucket(dataset_s3)
# Accessing dataset
ds <- open_dataset(s3_conn)

Remember that you can change the value of dataset_s3 to the S3 URL address for the dataset you want to access.

Note that if you do not have the arrow library installed in your machine, you will need to install it before running the code above. You can do so by running the following line: install.packages("arrow"). Alternatively, you can run refer to the Setting up your machine section below for instructions on how to install all packages used in this repository at once.

Instructions for Python users
# Loading pyarrow library to connect to S3 bucket
from pyarrow import parquet as pq
# Providing S3 URL address for dataset of interest
dataset_s3 = 's3://gbr-dms-data-public/aims-temp-loggers/data.parquet'
# Connecting to S3 bucket
ds = pq.ParquetDataset(dataset_s3)

Remember that you can change the value of dataset_s3 to the S3 URL address for the dataset you want to access.

Note that if you do not have the pyarrow package installed in your machine, you will not be able to run the code above. You can install it using a package manager such as pip or conda. Alternatively, you can run refer to the Setting up your machine section below for instructions on how to install all packages used in this repository at once.

Extracting tabular data from S3 bucket

Once you have connected to the S3 bucket, you do not have to download the entire dataset to your local machine to carry out your analysis. Instead, you can extract data from the dataset of interest based on one or more conditions. You can then load into memory only the relevant data needed to create summary tables, figures, or maps. We are including code snippets showing a simple data selection based on spatial and temporal conditions.

Instructions for R users

Once you have connected to the S3 bucket, you can use dplyr verbs to extract a subset of the data based on one or more conditions. Here, we assume that a dataset connection has already been established following instructions in the Connecting to S3 bucket section above and this dataset is stored in the ds variable. We will assume that our dataset has longitude, latitude, and time columns, and we will use them to extract data based on spatial and temporal conditions.

# Loading relevant libraries
library(dplyr)

# We will extract data for the year 2019 that includes Townsville and Cairns
ds_subset <- ds |>
  # First we apply a filter based on longitudes
  filter(longitude > 145.6 & longitude < 146.9) |>
  # Then we apply a filter based on latitudes
  filter(latitude > -19.3 & latitude < -16.8) |>
  # Finally, we apply a filter based on time
  filter(time >= "2019-01-01" & time <= "2019-12-31") |>
  # We could even select only the columns we need
  # We will assume that the dataset also has a column called 'site' and we want to select it
  select(longitude, latitude, time, site)

# We can now load the data into memory
ds_subset <- ds_subset |>
  collect()

You can change the values of the conditions above to extract data that is relevant for your needs. Other conditions may include extracting data based on a specific site, a specific depth range, or even a specific variable.

Instructions for Python users

Once you have connected to the S3 bucket, you can use the dask_geopandas package to connect to a dataset and extract a subset of the data based on one or more conditions. We will assume that our dataset has longitude, latitude, and time columns, and we will use them to extract data based on spatial and temporal conditions. We will use the AIMS Sea Surface Temperature Monitoring Program dataset as an example, but you can replace the S3 URL address with the one for the dataset you want to access.

# Loading relevant packages
import dask_geopandas as dgp

# We store the S3 URL address in a variable
dataset_s3 = 's3://gbr-dms-data-public/aims-temp-loggers/data.parquet'

# We will define a variable our conditions to extract data for the year 2019 that includes Townsville and Cairns
filter = [(lon > 145.6),
          (lon < 146.9),
          (lat > -19.3),
          (lat < -16.8),
          (time >= "2019-01-01"),
          (time <= "2019-12-31")]

# We will extract data for the year 2019 that includes Townsville and Cairns
ds_subset = dgp.read_parquet(dataset_s3,
                            # We can select the columns of our interest with the columns argument
                             columns = ['lon', 'lat', 'time', 'site', 'qc_val'],
                            # We can specify the column we want to use as index
                             index = 'fid',
                            # We can now apply our filters
                             filters = filter,
                            # We can connect anonimously because this is a public dataset
                             storage_options = {'anon': True})

# We can now load the data into memory
ds_subset = ds_subset.compute()

Extracting gridded data from S3 bucket

Gridded data is also available in the GBR DMS. This data is stored in Zarr format, which is a format that allows for efficient storage of array-based data. This data is also stored in S3 buckets, but the connection and extraction process is slightly different from the one described above for tabular data.

Instructions for R users

To make access of gridded data in R as easy as possible for users, we created a function called connect_dms_dataset, which you can find in the useful_functions.R script. This function takes the API address for the dataset of interest, the variable name of interest as arguments and returns a SpatRaster object. Additionally, you can provide spatial and temporal boundaries to extract data. We will use the NOAA Coral Reef Watch degree heating weeks dataset as an example, but you can replace the API address with the one for the dataset you want to access.

Note that you will need to head over to our dashboard: https://dashboard.reefdata.io/ to get a token before you can access the data. When you use the connect_dms_dataset function, you will be prompted to enter your token. If you do not have an account, you can contact us at [email protected].

#Loading useful_functions script
source("R_based_scripts/useful_functions.R")
#Loading and manipulating gridded data
library(terra)

#Defining API URL (obtained from STAC catalogue)
base_url <- "https://pygeoapi.reefdata.io/collections/noaa-crw-chs-dhw"

#Defining variable of interest (obtained from STAC catalogue)
variable_name <- "degree_heating_week"

#Connecting to DMS to extract data
ras_dhw <- connect_dms_dataset(base_url, variable_name,
                           #Temporal limits
                           start_time = "2023-01-01", end_time = "2023-01-07",
                           #Spatial limits
                           lon_limits = c(145.30, 146.90),
                           lat_limits = c(-17, -16.30))

#You can plot the raster to check that the data was extracted correctly
plot(ras_dhw)
Instructions for Python users

Instead of using dask_geopandas to connect to the S3 bucket and extract tabular data, we will use the s3fs package to connect and extract gridded data. We will use the NOAA Coral Reef Watch degree heating weeks dataset as an example, but you can replace the S3 URL address with the one for the dataset you want to access.

#Loading relevant packages
#Connecting to S3 bucket
import s3fs
#Loading and manipulating gridded data
import xarray as xr

#Storing the S3 URL address in a variable
coral_url = 's3://gbr-dms-data-public/noaa-crw-chs-dhw/data.zarr'

#Connecting to public bucket - No credentials required
s3_bucket = s3fs.S3FileSystem(anon = True)

#Loading data into memory
coral_ds = xr.open_dataset(s3fs.S3Map(root = coral_url, s3 = s3_bucket), engine = 'zarr')

Table of contents

Releases

No releases published

Packages

No packages published