Data

Soil and Landscape Grid National Soil Attribute Maps - Depth of Regolith (3" resolution) - Release 2

Commonwealth Scientific and Industrial Research Organisation
Wilford, John ; Searle, Ross ; Thomas, Mark ; Grundy, Mike
Viewed: [[ro.stat.viewed]] Cited: [[ro.stat.cited]] Accessed: [[ro.stat.accessed]]
ctx_ver=Z39.88-2004&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Adc&rfr_id=info%3Asid%2FANDS&rft_id=info:doi10.4225/08/55C9472F05295&rft.title=Soil and Landscape Grid National Soil Attribute Maps - Depth of Regolith (3 resolution) - Release 2&rft.identifier=https://doi.org/10.4225/08/55C9472F05295&rft.publisher=Commonwealth Scientific and Industrial Research Organisation&rft.description=This is Version 2 of the Depth of Regolith product of the Soil and Landscape Grid of Australia (produced 2015-06-01).\n\nThe Soil and Landscape Grid of Australia has produced a range of digital soil attribute products. The digital soil attribute maps are in raster format at a resolution of 3 arc sec (~90 x 90 m pixels). \n\nAttribute Definition: The regolith is the in situ and transported material overlying unweathered bedrock; \nUnits: metres; \nSpatial prediction method: data mining using piecewise linear regression; \nPeriod (temporal coverage; approximately): 1900-2013; \nSpatial resolution: 3 arc seconds (approx 90m); \nTotal number of gridded maps for this attribute:3; \nNumber of pixels with coverage per layer: 2007M (49200 * 40800); \nTotal size before compression: about 8GB; \nTotal size after compression: about 4GB; \nData license : Creative Commons Attribution 4.0 (CC BY); \nVariance explained (cross-validation): R^2 = 0.38; \nTarget data standard: GlobalSoilMap specifications; \nFormat: GeoTIFF.\nLineage: The methodology consisted of the following steps: (i) drillhole data preparation, (ii) compilation and selection of the environmental covariate raster layers and (iii) model implementation and evaluation.\n\nDrillhole data preparation: \nDrillhole data was sourced from the National Groundwater Information System (NGIS) database. This spatial database holds nationally consistent information about bores that were drilled as part of the Bore Construction Licensing Framework (http://www.bom.gov.au/water/groundwater/ngis/). The database contains 357,834 bore locations with associated lithology, bore construction and hydrostratigraphy records. This information was loaded into a relational database to facilitate analysis. \n\nRegolith depth extraction: \nThe first step was to recognise and extract the boundary between the regolith and bedrock within each drillhole record. This was done using a key word look-up table of bedrock or lithology related words from the record descriptions. 1,910 unique descriptors were discovered. Using this list of new standardised terms analysis of the drillholes was conducted, and the depth value associated with the word in the description that was unequivocally pointing to reaching fresh bedrock material was extracted from each record using a tool developed in C# code. \n\nThe second step of regolith depth extraction involved removal of drillhole bedrock depth records deemed necessary because of the “noisiness” in depth records resulting from inconsistencies we found in drilling and description standards indentified in the legacy database. \n\nOn completion of the filtering and removal of outliers the drillhole database used in the model comprised of 128,033 depth sites.\n\nSelection and preparation of environmental covariates\nThe environmental correlations style of DSM applies environmental covariate datasets to predict target variables, here regolith depth. Strongly performing environmental covariates operate as proxies for the factors that control regolith formation including climate, relief, parent material organisms and time.\n\nDepth modelling was implemented using the PC-based R-statistical software (R Core Team, 2014), and relied on the R-Cubist package (Kuhn et al. 2013). To generate modelling uncertainty estimates, the following procedures were followed: (i) the random withholding of a subset comprising 20% of the whole depth record dataset for external validation; (ii) Bootstrap sampling 100 times of the remaining dataset to produce repeated model training datasets, each time. The Cubist model was then run repeated times to produce a unique rule set for each of these training sets. Repeated model runs using different training sets, a procedure referred to as bagging or bootstrap aggregating, is a machine learning ensemble procedure designed to improve the stability and accuracy of the model. The Cubist rule sets generated were then evaluated and applied spatially calculating a mean predicted value (i.e. the final map). The 5% and 95% confidence intervals were estimated for each grid cell (pixel) in the prediction dataset by combining the variance from the bootstrapping process and the variance of the model residuals. Version 2 differs from version 1, in that the modelling of depths was performed on the log scale to better conform to assumptions of normality used in calculating the confidence intervals. The method to estimate the confidence intervals was improved to better represent the full range of variability in the modelling process. (Wilford et al, in press)\n&rft.creator=Wilford, John &rft.creator=Searle, Ross &rft.creator=Thomas, Mark &rft.creator=Grundy, Mike &rft.date=2024&rft.edition=v7&rft.coverage=westlimit=112.91246797222223; southlimit=-43.64247513888889; eastlimit=153.63996633333332; northlimit=-9.998309805555555; projection=WGS84&rft_rights=Creative Commons Attribution 4.0 International Licence https://creativecommons.org/licenses/by/4.0/&rft_rights=Data is accessible online and may be reused in accordance with licence conditions&rft_rights=All Rights (including copyright) CSIRO 2015.&rft_subject=TERN_Soils&rft_subject=TERN_Soils_DSM&rft_subject=Soil&rft_subject=TERN&rft_subject=Raster&rft_subject=Attributes&rft_subject=Depth of Regolith&rft_subject=Continental&rft_subject=Australia&rft_subject=DSM&rft_subject=Global Soil Map&rft_subject=spatial modelling&rft_subject=visible-near infrared spectroscopy&rft_subject=3-dimensional soil mapping&rft_subject=spatial uncertainty&rft_subject=Soil Maps&rft_subject=Digital Soil Mapping&rft_subject=SLGA&rft_subject=Soil sciences not elsewhere classified&rft_subject=Soil sciences&rft_subject=ENVIRONMENTAL SCIENCES&rft.type=dataset&rft.language=English Access the data

Licence & Rights:

Open Licence view details
CC-BY

Creative Commons Attribution 4.0 International Licence
https://creativecommons.org/licenses/by/4.0/

Data is accessible online and may be reused in accordance with licence conditions

All Rights (including copyright) CSIRO 2015.

Access:

Open view details

Accessible for free

Contact Information



Brief description

This is Version 2 of the Depth of Regolith product of the Soil and Landscape Grid of Australia (produced 2015-06-01).

The Soil and Landscape Grid of Australia has produced a range of digital soil attribute products. The digital soil attribute maps are in raster format at a resolution of 3 arc sec (~90 x 90 m pixels).

Attribute Definition: The regolith is the in situ and transported material overlying unweathered bedrock;
Units: metres;
Spatial prediction method: data mining using piecewise linear regression;
Period (temporal coverage; approximately): 1900-2013;
Spatial resolution: 3 arc seconds (approx 90m);
Total number of gridded maps for this attribute:3;
Number of pixels with coverage per layer: 2007M (49200 * 40800);
Total size before compression: about 8GB;
Total size after compression: about 4GB;
Data license : Creative Commons Attribution 4.0 (CC BY);
Variance explained (cross-validation): R^2 = 0.38;
Target data standard: GlobalSoilMap specifications;
Format: GeoTIFF.
Lineage: The methodology consisted of the following steps: (i) drillhole data preparation, (ii) compilation and selection of the environmental covariate raster layers and (iii) model implementation and evaluation.

Drillhole data preparation:
Drillhole data was sourced from the National Groundwater Information System (NGIS) database. This spatial database holds nationally consistent information about bores that were drilled as part of the Bore Construction Licensing Framework (http://www.bom.gov.au/water/groundwater/ngis/). The database contains 357,834 bore locations with associated lithology, bore construction and hydrostratigraphy records. This information was loaded into a relational database to facilitate analysis.

Regolith depth extraction:
The first step was to recognise and extract the boundary between the regolith and bedrock within each drillhole record. This was done using a key word look-up table of bedrock or lithology related words from the record descriptions. 1,910 unique descriptors were discovered. Using this list of new standardised terms analysis of the drillholes was conducted, and the depth value associated with the word in the description that was unequivocally pointing to reaching fresh bedrock material was extracted from each record using a tool developed in C# code.

The second step of regolith depth extraction involved removal of drillhole bedrock depth records deemed necessary because of the “noisiness” in depth records resulting from inconsistencies we found in drilling and description standards indentified in the legacy database.

On completion of the filtering and removal of outliers the drillhole database used in the model comprised of 128,033 depth sites.

Selection and preparation of environmental covariates
The environmental correlations style of DSM applies environmental covariate datasets to predict target variables, here regolith depth. Strongly performing environmental covariates operate as proxies for the factors that control regolith formation including climate, relief, parent material organisms and time.

Depth modelling was implemented using the PC-based R-statistical software (R Core Team, 2014), and relied on the R-Cubist package (Kuhn et al. 2013). To generate modelling uncertainty estimates, the following procedures were followed: (i) the random withholding of a subset comprising 20% of the whole depth record dataset for external validation; (ii) Bootstrap sampling 100 times of the remaining dataset to produce repeated model training datasets, each time. The Cubist model was then run repeated times to produce a unique rule set for each of these training sets. Repeated model runs using different training sets, a procedure referred to as bagging or bootstrap aggregating, is a machine learning ensemble procedure designed to improve the stability and accuracy of the model. The Cubist rule sets generated were then evaluated and applied spatially calculating a mean predicted value (i.e. the final map). The 5% and 95% confidence intervals were estimated for each grid cell (pixel) in the prediction dataset by combining the variance from the bootstrapping process and the variance of the model residuals. Version 2 differs from version 1, in that the modelling of depths was performed on the log scale to better conform to assumptions of normality used in calculating the confidence intervals. The method to estimate the confidence intervals was improved to better represent the full range of variability in the modelling process. (Wilford et al, in press)

Available: 2024-08-28

Data time period: 1900-01-01 to 2013-12-31

This dataset is part of a larger collection

153.63997,-9.99831 153.63997,-43.64248 112.91247,-43.64248 112.91247,-9.99831 153.63997,-9.99831

133.27621715277,-26.820392472222