Back to Abstract Index

Paper Number: 71

Output selection for Gaussian Processes classification in iron ore deposits

Silversides, K.L.1 and Melkumyan, A. 2

1Australian Centre for Field Robotics, University of Sydney, NSW 2006 Australia katherine.silversides@sydney.edu.au

___________________________________________________________________________

Modelling stratigraphic ore deposits, such as the banded iron formation hosted iron ore deposits in the Hamersley Ranges of Western Australia, requires detailed knowledge of the location of boundaries. Gaussian Processes (GPs) have been used to identify boundaries in the exploration hole geophysics [1] and chemical assays [2] from these deposits. The boundaries identified include marker shale based ore boundaries and the alluvial to bedded boundary. When using machine learning for feature classification, different techniques may have biases that can affect the results. Gaussian Process predictions are inclined towards the mean value, particularly if there is not sufficient information [3]. To examine the impact that this inclination has on feature identification, a typical Marra Mamba style iron ore deposit was chosen. The alluvial to bedded boundary at the top of the deposit has a signature that can be identified in the TiO2 and Al2O3 chemical assays.

Three GP models were trained using the same starting library, and output ranges of 0 to 1, -0.5 to 0.5 and -1 to 0. In each case the larger number is the output for positive identification of the signature. The GP mean is 0 for all tests. Therefore the outputs were pulled towards the negative, neutral and positive identification respectively. The libraries were progressively updated [1], diverging from each other and producing distinctly different final libraries. The results were compared to a manual geological interpretation (Table 1).

The accuracy and certainty varied between the models, with the 0 to 1 GP having the highest and the -1 to 0 GP having the lowest for both. The 0 to 1 GP also identified the greatest number of correct signatures. The 0 to 1 GP pulled the output towards a negative classification. Therefore the library had to include sufficient positive examples to identify the desired feature, but did not need to include all possible negative examples. This is an advantage due to the very high variability in the negative examples compared to the positive examples, which are variations of a single feature. In comparison, the -0.5 to 0.5 GP pulled all the outputs towards having an uncertain classification. This resulted in a high number of uncertain results as not all possible negative examples could be included in the training library. The -1 to 0 GP pulled the results towards having a positive classification. There were a large number of negative features and including so many dissimilar features in the same category resulted in difficulty training this GP model. Therefore, when using GPs for feature classification the outputs should be arranged so that the pull is against the desired feature. Then the library can be trained to specifically identify that feature as having an output away from the mean.

Table 1: Results for the GP models with different output ranges

-1 to 0
-0.5 to 0.5
0 to 1
No. of holes
Accuracy (%)
No. of holes
Accuracy (%)
No. of holes
Accuracy (%)
Certain holes
All
3649
86.8
4063
89.2
4895
88.7
Signature
2141
80.6
2623
87.2
3098
87.0
No Signature
1508
95.6
1440
92.8
1797
91.7
Uncertain holes
All
2285
66.2
1871
66.2
1039
65.8
Signature
1565
69.8
1082
63.4
591
71.2
No Signature
720
58.5
789
70.1
448
58.7
Overall accuracy
78.9
82.0
84.7
Percent of holes certain
61.5
68.5
82.5

[1] Silversides K et al. (2015) Comput Geosci 77:118-125

[2] Silversides KL and Melkumyan A (2015) Conf Proc 12th SEGJ

[3] Rasmussen CE and Williams CKI (2006) Gaussian Processes for Machine Learning. MIT Press, 248 pp