Supplementary material from "How many specimens make a sufficient training set for automated 3D feature extraction?" ...

Deep learning has emerged as a robust tool for automating feature extraction from 3D images, offering an efficient alternative to labour-intensive and potentially biased manual image segmentation methods. However, there has been limited exploration into the optimal training set sizes, including asse...

Full description

Bibliographic Details
Main Authors: Mulqueeney, James Michael, Searle-Barnes, Alex, Brombacher, Anieke, Sweeney, Marisa, Goswami, Anjali, Ezard, Thomas
Format: Article in Journal/Newspaper
Language:unknown
Published: The Royal Society 2024
Subjects:
Online Access:https://dx.doi.org/10.6084/m9.figshare.c.7288922.v1
https://rs.figshare.com/collections/Supplementary_material_from_How_many_specimens_make_a_sufficient_training_set_for_automated_3D_feature_extraction_/7288922/1
Description
Summary:Deep learning has emerged as a robust tool for automating feature extraction from 3D images, offering an efficient alternative to labour-intensive and potentially biased manual image segmentation methods. However, there has been limited exploration into the optimal training set sizes, including assessing whether artificial expansion by data augmentation can achieve consistent results in less time and how consistent these benefits are across different types of traits. In this study, we manually segmented 50 planktonic foraminifera specimens from the genus Menardella to determine the minimum number of training images required to produce accurate volumetric and shape data from internal and external structures. The results reveal unsurprisingly that deep learning models improve with a larger number of training images with eight specimens being required to achieve 95% accuracy. Furthermore, data augmentation can enhance network accuracy by up to 8.0%. Notably, predicting both volumetric and shape measurements for ...