| Literature DB >> 31831808 |
Gabriel Cifuentes-Alcobendas1, Manuel Domínguez-Rodrigo2.
Abstract
Accurate identification of bone surface modifications (BSM) is crucial for the taphonomic understanding of archaeological and paleontological sites. Critical interpretations of when humans started eating meat and animal fat or when they started using stone tools, or when they occupied new continents or interacted with predatory guilds impinge on accurate identifications of BSM. Until now, interpretations of Plio-Pleistocene BSM have been contentious because of the high uncertainty in discriminating among taphonomic agents. Recently, the use of machine learning algorithms has yielded high accuracy in the identification of BSM. A branch of machine learning methods based on imaging, computer vision (CV), has opened the door to a more objective and accurate method of BSM identification. The present work has selected two extremely similar types of BSM (cut marks made on fleshed an defleshed bones) to test the immense potential of artificial intelligence methods. This CV approach not only produced the highest accuracy in the classification of these types of BSM until present (95% on complete images of BSM and 88.89% of images of only internal mark features), but it also has enabled a method for determining which inconspicuous microscopic features determine successful BSM discrimination. The potential of this method in other areas of taphonomy and paleobiology is enormous.Entities:
Mesh:
Year: 2019 PMID: 31831808 PMCID: PMC6908723 DOI: 10.1038/s41598-019-55439-6
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Neural network model and parameters used for the present study (simple model).
| Layer (type) | Output Shape | Param # |
|---|---|---|
| conv2d_1 (Conv2D) | (None, 78, 398, 32) | 896 |
| max_pooling2d_1(MaxPooling2 | (None, 39, 199, 32) | 0 |
| conv2d_2 (Conv2D) | (None, 37, 197, 64) | 18496 |
| max_pooling2d_2 (MaxPooling2 | (None, 18, 98, 64) | 0 |
| conv2d_3 (Conv2D) | (None, 16, 96, 128) | 73856 |
| max_pooling2d_3 (MaxPooling2 | (None, 8, 48, 128) | 0 |
| conv2d_4 (Conv2D) | (None, 6, 46, 128) | 147584 |
| max_pooling2d_4 (MaxPooling2 | (None, 3, 23, 128) | 0 |
| flatten_1 (Flatten) | (None, 8832) | 0 |
| dense_1 (Dense) | (None, 512) | 4522496 |
| dense_2 (Dense) | (None, 1) | 513 |
Total params: 4,763,841.
Trainable params: 4,763,841.
Non-trainable params: 0.
Summary of the architecture of the Alexnet model.
| Layer (type) | Output Shape | Param # |
|---|---|---|
| conv2d_1 (Conv2D) | (None, 18, 98, 96) | 23424 |
| activation_1 (Activation) | (None, 18, 98, 96) | 0 |
| max_pooling2d_1 (MaxPooling2 | (None, 9, 49, 96) | 0 |
| batch_normalization_1 (Batch | (None, 9, 49, 96) | 384 |
| xconv2d_2 (Conv2D) | (None, 1, 41, 256) | 1990912 |
| activation_2 (Activation) | (None, 1, 41, 256) | 0 |
| max_pooling2d_2 (MaxPooling2 | (None, 1, 21, 256) | 0 |
| batch_normalization_2 (Batch | (None, 1, 21, 256) | 1024 |
| conv2d_3 (Conv2D) | (None, 1, 21, 384) | 98688 |
| activation_3 (Activation) | (None, 1, 21, 384) | 0 |
| batch_normalization_3 (Batch | (None, 1, 21, 384) | 1536 |
| conv2d_4 (Conv2D) | (None, 1, 21, 384) | 147840 |
| activation_4 (Activation) | (None, 1, 21, 384) | 0 |
| batch_normalization_4 (Batch | (None, 1, 21, 384) | 1536 |
| conv2d_5 (Conv2D) | (None, 1, 21, 256) | 98560 |
| activation_5 (Activation) | (None, 1, 21, 256) | 0 |
| max_pooling2d_3 (MaxPooling2 | (None, 1, 11, 256) | 0 |
| batch_normalization_5 (Batch | (None, 1, 11, 256) | 1024 |
| flatten_1 (Flatten) | (None, 2816) | 0 |
| dense_1 (Dense) | (None, 4096) | 11538432 |
| activation_6 (Activation) | (None, 4096) | 0 |
| dropout_1 (Dropout) | (None, 4096) | 0 |
| batch_normalization_6 (Batch | (None, 4096) | 16384 |
| dense_2 (Dense) | (None, 4096) | 16781312 |
| activation_7 (Activation) | (None, 4096) | 0 |
| dropout_2 (Dropout) | (None, 4096) | 0 |
| batch_normalization_7 | (Batch(None, 4096) | 16384 |
| dense_3 (Dense) | (None, 1000) | 4097000 |
| activation_8 (Activation) | (None, 1000) | 0 |
| dropout_3 (Dropout) | (None, 1000) | 0 |
| batch_normalization_8 | (Batch (None, 1000) | 4000 |
| dense_4 (Dense) | (None, 1) | 1001 |
| activation_9 (Activation) | (None, 1) | 0 |
Total params: 34,819,441.
Trainable params: 34,798,305.
Non-trainable params: 21,136.
Figure 1Selection of marks from both experiments showing overlapping similarities and contrasting differences. The upper half shows contrasting images of cut marks made on bones with meat (WM) and no meat (NM). The lower half shows images of visually indifferentiable cut marks made on bones with and without meat.
Figure 2Loss (y-axis) and epochs (x-axis) (lower) and accuracy (y-axis) and epochs (x-axis) (upper) of the control experiment comparing cut maks on fleshed bone made with the same five flakes that were used for imparting marks on defleshed bone and the new control flakes.
Figure 3Heat maps produced by the Grad CAM algorithm indicating discrimination areas important for correct classification of the complete image data set (obtained from the low-resolution experimental subsample). (a,b), Cut marks made on bones with meat. (c,d) cut marks made on defleshed bones. Red hat only show in the heat map denotes the most important features.
Figure 4Heat maps produced by the Grad CAM algorithm indicating discrimination areas important for correct classification of the high-resolution cropped images showing only the mark grooves. (A–C) Cut marks made on defleshed bones. (D-E), cut marks made on bones with meat. Red color in the heat map denotes the most important features.