Next Article in Journal
Physical Therapy Use and Associated Factors in Adults with and without Osteoarthritis—An Analysis of the Population-Based German Health Update Study
Next Article in Special Issue
Three-Dimensional Dental Analysis for Sex Estimation in the Italian Population: A Pilot Study Based on a Geometric Morphometric and Artificial Neural Network Approach
Previous Article in Journal
Safety Performance in Acute Medical Care: A Qualitative, Explorative Study on the Perspectives of Healthcare Professionals
Previous Article in Special Issue
Organizational Strategies for the Management of Intravenous Iron Therapy in Non-Hospitalized Settings: A Safe Opportunity to Implement Patient Blood Management in Italy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Use of Advanced Artificial Intelligence in Forensic Medicine, Forensic Anthropology and Clinical Anatomy

by
Andrej Thurzo
1,2,3,*,
Helena Svobodová Kosnáčová
2,4,
Veronika Kurilová
5,
Silvester Kosmeľ
6,
Radoslav Beňuš
3,7,
Norbert Moravanský
3,8,
Peter Kováč
3,9,
Kristína Mikuš Kuracinová
10,
Michal Palkovič
10,11 and
Ivan Varga
12
1
Department of Stomatology and Maxillofacial Surgery, Faculty of Medicine, Comenius University in Bratislava, 81250 Bratislava, Slovakia
2
Department of Simulation and Virtual Medical Education, Faculty of Medicine, Comenius University, Sasinkova 4, 81272 Bratislava, Slovakia
3
forensic.sk Institute of Forensic Medical Analyses Ltd., Boženy Němcovej 8, 81104 Bratislava, Slovakia
4
Department of Genetics, Cancer Research Institute, Biomedical Research Center, Slovak Academy Sciences, Dúbravská Cesta 9, 84505 Bratislava, Slovakia
5
Faculty of Electrical Engineering and Information Technology, Slovak University of Technology, Ilkovičova 3, 81219 Bratislava, Slovakia
6
Deep Learning Engineering Department at Cognexa, Faculty of Informatics and Information Technologies, Slovak University of Technology, Ilkovičova 2, 84216 Bratislava, Slovakia
7
Department of Anthropology, Faculty of Natural Sciences, Comenius University in Bratislava, Mlynská dolina Ilkovičova 6, 84215 Bratislava, Slovakia
8
Institute of Forensic Medicine, Faculty of Medicine, Comenius University in Bratislava, Sasinkova 4, 81108 Bratislava, Slovakia
9
Department of Criminal Law and Criminology, Faculty of Law Trnava University, Kollárova 10, 91701 Trnava, Slovakia
10
Institute of Pathological Anatomy, Faculty of Medicine, Comenius University in Bratislava, Sasinkova 4, 81108 Bratislava, Slovakia
11
Forensic Medicine and Pathological Anatomy Department, Health Care Surveillance Authority (HCSA), Sasinkova 4, 81108 Bratislava, Slovakia
12
Institute of Histology and Embryology, Faculty of Medicine, Comenius University in Bratislava, 81372 Bratislava, Slovakia
*
Author to whom correspondence should be addressed.
Submission received: 1 November 2021 / Revised: 10 November 2021 / Accepted: 10 November 2021 / Published: 12 November 2021
(This article belongs to the Special Issue New Trends in Forensic and Legal Medicine)

Abstract

:
Three-dimensional convolutional neural networks (3D CNN) of artificial intelligence (AI) are potent in image processing and recognition using deep learning to perform generative and descriptive tasks. Compared to its predecessor, the advantage of CNN is that it automatically detects the important features without any human supervision. 3D CNN is used to extract features in three dimensions where input is a 3D volume or a sequence of 2D pictures, e.g., slices in a cone-beam computer tomography scan (CBCT). The main aim was to bridge interdisciplinary cooperation between forensic medical experts and deep learning engineers, emphasizing activating clinical forensic experts in the field with possibly basic knowledge of advanced artificial intelligence techniques with interest in its implementation in their efforts to advance forensic research further. This paper introduces a novel workflow of 3D CNN analysis of full-head CBCT scans. Authors explore the current and design customized 3D CNN application methods for particular forensic research in five perspectives: (1) sex determination, (2) biological age estimation, (3) 3D cephalometric landmark annotation, (4) growth vectors prediction, (5) facial soft-tissue estimation from the skull and vice versa. In conclusion, 3D CNN application can be a watershed moment in forensic medicine, leading to unprecedented improvement of forensic analysis workflows based on 3D neural networks.

Graphical Abstract

1. Introduction

Conventional forensic analysis is based on forensic expert’s manual extraction of information. Forensic expert provides opinions established on medical and other fields of current knowledge combined with personal work experience. This is not only time-consuming, albeit frequently affected by subjective factors that are tough to overcome [1].
The main purpose of this paper is to analyze and introduce a very promising line of research applicable to forensic anthropology and various traditional sectors of forensic medicine. The application of artificial intelligence (AI) is a new trend in forensic medicine and a possible watershed moment for the whole forensic field [1,2,3,4,5,6].
This chapter paper explains basic terminology, principles and the current horizon of knowledge. The methodology chapter presents the novel clinical workflow based on implementing three-dimensional convolutional neural network (3D CNN) algorithms [7,8,9]. The input is full head cone-beam computer tomography scans (CBCT) in the Digital Imaging and Communications in Medicine format (DICOM) [9,10,11,12,13,14]. The methodology chapter describes technical data preparation for 3D CNN utilization in the following practical aspects from forensic medicine:
The result of this paper is a detailed guide for forensic scientists to implement features of 3D CNN to forensic research and analyses of their own (in five themes described above). This resulting practical concept—possible workflow shall be useful for any forensic expert interested in implementing this advanced artificial intelligence feature. This study is based on the worldwide review of 3D CNN use-cases that apply to clinical aspects of forensic medicine
This article’s secondary objective is to inspire forensic experts and approximate them to implement three-dimensional convolutional neural networks (3D CNN) in their forensic research in the fields of age, sex, face and growth determination.

1.1. Basic Terminology and Principles in Era of AI Enhanced Forensic Medicine

Artificial intelligence has brought new vigor to forensic medicine, but at the same time also some challenges. AI and forensic medicine are developing collaboratively and advanced AI implementation until now required extensive interdisciplinary cooperation. In the era of big data [3], forensic experts shall become familiar with these advanced algorithms and understand used technical terms.
For many forensic experts, the current benefits of advanced AI processes are still unknown. For example, automated AI algorithms for skull damage detection from CT [91] or soft-tissue prediction of a face from the skull [66,67,89,92] are still a mystery to many outstanding forensic scientists. Enabling them would catapult forensic research to a new era [1].
A Convolutional Neural Network (CNN) is a Deep Learning algorithm that can take in an input image, assign importance (learnable weights and biases) to various aspects/objects in the image, and differentiate one from the other.
CNN is an efficient recognition algorithm that is widely used in pattern recognition and image processing. It has many features such as simple structure, less training parameters and adaptability. CNN is a supervised type of Deep learning, most preferable used in image recognition and computer vision (Figure 1a,b).
Compared to its predecessors, the main advantage of CNN is that it automatically detects the crucial features without any human supervision. For example, given many pictures of cats and dogs, it learns distinctive features for each class. CNN is also computationally efficient.
3D CNN is used to extract features in 3 Dimensions or establish a relationship between 3 dimensions. A 3D CNN is simply the 3D equivalent: it takes as input a 3D volume or a sequence of 2D frames (e.g., CBCT scan).
In terms of Neural Networks and Deep Learning: Convolutions are filters (matrix/vectors) with learnable parameters used to extract low-dimensional features from input data. They have the property to preserve the spatial or positional relationships between input data points.
2D CNNs predict segmentation maps for DICOM slices in a single anatomical plane. 3D CNNs address this issue by using 3D convolutional kernels to make segmentation predictions for a volumetric patch of a scan (Figure 2).
In 3D convolution, a 3D filter can move in all 3-directions (height, width, channel of the image). At each position, the element-wise multiplication and addition provide one number. Since the filter slides through a 3D space, the output numbers are also arranged in a 3D space. The output is then 3D data.
The recognition of similar structures from the CBCT is based on their similar opacity on the X-ray classified by the Hounsfield scale. The process of defining ranges for particular tissues is called “thresholding”, which is prior to final the segmentation (Figure 3). Setting different thresholds for segmentation preprocessing step allows segmentation of different structures such as soft tissues (skin, airway, sinuses), nerves (inferior alveolar nerve, dental pulp), bones (mandible, maxilla or cervical vertebras) and many other (Figure 4).
The segmentation of original CBCT data can result in the definition of various 3D structures involved in 3D CNN training, or these 3D structures can serve as anchors for mapping another 3D scan, such as an intraoral optical scan or extraoral scan that includes texture. All these three data sources can be merged, and the 3D CNN network can work with unprecedented data that include wider face regions from face scan or morphological information on teeth and gums (Figure 5).

1.2. Overview of Used Artificial Intelligence for Forensic Age and Sex Determination

Traditional forensic analyses of age, gender and facial appearance are based on forensic expert manually acquiring information that provides the identification established on expert`s medical and biological knowledge and mathematical calculations [93,94,95]. In forensic outputs, the experiences of the investigator subjectivity and fatigue and emotions play a role [93,94,95]. To have forensic expert well trained on thousands of skulls of all possible ethnicities, would take a lifetime. Possible bias sourcing from fatigue, limited training dataset, emotional engagement or human calculation error cannot be absolutely eradicated with human forensic expert. Implementation of artificial intelligence (AI) can limit all these mentioned sources of possible bias. Machine learning works based on models that mimic neurons in the brain and can learn from experiences and solve complex problems. It is not influenced by subjective judgment; it does not become tired and does not use emotions and thus can work more efficiently [96,97,98].
AI usage is not without risks of undesired side effects. AI may become biased in the same way as a human forensic expert, depending on the source data used for AI training [99]. Obermeyer at el. found evidence of bias in a healthcare algorithm responsible for 200 million people, which systemically prevented almost 30% of eligible black patients from receiving additional care by giving lower risk scores to black patients than white patients with equal diagnoses [100].
Many studies in forensic science have been conducted in recent years, and some recent studies are beginning to focus on neural networks [101]. These studies were mainly aimed at determining the age and sex of postmortem skeletal remains and living people. Age and gender assessment active, used to identify victims, determine criminal liability or identify persons without legal documentation [8,102]. There is considerable interest in accelerating identification procedures, and experts are involved in machine learning in forensic procedures. They use X-ray images [103,104,105,106,107,108], MRI images [8,109], photography [90,110,111], CT scans [112,113,114,115,116,117] of the head or other bones such as the collarbone, femur, teeth, etc. and use databases to teach artificial intelligence to identify people’s age or gender. Pham et al. [113] examined age using the femur and mandible for neuronal networks. The femur could play a key role in predicting adulthood, especially the shape of the femoral head and bone densitometry. They used 814 whole-body post-mortem computed tomography (CT) scans to obtain results: 619 men, 195 women aged 20 to 70 years. They omitted subjects with fractures. Each CT output was in digital imaging and communication in medicine (DICOM) format [11,12]. The extracted femur and mandible data were preprocessed to create a 3D voxel inserted into a neural network model. Using this approach, the mean absolute error (MAE) of the mandible age identification was 7.07 years, and the MAE calculated from a femur age determination was 5.74 years. The combination of both approaches reached an excellent result—MAE = 5.15 years. CT scans were also used for learning and age determination in a study by Farhadian et al. [115]. AI determined the age learned from CT scans of 300 subjects aged 14 to 60 years of the canine teeth. In this study, they compared the methodology of neural networks with a regression model. The MAE for neural networks was 4.12 years, and the MAE for the regression model was 8.17 years, which demonstrated the higher accuracy of neural networks. Mauer et al. [102] aimed to develop a fully automated and computerized method for age estimation based on the knee’s 3D magnetic resonance imaging (MRI). They had 185 coronal and 404 sagittal MR images of Caucasian men aged 13 to 21 years. The best result obtained was a MAE of 0.67 ± 0.49 years and an accuracy of 90.9%. Here it can be seen that the group with a minor age variance more accurately determines the age of the individuals. A similar study was performed by Stern et al. performed a similar study [109] where 328 MR images were used for learning neural networks and subsequent age detection. Age was reported with a MAE of 0.37 ± 0.51 years for the age range of individuals ≤ 18 years.
Several research teams have tried neural network learning based on X-ray images [103,104,108]. Guo et al. [103] used 10,257 samples of dental orthopantomograms and, similar to Farhadian et al. [115], compared logistic regression linear models for each legal age limit (14, 16 and 18 years) with the neural network. The results showed that neural networks work better (linear regression models: 92.5%, 91.3% and 91.8% and neural networks: 95.9%, 95.4% and 92.3% success rate for age limits 14, 16 and 18 years). In Stepanovsky et al. [105] used 976 orthopantomography (662 men, 314 women) of people aged 2.7 to 20.5 years to learn neural networks. The results were very favorable, and the average absolute error (MAE) was below 0.7 years for both men and women. Vila-Blanco et al. [106] used landmarks on the mandible to search for patterns by neural networks. The age estimate reached an accuracy of 87.8%, and the MAE was only 1.57 years. De Tobel et al. [107] used panoramic molar panoramic radiographs to estimate age. The accuracy of the results was, on average, MAE = 0.5. Boedi et al. [108] later conducted a similar study with similar results. Li et al. [104] used 1875 X-ray images of the pelvis as a basis for evaluating bone age through deep learning. The age of the people whose X-rays were used to teach the model was 10 to 25 years. The performance of the model was MAE = 0.94 years.
More studies modelled gender determination using AI. Bewes et al. [42] used neural networks for this purpose with a detection accuracy of 95%. However, they trained them on 900 skull scans from CT scans. Oner et al. [114] achieved the same goal by using CT images of the sternum transmitted to the orthogonal plane for learning neural networks. They used 422 thin sections of thoracic CT scans (213 females, 209 males) with an age range of 27–60 years. The accuracy of gender prediction was 0.906, and the confidence interval of 94%. The success rate was higher than that achieved by linear models. Etli et al. [116] compared several methods in the study. They used CT scans with sacral and coccyx metric parameters of 480 patients. They used one-dimensional discriminant analysis, linear discriminant functional analysis, sequential analysis of discriminant function and multilayer perceptron neural networks. The maximum accuracy for each method was 67.1% for one-dimensional discriminant analysis, 82.5% for linear analysis of the discriminant function, 78.8% for sequential analysis of the discriminant function, and 86.3% for multilayer perceptron neural networks.
Gender classification was also discussed by Liew et al. [111]. The maximum accuracy for each method was 67.1% for one-dimensional discriminant analysis, 82.5% for linear analysis of the discriminant function, 78.8% for sequential analysis of the discriminant function, and 86.3% for multilayer perceptron neural networks. Gender classification was also discussed by Liew et al. [111]. They used grayscale images of 200 men and 200 women for analysis. The classification performance reached 98.75% and 99.38% in the facial databases SUMS and AT&T. To estimate the sex of infants in the study of Ortega et al. [110] used 2D photographs of the ilium of 135 individuals aged 5 months to 6 years were used. The accuracy was 59% compared to 61% for the specialist. In addition, Porto et al. [88] sought to determine the legal age of offenders at 14 and 18 years as Guo et al. [103]. They based on a database of photographs of 18,000 faces of men and women based on photo anthropometric indices from cephalometric landmarks marked and checked by forensic experts. The accuracy of age determination by neural networks was 0.72 with an age interval of 5 years and for the estimation of the age group higher than 0.93 and 0.83 for the threshold values of 14 and 18 years.
It is almost unbelievable how accurately neural networks can determine age or gender compared to commonly used methods. Therefore, we emphasize their use in forensic practice [9,46,50,117].
Regarding the Skeletal age estimation for forensic purposes, we consider ourselves useful for the direction of the 3D CNN on particular areas of the head and neck. Various experts published research on age estimation by measuring open apices in teeth, stage of teeth eruption, frequently of third molars or canine tooth/pulp ratio [6,17,18,20,21,23,24,25,27,29,31]. In general, teeth are frequently used for age assessment, but they are not the only structures in the skull to be considered. It is known that the shape of the frontal sinus can be an essential tool in personal forensic identification and is linked together with the cranial base to growth changes that can be evaluated [6,118]. Another typical location for skeletal age assessment in the head and neck X-ray diagnostics region is the stage of cervical vertebrae maturation [23,119]. Deep learning has been already implemented in this area [83]. Extensive research is published regarding skeletal age expert estimation Pinchi et al. [120,121,122,123,124,125,126] mainly combines dental and skeletal findings. If the 3D CNN fails to identify these valuable areas, we still have the opportunity to direct the focus on these areas.
Regarding forensic medico-legal aspects, the perspective on natural development estimated by AI algorithms is always relevant, especially in the situation of trauma or other damage that conflicted with this estimated development. AI is now used to evaluate CT scans of lungs and to predict the deterioration of COVID-19 patients in the emergency department [127,128,129].
In this case, 3D CNN algorithms can automatically evaluate not only hard-tissue structures and search for inapparent damage that could have been responsible for a sudden death incident [91,130].

1.3. Artificial Intelligence Implementation in 3D Cephalometric Landmark Identification

Analysis of complex cranial and facial structures is a domain of orthodontics. Historically they are fundamental for proper treatment planning, and they represent lines, angles, planes on the axilla-facial structures identifiable, especially on the X-ray (typically lateral X-ray). There is massive research regarding cephalometric parameters and their values. Observer defines the points, and their interobserver error are the main weakness of cephalometric analysis (Figure 6). Anthropometry in Forensic Medicine and Forensic Science is frequently used for sex and biological age determination and other purposes [129,130,131].
As the various cephalometric parameters (angles, ratios and distances) were well researched, and some are proven to be related to age, sex or growth, they are a frequent springboard for many research studies focused on facial parameters. Implementation of AI in cephalometric analysis has been published [132,133,134,135,136]. The question is whether the 3D CNN trained networks will find even better regions and soft- and hard-tissue features on CBCTs when autonomously searching for links between voxel structures and the age or sex. Either way, the reliable automatized 3D cephalometric algorithm precisely identifying particular points with extreme repeatability would be a helpful tool not intended to replace humans in cephalometric points identifications. However, the human error is impossible to cancel completely as the interobserver error.

1.4. Artificial Intelligence Implementation in Soft-Tissue Face Prediction from Skull and Vice Versa

Reconstruction of the face from the skull is an age-old desire of forensic experts. Current methods of not implementing AI are very limited. Prediction of soft tissues according to the hard tissues of the skull and vice versa can be significantly improved upon big-data training of 3D CNN with supplementary metadata about age, sex, BMI or ethnicity. New algorithms to perform facial reconstruction from a given skull has forensic application in helping the identification of skeletal remains when additional information is unavailable [29,64,66,67,68,69,70,72,73,85,86,88,89,92,137]. Implementation of 3D CNN can also unintentionally open pandora box of guided improving the morphology of the facial soft-tissues. From a socio-psychological standpoint, this is regarded as an important therapeutic goal in modern orthodontic treatments. Currently, many of the algorithms implemented in commercially available software present ability to predict profile changes grounded on the incorrect assumption that the amount of movement of hard-tissue and soft-tissue has a proportional relationship [82].
The beauty industry has seen rapid growth in multiple countries, and due to its applications in entertainment, the analysis and assessment of facial attractiveness have received attention from scientists, physicians, and artists because of digital media, plastic surgery, and cosmetics. An analysis of techniques is used to assess facial beauty that considers facial ratios and facial qualities as elements to predict facial beauty [81,82,138,139,140]. A popular and famous free app using AI is FaceApp, which uses neural networks to enhance, age or otherwise change 2D digital photos of users uploading them using this application (Figure 7). Using the 3D CNN approach was not yet implemented despite iPhones having a 3D lidar scanner to acquire a 3D soft-tissue scan of the user’s face. From a forensic aspect, this era of digital 2D face manipulation brought deep-fake videos and images. Detecting manipulated facial images and videos is an increasingly important topic in digital media forensics [118,141]. Any face can be used in the fake video, or unlimited numbers of nearly authentic pictures, including fake social media profiles, can be created. AI is used in forensic evaluation for facial forgery detection and manipulated region localization [118].
This paper most complex AI application is the final 5th theme—“Facial growth vectors prediction”. The authors of this paper addressed it for various reasons. Firstly, it is fundamentally different from the first four themes. Secondly, it requires the most complex implementation of AI strategies. To our knowledge, this is only the second paper in the world that handles the problem of facial growth prediction with ML methods and absolutely the first paper to consider a 3D CNN for facial growth predictions.
Prediction of natural growth is compared to typically forensic topics such as human remains reconstruction and identification or age and sex determination rather less familiar topic. Mainly because despite numerous research attempts to predict facial growth, a satisfactory method has not been established yet, and the problem still poses a challenge for medical experts [142,143,144]. Predicting natural growth and later ageing is relevant for orthodontic therapy planning and from a forensic aspect. Any damage to the head and neck region that would affect otherwise natural growth or simple ageing could be evaluated. The effect of such a trauma could be in the future forensically quite accurately evaluated.
In 1971 Hirschfeld and Moyers published an article named “Prediction of craniofacial growth: the state of the art” [144]. This was one of the first attempts for facial growth predictions. The authors concluded that there are many reasons why they fail to predict craniofacial growth, and some they named persisted until today. They expressed doubts that we have not always measured the right thing. They also pointed out the lack of biological meaning for many traditional cephalometric measures. They have also pointed to the heritability of attained growth in the face and predicted the future importance of craniofacial genetics. The future that comes proved them correct in many aspects. Since these first attempts to predict the facial growth direction over half of a century ago, we did not become much better in facial growth prediction [142]. The complexity of the problem is challenging.
The only study that was focused on the prediction of the facial growth direction with Machine Learning methods and has been published so far is a paper with its pre-print [90,145] from 2021 by Stanislaw Kazmierczak et al. The outcomes of this paper are not impressive regarding facial growth prediction, albeit inspiring in the method of evaluation. The authors of this novel paper [94] performed feature selection and pointed out the attribute that plays a central role in facial growth. Then they performed data augmentation (DA) methods. This study is discussed in more detail later in this paper.

2. 3D Convolutional Neural Networks and Methods of Their Use in Forensic Medicine

2.1. Hardware and Software Used

CBCT scans analyzed for this paper were made on one machine: i-CAT™ FLX V17 with the Field of View (FOV) of 23 cm × 17 cm with technical parameters and settings Table 1.
Medical software used for DICOM data processing and analysis was Invivo™ 6 from Anatomage Inc., Silicon Valley, Thomas Road Suite 150, Santa Clara, CA 95054, USA.
Software for the AI solution base we have used the Python programming language along with 3 deep learning libraries—TensorFlow 2, PyTorch and MONAI. As for the hardware, the whole AI system is powered by multiple GPUs.

2.2. Main Tasks Definitions

Task 1—Age estimation from whole 3D CT scan image
Definition: the task is to estimate the approximate age of a person from a whole head 3D CBCT scan
Proposed method: build regression model represented by a 3D deep neural network that has the current state of the art network architecture as a backbone
Metrics: Mean Absolute Error (MAE) and Mean Squared Error (MSE) (see Section Evaluation)
Task 2—Sex classification from thresholded soft and hard tissues
Definition: the task is to classify input 3D CBCT scans (whole head or experimentally segmented parts) into one of 2 predefined categories—female and male
Proposed method: build classification model represented by 3D deep neural network based on convolutional layers and outputs class probabilities for both targets
Metrics: Accuracy and Confusion Matrix (CM) (other metrics such as precision, recall and F1 score will be evaluated in a later phase)
Task 3—Automatization of cephalometric measurements
Definition: the task is to create an automated system able to tag cephalometric landmarks on whole head 3D CT scan
Proposed method: build object detection model based on 3D neural network that estimates cephalometric measurements automatically
Metrics: Mean Absolute Error (MAE) and Mean Squared Error (MSE) (see Section Evaluation)
Task 4—Soft-tissue face prediction from skull and vice versa
Definition: the task is to create an automated system that predicts the distance of the face surface from the bone surface according to the estimated age and sex. 3D CNN to be trained on whole-head CBCTs of soft-tissue and hard-tissue pairs. *CBCTs with trauma and other unnatural deformations shall be excluded.
Proposed method: build a generative model based on Generative Adversarial Network that synthesizes both soft and hard tissues
Metrics: the slice-wise Frechet Inception Distance (see Section Evaluation)
Task 5—Facial growth prediction
Definition: the task is to create an automated system that predicts future morphological change in defined time for the face’s hard- and soft tissues. This shall be based on two CBCT input scans of the same individual in two different time points. The second CBCTs must not be deformed with therapy affecting morphology or unnatural event. This already defines the extremely challenging condition. There is a high possibility of insufficient datasets and the necessity of multicentric cooperation for successful training of 3D CNN on this task.
Proposed method: In this final complex task, the proposed method builds on previous tasks. We strongly recommend adding metadata layers on gender, biological age and especially genetics or letting the CNN determine them by itself. We suggest disregarding the established cephalometric points, lines, angles and plains as these were defined in regards to lateral X-ray, emphasising good contrast of the bone structures with high reproducibility of the point and not necessarily with focus on particular structures most affected by growth. We suggest letting3D CNN establish its observations and focus areas.
We also suggest allowing 3D CNN analysis of genetic predisposition in a smart way: by analysis of possibly CBCT of the biological parents or preferably non-invasive face-scan providing at least facial shell data.

2.3. The Data Management

The processing of data in deep learning is crucial for the sufficient result of any neural network. Currently, most of the implementations depend on the dominant model-centric approach to AI, which means that developers spend most of their time improving neural networks.
For medical images, various preprocessing steps are recommended. In most cases, the initial steps are following (Figure 8):
  • Loading DICOM files—the proper way of loading the DICOM file ensures that we will not lose the exact quality
  • Pixel values to Hounsfield Units alignment—the Hounsfield Unit (HU) measures radiodensity for each body tissue. The Hounsfield scale that determines the values for various tissues usually ranges from −1000 HU to +3000 HU, and therefore, this step ensures that the pixel values for each CT scan do not exceed these thresholds.
  • Resampling to isomorphic resolution—the distance between consecutive slices in each CT scan defines the slice thickness. This would mean a nontrivial challenge for the neural network. The thickness depends on the CT device setup, and therefore it is necessary to create equally spaced slices.
  • [Optional] Specific part segmentation—each tissue corresponds to a specific range in the Hounsfield units scale, and in some cases, we can segment out specific parts of the CT scan by thresholding the image.
  • Normalization and zero centering—these two steps ensure that the input data that are feed into the neural network are normalized into [0, 1] interval (normalization) and are zero centered (achieved by subtracting the mean value of the image pixel values).
Preprocessing the image dataset before feeding the CNN or other classifiers is important for all imaging modalities. Several preprocessing steps are recommended for the medical images before they are fed as input to the deep neural network model, such as (1) artefact removal, (2) normalization, (3) slice timing correction (STC), (4) image registration and (5) bias field correction. While all the steps (1) to (5), help in acquiring reliable results, STC and image registration are very important in the case of 3D medical images (especially nMR and CT images). Artefact removal and normalization are the most performed preprocessing steps across the modalities [146].

2.4. Dataset Specification

This study comprises approximately 500 iCAT CBCT scans of human heads. Each CBCT scan has the spatial resolution of 768 × 768 pixels and the default device pixel spacing is [0.3 × 0.3 × 0.3] millimeters.
The subjects are split by sex, with the ratio of 6:4 for female/male ranging from 8 to 72 years. The majority (90%) of the subjects are between 18 and 36 years.
These dataset parameters were used in suggested considerations for 3D CNN applications concepts stated in Section 2.2 Main tasks definitions.

2.5. Deep Learning Approach

2.5.1. Age Estimation Using 3D Deep Neural Networks

In recent research [7,14] AI-based age estimation has proven to be a successful competitor to classical approaches from forensic medicine. The aim of this study is to create an automated system for age estimation from 3D cranial CT scans. There is an expectation that particular parts of the skull have a decisive impact on the final prediction, and therefore we propose a solution that includes two stages:
Age estimation from dense tissue layer—we use whole skull CT scan as an input into the 3D convolutional neural network, which would serve as a regression model that estimates the continuous values of age for each CT scan separately.
[Experimental] Visualization of network activations that represent regions of interest—neural network’s intermediate layers often serve as a an excellent explaining tool in order to find visual explanation heat maps [113] that highlight regions that affect neural network the most.
As for the specific neural network architecture, we derive the backbone part from the current state of the art research. We primarily consider the EfficientNet [147] and DenseNet [148] with their implementations adapted to 3D inputs.
Both architectures base includes convolutional layers that serve as feature extraction blocks to obtain specific indicators from input x represented as a loaded DICOM image. These extracted feature maps are then forwarded to a fully-connected layer that outputs the single age estimation value:
ŷ = F C(CL(x))
where CL is an intermediate block consisting of convolutional layers, FC is a fully-connected top part of the network that outputs a single floating-point value.

2.5.2. Sex Classification Using 3D Deep Neural Networks

The determination of sex from human remains is a challenging task in various fields such as archeology, physical anthropology and forensics because there is no proven method that exactly leads to correct classification.
The use of AI in this field is highly desirable as manual determination is often very complex and time-consuming. The objectiveness of the deep learning approach can also eliminate human bias leading to reliable software products.
The sex classification is carried out similarly to the previous age estimation approach, but this task´s objective is to classify the final outputs from the neural network into 2 classes—female and male. For this purpose, we use the softmax activation function as a last operation to obtain class probabilities for both targets. The computation is following:
ŷ = α (FC(CL(x)))
where CL and FC represent the convolutional and fully-connected blocks of the neural network.
The experimental part would include the input x consisting of 2 separate inputs—one will be the segmented skull and the other will be the segmented soft tissue (skin) which is achieved by setting different thresholds for segmentation preprocessing step.

2.5.3. Automatization of Cephalometric Analysis

The cephalometric analysis aims to set landmarks of CT(CBCT) scans which serve as an important factor in the alignment of a skull. These measurements can also be used as surgery planning parameters or pre-and post-surgery comparisons [149,150].
The idea behind this approach is to use 3D convolutional neural networks for fully automated cephalometric analysis. Networks aim to output probabilistic estimations for each cephalometric landmark and then create a projection of these estimations into a real skull CT scan (Figure 9).
Two approaches come into consideration:
  • Landmarks estimation in whole CT scan image—in this approach, the probability estimation for all landmarks is assigned for each pixel in the CT scan
  • Landmarks estimation for selected regions of interest—assuming that each landmark corresponds to a specific area we could add another preprocessing step—slice cut where each slice would be a template-based region fed into a neural network. We can determine the expected landmark detection for each slice independently, which should help in the final model performance

2.5.4. Neural Networks Architectures and Clinical Data Pre-Processing

Recently, CNNs have been successfully applied in widespread medical image analysis and achieved significant benefits [9,59,115,141,151]. We investigated the design of a 3D CNN with backbones based on Resnet, MobileNet, and SqueezeNet models, which have proven to be the most efficient and widely used in various applications. One of the preferable architectures was based on 3D Resnet34 for the mandible segmentation in research of Pham et al. 2021 [113].
We have considered various approaches:
  • Use whole 3D CT scan as an input into the neural network and output one value for age estimation as floating value and one for sex classification as a binary value.
  • Segment out the mandible and use it as input into the neural network. Output is the same as in the previous task.
  • (experimental) Use a whole 3D CT scan to input into the neural network and output multiple values representing specific skull features (as discussed at the meeting last week). Then use these values as an input into another machine learning model to estimate age and gender.
Suppose we take an example of mandible segmentation from DICOM. The first step is to have DICOM files loaded and then, added any missing metadata; particularly, the slice thickness, that is, the pixel size in the Z direction, which was obtained from the DICOM file. The unit of measurement in CBCT scans is the Hounsfield Unit (HU), which is a measure of radiodensity. Thus, HU shall be converted to pixel values. Subsequently, it shall be resampled to an isomorphic resolution to remove the scanner resolution. The slice thickness refers to the distance between consecutive slices (when viewing a 3D image as a collection of 2D slices) and varies between scans.
The final preprocessing step is bone segmentation and pixel normalization. Mandible bone extraction is complex because the surrounding bone has to be removed. An image binary thresholding and morphological opening operation for each slice shall be applied.
The morphological opening operation is an essential technique in image processing, achieved by erosion and the dilation of an image. This technique helps to remove small objects while retaining more significant parts from an image. To obtain the mandible bone part, the largest areas after morphological opening shall be kept. Finally, all the slices shall be stacked together to obtain the mandible voxels.

2.6. Evaluation

All approaches are evaluated in a classical machine learning manner—the dataset is split into three parts train, validation and test split. The test split mainly serves as a benchmarking set in order to compare our results with other approaches.

2.6.1. Regression Models

When dealing with regression models in the deep learning field, we usually take into consideration two main regression metrics—Mean Absolute Error (MAE) and Mean Squared Error (MSE). Both metrics calculate the error between predicted y and ground truth labels denoted as y.
MAE is defined as the mean of the sum of absolute differences between y and ŷ for each pixel:
MAE = 1 n y ŷ  
while MSE is defined as mean of the squares of the errors, where error is defined as difference between y and ŷ:
MSE = 1 n y ŷ   2
the regression tasks are primarily related to Task 1—age estimation and Task 3—automated cephalometric analysis.

2.6.2. Classification Models

In order to evaluate the classification task, which in our case is represented by Task 2—sex classification, we need to consider the current distribution of male and female samples in our dataset. As the distribution is approximately 6:4 (almost equal), we can calculate the overall accuracy and corresponding confusion matrix (CM). In the later phase, we can also test other metrics such as precision, recall or F1 score.
The calculation of accuracy is defined just as the number of correct predictions divided by the total number of predictions. More interesting for use would be the CM. It is a tabular visualization of a model prediction for each class separately (Figure 10).

3. Resulting Summary of Proposed Approach for Utilization of 3D CNN in Investigated Aspects of Forensic Medicine

This chapter is presenting summary outcome from the detailed research in previous sections of this paper. Investigation of 3D CNN modalities, their features, advantages and disadvantages and also clinical requirements for implementation in the field of forensic medicine has led to these proposed designs (guide) of future forensic research based on 3D CNN analyses.
Table 2 presents condensed summary of recommended approach for 3D CNN implementations in various forensic topics. Expected input data is the minimal dataset of 500 full-head CBCT scans, described in more detail in previous sections.

4. Discussion

The authors of this paper have no doubts that 3D CNN, as another evolutionary step in advanced AI, will be with practical implementation a watershed moment in forensic medicine fields dealing with morphological aspects.
With considered data input as CT or CBCT (DICOM), the implementation of 3D CNN algorithms opens unique opportunities in areas of:
  • Biological age determination
  • Sex determination
  • Automatized, precise and reliable:
    3D cephalometric analysis of soft and hard tissues
    3D face prediction from the skull (soft-tissues) and vice versa
    Search for hidden damage in post-mortem high-resolution CT images
    Asymmetry and disproportionality evaluation
  • Predictions of:
    Hard-tissue and soft tissue growth
    Aging in general
    Ideal face proportions respecting golden ratio proportions
  • 3D reconstructions of:
    Missing parts of the skull or face
    3D dental fingerprints for identification with 2D dental records
First clinical applications of 3D CNN have shown [91,113,115,126,150] that the algorithms can be successfully used in CT analysis and identifications of specific diseases such as Alzheimer or COVID19 as these have a specific representation on the X-ray. With a high probability bordering on certainty, the future development of advanced 3D CNN will result in sophisticated automatized algorithms processing 3D diagnostic data similarly to the trained human eye of the forensic expert. These algorithms will automatically process 3D diagnostic data such as CT or NMR, searching for patterns they were trained to see. They will recognize unseen details of hidden damage or representations of rare diseases when trained to do so. In the next level, they will approximate the finding to become an ultimate autopsy tool for even unknown diseases [36,113,126,152].
The limitation of this paper is that practical examination of the proposed directions for 3D CNN implementations will require some time. Currently, there are many different 3D CNN in development, and actually, this is where most of the research activity is carried out [151,153,154,155].
Another limitation of this study is the high level of dynamics of research and development in this field of advanced AI implementations. The velocity in training the 3D CNN is high, and it is possible that a better approach can be recognized in the process.
Interesting limitation of 3D CNN usage is the known fact [99] the any AI may become biased in the same way as a human forensic expert does and not only in the context of the criminal trial. This depends on the source data used for AI training [99] and is elaborated in more context in Section 1.2. On the other hand, in many forensic cases we need to achieve highest probabilities on the boundary with certainty. Here a respected and internationally recognized algorithm might become a useful tool for achieving an unprecedented levels of probability superior to human evaluation. However, this development is a possibility, not certainty.
The final limitation of implementing the suggested designs for 3D CNN implementation for forensic researchers is the physical and legal availability of big data necessary for 3D CNN training. This can be solved with multicentric cooperation.
There already exist many CNN processing DICOM data and are available for use [11,12,14]. Researchers this year have already achieved significant milestones in multiclass CBCT image segmentation for orthodontics with Deep Learning. They trained and validated a mixed-scale dense convolutional neural network for multiclass segmentation of the jaw, the teeth, and the background in CBCT scans [153]. This study showed that multiclass segmentation of jaw and teeth was accurate, and its performance was comparable to binary segmentation. This is important because this strongly reduces the time required to segment multiple anatomic structures in CBCT scans.
In our efforts, we have faced the issue of CBCT scan distortion caused by metal artefacts (mostly by amalgam dental fillings). Fortunately, a novel coarse-to-fine segmentation framework was recently published based on 3D CNN and recurrent SegUnet for mandible segmentation in CBCT scans. Moreover, the experiments indicate that the proposed algorithm can provide more accurate and robust segmentation results for different imaging techniques compared to the state-of-the-art models with respect to these three datasets [156].
As there already exists a fully automated method for 3D individual tooth identification and segmentation from dental CBCT [154], these algorithms can be combined.
The most complex area covered by this paper is a 3D prediction of growth and in a wider perspective of ageing. It is known that this process is laden with various variables including hormonal (sex) [142,143,157,158,159] and functional aspects (bad habits) [160,161,162], as well as genetics [163,164,165,166].
The only published study focused on predicting the facial growth direction with the implementation of Machine Learning methods is from 2021 Kazmierczak et al. 2021 [90,145]. The outcomes of this paper are limited in regards to facial growth prediction. The authors of this original paper did feature selection and pointed the attribute that plays a central role in facial growth. Then they performed data augmentation (DA) methods.
The principal weakness of this study is not the method but probably the input. The authors used only 2D lateral X-rays of various qualities and sizes. In addition, the evaluation was performed only in one 2D projection. The researchers focused on the angle between the Sella—Nasion line and the Mandibular plane formed by connecting the point gonion to gnathion at the inferior border of the mandible. They engaged an orthodontic expert to identify approximately 20 characteristic anatomic landmarks on LC to assess a subject. These were assessed manually on the lateral cephalogram. Some of the landmarks define angles which, from the clinical perspective, have special significance. As far as facial growth direction is concerned, there are no standardized measurements available in the literature to evaluate. The focus of supervised ML with a concentration on established cephalometric parameters might be wrong. It is the fact that they were originally chosen as well distinguished points on lateral X-ray with a priority of high reproducibility. So as considered by Hirschfeld and Moyers more than 50 years ago, we might be looking in the wrong places. Prediction of the change of SN/MP also oversimplifies the problem. The questions from the past remain, and facial growth prediction remains a complex mystery. The application of 3D CNN on this most complex task is described in more technical details and discussed later in the context of addressing other variables such as biological age, sex and genetics.

5. Conclusions

In conclusion, we can assume that the 3D CNN, as an advanced AI feature, will shift the paradigm in all areas researched in this paper. Forensic experts are now guided to step into the era of artificial intelligence as a helpful tool for research and possibly even future routine forensic analyses. Proposed methods and metrics for 3D CNN application on particular forensic topics (Biological age determination, Sex determination, 3D cephalometric analysis and Face prediction from skull), summarized in resulting Table 2, can be used as the initial guide. Forensic 3D reconstructions using artificial intelligence will be new, exciting and practically usable methods.
The implementation of advanced AI still requires interdisciplinary cooperation, albeit, with understanding, it can be used to crack unsolved mysteries. It definitely is not a trend that can be ignored.

Author Contributions

Conceptualization, A.T., H.S.K., V.K., S.K., R.B., N.M., P.K., K.M.K., M.P. and I.V.; methodology, A.T., H.S.K., V.K., S.K., R.B., N.M., P.K., K.M.K., M.P. and I.V.; software, A.T., H.S.K., S.K. and I.V.; validation, A.T., H.S.K., V.K., S.K. and I.V.; formal analysis, A.T., H.S.K., V.K., S.K. and I.V.; investigation, A.T., H.S.K., V.K., S.K. and I.V.; resources, A.T.; data curation, A.T.; writing—original draft preparation, A.T., H.S.K., V.K., S.K., R.B., N.M., P.K., K.M.K., M.P. and I.V.; writing—review and editing, A.T., H.S.K., V.K., S.K., R.B., N.M., P.K., K.M.K., M.P. and I.V.; visualization, A.T.; supervision, A.T., H.S.K., S.K. and I.V.; project administration, A.T.; funding acquisition, A.T., H.S.K., S.K. and I.V.; All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the KEGA grant agency of the Ministry of Education, Science, Research, and Sport of the Slovak Republic (Grant No. 081UK-4/2021).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and no approval was necessary by the Ethics Committee. Ethical review and approval were waived for this study, due to the fact that no experimental materials or approaches were used.

Informed Consent Statement

Written informed consent was obtained from all subjects involved in the study.

Data Availability Statement

We fully adhere to Data Availability Statements in section “MDPI Research Data Policies” at https://0-www-mdpi-com.brum.beds.ac.uk/ethics (accessed on 1 November 2021).

Acknowledgments

We acknowledge technological support of Cognexa software company for support and digital dental lab infrastructure of 3Dent Medical s.r.o company as well as dental clinic Sangre Azul s.r.o.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

3D CNNThree-Dimensional Convolutional Neural Network
AIArtificial Intelligence
CBCTCone-Beam Computer Tomography
CMConfusion Matrix
DICOMCommunications in Medicine Format
DAData Augmentation
FOVField of View
HUHounsfield Unit
MLMachine Learning
MAEMean Absolute Error
MSEMean Squared Error
nMRNuclear Magnetic Resonance
STCSlice Timing Correction

References

  1. Ya-ting, F.A.; Qiong, L.A.; Tong, X.I. New Opportunities and Challenges for Forensic Medicine in the Era of Artificial Intelligence Technology. J. Forensic Med. 2020, 36, 77. [Google Scholar] [CrossRef]
  2. Khanagar, S.B.; Vishwanathaiah, S.; Naik, S.; Al-Kheraif, A.A.; Devang Divakar, D.; Sarode, S.C.; Bhandi, S.; Patil, S. Application and Performance of Artificial Intelligence Technology in Forensic Odontology—A Systematic Review. Leg. Med. 2021, 48, 101826. [Google Scholar] [CrossRef]
  3. Lefèvre, T. Big Data in Forensic Science and Medicine. J. Forensic Leg. Med. 2018, 57, 1–6. [Google Scholar] [CrossRef] [PubMed]
  4. Pathak, M.; Narang, H. Application of Artificial Intelligence in the Field of Forensic Medicine. Indian J. Forensic Med. Toxicol. 2021, 15. [Google Scholar] [CrossRef]
  5. Franke, K.; Srihari, S.N. Computational Forensics: An Overview. In Computational Forensics; Springer: Berlin/Heidelberg, Germany, 2008; Volume 5158, pp. 1–10. [Google Scholar] [CrossRef]
  6. Cossellu, G.; de Luca, S.; Biagi, R.; Farronato, G.; Cingolani, M.; Ferrante, L.; Cameriere, R. Reliability of Frontal Sinus by Cone Beam-Computed Tomography (CBCT) for Individual Identification. Radiol. Med. 2015, 120, 1130–1136. [Google Scholar] [CrossRef]
  7. der Mauer, M.A.; Well, E.J.; Herrmann, J.; Groth, M.; Morlock, M.M.; Maas, R.; Säring, D. Automated Age Estimation of Young Individuals Based on 3D Knee MRI Using Deep Learning. Int. J. Leg. Med. 2020, 135, 649–663. [Google Scholar] [CrossRef]
  8. Armanious, K.; Abdulatif, S.; Bhaktharaguttu, A.R.; Küstner, T.; Hepp, T.; Gatidis, S.; Yang, B. Organ-Based Chronological Age Estimation Based on 3D MRI Scans. In Proceedings of the 28th European Signal Processing Conference, Amsterdam, The Netherlands, 24–28 August 2021; pp. 1225–1228. [Google Scholar] [CrossRef]
  9. Souadih, K.; Belaid, A.; Salem, D. Automatic Segmentation of the Sphenoid Sinus in CT-Scans Volume with DeepMedics 3D CNN Architecture. Med. Technol. J. 2019, 3, 334–346. [Google Scholar] [CrossRef]
  10. Dubey, C.; Raj, S.; Munuswamy, S.; Katta, A.; Chetty, G. Semantic Segmentation Using Deep Neural Networks in Medicine—A Survey. ISO 2018, 3, 39–48. [Google Scholar]
  11. Du, R.; Vardhanabhuti, V. 3D-RADNet: Extracting Labels from DICOM Metadata for Training General Medical Domain Deep 3D Convolution Neural Networks. Proc. Mach. Learn. Res. 2020, 121, 174–192. [Google Scholar]
  12. Fajar, A.; Sarno, R.; Fatichah, C.; Fahmi, A. Reconstructing and Resizing 3D Images from DICOM Files. J. King Saud Univ. Comput. Inf. Sci. 2020. [Google Scholar] [CrossRef]
  13. da Silva, R.D.C.; Jenkyn, T.; Carranza, V. Convolutional Neural Networks and Geometric Moments to Identify the Bilateral Symmetric Midplane in Facial Skeletons from CT Scans. Biology 2021, 10, 182. [Google Scholar] [CrossRef] [PubMed]
  14. Singh, S.P.; Wang, L.; Gupta, S.; Goli, H.; Padmanabhan, P.; Gulyás, B. 3D Deep Learning on Medical Images: A Review. Sensors 2020, 20, 5097. [Google Scholar] [CrossRef]
  15. Armanious, K.; Abdulatif, S.; Shi, W.; Salian, S.; Küstner, T.; Weiskopf, D.; Hepp, T.; Gatidis, S.; Yang, B. Age-Net: An MRI-Based Iterative Framework for Biological Age Estimation. arXiv 2009, arXiv:10765.14-37. [Google Scholar] [CrossRef]
  16. Sajedi, H.; Pardakhti, N. Age Prediction Based on Brain MRI Image: A Survey. J. Med Syst. 2019, 43, 1–30. [Google Scholar] [CrossRef]
  17. de Luca, S.; Alemán, I.; Bertoldi, F.; Ferrante, L.; Mastrangelo, P.; Cingolani, M.; Cameriere, R. Age Estimation by Tooth/Pulp Ratio in Canines by Peri-Apical X-Rays: Reliability in Age Determination of Spanish and Italian Medieval Skeletal Remains. J. Archaeol. Sci. 2010, 37, 3048–3058. [Google Scholar] [CrossRef] [Green Version]
  18. Cameriere, R.; Ferrante, L.; Cingolani, M. Age Estimation in Children by Measurement of Open Apices in Teeth. Int. J. Leg. Med. 2006, 120, 49–52. [Google Scholar] [CrossRef] [Green Version]
  19. Hostiuc, S.; Diaconescu, I.; Rusu, M.C.; Negoi, I. Age Estimation Using the Cameriere Methods of Open Apices: A Meta-Analysis. Healthcare 2021, 9, 237. [Google Scholar] [CrossRef] [PubMed]
  20. Cameriere, R.; de Luca, S.; Egidi, N.; Bacaloni, M.; Maponi, P.; Ferrante, L.; Cingolani, M. Automatic Age Estimation in Adults by Analysis of Canine Pulp/Tooth Ratio: Preliminary Results. J. Forensic Radiol. Imaging 2015, 3, 61–66. [Google Scholar] [CrossRef]
  21. Rai, B.; Kaur, J.; Cingolani, M.; Ferrante, L.; Cameriere, R. Age Estimation in Children by Measurement of Open Apices in Teeth: An Indian Formula. Int. J. Leg. Med. 2010, 124, 237–241. [Google Scholar] [CrossRef] [Green Version]
  22. Cameriere, R.; Ferrante, L. Canine Pulp Ratios in Estimating Pensionable Age in Subjects with Questionable Documents of Identification. Forensic Sci. Int. 2011, 206, 132–135. [Google Scholar] [CrossRef] [PubMed]
  23. Cameriere, R.; Giuliodori, A.; Zampi, M.; Galić, I.; Cingolani, M.; Pagliara, F.; Ferrante, L. Age Estimation in Children and Young Adolescents for Forensic Purposes Using Fourth Cervical Vertebra (C4). Int. J. Leg. Med. 2015, 129, 347–355. [Google Scholar] [CrossRef] [PubMed]
  24. Cameriere, R.; Pacifici, A.; Pacifici, L.; Polimeni, A.; Federici, F.; Cingolani, M.; Ferrante, L. Age Estimation in Children by Measurement of Open Apices in Teeth with Bayesian Calibration Approach. Forensic Sci. Int. 2016, 258, 50–54. [Google Scholar] [CrossRef] [PubMed]
  25. de Micco, F.; Martino, F.; Velandia Palacio, L.A.; Cingolani, M.; Campobasso, C. pietro Third Molar Maturity Index and Legal Age in Different Ethnic Populations: Accuracy of Cameriere’s Method. Med. Sci. Law 2021, 61, 105–112. [Google Scholar] [CrossRef]
  26. Scendoni, R.; Cingolani, M.; Giovagnoni, A.; Fogante, M.; Fedeli, P.; Pigolkin, Y.I.; Ferrante, L.; Cameriere, R. Analysis of Carpal Bones on MR Images for Age Estimation: First Results of a New Forensic Approach. Forensic Sci. Int. 2020, 313. [Google Scholar] [CrossRef]
  27. Cameriere, R.; Santoro, V.; Roca, R.; Lozito, P.; Introna, F.; Cingolani, M.; Galić, I.; Ferrante, L. Assessment of Legal Adult Age of 18 by Measurement of Open Apices of the Third Molars: Study on the Albanian Sample. Forensic Sci. Int. 2014, 245, 205.e1–205.e5. [Google Scholar] [CrossRef]
  28. de Luca, S.; Mangiulli, T.; Merelli, V.; Conforti, F.; Velandia Palacio, L.A.; Agostini, S.; Spinas, E.; Cameriere, R. A New Formula for Assessing Skeletal Age in Growing Infants and Children by Measuring Carpals and Epiphyses of Radio and Ulna. J. Forensic Leg. Med. 2016, 39, 109–116. [Google Scholar] [CrossRef]
  29. Cameriere, R.; de Luca, S.; Biagi, R.; Cingolani, M.; Farronato, G.; Ferrante, L. Accuracy of Three Age Estimation Methods in Children by Measurements of Developing Teeth and Carpals and Epiphyses of the Ulna and Radius. J. Forensic Sci. 2012, 57, 1263–1270. [Google Scholar] [CrossRef]
  30. Cameriere, R.; de Luca, S.; de Angelis, D.; Merelli, V.; Giuliodori, A.; Cingolani, M.; Cattaneo, C.; Ferrante, L. Reliability of Schmeling’s Stages of Ossification of Medial Clavicular Epiphyses and Its Validity to Assess 18 Years of Age in Living Subjects. Int. J. Leg. Med. 2012, 126, 923–932. [Google Scholar] [CrossRef]
  31. Cameriere, R.; de Luca, S.; Cingolani, M.; Ferrante, L. Measurements of Developing Teeth, and Carpals and Epiphyses of the Ulna and Radius for Assessing New Cut-Offs at the Age Thresholds of 10, 11, 12, 13 and 14 Years. J. Forensic Leg. Med. 2015, 34, 50–54. [Google Scholar] [CrossRef]
  32. Afrianty, I.; Nasien, D.; Kadir, M.R.A.; Haron, H. Backpropagation Neural Network for Sex Determination from Patella in Forensic Anthropology. In Advances in Computer Science and its Applications; Springer: Berlin/Heidelberg, Germany, 2014; Volume 279, pp. 723–728. [Google Scholar] [CrossRef]
  33. Pötsch-Schneider, L.; Endris, R.; Schmidt, H. Discriminant Analysis of the Mandible for Sex Determination. Z. Fur Rechtsmedizin. J. Leg. Med. 1985, 94, 21–30. [Google Scholar] [CrossRef]
  34. el Morsi, D.A.; Gaballah, G.; Tawfik, M.W. Sex Determination in Egyptian Population from Scapula by Computed Tomography. J Forensic Res 2017, 8, 376. [Google Scholar] [CrossRef]
  35. Kalmey, J.; Rathbun, T. Sex Determination by Discriminant Function Analysis of the Petrous Portion of the Temporal Bone. J. Forensic Sci. 1996, 41, 865–867. [Google Scholar] [CrossRef] [PubMed]
  36. Sanchez, L.; Grajeda, C.; Baggili, I.; Hall, C. A Practitioner Survey Exploring the Value of Forensic Tools, AI, Filtering, & Safer Presentation for Investigating Child Sexual Abuse Material (CSAM). Digit. Investig. 2019, 29, S124–S142. [Google Scholar] [CrossRef]
  37. Teixeira, W.R. Sex Identification Utilizing the Size of the Foramen Magnum. Am. J. Forensic Med. Pathol. 1982, 3, 203–206. [Google Scholar] [CrossRef]
  38. Randhawa, K.; Narang, R.S.; Arora, P.C. Study of the effect of age changes on lip print pattern and its reliability in sex determination. J. Forensic Odonto-Stomatol. 2011, 29, 45. [Google Scholar]
  39. Bewes, J.; Low, A.; Morphett, A.; Pate, F.D.; Henneberg, M. Artificial Intelligence for Sex Determination of Skeletal Remains: Application of a Deep Learning Artificial Neural Network to Human Skulls. J. Forensic Leg. Med. 2019, 62, 40–43. [Google Scholar] [CrossRef]
  40. Regina Silva Lucena dos Santos, E.; Paulo Feitosa de Albuquerque, P.; Virgínio de Albuquerque, P.; Duarte Ribeiro de Oliveira, B.; Caiaffo, V. Determination of Sex Based on the Morphometric Evaluation of the Proximal Tibia Determinación Del Sexo Basada En La Evaluación Morfométrica de La Tibia Proximal. Int. J. Morphol 2018, 36, 104–108. [Google Scholar] [CrossRef]
  41. Hwang, H.-W.; Park, J.-H.; Moon, J.-H.; Yu, Y.; Kim, H.; Her, S.-B.; Srinivasan, G.; Aljanabi, M.N.A.; Donatelli, R.E.; Lee, S.-J. Automated Identification of Cephalometric Landmarks: Part 2-Might It Be Better Than Human? Angle Orthod. 2020, 90, 69–76. [Google Scholar] [CrossRef] [Green Version]
  42. Lee, S.M.; Kim, H.P.; Jeon, K.; Lee, S.-H.; Seo, J.K. Automatic 3D Cephalometric Annotation System Using Shadowed 2D Image-Based Machine Learning. Phys. Med. Biol. 2019, 64, 055002. [Google Scholar] [CrossRef] [PubMed]
  43. Montúfar, J.; Romero, M.; Scougall-Vilchis, R.J. Hybrid Approach for Automatic Cephalometric Landmark Annotation on Cone-Beam Computed Tomography Volumes. Am. J. Orthod. Dentofac. Orthop. 2018, 154, 140–150. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Montúfar, J.; Romero, M.; Scougall-Vilchis, R.J. Automatic 3-Dimensional Cephalometric Landmarking Based on Active Shape Models in Related Projections. Am. J. Orthod. Dentofac. Orthop. 2018, 153, 449–458. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Codari, M.; Caffini, M.; Tartaglia, G.M.; Sforza, C.; Baselli, G. Computer-Aided Cephalometric Landmark Annotation for CBCT Data. Int. J. Comput. Assist. Radiol. Surg. 2016, 12, 113–121. [Google Scholar] [CrossRef] [Green Version]
  46. Baksi, S.; Freezer, S.; Matsumoto, T.; Dreyer, C. Accuracy of an Automated Method of 3D Soft Tissue Landmark Detection. Eur. J. Orthod. 2020. [Google Scholar] [CrossRef]
  47. Ed-Dhahraouy, M.; Riri, H.; Ezzahmouly, M.; Bourzgui, F.; el Moutaoukkil, A. A New Methodology for Automatic Detection of Reference Points in 3D Cephalometry: A Pilot Study. Int. Orthod. 2018, 16, 328–337. [Google Scholar] [CrossRef]
  48. Croquet, B.; Matthews, H.; Mertens, J.; Fan, Y.; Nauwelaers, N.; Mahdi, S.; Hoskens, H.; el Sergani, A.; Xu, T.; Vandermeulen, D.; et al. Automated Landmarking for Palatal Shape Analysis Using Geometric Deep Learning. Orthod. Craniofac. Res. 2021. [Google Scholar] [CrossRef] [PubMed]
  49. Lachinov, D.; Getmanskaya, A.; Turlapov, V. Cephalometric Landmark Regression with Convolutional Neural Networks on 3D Computed Tomography Data. Pattern Recognit. Image Anal. 2020, 30, 512–522. [Google Scholar] [CrossRef]
  50. Dot, G.; Rafflenbeul, F.; Arbotto, M.; Gajny, L.; Rouch, P.; Schouman, T. Accuracy and Reliability of Automatic Three-Dimensional Cephalometric Landmarking. Int. J. Oral Maxillofac. Surg. 2020, 49, 1367–1378. [Google Scholar] [CrossRef]
  51. Bermejo, E.; Taniguchi, K.; Ogawa, Y.; Martos, R.; Valsecchi, A.; Mesejo, P.; Ibáñez, O.; Imaizumi, K. Automatic Landmark Annotation in 3D Surface Scans of Skulls: Methodological Proposal and Reliability Study. Comput. Methods Programs Biomed. 2021, 210, 106380. [Google Scholar] [CrossRef]
  52. Lee, J.-H.; Yu, H.-J.; Kim, M.; Kim, J.-W.; Choi, J. Automated Cephalometric Landmark Detection with Confidence Regions Using Bayesian Convolutional Neural Networks. BMC Oral Health 2020, 20, 270. [Google Scholar] [CrossRef] [PubMed]
  53. Kang, S.H.; Jeon, K.; Kang, S.-H.; Lee, S.-H. 3D Cephalometric Landmark Detection by Multiple Stage Deep Reinforcement Learning. Sci. Rep. 2021, 11, 17509. [Google Scholar] [CrossRef] [PubMed]
  54. Juneja, M.; Garg, P.; Kaur, R.; Manocha, P.; Prateek; Batra, S.; Singh, P.; Singh, S.; Jindal, P. A Review on Cephalometric Landmark Detection Techniques. Biomed. Signal Process. Control 2021, 66, 102486. [Google Scholar] [CrossRef]
  55. Yun, H.S.; Hyun, C.M.; Baek, S.H.; Lee, S.-H.; Seo, J.K. Automated 3D Cephalometric Landmark Identification Using Computerized Tomography. arXiv 2020, arXiv:2101.05205. [Google Scholar]
  56. Silva, T.P.; Hughes, M.M.; Menezes, L.D.; de Melo, M.D.; Takeshita, W.M.; Freitas, P.H. Artificial Intelligence-Based Cephalometric Landmark Annotation and Measurements According to Arnett’s Analysis: Can We Trust a Bot to Do That? Dentomaxillofac. Radiol. 2021, 20200548. [Google Scholar] [CrossRef]
  57. Yun, H.S.; Jang, T.J.; Lee, S.M.; Lee, S.-H.; Seo, J.K. Learning-Based Local-to-Global Landmark Annotation for Automatic 3D Cephalometry. Phys. Med. Biol. 2020, 65, 085018. [Google Scholar] [CrossRef]
  58. Kang, S.H.; Jeon, K.; Kim, H.-J.; Seo, J.K.; Lee, S.-H. Automatic Three-Dimensional Cephalometric Annotation System Using Three-Dimensional Convolutional Neural Networks: A Developmental Trial. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2019, 8, 210–218. [Google Scholar] [CrossRef]
  59. Kute, R.S.; Vyas, V.; Anuse, A. Association of Face and Facial Components Based on CNN and Transfer Subspace Learning for Forensics Applications. SN Comput. Sci. 2020, 1, 1–16. [Google Scholar] [CrossRef]
  60. Knyaz, V.A.; Maksimov, A.A.; Novikov, M.M.; Urmashova, A.V. Automatic Anthropological Landmarks Recognition and Measurements. ISPAr 2021, 4421, 137–142. [Google Scholar] [CrossRef]
  61. Guyomarc, P.; Dutailly, B.; Charton, J.; Santos, F.; Desbarats, P. Helene Coqueugniot Anthropological Facial Approximation in Three Dimensions (AFA3D): Computer-Assisted Estimation of the Facial Morphology Using Geometric Morphometrics. J Forensic Sci. 2014, 1502–1516. [Google Scholar] [CrossRef]
  62. Barbero-García, I.; Pierdicca, R.; Paolanti, M.; Felicetti, A.; Lerma, J.L. Combining Machine Learning and Close-Range Photogrammetry for Infant’s Head 3D Measurement: A Smartphone-Based Solution. Measurement 2021, 182, 109686. [Google Scholar] [CrossRef]
  63. Wen, Y.; Mingquan, Z.; Pengyue, L.; Guohua, G.; Xiaoning, L.; Kang, L. Ancestry Estimation of Skull in Chinese Population Based on Improved Convolutional Neural Network. In Proceedings of the 2020 IEEE International Conference on Bioinformatics and Biomedicine, BIBM, Online Conference, 16–19 December 2020; pp. 2861–2867. [Google Scholar] [CrossRef]
  64. Liu, X.; Zou, Y.; Kuang, H.; Ma, X. Face Image Age Estimation Based on Data Augmentation and Lightweight Convolutional Neural Network. Symmetry 2020, 12, 146. [Google Scholar] [CrossRef] [Green Version]
  65. Fatima, A.; Shahid, A.R.; Raza, B.; Madni, T.M.; Janjua, U.I. State-of-the-Art Traditional to the Machine- and Deep-Learning-Based Skull Stripping Techniques, Models, and Algorithms. J. Digit. Imaging 2020, 33, 1443–1464. [Google Scholar] [CrossRef]
  66. Tu, P.; Book, R.; Liu, X.; Krahnstoever, N.; Adrian, C.; Williams, P. Automatic Face Recognition from Skeletal Remains. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007. [Google Scholar] [CrossRef]
  67. Curtner, R.M. Predetermination of the Adult Face. Am. J. Orthod. 1953, 39, 201–217. [Google Scholar] [CrossRef]
  68. Albert, A.M.; Ricanek, K.; Patterson, E. A Review of the Literature on the Aging Adult Skull and Face: Implications for Forensic Science Research and Applications. Forensic Sci. Int. 2007, 172, 1–9. [Google Scholar] [CrossRef]
  69. Zhang, D.; Wu, Z.; Wang, X.; Lv, C.; Liu, N. 3D Skull and Face Similarity Measurements Based on a Harmonic Wave Kernel Signature. Vis. Comput. 2020, 37, 749–764. [Google Scholar] [CrossRef]
  70. Jain, A.; Singh, R.; Vatsa, M. Face Recognition for Forensics. Available online: https://repository.iiitd.edu.in/xmlui/handle/123456789/593?show=fullhttps://repository.iiitd.edu.in/xmlui/handle/123456789/593?show=full (accessed on 1 November 2021).
  71. Ma, Q.; Kobayashi, E.; Fan, B.; Nakagawa, K.; Sakuma, I.; Masamune, K.; Suenaga, H. Automatic 3D Landmarking Model Using Patch-Based Deep Neural Networks for CT Image of Oral and Maxillofacial Surgery. Int. J. Med Robot. Comput. Assist. Surg. 2020, 16, e2093. [Google Scholar] [CrossRef] [Green Version]
  72. Tin, M.L. Machine, Discourse and Power: From Machine Learning in Construction of 3D Face to Art and Creativity. Adv. Intell. Syst. Comput. 2020, 517–523. [Google Scholar] [CrossRef]
  73. Knyaz, V.A.; Kniaz, V.V.; Novikov, M.M.; Galeev, R.M. Machine learning for approximating unknown face—Proquest. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2020, 43, 857–862. [Google Scholar] [CrossRef]
  74. Rajesh Kumar, B. Region of interest feature extraction in facial expressions with convolutional neural network classification. Ctact J. Data Sci. Mach. Learn. 2020, 2, 149–152. [Google Scholar]
  75. Mashouri, P.; Skreta, M.; Phillips, J.; McAllister, D.; Roy, M.; Senkaiahliyan, S.; Brudno, M.; Singh, D. 3D Photography Based Neural Network Craniosynostosis Triaging System. Proc. Mach. Learn. Res. 2020, 136, 226–237. [Google Scholar]
  76. Lo, L.J.; Yang, C.T.; Ho, C.T.; Liao, C.H.; Lin, H.H. Automatic Assessment of 3-Dimensional Facial Soft Tissue Symmetry Before and After Orthognathic Surgery Using a Machine Learning Model: A Preliminary Experience. Ann. Plast. Surg. 2021, 86, S224–S228. [Google Scholar] [CrossRef]
  77. Hung, K.; Yeung, A.W.K.; Tanaka, R.; Bornstein, M.M. Current Applications, Opportunities, and Limitations of AI for 3D Imaging in Dental Research and Practice. Int. J. Environ. Res. Public Health 2020, 17, 4424. [Google Scholar] [CrossRef]
  78. Dalvit Carvalho da Silva, R.; Jenkyn, T.R.; Carranza, V.A. Development of a Convolutional Neural Network Based Skull Segmentation in MRI Using Standard Tesselation Language Models. J. Pers. Med. 2021, 11, 310. [Google Scholar] [CrossRef] [PubMed]
  79. Bekhouche, S.E.; Dornaika, F.; Benlamoudi, A.; Ouafi, A.; Taleb-Ahmed, A. A Comparative Study of Human Facial Age Estimation: Handcrafted Features vs. Deep Features. Multimed. Tools Appl. 2020, 79, 26605–26622. [Google Scholar] [CrossRef]
  80. Punyani, P.; Gupta, R.; Kumar, A. Neural Networks for Facial Age Estimation: A Survey on Recent Advances. Artif. Intell. Rev. 2019, 53, 3299–3347. [Google Scholar] [CrossRef]
  81. Liu, H.; Lu, J.; Feng, J.; Zhou, J. Label-Sensitive Deep Metric Learning for Facial Age Estimation. IEEE Trans. Inf. Forensics Secur. 2018, 13, 292–305. [Google Scholar] [CrossRef]
  82. Lin, H.H.; Lo, L.J.; Chiang, W.C. A Novel Assessment Technique for the Degree of Facial Symmetry before and after Orthognathic Surgery Based on Three-Dimensional Contour Features Using Deep Learning Algorithms. PervasiveHealth Pervasive Comput. Technol. Healthc. 2019, 170–173. [Google Scholar] [CrossRef]
  83. Makaremi, M.; Lacaule, C.; Mohammad-Djafari, A. Deep Learning and Artificial Intelligence for the Determination of the Cervical Vertebra Maturation Degree from Lateral Radiography. Entropy 2019, 21, 1222. [Google Scholar] [CrossRef] [Green Version]
  84. Ramanathan, N.; Chellappa, R.; Biswas, S. Computational Methods for Modeling Facial Aging: A Survey. J. Vis. Lang. Comput. 2009, 20, 131–144. [Google Scholar] [CrossRef]
  85. Dou, P.; Shah, S.K.; Kakadiaris, I.A. End-To-End 3D Face Reconstruction With Deep Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5908–5917. [Google Scholar]
  86. Fu, Y.; Guo, G.; Huang, T.S. Age Synthesis and Estimation via Faces: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1955–1976. [Google Scholar] [CrossRef]
  87. Sai, P.K.; Wang, J.G.; Teoh, E.K. Facial Age Range Estimation with Extreme Learning Machines. Neurocomputing 2015, 149, 364–372. [Google Scholar] [CrossRef]
  88. Porto, L.F.; Lima, L.N.; Franco, A.; Pianto, D.; Machado, C.E.; de Barros Vidal, F. Estimating Sex and Age from a Face: A Forensic Approach Using Machine Learning Based on Photo-Anthropometric Indexes of the Brazilian Population. Int. J. Leg. Med. 2020, 134, 2239–2259. [Google Scholar] [CrossRef] [PubMed]
  89. Ramanathan, N.; Chellappa, R.; Biswas, S. Age Progression in Human Faces: A Survey. J. Vis. Lang. Comput. 2009, 15, 3349–3361. [Google Scholar]
  90. Kaźmierczak, S.; Juszka, Z.; Vandevska-Radunovic, V.; Maal, T.J.; Fudalej, P.; Mańdziuk, J. Prediction of the Facial Growth Direction Is Challenging. Available online: https://arxiv.org/pdf/2110.02316.pdf (accessed on 1 November 2021).
  91. Mangrulkar, A.; Rane, S.B.; Sunnapwar, V. Automated Skull Damage Detection from Assembled Skull Model Using Computer Vision and Machine Learning. Int. J. Inf. Technol. 2021, 13, 1785–1790. [Google Scholar] [CrossRef]
  92. Tan, J.S.; Liao, I.Y.; Venkat, I.; Belaton, B.; Jayaprakash, P.T. Computer-Aided Superimposition via Reconstructing and Matching 3D Faces to 3D Skulls for Forensic Craniofacial Identifications. Vis. Comput. 2019, 36, 1739–1753. [Google Scholar] [CrossRef]
  93. Schmeling, A.; Dettmeyer, R.; Rudolf, E.; Vieth, V.; Geserick, G. Forensic Age Estimation: Methods, Certainty, and the Law. Dtsch. Aerzteblatt Online 2016, 113, 44–50. [Google Scholar] [CrossRef] [Green Version]
  94. Franklin, D. Forensic Age Estimation in Human Skeletal Remains: Current Concepts and Future Directions. Leg. Med. 2010, 12, 1–7. [Google Scholar] [CrossRef]
  95. Kotěrová, A.; Navega, D.; Štepanovský, M.; Buk, Z.; Brůžek, J.; Cunha, E. Age Estimation of Adult Human Remains from Hip Bones Using Advanced Methods. Forensic Sci. Int. 2018, 287, 163–175. [Google Scholar] [CrossRef]
  96. Chia, P.Y.; Coleman, K.K.; Tan, Y.K.; Ong, S.W.X.; Gum, M.; Lau, S.K.; Lim, X.F.; Lim, A.S.; Sutjipto, S.; Lee, P.H.; et al. Detection of Air and Surface Contamination by SARS-CoV-2 in Hospital Rooms of Infected Patients. Nat. Commun. 2020, 11, 2800. [Google Scholar] [CrossRef]
  97. Singh Sankhla, M.; Kumar, R.; Jadhav, E.B. Artificial Intelligence: Advancing Automation in Forensic Science & Criminal Investigation. J. Seybold Rep. 2020, 15, 2064–2075. [Google Scholar]
  98. Yadav, J.; Kaur, A. Artificial Neural Network Implementation In Forensic Science. Eur. J. Mol. Clin. Med. 2020, 7, 5935–5939. [Google Scholar]
  99. Livingston, M. Preventing Racial Bias in Federal AI. JSPG 2020, 16. [Google Scholar] [CrossRef]
  100. Obermeyer, Z.; Powers, B.; Vogeli, C.; Mullainathan, S. Dissecting Racial Bias in an Algorithm Used to Manage the Health of Populations. Science 2019, 366, 447–453. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  101. Jarrett, A.; Choo, K.R. The Impact of Automation and Artificial Intelligence on Digital Forensics. WIREs Forensic Sci. 2021. [Google Scholar] [CrossRef]
  102. Sykes, L.; Bhayat, A.; Bernitz, H. The Effects of the Refugee Crisis on Age Estimation Analysis over the Past 10 Years: A 16-Country Survey. Int. J. Environ. Res. Public Health 2017, 14, 630. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  103. Guo, Y.C.; Han, M.; Chi, Y.; Long, H.; Zhang, D.; Yang, J.; Yang, Y.; Chen, T.; Du, S. Accurate Age Classification Using Manual Method and Deep Convolutional Neural Network Based on Orthopantomogram Images. Int. J. Leg. Med. 2021, 135, 1589–1597. [Google Scholar] [CrossRef]
  104. Li, Y.; Huang, Z.; Dong, X.; Liang, W.; Xue, H.; Zhang, L.; Zhang, Y.; Deng, Z. Forensic Age Estimation for Pelvic X-ray Images Using Deep Learning. Eur. Radiol. 2019, 29, 2322–2329. [Google Scholar] [CrossRef]
  105. Štepanovský, M.; Ibrová, A.; Buk, Z.; Velemínská, J. Novel Age Estimation Model Based on Development of Permanent Teeth Compared with Classical Approach and Other Modern Data Mining Methods. Forensic Sci. Int. 2017, 279, 72–82. [Google Scholar] [CrossRef]
  106. Vila-Blanco, N.; Varas-Quintana, P.; Aneiros-Ardao, Á.; Tomás, I.; Carreira, M.J. Automated Description of the Mandible Shape by Deep Learning. Int. J. Comput. Assist. Radiol. Surg. 2021, 1–10. [Google Scholar] [CrossRef]
  107. De Tobel, J.; Radesh, P.; Vandermeulen, D.; Thevissen, P.W. An Automated Technique to Stage Lower Third Molar Development on Panoramic Radiographs for Age Estimation: A Pilot Study. J. Forensic Odonto-Stomatol. 2017, 35, 42–54. [Google Scholar]
  108. Merdietio Boedi, R.; Banar, N.; De Tobel, J.; Bertels, J.; Vandermeulen, D.; Thevissen, P.W. Effect of Lower Third Molar Segmentations on Automated Tooth Development Staging Using a Convolutional Neural Network. J. Forensic Sci. 2020, 65, 481–486. [Google Scholar] [CrossRef] [PubMed]
  109. Štern, D.; Payer, C.; Urschler, M. Automated Age Estimation from MRI Volumes of the Hand. Med. Image Anal. 2019, 58. [Google Scholar] [CrossRef] [PubMed]
  110. Ortega, R.F.; Irurita, J.; Campo, E.J.E.; Mesejo, P. Analysis of the Performance of Machine Learning and Deep Learning Methods for Sex Estimation of Infant Individuals from the Analysis of 2D Images of the Ilium. Int. J. Leg. Med. 2021, 2659–2666. [Google Scholar] [CrossRef] [PubMed]
  111. Liew, S.S.; Khalil-Hani, M.; Ahmad Radzi, S.; Bakhteri, R. Gender Classification: A Convolutional Neural Network Approach. Turk. J. Electr. Eng. Comput. Sci. 2016, 24, 1248–1264. [Google Scholar] [CrossRef]
  112. Kasar, M.M.; Bhattacharyya, D.; Kim, T. Face Recognition Using Neural Network: A Review. Int. J. Secur. Its Appl. 2016, 10, 81–100. [Google Scholar] [CrossRef]
  113. Pham, C.V.; Lee, S.J.; Kim, S.Y.; Lee, S.; Kim, S.H.; Kim, H.S. Age Estimation Based on 3D Post-Mortem Computed Tomography Images of Mandible and Femur Using Convolutional Neural Networks. PLoS ONE 2021, 16, e0251388. [Google Scholar] [CrossRef]
  114. Oner, Z.; Turan, M.K.; Oner, S.; Secgin, Y.; Sahin, B. Sex Estimation Using Sternum Part Lenghts by Means of Artificial Neural Networks. Forensic Sci. Int. 2019, 301, 6–11. [Google Scholar] [CrossRef] [PubMed]
  115. Farhadian, M.; Salemi, F.; Saati, S.; Nafisi, N. Dental Age Estimation Using the Pulp-to-Tooth Ratio in Canines by Neural Networks. Imaging Sci. Dent. 2019, 49, 19–26. [Google Scholar] [CrossRef]
  116. Etli, Y.; Asirdizer, M.; Hekimoglu, Y.; Keskin, S.; Yavuz, A. Sex Estimation from Sacrum and Coccyx with Discriminant Analyses and Neural Networks in an Equally Distributed Population by Age and Sex. Forensic Sci. Int. 2019, 303, 109955. [Google Scholar] [CrossRef]
  117. Chen, X.; Lian, C.; Deng, H.H.; Kuang, T.; Lin, H.Y.; Xiao, D.; Gateno, J.; Shen, D.; Xia, J.J.; Yap, P.T. Fast and Accurate Craniomaxillofacial Landmark Detection via 3D Faster R-CNN. IEEE Trans. Med. Imaging 2021. [Google Scholar] [CrossRef] [PubMed]
  118. Iyer, T.J.; Rahul, K.; Nersisson, R.; Zhuang, Z.; Joseph Raj, A.N.; Refayee, I. Machine Learning-Based Facial Beauty Prediction and Analysis of Frontal Facial Images Using Facial Landmarks and Traditional Image Descriptors. Comput. Intell. Neurosci. 2021, 2021. [Google Scholar] [CrossRef]
  119. Baccetti, T.; Franchi, L.; McNamara, J.A. The Cervical Vertebral Maturation (CVM) Method for the Assessment of Optimal Treatment Timing in Dentofacial Orthopedics. Semin. Orthod. 2005, 11, 119–129. [Google Scholar] [CrossRef]
  120. Corradi, F.; Pinchi, V.; Barsanti, I.; Garatti, S. Probabilistic Classification of Age by Third Molar Development: The Use of Soft Evidence. J. Forensic Sci. 2013, 58, 51–59. [Google Scholar] [CrossRef] [Green Version]
  121. Focardi, M.; Pinchi, V.; de Luca, F.; Norelli, G.-A. Age Estimation for Forensic Purposes in Italy: Ethical Issues. Int. J. Leg. Med. 2014, 128, 515–522. [Google Scholar] [CrossRef]
  122. Pinchi, V.; Norelli, G.A.; Caputi, F.; Fassina, G.; Pradella, F.; Vincenti, C. Dental Identification by Comparison of Antemortem and Postmortem Dental Radiographs: Influence of Operator Qualifications and Cognitive Bias. Forensic Sci. Int. 2012, 222, 252–255. [Google Scholar] [CrossRef]
  123. Pinchi, V.; Pradella, F.; Buti, J.; Baldinotti, C.; Focardi, M.; Norelli, G.A. A New Age Estimation Procedure Based on the 3D CBCT Study of the Pulp Cavity and Hard Tissues of the Teeth for Forensic Purposes: A Pilot Study. J. Forensic Leg. Med. 2015, 36, 150–157. [Google Scholar] [CrossRef]
  124. Pinchi, V.; Norelli, G.-A.; Pradella, F.; Vitale, G.; Rugo, D.; Nieri, M. Comparison of the Applicability of Four Odontological Methods for Age Estimation of the 14 Years Legal Threshold in a Sample of Italian Adolescents. J. Forensic Odonto-Stomatol. 2012, 30, 17. [Google Scholar]
  125. Pinchi, V.; de Luca, F.; Focardi, M.; Pradella, F.; Vitale, G.; Ricciardi, F.; Norelli, G.A. Combining Dental and Skeletal Evidence in Age Classification: Pilot Study in a Sample of Italian Sub-Adults. Leg. Med. 2016, 20, 75–79. [Google Scholar] [CrossRef] [Green Version]
  126. Pinchi, V.; de Luca, F.; Ricciardi, F.; Focardi, M.; Piredda, V.; Mazzeo, E.; Norelli, G.A. Skeletal Age Estimation for Forensic Purposes: A Comparison of GP, TW2 and TW3 Methods on an Italian Sample. Forensic Sci. Int. 2014, 238, 83–90. [Google Scholar] [CrossRef]
  127. Li, L.; Qin, L.; Xu, Z.; Yin, Y.; Wang, X.; Kong, B.; Bai, J.; Lu, Y.; Fang, Z.; Song, Q.; et al. Using Artificial Intelligence to Detect COVID-19 and Community-Acquired Pneumonia Based on Pulmonary CT: Evaluation of the Diagnostic Accuracy. Radiology 2020, 296, E65–E71. [Google Scholar] [CrossRef]
  128. Sessa, F.; Bertozzi, G.; Cipolloni, L.; Baldari, B.; Cantatore, S.; D’Errico, S.; di Mizio, G.; Asmundo, A.; Castorina, S.; Salerno, M.; et al. Clinical-Forensic Autopsy Findings to Defeat COVID-19 Disease: A Literature Review. J. Clin. Med. 2020, 9, 2026. [Google Scholar] [CrossRef]
  129. Shamout, F.E.; Shen, Y.; Wu, N.; Kaku, A.; Park, J.; Makino, T.; Jastrzębski, S.; Witowski, J.; Wang, D.; Zhang, B.; et al. An Artificial Intelligence System for Predicting the Deterioration of COVID-19 Patients in the Emergency Department. NPJ Digit. Med. 2021, 4, 80. [Google Scholar] [CrossRef] [PubMed]
  130. Sessa, F.; Esposito, M.; Messina, G.; Di Mizio, G.; Di Nunno, N.; Salerno, M. Sudden Death in Adults: A Practical Flow Chart for Pathologist Guidance. Healthcare 2021, 9, 870. [Google Scholar] [CrossRef]
  131. Ali, A.R.; Al-Nakib, L.H. The Value of Lateral Cephalometric Image in Sex Identification. J. Baghdad Coll. Dent. 2013, 325, 1–5. [Google Scholar] [CrossRef]
  132. Patil, K.R.; Mody, R.N. Determination of Sex by Discriminant Function Analysis and Stature by Regression Analysis: A Lateral Cephalometric Study. Forensic Sci. Int. 2005, 147, 175–180. [Google Scholar] [CrossRef] [PubMed]
  133. Krishan, K. Anthropometry in Forensic Medicine and Forensic Science-Forensic Anthropometry. Internet J. Forensic Sci. 2007, 2, 95–97. [Google Scholar]
  134. Kim, M.-J.; Liu, Y.; Oh, S.H.; Ahn, H.-W.; Kim, S.-H.; Nelson, G. Automatic Cephalometric Landmark Identification System Based on the Multi-Stage Convolutional Neural Networks with CBCT Combination Images. Sensors 2021, 21, 505. [Google Scholar] [CrossRef]
  135. Leonardi, R.; lo Giudice, A.; Isola, G.; Spampinato, C. Deep Learning and Computer Vision: Two Promising Pillars, Powering the Future in Orthodontics. Semin. Orthod. 2021, 27, 62–68. [Google Scholar] [CrossRef]
  136. Bouletreau, P.; Makaremi, M.; Ibrahim, B.; Louvrier, A.; Sigaux, N. Artificial Intelligence: Applications in Orthognathic Surgery. J. Stomatol. Oral Maxillofac. Surg. 2019, 120, 347–354. [Google Scholar] [CrossRef]
  137. Hung, K.; Montalvao, C.; Tanaka, R.; Kawai, T.; Bornstein, M.M. The Use and Performance of Artificial Intelligence Applications in Dental and Maxillofacial Radiology: A Systematic Review. Dentomaxillofac. Radiol. 2019, 49. [Google Scholar] [CrossRef] [PubMed]
  138. Schwendicke, F.; Chaurasia, A.; Arsiwala, L.; Lee, J.-H.; Elhennawy, K.; Jost-Brinkmann, P.-G.; Demarco, F.; Krois, J. Deep Learning for Cephalometric Landmark Detection: Systematic Review and Meta-Analysis. Clin. Oral Investig. 2021, 25, 4299–4309. [Google Scholar] [CrossRef]
  139. Koberová, K.; Thurzo, A.; Dianišková, S. Evaluation of Prevalence of Facial Asymmetry in Population According to the Analysis of 3D Face-Scans. Lek. Obz. 2015, 64, 101–104. [Google Scholar]
  140. Tanikawa, C.; Yamashiro, T. Development of Novel Artificial Intelligence Systems to Predict Facial Morphology after Orthognathic Surgery and Orthodontic Treatment in Japanese Patients. Sci. Rep. 2021, 11, 15853. [Google Scholar] [CrossRef] [PubMed]
  141. Cao, K.; Choi, K.; Jung, H.; Duan, L. Deep Learning for Facial Beauty Prediction. Information 2020, 11, 391. [Google Scholar] [CrossRef]
  142. Dang, H.; Liu, F.; Stehouwer, J.; Liu, X.; Jain, A.K. On the Detection of Digital Face Manipulation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 5781–5790. [Google Scholar]
  143. Kittler, J.; Koppen, P.; Kopp, P.; Huber, P.; Ratsch, M. Conformal Mapping of a 3d Face Representation onto a 2D Image for CNN Based Face Recognition. In Proceedings of the 2018 International Conference on Biometrics (ICB), Gold Coast, Australia, 20–23 February 2018; pp. 124–131. [Google Scholar] [CrossRef]
  144. Turchetta, B.J.; Fishman, L.S.; Subtelny, J.D. Facial Growth Prediction: A Comparison of Methodologies. Am. J. Orthod. Dentofac. Orthop. 2007, 132, 439–449. [Google Scholar] [CrossRef] [PubMed]
  145. Hirschfeld, W.J.; Moyers, R.E. Prediction of Craniofacial Growth: The State of the Art. Am. J. Orthod. 1971, 60, 435–444. [Google Scholar] [CrossRef] [Green Version]
  146. Kaźmierczak, S.; Juszka, Z.; Fudalej, P.; Mańdziuk, J. Prediction of the Facial Growth Direction with Machine Learning Methods. arXiv 2021, arXiv:2106.10464. [Google Scholar]
  147. Yang, C.; Rangarajan, A.; Ranka, S. Visual Explanations From Deep 3D Convolutional Neural Networks for Alzheimer’s Disease Classification. AMIA Annu. Symp. Proc. 2018, 2018, 1571. [Google Scholar] [PubMed]
  148. Tan, M.; Le, Q. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  149. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR, Honolulu, HI, USA, 21–26 July 2016; pp. 2261–2269. [Google Scholar]
  150. de Jong, M.A.; Gül, A.; de Gijt, J.P.; Koudstaal, M.J.; Kayser, M.; Wolvius, E.B.; Böhringer, S. Automated Human Skull Landmarking with 2D Gabor Wavelets. Phys. Med. Biol. 2018, 63. [Google Scholar] [CrossRef]
  151. Arık, S.; Ibragimov, B.; Xing, L. Fully Automated Quantitative Cephalometry Using Convolutional Neural Networks. J. Med Imaging 2017, 4, 014501. [Google Scholar] [CrossRef] [PubMed]
  152. Liu, Q.; Deng, H.; Lian, C.; Chen, X.; Xiao, D.; Ma, L.; Chen, X.; Kuang, T.; Gateno, J.; Yap, P.-T.; et al. SkullEngine: A Multi-Stage CNN Framework for Collaborative CBCT Image Segmentation and Landmark Detection. In International Workshop on Machine Learning in Medical Imaging; Springer: Cham, The Netherlands, 2021; pp. 606–614. [Google Scholar] [CrossRef]
  153. Sessa, F.; Salerno, M.; Pomara, C.; Sessa, F.; Salerno, M.; Pomara, C. Autopsy Tool in Unknown Diseases: The Experience with Coronaviruses (SARS-CoV, MERS-CoV, SARS-CoV-2). Medicina 2021, 57, 309. [Google Scholar] [CrossRef] [PubMed]
  154. Wang, H.; Minnema, J.; Batenburg, K.J.; Forouzanfar, T.; Hu, F.J.; Wu, G. Multiclass CBCT Image Segmentation for Orthodontics with Deep Learning. J. Dent. Res. 2021, 100, 943–949. [Google Scholar] [CrossRef]
  155. Jang, T.J.; Kim, K.C.; Cho, H.C.; Seo, J.K. A Fully Automated Method for 3D Individual Tooth Identification and Segmentation in Dental CBCT. IEEE Trans. Pattern Anal. Mach. Intell. 2021. [Google Scholar] [CrossRef]
  156. Qiu, B.; van der Wel, H.; Kraeima, J.; Glas, H.H.; Guo, J.; Borra, R.J.H.; Witjes, M.J.H.; van Ooijen, P.M.A. Robust and Accurate Mandible Segmentation on Dental CBCT Scans Affected by Metal Artifacts Using a Prior Shape Model. J. Pers. Med. 2021, 11, 364. [Google Scholar] [CrossRef] [PubMed]
  157. Qiu, B.; van der Wel, H.; Kraeima, J.; Glas, H.H.; Guo, J.; Borra, R.J.H.; Witjes, M.J.H.; van Ooijen, P.M.A. Mandible Segmentation of Dental CBCT Scans Affected by Metal Artifacts Using Coarse-to-Fine Learning Model. J. Pers. Med. 2021, 11, 560. [Google Scholar] [CrossRef] [PubMed]
  158. Funatsu, M.; Sato, K.; Mitani, H. Effects of Growth Hormone on Craniofacial GrowthDuration of Replacement Therapy. Angle Orthod. 2006, 76, 970–977. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  159. Cantu, G.; Buschang, P.H.; Gonzalez, J.L. Differential Growth and Maturation in Idiopathic Growth-Hormone-Deficient Children. Eur. J. Orthod. 1997, 19, 131–139. [Google Scholar] [CrossRef] [Green Version]
  160. Konfino, R.; Pertzelan, A.; Laron, Z. Cephalometric Measurements of Familial Dwarfism and High Plasma Immunoreactive Growth Hormone. Am. J. Orthod. 1975, 68, 196–201. [Google Scholar] [CrossRef]
  161. Grippaudo, C.; Paolantonio, E.G.; Antonini, G.; Saulle, R.; la Torre, G.; Deli, R. Association between Oral Habits, Mouth breathing and Malocclusion. Acta Otorhinolaryngol. Italy 2016, 36, 386. [Google Scholar] [CrossRef]
  162. Ranggang, B.M.; Armedina, R.N. CrossMark Comparison of Parents Knowledge of Bad Habits and the Severity Maloclusion of Children in Schools with Different Social Levels. J. Dentomaxillofacial Sci. 2020, 5, 48–51. [Google Scholar] [CrossRef]
  163. Arcari, S.R.; Biagi, R. Oral Habits and Induced Occlusal-Skeletal Disarmonies. J. Plast. Dermatol. 2015, 11, 1. [Google Scholar]
  164. Hammond, P.; Hannes, F.; Suttie, M.; Devriendt, K.; Vermeesch, J.R.; Faravelli, F.; Forzano, F.; Parekh, S.; Williams, S.; McMullan, D.; et al. Fine-Grained Facial Phenotype–Genotype Analysis in Wolf–Hirschhorn Syndrome. Eur. J. Hum. Genet. 2011, 20, 33–40. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  165. Hartsfield, J.K. The Importance of Analyzing Specific Genetic Factors in Facial Growth for Diagnosis and Treatment Planning. Available online: https://www.researchgate.net/profile/James-Hartsfield/publication/235916088_The_Importance_of_Analyzing_Specific_Genetic_Factors_in_Facial_Growth_for_Diagnosis_and_Treatment_Planning/links/00b7d5141dd8fd2d82000000/The-Importance-of-Analyzing-Specific-Genetic-Factors-in-Facial-Growth-for-Diagnosis-and-Treatment-Planning.pdf (accessed on 1 November 2021).
  166. Hartsfield, J.K.; Morford, L.A.; Otero, L.M.; Fardo, D.W. Genetics and Non-Syndromic Facial Growth. J. Pediatr. Genet. 2015, 2, 9–20. [Google Scholar] [CrossRef] [Green Version]
Figure 1. (a) Recognition of objects. Try, using your imagination, to recognize the objects on the three blurred variants of the same anatomical slice. Convolutional Neural Networks (CNNs) work similar to our visual brain when trying to recognize these objects. (b) Our recognition of objects on the picture is significantly improved when more layers—slices are added thus providing further context with the 3rd dimension. In the top row is recognizable intersection of the mandible and vertebra and on the lower row is recognizable slice of the face. 3D CNN recognition is similarly improved with providing context of depth.
Figure 1. (a) Recognition of objects. Try, using your imagination, to recognize the objects on the three blurred variants of the same anatomical slice. Convolutional Neural Networks (CNNs) work similar to our visual brain when trying to recognize these objects. (b) Our recognition of objects on the picture is significantly improved when more layers—slices are added thus providing further context with the 3rd dimension. In the top row is recognizable intersection of the mandible and vertebra and on the lower row is recognizable slice of the face. 3D CNN recognition is similarly improved with providing context of depth.
Healthcare 09 01545 g001
Figure 2. The comparison of 2D CNN (above) and 3D CNN (below). 3D CNN works with 3rd dimension and can reconstruct shapes from the CBCT 2D slides. The sequence of 2D pictures where the 3rd dimension is time, we speak of a common video sequence that can be a subject of 3D CNN analysis too.
Figure 2. The comparison of 2D CNN (above) and 3D CNN (below). 3D CNN works with 3rd dimension and can reconstruct shapes from the CBCT 2D slides. The sequence of 2D pictures where the 3rd dimension is time, we speak of a common video sequence that can be a subject of 3D CNN analysis too.
Healthcare 09 01545 g002
Figure 3. The example of the process of defining ranges for particular visualized tissues is called “thresholding”.
Figure 3. The example of the process of defining ranges for particular visualized tissues is called “thresholding”.
Healthcare 09 01545 g003
Figure 4. The examples of the segmentation process on the CBCT data based on defining ranges for particular tissues thus defining 3D structures such as airway, nerve canal, face surface or bone structures.
Figure 4. The examples of the segmentation process on the CBCT data based on defining ranges for particular tissues thus defining 3D structures such as airway, nerve canal, face surface or bone structures.
Healthcare 09 01545 g004
Figure 5. The example of 3D data augmentation in a sense of mapping another 3D scans on the segmented structures. Facial 3D scan with texture mapped on the segmented face surface from CBCT and intraoral scan of teeth and gums mapped on tooth surfaces from the CBCT. Finally merged into complex set of 3D models. Training of 3D CNN with such a complex 3D virtualized model has never been performed before and is worth a consideration.
Figure 5. The example of 3D data augmentation in a sense of mapping another 3D scans on the segmented structures. Facial 3D scan with texture mapped on the segmented face surface from CBCT and intraoral scan of teeth and gums mapped on tooth surfaces from the CBCT. Finally merged into complex set of 3D models. Training of 3D CNN with such a complex 3D virtualized model has never been performed before and is worth a consideration.
Healthcare 09 01545 g005
Figure 6. Example of 3D cephalometric analysis where orthodontist identifies more than 50 points and the hard- and soft-tissues analyzed. Humans chose these points as the most reproducible on X-ray. These might not be ideal representatives of head and neck structures linked with biological ageing or sexual dimorphism.
Figure 6. Example of 3D cephalometric analysis where orthodontist identifies more than 50 points and the hard- and soft-tissues analyzed. Humans chose these points as the most reproducible on X-ray. These might not be ideal representatives of head and neck structures linked with biological ageing or sexual dimorphism.
Healthcare 09 01545 g006
Figure 7. Example of CNN use of the FaceApp AI application to render the face mapped on CBCT to look younger or older. The algorithms changed just the texture and not the 3D mask, however this is probably only a matter of time. 2D face morphing based on AI or face swapping in popular videos are available and popular already a couple of years. Original face is the 2nd one.
Figure 7. Example of CNN use of the FaceApp AI application to render the face mapped on CBCT to look younger or older. The algorithms changed just the texture and not the 3D mask, however this is probably only a matter of time. 2D face morphing based on AI or face swapping in popular videos are available and popular already a couple of years. Original face is the 2nd one.
Healthcare 09 01545 g007
Figure 8. For medical images, there are various preprocessing steps that are recommended.
Figure 8. For medical images, there are various preprocessing steps that are recommended.
Healthcare 09 01545 g008
Figure 9. Pipeline from pre-processed CBCT scans to prediction on 3D CNN.
Figure 9. Pipeline from pre-processed CBCT scans to prediction on 3D CNN.
Healthcare 09 01545 g009
Figure 10. Confusion matrix for 2 classes image classification—Cat and Non-Cat. Each row corresponds to the predicted class from neural network output. In case of a class Cat 90 samples were correctly classified as Cat and 60 samples were incorrectly classified as Non-Cat.
Figure 10. Confusion matrix for 2 classes image classification—Cat and Non-Cat. Each row corresponds to the predicted class from neural network output. In case of a class Cat 90 samples were correctly classified as Cat and 60 samples were incorrectly classified as Non-Cat.
Healthcare 09 01545 g010
Table 1. Full-head CBCT scans were mate with i-CAT™ FLX V17 with these settings.
Table 1. Full-head CBCT scans were mate with i-CAT™ FLX V17 with these settings.
ParameterSetting
Sensor Type Amorphous Silicon Flat Panel Sensor with Csl Scintillator
Grayscale Resolution16-bit
Voxel Size0.3 mm,
CollimationElectronically controlled fully adjustable collimation
Scan Time17.8 s
Exposure TypePulsed
Field-of-View23 cm × 17 cm
Reconstruction ShapeCylinder
Reconstruction TimeLess than 30 s
OutputDICOM
Patient PositionSeated
Table 2. Guide of recommended designs for 3D CNN implementations in various forensic topics.
Table 2. Guide of recommended designs for 3D CNN implementations in various forensic topics.
Area of Forensic ResearchProposed MethodMetrics
Biological age determinationRegression model by 3D deep CNNMAE, MSE
Sex determinationDeep 3D CNN—conv.layers and outputs class probabilities for both targetsCM such as precision, recall and F1 score
3D cephalometric analysisObject detection model on 3D CNN that auto.estimates cephalom.measurements MAE, MSE
Face prediction from skullmodel on Generative Adversarial Network that synthesize soft/hard tissuesslice-wise Frechet Inception Distance
Facial growth predictionBased on methods stated above 1another 1
1 Method and metrics are not proposed from the current state of knowledge for Facial growth prediction and need further consideration upon clinical experience from 3D CNN applications.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Thurzo, A.; Kosnáčová, H.S.; Kurilová, V.; Kosmeľ, S.; Beňuš, R.; Moravanský, N.; Kováč, P.; Kuracinová, K.M.; Palkovič, M.; Varga, I. Use of Advanced Artificial Intelligence in Forensic Medicine, Forensic Anthropology and Clinical Anatomy. Healthcare 2021, 9, 1545. https://0-doi-org.brum.beds.ac.uk/10.3390/healthcare9111545

AMA Style

Thurzo A, Kosnáčová HS, Kurilová V, Kosmeľ S, Beňuš R, Moravanský N, Kováč P, Kuracinová KM, Palkovič M, Varga I. Use of Advanced Artificial Intelligence in Forensic Medicine, Forensic Anthropology and Clinical Anatomy. Healthcare. 2021; 9(11):1545. https://0-doi-org.brum.beds.ac.uk/10.3390/healthcare9111545

Chicago/Turabian Style

Thurzo, Andrej, Helena Svobodová Kosnáčová, Veronika Kurilová, Silvester Kosmeľ, Radoslav Beňuš, Norbert Moravanský, Peter Kováč, Kristína Mikuš Kuracinová, Michal Palkovič, and Ivan Varga. 2021. "Use of Advanced Artificial Intelligence in Forensic Medicine, Forensic Anthropology and Clinical Anatomy" Healthcare 9, no. 11: 1545. https://0-doi-org.brum.beds.ac.uk/10.3390/healthcare9111545

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop