Skip to main content

Beyond Human Perception: Sexual Dimorphism in Hand and Wrist Radiographs Is Discernible by a Deep Learning Model


Despite the well-established impact of sex and sex hormones on bone structure and density, there has been limited description of sexual dimorphism in the hand and wrist in the literature. We developed a deep convolutional neural network (CNN) model to predict sex based on hand radiographs of children and adults aged between 5 and 70 years. Of the 1531 radiographs tested, the algorithm predicted sex correctly in 95.9% (κ = 0.92) of the cases. Two human radiologists achieved 58% (κ = 0.15) and 46% (κ = − 0.07) accuracy. The class activation maps (CAM) showed that the model mostly focused on the 2nd and 3rd metacarpal base or thumb sesamoid in women, and distal radioulnar joint, distal radial physis and epiphysis, or 3rd metacarpophalangeal joint in men. The radiologists reviewed 70 cases (35 females and 35 males) labeled with sex along with heat maps generated by CAM, but they could not find any patterns that distinguish the two sexes. A small sample of patients (n = 44) with sexual developmental disorders or transgender identity was selected for a preliminary exploration of application of the model. The model prediction agreed with phenotypic sex in only 77.8% (κ = 0.54) of these cases. To the best of our knowledge, this is the first study that demonstrated a machine learning model to perform a task in which human experts could not fulfill.


A picture may be worth a thousand words, but describing it should be much more concise. When radiologists interpret medical images, they look for specific features associated with disease and may overlook certain features, either intentionally or unconsciously, that are not obviously indicative of pathology. This practice leads to more efficient image interpretation; however, there are likely to be additional clinically relevant imaging features that are beyond current human radiologist visual discernment.

Computational models that used machine learning, especially deep learning, has shown remarkable performance in medical image analysis over the past few years in a variety of tasks including classifying skin cancer [1] and predicting cardiovascular risk [2]. However, most of the studies that leverage deep learning for medical image analysis focus on a replicating a task already performed by humans. In addition to mimicking humans, machine learning also offers the potential of identifying significant imaging features that are beyond a radiologist’s visual search pattern, and perhaps enhance the diagnostic utility of medical images.

Hand radiographs are widely used for assessment of skeletal age because the examination is straightforward to perform with minimal radiation exposure and include multiple bones in a single view. Bone age assessment (BAA) is used to find abnormalities in skeletal development, monitoring growth hormone therapy, diagnosis of endocrine disorders, predicting adult height, and planning surgery of the long bones or vertebral column [3]. Many children with precocious puberty, absent secondary sexual characteristics, or short stature undergo BAA, as bone maturation is closely related to sexual development. In this subset of patients for whom hormone replacement treatment, future fertility, and psychosocial support are important issues, information about sexual development will add important value to this simple imaging process.

To the best of our knowledge, there has not been reported tools or parameters by which sex can be reliably identified from hand radiographs in children. Because a patient’s sex is provided for radiologists to interpret bone age radiographs, there is no established set of sex-specific radiographic features. In this study, we developed a deep learning system that analyzes hand radiographs of individuals of 5 to 70 years of age and predicts sex. Two radiologists reviewed the system output to describe the patterns used to distinguish sexes. To find at which age it becomes evident, we analyzed the system output stratified by age. In addition, we tested the system on separately selected radiographs to explore how the deep learning–based sexual dimorphism is expressed in people with special conditions.

Materials and Methods

Study Subjects

The institutional review board of the Massachusetts General Hospital approved the study protocol and waived the requirement for informed consent based on the substantial difficulty to acquire consent and the minimal risk to the study subjects. All study data were retrieved from an institutional registry which only includes data from patients who agreed with use of their data for research purposes. Previously identified pediatric left-hand radiographs of 4278 females and 4047 males were first included in the current study data [4]. Additionally, left-hand radiographs of male and female adults of age 19 or older were collected from the institutional research database. Among all left-hand radiographs in the database, we found 2282 cases that are reported as normal in radiology reports. In total, we compiled a dataset of 10,607 (5459 females and 5148 males) radiographs of the left hand and wrist from patients aged 5 years to 70 years.

Next, we made a list of patients who ever had assigned with any of the following international classification of diseases (ICD) codes: androgen insensitivity syndrome (ICD-9 259.5, ICD-10 E34.5), congenital adrenal hyperplasia (ICD-9 255.2, ICD-10 E25), chromosomal anomalies (ICD-9 758.6, 758.7, 758.8; ICD-10 Q96, Q97, Q98, Q99), and gender identity disorders (ICD-9 302.85, ICD-10 F64). The list contained 2189 patients, from which we found 444 left-hand radiographs. From the initially compiled dataset, we excluded 289 radiographs included in this list to avoid potential impacts of these conditions on sexual dimorphism in hand radiographs. The remaining 10,318 (5305 females and 5013 males) radiographs were used for training (7251), validation (1536), and testing (1531). For age-stratified analysis, we categorized the radiographs by age: by 1 year from 5 to 19 years old, 20 to 29, 30 to 39, 40 to 49, and 50 and older. The age and sex distribution of the subjects included in the final dataset is described in Fig. 1. The age ranged from 5 to 70 years with the median of 12. Lastly, we found 155 patients in the ICD-based list whose left-hand radiographs were available. Among these 155 patients, chart review by a physician board certified in internal medicine identified 44 patients who are confirmed to have the conditions.

Fig. 1
figure 1

Age and sex distribution of the study subjects included in the final dataset. The numbers on top of each bar indicates the number of radiographs in each age category. The number on top of the red portion of each bar indicates the percentage of females in each age category. The total numbers of each sex are shown at the top-right corner

Data Preprocessing

The radiographs varied considerably in intensity, contrast, image resolution, and existence of artifacts. To allow deep learning models to learn salient features, a preprocessing pipeline (Fig. 2) was implemented based on the previously developed module and modified by replacing the conventional detection CNN with a newly trained segmentation CNN [4]. The network architecture (FCN-2S) that performed best to segment skeletal muscle regions at the level of third lumbar (L3) vertebral body was trained and validated on the preliminarily compiled datasets for segmenting regions of the hand and wrist [4, 5]. The new segmentation CNN achieved the mean of intersection over union of 0.95. The preprocessing module first normalizes radiographs to have a uniform size (512 × 512 pixels) with preserving their aspect ratios, then segments a region of the hand and wrist and removes extraneous objects such as annotation markers and collimation. Subsequently, contrast limited adaptive histogram equalization (CLAHE) with default settings [6] was applied to the segmented and normalized images for contrast enhancement.

Fig. 2
figure 2

Data preprocessing pipeline. An overview of data preprocessing engine that normalizes radiographs to have a uniform size of 512 × 512 pixels, segments a region of the hand and wrist using a segmentation CNN, and enhances image contrast using contrast limited adaptive histogram equalization (CLAHE)

Model Development and Network Training

The standardized images that went through the preprocessing engine were passed to a deep CNN (VGG16) [7] for bone sex classification. VGG16 is one of the CNNs validated in ImageNet Large Scale Visual Recognition Competition (ILSVRC) [8] for its decent classification performance. The CNN was pretrained on ImageNet [9] (a 1.28-M training images with 1000 class labels), modified by replacing the fully connected layers with a global average pooling [10], a fully connected, and a sigmoid layer, and then fine-tuned on our train dataset. We trained the classification and segmentation models for 100 epochs using a mini-batch stochastic gradient descent (SGD) with 0.9 Nesterov [11] momentum and a batch size of 64. A base learning rate of 10−3 and a weight decay of 5 × 10−5 were used for training the classification CNN, and a base learning rate of 10−10 and a weight decay of 10−12 were used for training the segmentation CNN. The base learning rates were decreased by a factor of 10 every 33 epochs for stable convergence of training loss function. The best models were selected based on validation losses. Keras (version 2.1.1) with a Tensorflow backend (version 1.3.0) were used as deep learning framework to develop models, and an NVIDIA Devbox (Santa Clara, CA) equipped with four TITAN X GPUs with 12 GB of memory per GPU was utilized to perform all experiments.


We investigated the intermediate features learned by the CNN using t distributed Stochastic Neighbor Embedding (t-SNE) [12]. Each test image was presented to the trained CNN to obtain the corresponding 512-dimensional features from the last convolutional layer. The high dimensional representation was converted into the 2-dimensional data, and the lower dimensional features were then visualized as shown in Fig. 3.

Heat Map and Atlas Generation

Two visualization techniques were utilized to determine salient features that the model used for bone sex classification. First, class activation mapping [13] (CAM) technique was applied to the trained models to generate attention maps that highlight significant pixels for model predictions. Second, a set of training images most relevant to a given test case were retrieved from an atlas that consists of visual depictions of important features of each bone sex. The atlas was created by feeding all training images through the trained classification model, tracking all activation values of feature maps at the last convolutional layer, and keeping training images that caused highest activations on each feature map. During inference, gradients of a predicted output with regard to individual feature maps at the last convolutional layer were calculated via backpropagation, and the associated training images and attention maps with highest gradient were retrieved as the prediction basis of the model for a given case.

Radiologist Evaluation

To test if human radiologists can see the sexual difference in hand radiographs, we evaluated radiologist performance on this task. Radiologist A, a diagnostic radiology board-certified physician with 15-year experience, and radiologist B in the third year of radiology residency, were blind-tested to predict sex from randomly selected 50 (24 females and 26 males) left-hand radiographs, for which the system predicted sex correctly with 100% confidence. DICOM files from which all patient information was removed were shown to the radiologists. The radiologists independently recorded their sex prediction for each case in a spreadsheet, and the result was compared with the phenotypic sex found in the electronic health record (EHR). After the feature description and attention localization process, the radiologists were tested again on the same 50 cases to assess if they learned any patterns during the process.

Feature Description and Attention Localization

To find the features that distinguish the two sexes, the radiologists independently reviewed randomly selected radiographs with corresponding heat map and atlas, for which the model predicted sex correctly with 100% confidence. Radiologist A reviewed 70 cases (35 females and 35 males) and radiologist B reviewed 120 cases (60 females and 60 males) including the 70 cases reviewed by radiologist A. The cases used for the blind test were excluded when selecting these review cases. After the independent review, the two radiologists discussed their findings with each other. In addition, to localize the model attention, they checked the anatomical location of the heat maps of 50 cases using a customized spreadsheet. The localization was independently annotated, and radiologists were allowed to mark as many locations as they see. We counted the frequency of each location marked by either of the radiologists, separately for females and males. If two radiologists marked the same location in a case, the location was counted twice.

Statistical Analysis

Stata version 14.2 (StataCorp, College Station, Texas, USA) was used for statistical analysis. Fisher’s exact test was used for the comparison of accuracy among age groups. The accuracy of the model and the radiologists were presented both as percent accuracy and Cohen’s kappa coefficient (κ) [14]. Cohen’s kappa was calculated using the formula κ = ( po − pe)/(1 − pe), where po = observed agreement among raters and pe = expected agreement = \( \frac{1}{N^2}{\sum}_k{n}_{k1}{n}_{k2} \) for categories k, number of items N, and nki the number of times rater i predicted category k.


Model Accuracy

A deep learning model was developed and evaluated on the 1531 radiographs in the test dataset. The percent accuracies stratified by age groups are summarized in Fig. 3. The overall accuracy of the model on the test dataset was 95.9% and the κ = 0.918 (95% CI 0.898–0.938). The accuracies were not significantly different across all age groups (p = 0.232) as well as between females and males (p = 0.946).

Fig. 3
figure 3

Age-stratified test accuracies. Test accuracies were shown as percent accuracy and stratified by age. The black dotted line indicates the overall accuracy across all 1531 radiographs in the test dataset

Radiologist Performance

On the initial test, radiologist A showed 58% accuracy (κ = 0.152, 95% CI = − 0.117–0.421). Radiologist B showed 46% accuracy (κ = − 0.077, 95% CI = − 0.351–0.198). On the second test after the feature description and attention localization process, the accuracy of radiologist A remained at 58% (κ = 0.160, 95% CI = − 0.113–0.433) and that of radiologist B was 42% (κ = − 0.164, 95% CI = − 0.437–0.109). The intra-reader agreement was 64% for radiologist A (κ = 0.280, 95% CI = 0.022–0.538) and 76% for radiologist B (κ = 0.523, 95% CI = 0.291–0.756). The results of radiologist performance assessment are presented in Table 1.

Table 1 Radiologist accuracies to predict sex from 50 randomly selected hand radiographs. Accuracies are presented as percent accuracy and Cohen’s kappa between radiologist prediction and phenotypic sex found in electronic health record. Intra-reader agreement measures agreement between the two test results by the same radiologist. Inter-reader agreement measures agreement between the two radiologists during the same test

Feature Description and Attention Localization

The two radiologists did not find any consistent patterns that distinguish males and females in the hand radiographs. For the 50 cases for which the attention maps were annotated, each radiologist marked from 1 to 6 locations in each case. Collectively in all cases, there were 159 annotation counts for the 25 females and 103 annotation counts for the 25 males. In females, the second carpometacarpal joint was the most frequently localized (30 counts) followed by the third carpometacarpal joint (24 counts) and the thumb sesamoid (18 counts). In males, the most frequently localized region was the distal radioulnar joint (31 counts), the radial physis and epiphysis (13 counts), and the third metacarpophalangeal joint (9 counts). Figure 4 shows the representative images with overlaying heat maps in males and females as well as a t-SNE visualization.

Fig. 4
figure 4

t-SNE visualization of the representations from the last convolutional layer of the model for bone sex classification. Here, we show how the algorithm clusters males and females. Radiographs with attention maps are linked to the corresponding points

Preliminary Test on Selected Radiographs

Among the 44 patients selected by ICD code screening and chart review, 17 patients had congenital adrenal hyperplasia (9 females and 8 males), 6 females had Turner’s syndrome, 5 males had Klinefelter’s syndrome, 1 female had complete androgen insensitivity syndrome (CAIS) with 46, XY, 4 males had mixed gonadal dysgenesis, 8 individuals had female-to-male (FtM) transgender identity, and 3 individuals had male-to-female (MtF) identity. The overall accuracy of the model on the radiographs from these individuals was 77.8% (κ = 0.538, 95% CI = 0.288–0.788). Table 2 shows the agreement between model prediction and phenotypic sex.

Table 2 Agreement between algorithm prediction and phenotypic sex in a cohort with selected conditions. The ground truth for transgender individuals was set as the sex assigned at birth. Numbers in the parenthesis denotes (number of correctly predicted case)/(number of total case)


In this work, we developed an algorithm that accurately and reproducibly (κ > 0.9) identifies sex from hand radiographs and assessed the radiologists’ ability to do the same. Although it was generally agreed among radiologists that sex cannot be reliably determined by visual inspection of hand radiographs, we formally demonstrated this by showing very low agreement (κ < 0.2) between radiologist prediction and phenotypic sex even after a review session. The low inter-reader agreement (κ < 0.1) also indicates the near-randomness of sex prediction by radiologists. To the best of our knowledge, no previous research has shown that human radiologists cannot attain the performance of a machine learning algorithm for identifying a feature in radiographs. The results of our study show that radiographs contain more information than that is currently being utilized for clinical interpretation.

Previous studies about skeletal sexual dimorphism have shown differences between men and women. Men have bigger and stronger bones compared to women, and the difference is established mostly during puberty [15]. Based on such difference, automated sex estimation methods in adults have been developed using computed tomography (CT) scans and 3D imaging of the pelvis [16] and skull [17]. However, these methods were developed primarily for forensic anthropological analysis and use CT scans, limiting the more widespread use due to high radiation and cost. In hand and wrist, the most widely known feature is the digit ratio (2D:4D) that is frequently used as a biomarker for perinatal sex hormone exposure. However, despite numerous studies that investigated the subject, the association between digit ratio and perinatal sex hormone is still not confirmed [18]. In addition, there is substantial overlap between the distributions of digit ratio in males and females, thus any inference based on an individual’s digit ratio would be inaccurate [19, 20]. So far, the most prominent sexual difference in the skeletal structure of the hand and wrist is the size and volume of the bones [21]. Using this difference, a study that examined digital hand radiographs to construct a statistical classification model achieved an accuracy of 91% [22]. However, this study only included subjects that are over 18 years of age and requires measurement of 8 variables in the hand and wrist bones. This method not only requires the tedious process of measuring multiple points on radiographs but also is not applicable for children. We believe that our model is the first to classify sex using hand radiographs of all ages, regardless of the method.

Recent work that used deep learning as the primary method also showed the potential for detecting features that have not been established to be extracted from various medical images. In a study that used retinal fundus photographs to predict cardiovascular risk factors [2], the model accurately predicted gender, which has not been considered to be identifiable from retinal images. In another study, researchers showed that a machine learning approach can reliably predict cell nuclei, cell viability, cell type, and subcellular process type from transmitted light images [23]. These features require fluorescent labels for human scientists to detect. Although these studies did not investigate the capability of human experts to perform the same task, they also suggest that deep learning algorithms can be developed to find features that are not seen by humans.

In addition to showing the capability of seeing patterns not perceived by human experts, our model proposes a new indicator of skeletal gender, i.e., bone sex. It can be used to make more robust models for bone age assessment and other tasks that use sex as a key factor. Using this model, we can explore the association of bone sex and various clinical outcomes such as bone maturation patterns or development and prognosis of musculoskeletal diseases. This extends the utility of bone age radiographs beyond chronological age determination to potential assessment of sexual development or sex-determining hormone exposure. The persistent use of digit ratios to evaluate cancer risks [24, 25], development of mental disorders [26, 27], and sporting performance [28], despite its questionable accuracy for these tasks, underscores the need for such a biomarker. With further research that clarifies the factors associated with bone sex, especially in patients whose bone sex is discordant with their phenotypic sex, it could be used to detect early signs of certain conditions.

To explore the potential for the clinical implication of bone sex, we tested the model on a selected group of patients with a condition that could be associated with abnormal sexual development or disrupted sex hormone exposure. Numerous studies observed significant association between digit ratio with sex hormone levels, congenital adrenal hyperplasia, homosexuality [29], or transgender identity [30]. In addition, it is widely known that sex chromosome abnormality and congenital adrenal hyperplasia affect bone metabolism [31,32,33]. Although we could not draw a conclusion from this data because of the small sample size and possible selection bias, the lower accuracy compared to that in the test cohort implies the possibility that the discrepancy between model prediction and phenotypic sex reflects the altered skeletal development associated with these conditions. Further research in a well-defined cohort will provide more insight into the clinical use of bone sex.

One limitation of this study is that we only included left-hand radiographs to develop and validate the model. Although this study was focused more on the ability of a deep learning model to discover a pattern that has previously not been described, only including left-hand images could limit the use of the model in the clinical practice and other research studies. While it is the standard procedure to use the left hand for bone age assessment [34, 35], some studies suggest that sex differences in digit ratio are more profound in the right hand [36]. Further studies that include the right hand for testing and development will render a more comprehensive model to be clinically used.

An additional limitation is our inability to identify radiographic features that are sex-specific, despite the availability of CNN heat maps and atlas. This is a universal issue with deep learning as the inner-workings of the resulting algorithm are not completely understood. We used the heat maps and the atlas to assist the radiologists in identifying distinguishing features, to no avail. We believe this is an important result of our work that highlights the power of artificial intelligence to go beyond the limits of human visual perception. Further work is needed to identify sex-specific radiographic features and teach them to human users. To use the additional information obtained by using this innovative technology, clinicians must be cautious and understand how deep learning algorithms work.


We developed a deep learning model that distinguishes males from females based on hand and wrist radiographs, a task that human radiologists failed to reproduce. The current study shows that deep learning can be used to identify patterns that are beyond human perception.


  1. Esteva A, Kuprel B, Novoa RA, Ko J, Swetter SM, Blau HM, Thrun S: Dermatologist-level classification of skin cancer with deep neural networks. Nature 542:115–118, 2017

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  2. Poplin R, Varadarajan AV, Blumer K, Liu Y, McConnell MV, Corrado GS, Peng L, Webster DR: Prediction of cardiovascular risk factors from retinal fundus photographs via deep learning. Nat Biomed Eng 2:158–164, 2018

    Article  PubMed  Google Scholar 

  3. Gertych A, Zhang A, Sayre J, Pospiech-Kurkowska S, Huang H: Bone age assessment of children using a digital hand atlas. Comput Med Imaging Graph 31:322–331, 2007

    Article  PubMed  PubMed Central  Google Scholar 

  4. Lee H, Tajmir S, Lee J, Zissen M, Yeshiwas BA, Alkasab TK, Choy G, Do S: Fully automated deep learning system for bone age assessment. J Digit Imaging 30(4):427–441, 2017

    Article  PubMed  PubMed Central  Google Scholar 

  5. Lee H, Troschel FM, Tajmir S, Fuchs G, Mario J, Fintelmann FJ, Do S: Pixel-level deep segmentation: artificial intelligence quantifies muscle on computed tomography for body morphometric analysis. J Digit Imaging 30(4):487–98, 2017

    Article  PubMed  PubMed Central  Google Scholar 

  6. Zuiderveld K: Contrast limited adaptive histogram equalization. San Diego: Academic Press Professional, Inc., 1994

    Book  Google Scholar 

  7. Simonyan K, Zisserman A: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:14091556, 2014. Available at

  8. Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M, Berg AC, Fei-Fei L: Imagenet large scale visual recognition challenge. International Journal of Computer Vision 115:211–252, 2015

    Article  Google Scholar 

  9. Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L: Imagenet: A large-scale hierarchical image database. IEEE Conference on Computer Vision and Pattern Recognition, 2009:248–255, 2009

  10. Lin M, Chen Q, Yan S: Network in network. arXiv preprint arXiv:1312.4400, 2013. Available at

  11. Nesterov Y: A method for unconstrained convex minimization problem with the rate of convergence O (1/k^ 2). Doklady AN USSR 269:543–547, 1983

    Google Scholar 

  12. Maaten L, Hinton G: Visualizing data using t-SNE. J Mach Learn Res 9:2579–2605, 2008

    Google Scholar 

  13. Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D: Grad-CAM: visual explanations from deep networks via gradient-based localization. In: ICCV, Venice, Italy, 2017, pp 618–626

  14. Cohen J: A coefficient of agreement for nominal scales. Educ Psychol Meas 20:37–46, 1960

    Article  Google Scholar 

  15. Seeman E: Sexual dimorphism in skeletal size, density, and strength, J Clin Endocrinol Metab. 86:4576–4584, 2001

  16. Decker SJ, Davy-Jow SL, Ford JM, Hilbelink DR: Virtual determination of sex: Metric and nonmetric traits of the adult pelvis from 3D computed tomography models. J Forensic Sci 56:1107–1114, 2011

    Article  PubMed  Google Scholar 

  17. Luo L, Wang M, Tian Y, Duan F, Wu Z, Zhou M, Rozenholc Y: Automatic sex determination of skulls based on a statistical shape model. Comput Math Methods Med 2013:251628, 2013

    Article  PubMed  PubMed Central  Google Scholar 

  18. Richards G: What is the evidence for a link between digit ratio (2D:4D) and direct measures of prenatal sex hormones? Early Hum Dev 113:71–72, 2017

    Article  CAS  PubMed  Google Scholar 

  19. Muller DC, Giles GG, Bassett J, Morris HA, Manning JT, Hopper JL, English DR, Severi G: Second to fourth digit ratio (2D: 4D) and concentrations of circulating sex hormones in adulthood. Reprod Biol Endocrinol 9(57), 2011

  20. Barrett CK, Case DT: Use of 2D:4D digit ratios to determine sex. J Forensic Sci 59:1315–1320, 2014

    Article  PubMed  Google Scholar 

  21. Crisco JJ, Coburn JC, Moore DC, Upal MA: Carpal bone size and scaling in men versus in women. J Hand Surg Am 30:35–42, 2005

    Article  PubMed  Google Scholar 

  22. DeSilva R, Flavel A, Franklin D: Estimation of sex from the metric assessment of digital hand radiographs in a Western Australian population. Forensic Sci Int 244:314 e311–314 e317, 2014

    Article  Google Scholar 

  23. Christiansen EM et al.: In silico labeling: Predicting fluorescent labels in unlabeled images. Cell 173:792–803. e719, 2018

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  24. Muller DC, Baglietto L, Manning JT, McLean C, Hopper JL, English DR, Giles GG, Severi G: Second to fourth digit ratio (2D:4D), breast cancer risk factors, and breast cancer risk: A prospective cohort study. Br J Cancer 107:1631–1636, 2012

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  25. Bunevicius A: The Association of Digit Ratio (2D : 4D) with Cancer: A Systematic Review and Meta-Analysis. Dis Markers 2018:7698193, 2018

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  26. Mackus M, de Kruijff D, Otten LS, Kraneveld AD, Garssen J, Verster JC: The 2D : 4D Digit Ratio as a Biomarker for Autism Spectrum Disorder. Autism Res Treat 2017:1048302, 2017

    CAS  PubMed  PubMed Central  Google Scholar 

  27. Collinson SL et al.: Increased ratio of 2nd to 4th digit (2D:4D) in schizophrenia. Psychiatry Res 176:8, 2010–12

  28. Kim TB, Kim KH: Why is digit ratio correlated to sports performance? J Exerc Rehabil 12:515–519, 2016

    Article  PubMed  PubMed Central  Google Scholar 

  29. Putz DA, Gaulin SJ, Sporter RJ, McBurney DH: Sex hormones and finger length: What does 2D: 4D indicate? Evol Hum Behav 25:182–199, 2004

    Article  Google Scholar 

  30. Voracek M, Kaden A, Kossmeier M, Pietschnig J, Tran US: Meta-Analysis Shows Associations of Digit Ratio (2d:4d) and Transgender Identity Are Small at Best. Endocr Pract 24:386–390, 2018

    Article  PubMed  Google Scholar 

  31. Breuil V, Euller-Ziegler L: Gonadal dysgenesis and bone metabolism. Joint Bone Spine 68:26–33, 2001

    Article  CAS  PubMed  Google Scholar 

  32. Falhammar H, Filipsson Nystrom H, Wedell A, Brismar K, Thoren M: Bone mineral density, bone markers, and fractures in adult males with congenital adrenal hyperplasia. Eur J Endocrinol 168:331–341, 2013

    Article  CAS  PubMed  Google Scholar 

  33. Bechtold S, Beyerlein A, Bonfig W, Dalla Pozza R, Putzker S, Otto R, Schmidt H, Schwarz HP: Sexual difference in bone geometry of adult patients with classical congenital adrenal hyperplasia: Data using peripheral quantitative computed tomography. Horm Res Paediatr 82:171–178, 2014

    Article  CAS  PubMed  Google Scholar 

  34. Tanner J, Whitehouse R, Marshall W, Healy M: Assessment of Skeletal Maturity and Prediction of Adult Stature (TW2 Method). London: Academic Press, 1975

    Google Scholar 

  35. Greulich WW, Pyle SI, Todd TW: Radiographic atlas of skeletal development of the hand and wrist. Stanford: Stanford University Press, 1959

  36. Hönekopp J, Watson S: Meta-analysis of digit ratio 2D: 4D shows greater sex difference in the right hand. Am J Hum Biol 22:619–630, 2010

    Article  PubMed  Google Scholar 

Download references

Author information

Authors and Affiliations


Corresponding author

Correspondence to Synho Do.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Yune, S., Lee, H., Kim, M. et al. Beyond Human Perception: Sexual Dimorphism in Hand and Wrist Radiographs Is Discernible by a Deep Learning Model. J Digit Imaging 32, 665–671 (2019).

Download citation

  • Published:

  • Issue Date:

  • DOI:


  • Machine learning
  • Artificial intelligence
  • Sexual dimorphism
  • Sexual development
  • Bone development