Next Article in Journal
Developments in Electrical-Property Tomography Based on the Contrast-Source Inversion Method
Next Article in Special Issue
Analysis of Image Feature Characteristics for Automated Scoring of HER2 in Histology Slides
Previous Article in Journal
Origami Lesion-Targeting Device for CT-Guided Interventions
Previous Article in Special Issue
Macrosight: A Novel Framework to Analyze the Shape and Movement of Interacting Macrophages Using Matlab®
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparative Study on Local Binary Patterns for Mammographic Density and Risk Scoring †

Department of Computer Science, Aberystwyth University, Aberystwyth SY23 3DB, UK
*
Authors to whom correspondence should be addressed.
This paper is an extended version of our paper published in Annual Conference on Medical Image Understanding and Analysis, Southampton, UK, 9–11 July 2018.
Submission received: 30 November 2018 / Revised: 25 January 2019 / Accepted: 28 January 2019 / Published: 1 February 2019
(This article belongs to the Special Issue Medical Image Understanding and Analysis 2018)

Abstract

:
Breast density is considered to be one of the major risk factors in developing breast cancer. High breast density can also affect the accuracy of mammographic abnormality detection due to the breast tissue characteristics and patterns. We reviewed variants of local binary pattern descriptors to classify breast tissue which are widely used as texture descriptors for local feature extraction. In our study, we compared the classification results for the variants of local binary patterns such as classic LBP (Local Binary Pattern), ELBP (Elliptical Local Binary Pattern), Uniform ELBP, LDP (Local Directional Pattern) and M-ELBP (Mean-ELBP). A wider comparison with alternative texture analysis techniques was studied to investigate the potential of LBP variants in density classification. In addition, we investigated the effect on classification when using descriptors for the fibroglandular disk region and the whole breast region. We also studied the effect of the Region-of-Interest (ROI) size and location, the descriptor size, and the choice of classifier. The classification results were evaluated based on the MIAS database using a ten-run ten-fold cross validation approach. The experimental results showed that the Elliptical Local Binary Pattern descriptors and Local Directional Patterns extracted most relevant features for mammographic tissue classification indicating the relevance of directional filters. Similarly, the study showed that classification of features from ROIs of the fibroglandular disk region performed better than classification based on the whole breast region.

1. Introduction

It is estimated that one in eight women have the chance of getting breast cancer in their life time [1]. Cancer mortality rates show a slight decline compared to 2012 (15.2/100,000) and the predicted 2020 rate is 13.4/100,000 in Europe [2]. While there is no clear evidence of any single factor which caused the decreasing mortality rate in recent years, the most likely cause has been early diagnosis, treatment and care over the last few decades [2,3,4,5]. Though physical examination is recommended, with this it is difficult to determine breast cancer in its early stages. A variety of medical imaging modalities help in early diagnosis of breast cancer; e.g., mammography, MRI, ultrasound and tomosynthesis. Irrespective of advanced imaging modalities, mammography is still considered as the golden standard for breast screening programs. Mammography provides the radiologist with a visualization of the internal tissue structure of the breast along with information of the amount of glandular and connective tissue relative to the fatty tissue in the breast [6,7,8]. Mammographic density, the relative amount of radiodense tissue, has been considered as a strong risk factor for developing breast cancer (together with gender, age, gene-mutations and family history) [9]. According to many studies, women with high breast density have a higher tendency to develop breast cancer with a two to six-fold increased risk compared to women with low breast density [6,7,10,11,12]. In addition, studies have shown that the sensitivity performance of Computer-Aided Detection (CAD) systems to localize breast abnormalities decreases with increased breast density [13,14]. Glandular tissue in mammographic images is represented as brighter areas while the darker regions represent fatty tissue.
There are different breast classification schemes in the literature based on the breast tissue features used. It has been shown that there is a strong correlation between the breast parenchymal patterns/density and the risk of developing breast cancer and Wolfe et al. classified breast density into five categories [15]. Thereafter, Boyd et al. [16] classified density into six groups based on the proportion of fibroglandular tissue in the breast region. Similar to Wolfe, Tabár et al. [17] classified mammographic images based on the parenchyma rather than on the proportion of breast density. In our study, we used the MIAS dataset, and we classified the mammograms into three categories; Fatty (F), Glandular (G) and Dense (D), based on the tissue structure. As the MIAS dataset provide a three class density classification (F, G, D) as a ground truth and is publically available, we use it in our work. Using three classes helps in controlling the inter and intra- validation variation effect while classifying the mammographic images by different radiologist [18,19]. Example of mammographic images with different tissue type from MIAS dataset are shown in Figure 1. To standardize the mammography density reporting and to minimize the confusion in density interpretations, the American College of Radiology’s Breast Imaging Reporting and Data System (Birads) was developed as a benchmark and quality assurance tool. According to the American College of Radiology (ACR), using the latest (version 5) classification and evaluation of breast tissues used four categories [20]. Similarly, The mammographic density used to be classified into two categories as fatty and dense where the BIRADS I and II considered as fatty while BIRADS III and IV categorized as dense. In order to show the relationship between each category, Muhimmah et al. [21] have performed a comparative study based on mammographic risk assessment.
The classification of mammographic images based on tissue type density and mammographic risk estimation was initiated by Wolfe [15], which was followed by the development of different automatic techniques for breast tissue classification. Developing automatic methods for breast density estimation and classification are appropriate as the sensitivity for detection in mammographic images decreases with the increase in breast density [22]. Most CAD systems have used either segmented breast tissue or a pre-selected breast tissue Region-of-Interest (ROI) for further feature extraction and classification. Diverse features have been used such as histograms [23,24,25], intensity distribution [26], texture based approaches [27,28,29]. While Oliver et al. [6,29] used features based on texture and morphology of tissue patterns, Mustra et al. [30] used GLCM features, Petroudi et al. [12] used textons and statistical features to capture mammographic appearance, Vallez et al. [31] focused on a novel weighted voting tree classification scheme while Bovis et al. [32] used texture features from spatial grey level dependency matrices (SGLD), Fourier power spectrum, Law’s texture measures, and discrete wavelet transform for breast density classification. A detailed study on the datasets and density classification is shown in Table 1 and [33] as the datasets, number of images chosen by researchers are different. With the advent of deep learning in image processing and classification, the technique has been gaining more attention in medical image processing [34,35] worldwide for segmentation and classification processes. Various studies have been performed using deep learning for density classification resulting in promising classification models [36,37,38,39]. But most of the deep learning methods focused on binary classification (fatty or dense) of mammographic breast density.
From the above, it is clear that texture has played a role in density segmentation/classification. Ojala et al. [52] introduced a powerful and computationally simple rotation invariant generic texture classification based on Local Binary Patterns (LBP). LBP and its variants have proven to be useful [29,53] in various medical image analysis applications for extracting local texture features. Zwiggelaar et al. [29] combined Local Binary Pattern (LBP) texture features and texture features extracted from grey level co-occurrence matrices to classify mammograms. Later, Chen et al. [48] performed a comparative study on the performance of LBP, local grey-level appearance (LGA), textons and basic image features obtaining accuracies of 59%, 72%, 75% and 70%, respectively using the whole breast region from the MIAS dataset. We are building on this and we review the use of various LBP variants for density classification.
We have used Elliptical Local Binary Pattern (ELBP) [53], Uniform ELBP, Mean-Elliptical Local Binary Patterns (M-ELBP) [51], Local Binary Pattern (LBP) and LDP (Local Directional Patterns) [54] to capture intrinsic and detailed micro-pattern features from the mammographic images for breast tissue classification into fatty, glandular and dense breast tissue.
In addition to our initial work in [55], the efficiency of M-ELBP texture descriptor was analyzed and compared with other similar LBP variants (LBP, ELBP and Uniform-ELBP). The previous study [55] showed that the texture descriptors like ELBP, U-ELBP and M-ELBP performed better than LBP showing that anisotropic information extracted from breast tissue was potentially more important than a roatation invariant texture descriptor like LBP. A further extension is the inclusion of the Directional texture descriptor LDP [54] is included for analyzing the potentiality of using directionality features of texture descriptors in estimating mammographic breast density.
A further extension to the previous work (where only the ROIs from the fibroglandular region with size equal to 256 × 256 were considered for classification), this study covered the variation in density classification accuracy while selecting ROIs from different regions of the fibroglandular disk region and is compared to the whole breast.
In addition, the suitable size of ROIs for classification is evaluated by comparing ROIs of different sizes from the fibroglandular disk region. So in summary, the proposed study focused on the difference in the classification accuracy for LBP variants variation in size and location of ROI, selecting whole breast and ROI, the role of machine learning algorithms, the effect of descriptor size, and the role of directional filters in extracting more multidimensional complex texture features for density classification.

2. Methodology

Breast ROIs have been used for texture analysis in detecting and classifying different breast abnormalities [56,57]. The analysis and classification of breast density depends upon the breast size and location of the ROI [58,59]. To analyse the effect of ROI location, Li et al. [58] performed a risk estimation on breast parenchymal patterns on breast ROIs at different locations behind the central nipple area using texture features. The study showed a significant decrease in performance when the ROI location moved backwards from the central region. Similarly, Sharma et al. [59] completed a study to analyze the effect of the size of the ROIs for classification and concluded that larger ROIs (above 200 × 200 pixels) did decrease the classification accuracy due to inclusion of more tissue and smaller ROIs were affected by lose of important tissue pattern features.

2.1. Pre-Processing

The mammographic images are pre-processed to remove the pectoral muscle and other artifacts (e.g., labels; see Figure 1) to extract the main breast region for processing [60]. The method, developed by the Chen and Zwiggelaar [60] separated the background region containing the annotations, labels and frames from the whole breast as the initial step. Subsequently, the method used histogram thresholding, contour growing and polynomial fitting method to remove the pectoral muscle from the breast tissue region. As most of the dense tissue and parenchymal patterns are situated within the breast fibroglandular disk area (the central region behind the nipple), we extracted ROIs from that region. The ROI was extracted ([51] for details) from each image in the MIAS database [61] with size equal to 256 × 256 pixels as shown in Figure 2. The longest vertical and horizontal distance within the breast was estimated and the intersection point of the both lines was considered as the central point of the fibroglandular region for extracting the ROIs (see Figure 2). With this intersection point as the center, the ROIs of size 256 × 256 were extracted. Noise reduction was performed on the extracted ROI using a median filter of 3 × 3 size.

2.2. Feature Extraction and Feature Selection

Ojala et al. [52] proposed a local texture descriptor, Local Binary Pattern (LBP) for texture feature extraction. Besides the simplicity and robustness to variation in image intensity levels, its computational efficiency offers a wide range of application in texture feature extraction. In the classical LBP developed by Ojala et al., for each central pixel ( x c , y c ) in the input image with a grey level value g c , its LBP value is estimated by comparing the g c value with the grey level values of pixels at R distance within its surrounding P neighbourhood pixels following the pixels along a circular path either clockwise or counter-clockwise (see in Figure 3a). If the central pixel value ( x c , y c ) has a higher grey level value than the neighbouring pixel P i , the neighbour pixel is assigned a value 0 else 1 giving a P-bit binary number. Using ROI level histogram information provides robustness, although it removes localisation, which prevents detailed segmentation information. Although this can be addressed by using small local ROIs/windows, but this is seen as future work (the histogram of LBP values for the whole ROI was calculated). For a neighbourhood of eight, LBP can generate 256 different binary patterns. The L B P P , R is calculated as follows:
L B P P , R ( x c , y c ) = i = 1 P s ( g i P , R g c ) 2 i 1
where s ( x ) is defined as
s ( x ) = 1 , if x > = 0 0 , if x < 0
A circular neighbourhood is an advantage for a texture descriptor but there are applications were the anisotropic texture information have more potential in distinguishing objects. A circular-like topologies like an ellipse helps in attaining features from different directions [62]. For Elliptical Local Binary Patterns (ELBP), for each central pixel ( x c , y c ) , we consider the neighbouring pixels P which lie on an ellipse (with radius R 1 and R 2 ) which was calculated as:
E L B P P , R 1 , R 2 ( x c , y c ) = i = 1 P s ( g i P , R 1 , R 2 g c ) 2 i 1
where the ith neighbouring pixel coordinate of ( x c , y c ) , is computed as:
a n g l e s t e p = 2 π P
x i = x c + R 1 cos ( ( i 1 ) a n g l e s t e p )
y i = y c R 2 sin ( ( i 1 ) a n g l e s t e p )
Compared to the LBP patterns, the ELBP descriptors help in extracting more specific spatial features from the mammographic images as it could extract additional directional features at different orientations covering additional micro patterns. For the computation of the Mean-Elliptical Local Binary Patterns (M-ELBP), we consider the mean intensity values ( g ¯ i ) around each neighbouring pixel P i which lie on an ellipse (with radius R 1 and R 2 ) as
M-ELBP P , R 1 , R 2 ( x c , y c ) = i = 1 P s ( g ¯ i P , R 1 , R 2 g c ) 2 i 1
The mean intensity value around each neighborhood pixels (a local window of size 3 × 3 was used here) is compared with the central pixel to create a binary pattern as in LBP or ELBP.
We used the ELBP/M-ELBP descriptors at eight different orientations generating an eight-bit pattern for each central pixel. The experimental results by [63], indicated that extracting tissue features from multiple orientations performed better in determining the mammographic tissue class due to the complex and multidimensional appearance of tissue patterns in the breast. When ( R 1 = R 2 ) , the ELBP/M-ELBP reduces to the LBP descriptor. To extract additional features, we modified the ELBP operator to the mean-ELBP (M-ELBP) where the intensity features at different orientations were attained along with the texture features [51]. A detailed overview of the various method is described in Figure 3. Compared to the LBP descriptor, the elliptical descriptors could extract more features when applied to ROIs at different orientations θ = { 0 , 45 , 90 , 135 , 180 , 225 , 270 and 315 } . Figure 4 shows the effect of horizontal and vertical M-ELBP on the extracted image ROI for P=8 at an orientation θ = 0 and θ = 90 . The procedure is repeated for all orientations to extract M-ELBP pattern features. Figure 5 summarizes the classification of breast tissue using the ELBP/M-ELBP descriptor variants at eight different orientations with an elliptical radius of ( R 1 = 4 , R 2 = 7 ) for a neighbourhood size of eight pixel. Subsequently, the histogram labels for the ELBP/M-ELBP descriptors at all orientations are concatenated to generate the texture feature vector for the ROIs.
A useful modification to the LBP operator was the introduction of the uniform patterns [64] concept to the LBP which can reduce the length of the feature vector from 256 to 59 for an 8- bit binary pattern. The uniform pattern concept was developed from the fact that certain binary patterns occur commonly in texture images. If a binary pattern contains at most two transitions of 0->1 or 1->0, the pattern is called uniform. So, while constructing histogram, the uniform patterns have separate bins with a single bin assigned to all other non-uniform patterns.
Due to the feature extraction at multiple orientations, the feature vector length when concatenated will be large (each orientation for the ELBP descriptor generated 256 size histogram) and the probability of correlation between features increases. To retain only prominent features and to reduce the computational cost due to high dimensionality, feature selection was performed on the extracted feature vector. A correlation based feature subset (CFS) selection method [65] with a best first search method was used. Highly correlated feature subsets that are highly correlated with the group class while having lower inter correlation among feature subsets are preferred for attribute selection. This calculated the individual predictive ability of each attribute/feature in the dataset along with the redundancy between each feature. The most prominent selected features were then fed into various classifiers for classification of mammographic breast tissue.
Additionally, to compare the directional feature extraction capability of the M-ELBP descriptor and the importance of directional texture feature descriptors for extracting mammographic parenchymal features, we compute the classification accuracy of the Local Directional Pattern (LDP) [66]. The LDP produces an eight-bit binary pattern similar to LBP for a neighbourhood of eight. Unlike LBP, the LDP patterns are computed by comparing the relative edge response values of each pixel at different orientations. In order to get the edge responses, the eight directional edge response values of a particular pixel were calculated using Kirsch masks ( ( M 0 M 7 , also known as the Kirsch compass kernel which is a non-linear edge detector to find the maximum edge strength in eight predetermined compass directions with 45 degree increments). The edge responses are not equally important in all directions ( m 0 , m 1 , ... , m 7 ) as the corners and edges show high responses for particular directions. So the top k values from | m j | were selected and defined as ’1’ while the other ( 8 k ) bits were set to ’0’.

2.3. Extraction of ROIs

Subsequently, to study the effect of classification using the whole breast and different sized ROIs, varying sizes of ROIs were extracted ranging from 256 × 256 pixels, 200 × 200 pixels and 100 × 100 pixels. All the ROIs were selected randomly from the 256 × 256 pixels ROI region (used for descriptor comparison) selected from the fibroglandular disk region.
Similarly, to study the effect of the descriptor size on tissue classification, we performed the experiment as described in Figure 5 with varying size of R 1 and R 2 ( R 1 = { 2 , 4 , 6 , 12 } and R 2 = { 5 , 7 , 9 , 15 } ). For each elliptical size (varying sizes of R 1 , R 2 ) with a neighborhood of eight, the feature extraction is done for eight different orientations. The extracted features at different orientations are concatenated and feature selection (extracting the most prominent features) is done as in Figure 5. Tahmassebi et al. [67] showed the importance of ranking and selecting the best extracted parameters while considering medical image data.

3. Experimental Results

The evaluation presented in this paper was performed on the Mammographic Image Analysis Society (MIAS) dataset [61]. The database contains 322 mammographic images of 161 patients with mediolateral-oblique views. Each mammogram is classified into three distinct groups as either Fatty (F), Fatty-Glandular(G) or Dense-Glandular (D) based on the tissue pattern. We perform classification on 321 images (one image was not available for historical reasons) with 106 fatty cases, 104 fatty-glandular and 111 dense-glandular cases.

3.1. Comparison between LBP Variants

For the comparison of classification by different Local Binary Pattern variants, we performed LBP, ELBP, uniform-ELBP and M-ELBP and LDP on ROIs of size 256 × 256 from the fibroglandular region with neighbourhood size equal to eight. While LBP extracted features only from a circular neighbourhood, the ELBP was able to extract more structural and spatial features at different orientations extracting multidimensional micro-pattern features of the breast tissue. In-order to incorporate the intensity features along with the textural features into the histogram, we used M-ELBP. Similarly, the effect of Uniform patterns in tissue classification was estimated using Uniform ELBP [68]. To keep the consistency of classification accuracy due to the classifier effect, we used the Bayesian Network for every descriptor with 10-fold cross validation.
Table 2 shows the classification results on the ROIs from the MIAS database using LBP. The approach gave a classification accuracy equal to 70.0%. Since the LBP operator considered a circular pattern of the neighbourhood, it could not capture the directional features of the breast tissue due to its rotation invariant feature.
Table 3 shows that the classification accuracy for the ELBP descriptor has improved to 75.0% compared to LBP showing that ELBP can perform better by extracting additional multidimensional features from different directions. In-order to study the effect of uniform patterns in the classification of density, we performed Uniform-ELBP and found results similar to ELBP giving an accuracy of 74.0% as shown in Table 4.
To understand the effect of intensity features in mammographic breast tissue classification, our proposed variant of ELBP [51], M-ELBP was used which classified the mammographic images from the MIAS dataset into BIRADS classes obtaining a classification accuracy of 74%. Table 5 shows that the classification accuracy for this ELBP variant (M-ELBP) improved to 79.8% while including intensity features to texture pattern in breast tissue classification.
Similarly, to analyze the role of directional texture descriptor for mammographic density estimation, we used the LDP [54] operator on the same set of ROIs with the Bayesian Network. The classification results in Table 6 show an accuracy of 75.4%, which was lower than the M-ELBP descriptor for the Bayesian Network classifier but shows the strength of pixel directional response in feature extraction.
When considering Table 2, Table 3, Table 4, Table 5 and Table 6, it seems there is a trend for over-estimation, which could be due to the wide variation in glandular tissue appearance within the dataset and the ROI selection process. For the classification results for M-ELBP (see Table 5) it should be noted that this is the only approach which does not have any cases which are mis-classified by more than one class (so no Fatty cases were classified as Dense and visa versa), which could be caused by the incorporation of texture features along with the mean intensity values of the neighbourhood making it more robust to the tissue pattern and intensity variations. In addition, it should be noted that the LBP based approach performs worse for the Dense class (which is probably caused by the rotational invariant aspects), whilst both ELBP and uniform-ELBP perform less on Glandular cases (which could again be caused by the wide range of variation in the Glandular class and the effects of noise). We see a more detailed analysis of these aspects as future work, which could be based on an in-dept analysis of the feature space.
In order to analyze the significance of the test for different texture descriptors [69], we perform a ten-run ten-fold cross validation for all the texture descriptors. Table 7 shows that the Bayesian Network yielded an accuracy of 69.44 ± 0.92 , 75.41 ± 1.05 , 72.86 ± 1.06 , 77.38 ± 1.06 and 74.92 ± 0.67 for LBP, ELBP, U-ELBP, M-ELBP and LDP, respectively. Table 7 also indicates if differences were significant (at the 0.05 level).

3.2. Study on Classifier Effect on Classifications

Machine learning approaches have a great impact on medical image classification and can provide valuable predictive information to guide treatment decisions [70]. In order to understand the classifier effect on mammographic density estimation accuracy, we computed the classification using different classifiers for the same feature set. The classification accuracy illustrated by Table 7 shows the variation in classification accuracy for a 10 run 10 fold cross validation for different classifiers showing the importance of feature selection and classifier effect on density classification. The classifiers chosen for our experimental study are Bayesian Network, K-Nearest neighbor classifier, Support Vector Machine, and Random Forest [71,72]. Additionally, to test the statistical significance of classification accuracy, we have performed a unpaired t-test for a significance level of p = 0.05 between descriptors and classifiers (see in Table 7) taking best classification result by the respective classifier for each descriptor as base line.
Similarly, the area under the ROC curve ( A z ) [73] was calculated to compare the efficiency of classifiers for density estimation. It represents how good a classifier can distinguish between different groups. Table 8 shows results using stratified 10 run 10 fold cross validation for A z . On comparing the classification accuracy and A z results from Table 7 and Table 8, LDP and M-ELBP show a balanced result for classification accuracy by same classifiers showing the potential of directional filters for mammogram density classification. Similarly, Figure 6 represents the ROC curve for M-ELBP descriptor using the Bayesian classifier.
From the experimental studies, the M-ELBP and LDP performed better in classifying the mammographic breast tissue showing the importance of directional filters in feature extraction. In-order to avoid the classifier effect on classification accuracy, we use the Bayesian network for further experiments in the paper.

3.3. Study on Descriptor Size

To compare the effect of descriptor size on classification, the M-ELBP operator with varying size of R 1 = { 2 , 4 , 6 , 12 } and R 2 = { 5 , 7 , 9 , 15 } were applied for the ROI size equal to 256 × 256 pixels. In order to compute the stability of the approach we used a 10 run 10 fold cross validation scheme using a Bayesian network. The results showed that the descriptor classification remained stable with the change in size. Figure 7 shows the classification accuracy ( C A % ) to be 77.82 ± 0.68 , 77.66 ± 1.01 , 78.01 ± 0.57 and 76.85 ± 0.73 . The figure represents the classification accuracy obtained by the Bayesian Network classifier for each run (N = number of runs) for a 10-fold cross validation scheme. Similarly, from Figure 8, the area under the ROC curve( A z ) shows similar results as 0.9204 ± 0.0010 , 0.9181 ± 0.0011 , 0.9205 ± 0.0015 and 0.9121 ± 0.0022 for the M-ELBP descriptor of size R 1 = { 2 , 4 , 7 , 12 } , R 2 = { 5 , 7 , 9 , 15 } with a neighbourhood of eight pixels, respectively.

3.4. Effect of ROI Size and Location in Classification

Further experiments were done to estimate the effect of the ROI size and location in determining the classification accuracy. We estimate the classification accuracy while using the whole breast, or ROI sizes varying from 100 × 100 , 200 × 200 to 256 × 256 pixels were selected from the fibroglandular disk region. The classification results for the M-ELBP descriptor with size R = { 4 , 7 } and P = 8 using the Bayesian Network classifier for 10 run 10- fold cross validation. Figure 9 and Figure 10 shows the variation in classification accuracy (N indicate a distinct run) while the whole breast region ( C A % = 65.79 ± 0.70 , A z = 0.83 ± 0.003 ) or an ROI of size 100 × 100 pixels ( C A % = 70.28 ± 1.16 , A z = 0.86 ± 0.003 ) are selected. The most appropriate ROI size for this study was found to be between 200 × 200 pixels ( C A % = 77.41 ± 0.72 , A z = 0.91 ± 0.002 ) to 256 × 256 pixels ( C A % = 77.39 ± 1.04 , A z = 0.92 ± 0.001 ) from the fibroglandular region due to the presence of the appropriate parenchymal patterns. This showed that if the ROI size is small, classification can be affected due to minimum dominant features. Similarly, if the ROI size is too large or include irrelevant regions like tissue near the pectoral muscle or the breast boundary regions can result in extracting less discriminant features. In case of choosing the whole breast, the classification can be affected by irrelevant data resulting in a poor classification accuracy as it extracted multi-class texture information making it difficult for discriminating the prominent features for each density class. So the study showed that choosing the correct ROI size and location can lead to improved density classification inline with similar studies in the field [58,59].

4. Discussions and Conclusions

Due to variations in datasets, classifiers, number of images used by researchers, difference in density classes a direct comparison of our results with existing methods is difficult. So the state of art with the same dataset and density classes are considered for comparison with our studies. Muhimmah et al. [23] classified breast density into 3 classes using multi resolution histogram information using an SVM classifier obtaining an accuracy of 77.57%. Similarly, Subashini et al. [43] used 43 images from the MIAS dataset using a segmented breast tissue approach and attained an accuracy of 95.55% with a SVM classifier using statistical features. But the good classification results could be due to the limited number of images selected and so the classification results would probably decrease with whole database. Blot et al. [41] used background texture information for 265 images and obtained a CA of 65.0% using a KNN classifier. Zwiggelaar et al. [24] used gray level histogram features for three class density classification on MIAS dataset with a subset of 312 images and attained an accuracy of 71.5%. Compared to the methods used for breast density classification, the Local Binary Pattern variants such as traditional LBP, ELBP, U-ELBP, M-ELBP, LDP obtained CA value of 69.44 ± 0.92 , 75.41 ± 1.05 , 73.29 ± 0.64 , 77.38 ± 1.06 , 75.95 ± 0.96 respectively (see Table 7). Comparison with the results in Table 1 shows that the Local Binary Pattern variants are competitive in extracting relevant features for breast density classification. Thus our results are comparable with the state of art while considering the number of images (321) and the density class used for classification.
It is evident from the Table 7 and Table 8 that though circular neighborhood provide rotation invariance as in the traditional LBP, elliptical topologies could extract more anisotropic features from an image especially for images which do not have a specific pattern like mammogram tissue which are complex in nature. So here, the elliptical texture descriptors performed better compared to traditional LBP information. Additionally, elliptical descriptors performed better compared to traditional LBP variants as it extract more complex texture patterns due to its directionality nature along with the possibility of multi-scale texture analysis. Similarly, the results showed that incorporating image intensity features like the mean with texture features (as in M-ELBP) improved the feature vector strength in distinguishing different patterns for a complex image. Similarly, results from the LDP descriptor showed the potentiality of directional filters and importance of analyzing the strength of edge and corner features of pixels in an image in classifying patterns. Unlike traditional LBP, both M-ELBP and LDP is better dealing with noise in the image as it is not directly comparing the gray values (contrast) between pixels. Therefore it helped in extracting the important parameters for classification.
Additionally, this study concentrated on analysing the effect of selecting the region of interest (ROI) from mammographic images for breast tissue classification. The study showed that selecting the size of the ROIs at 200 × 200 pixels and 256 × 256 pixels taken from fibroglandular disk region were most appropriate. This showed similar result to the study performed by Li et al. [58] where ROIs were selected at different distances from the breast nipple. A similar study was performed by Rampun et al. [63] showing the importance of choosing the ROIs from the fibroglandular region for density classification.
Further results on the selection of the ROI size showed that smaller ROIs can lead to decreased classification due to lack of texture patterns extracted and very large ROIs can include irrelevant features, again resulting in decreased classification. The ROI size indicated is comparable with the results shown by Sharma et al. [59].
For future work, we will focus on developing a method which extracts additional multidimensional directional pattern features by focusing on the edge responses of each pixel and gradient changes. Additionally, the classification accuracy of boosting algorithms like XGBoost based on gradient boosting will be analyzed [74]. Similarly, combining handcrafted features focusing on pattern responses with deep learning approaches for feature extraction and classification can be an advancement in the field of breast density classification. Although several studies [36,37,38,39] have focused on deep learning features for mammographic density classification, a comparison of handcrafted features (focusing on intensity and local patterns) to deep learning feature extraction and classification is a potential area of investigation.

Author Contributions

In this study, M.G. did all the experiments and evaluations discussed. R.Z. supervised the project and contributed equally for the preparation of final version of paper.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ferlay, J.; Soerjomataram, I.; Dikshit, R.; Eser, S.; Mathers, C.; Rebelo, M.; Parkin, D.M.; Forman, D.; Bray, F. Cancer incidence and mortality worldwide: Sources, methods and major patterns in GLOBOCAN 2012. Int. J. Cancer 2015, 136, E359–E386. [Google Scholar] [CrossRef] [PubMed]
  2. Carioli, G.; Malvezzi, M.; Rodriguez, T.; Bertuccio, P.; Negri, E.; La Vecchia, C. Trends and predictions to 2020 in breast cancer mortality in Europe. Breast 2017, 36, 89–95. [Google Scholar] [CrossRef] [PubMed]
  3. Day, N.; Williams, D.; Khaw, K. Breast cancer screening programmes: The development of a monitoring and evaluation system. Br. J. Cancer 1989, 59, 954. [Google Scholar] [CrossRef] [PubMed]
  4. Autier, P.; Héry, C.; Haukka, J.; Boniol, M.; Byrnes, G. Advanced breast cancer and breast cancer mortality in randomized controlled trials on mammography screening. J. Clin. Oncol. 2009, 27, 5919–5923. [Google Scholar] [CrossRef] [PubMed]
  5. Kuhl, C.; Kuhn, W.; Schild, H. Management of women at high risk for breast cancer: New imaging beyond mammography. Breast 2005, 14, 480–486. [Google Scholar] [CrossRef] [PubMed]
  6. Oliver, A.; Freixenet, J.; Marti, R.; Pont, J.; Pérez, E.; Denton, E.R.; Zwiggelaar, R. A novel breast tissue density classification methodology. IEEE Trans. Inf. Technol. Biomed. 2008, 12, 55–65. [Google Scholar] [CrossRef] [PubMed]
  7. McCormack, V.A.; dos Santos Silva, I. Breast density and parenchymal patterns as markers of breast cancer risk: A meta-analysis. Cancer Epidemiol. Prev. Biomark. 2006, 15, 1159–1169. [Google Scholar] [CrossRef]
  8. George, M.; Rampun, A.; Denton, E.; Zwiggelaar, R. Mammographic ellipse modelling towards birads density classification. In Proceedings of the International Workshop on Digital Mammography, Malmo, Sweden, 19–22 June 2016; pp. 423–430. [Google Scholar]
  9. Kallenberg, M.; Petersen, K.; Nielsen, M.; Ng, A.Y.; Diao, P.; Igel, C.; Vachon, C.M.; Holland, K.; Winkel, R.R.; Karssemeijer, N.; et al. Unsupervised deep learning applied to breast density segmentation and mammographic risk scoring. IEEE Trans. Med Imaging 2016, 35, 1322–1331. [Google Scholar] [CrossRef]
  10. Boyd, N.F.; Martin, L.J.; Bronskill, M.; Yaffe, M.J.; Duric, N.; Minkin, S. Breast tissue composition and susceptibility to breast cancer. J. Natl. Cancer Inst. 2010, 102, 1224–1237. [Google Scholar] [CrossRef]
  11. Byng, J.W.; Boyd, N.; Fishell, E.; Jong, R.; Yaffe, M.J. The quantitative analysis of mammographic densities. Phys. Med. Biol. 1994, 39, 1629. [Google Scholar] [CrossRef]
  12. Petroudi, S.; Brady, M. Breast density segmentation using texture. In Proceedings of the International Workshop on Digital Mammography, Malmo, Sweden, 19–22 June 2016; pp. 609–615. [Google Scholar]
  13. Ho, W.; Lam, P. Clinical performance of computer-assisted detection (CAD) system in detecting carcinoma in breasts of different densities. Clin. Radiol. 2003, 58, 133–136. [Google Scholar] [CrossRef] [PubMed]
  14. Obenauer, S.; Sohns, C.; Werner, C.; Grabbe, E. Impact of breast density on computer-aided detection in full-field digital mammography. J. Digit. Imaging 2006, 19, 258. [Google Scholar] [CrossRef] [PubMed]
  15. Wolfe, J.N. Risk for breast cancer development determined by mammographic parenchymal pattern. Cancer 1976, 37, 2486–2492. [Google Scholar] [CrossRef] [Green Version]
  16. Boyd, N.; Byng, J.; Jong, R.; Fishell, E.; Little, L.; Miller, A.; Lockwood, G.; Tritchler, D.; Yaffe, M.J. Quantitative classification of mammographic densities and breast cancer risk: Results from the Canadian National Breast Screening Study. J. Natl. Cancer Inst. 1995, 87, 670–675. [Google Scholar] [CrossRef] [PubMed]
  17. Tabár, L.; Dean, P.B. Mammographic parenchymal patterns: Risk indicator for breast cancer? JAMA 1982, 247, 185–189. [Google Scholar] [CrossRef] [PubMed]
  18. Warfield, S.K.; Zou, K.H.; Wells, W.M. Validation of image segmentation by estimating rater bias and variance. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Athens, Greece, 17–21 October 2006; pp. 839–847. [Google Scholar]
  19. Kauppi, T.; Kamarainen, J.K.; Lensu, L.; Kalesnykiene, V.; Sorri, I.; Kälviäinen, H.; Uusitalo, H.; Pietilä, J. Fusion of multiple expert annotations and overall score selection for medical image diagnosis. In Proceedings of the Scandinavian Conference on Image Analysis, Oslo, Norway; 2009; pp. 760–769. [Google Scholar]
  20. American College of Radiology, B.R.C. Breast Imaging Reporting and Data System; American College of Radiology: Reston, VA, USA, 1998. [Google Scholar]
  21. Muhimmah, I.; Oliver, A.; Denton, E.R.; Pont, J.; Pérez, E.; Zwiggelaar, R. Comparison between Wolfe, Boyd, BI-RADS and Tabár based mammographic risk assessment. In Proceedings of the International Workshop on Digital Mammography, Malmo, Sweden, 19–22 June 2006; pp. 407–415. [Google Scholar]
  22. Sickles, E.A. Wolfe mammographic parenchymal patterns and breast cancer risk. Am. J. Roentgenol. 2007, 188, 301–303. [Google Scholar] [CrossRef] [PubMed]
  23. Muhimmah, I.; Zwiggelaar, R. Mammographic density classification using multiresolution histogram information. In Proceedings of the International Special Topic Conference on Information Technology in Biomedicine, Epirus, Greece, 26–28 October 2006; pp. 26–28. [Google Scholar]
  24. Zwiggelaar, R.; Muhimmah, I.; Denton, E. Mammographic density classification based on statistical gray-level histogram modelling. In Proceedings of the Medical Image Understanding and Analysis Conference, Bristol, UK; 2005; pp. 183–186. [Google Scholar]
  25. Hadjidemetriou, E.; Grossberg, M.D.; Nayar, S.K. Multiresolution histograms and their use for recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 831–847. [Google Scholar] [CrossRef]
  26. Kallenberg, M.G.; Lokate, M.; Van Gils, C.H.; Karssemeijer, N. Automatic breast density segmentation: An integration of different approaches. Phys. Med. Biol. 2011, 56, 2715. [Google Scholar] [CrossRef]
  27. Zhou, C.; Chan, H.P.; Petrick, N.; Helvie, M.A.; Goodsitt, M.M.; Sahiner, B.; Hadjiiski, L.M. Computerized image analysis: Estimation of breast density on mammograms. Med. Phys. 2001, 28, 1056–1069. [Google Scholar] [CrossRef]
  28. He, W.; Denton, E.R.; Zwiggelaar, R. Mammographic segmentation and risk classification using a novel binary model based bayes classifier. In Proceedings of the International Workshop on Digital Mammography, Philadelphia, PA, USA, 8–11 July 2012; pp. 40–47. [Google Scholar]
  29. Oliver, A.; Freixenet, J.; Zwiggelaar, R. Automatic classification of breast density. In Proceedings of the IEEE International Conference on Image Processing 2005, Genova, Italy, 14 September 2005; Volume 2. [Google Scholar]
  30. Muštra, M.; Grgić, M.; Delač, K. Breast density classification using multiple feature selection. Automatika 2012, 53, 362–372. [Google Scholar] [CrossRef]
  31. Vállez, N.; Bueno, G.; Déniz, O.; Dorado, J.; Seoane, J.A.; Pazos, A.; Pastor, C. Breast density classification to reduce false positives in CADe systems. Comput. Methods Programs Biomed. 2014, 113, 569–584. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Bovis, K.; Singh, S. Classification of mammographic breast density using a combined classifier paradigm. In Proceedings of the 4th International Workshop on Digital Mammography, Bremen, Germany, 22–25 June 2002; pp. 177–180. [Google Scholar]
  33. He, W.; Juette, A.; Denton, E.R.; Oliver, A.; Martí, R.; Zwiggelaar, R. A review on automatic mammographic density and parenchymal segmentation. Int. J. Breast Cancer 2015, 2015. [Google Scholar] [CrossRef] [PubMed]
  34. Tahmassebi, A.; Gandomi, A.; McCann, I.; Schulte, M.; Goudriaan, A.; Meyer-Baese, A. Deep learning in medical imaging: fMRI big data analysis via convolutional neural networks. Proc. Pract. Exp. Adv. Res. Comput. ACM 2018. [Google Scholar] [CrossRef]
  35. Hamidinekoo, A.; Suhail, Z.; Qaiser, T.; Zwiggelaar, R. Investigating the effect of various augmentations on the input data fed to a convolutional neural network for the task of mammographic mass classification. In Proceedings of the Annual Conference on Medical Image Understanding and Analysis, Edinburgh, UK, 11–13 July 2017; pp. 398–409. [Google Scholar]
  36. Wu, N.; Geras, K.J.; Shen, Y.; Su, J.; Kim, S.G.; Kim, E.; Wolfson, S.; Moy, L.; Cho, K. Breast density classification with deep convolutional neural networks. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 6682–6686. [Google Scholar]
  37. Wainer, J. Breast Density Classification with Convolutional Neural Networks. In Proceedings of the Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications: 21st Iberoamerican Congress, Lima, Peru, 8–11 November 2016; Volume 10125, p. 101. [Google Scholar]
  38. Ciritsis, A.; Rossi, C.; Vittoria De Martini, I.; Eberhard, M.; Marcon, M.; Becker, A.S.; Berger, N.; Boss, A. Determination of mammographic breast density using a deep convolutional neural network. Br. J. Radiol. 2018, 91. [Google Scholar] [CrossRef] [PubMed]
  39. Mohamed, A.A.; Berg, W.A.; Peng, H.; Luo, Y.; Jankowitz, R.C.; Wu, S. A deep learning method for classifying mammographic breast density categories. Med. Phys. 2018, 45, 314–321. [Google Scholar] [CrossRef] [PubMed]
  40. Karssemeijer, N. Automated classification of parenchymal patterns in mammograms. Phys. Med. Biol. 1998, 43, 365. [Google Scholar] [CrossRef] [PubMed]
  41. Blot, L.; Zwiggelaar, R. Background texture extraction for the classification of mammographic parenchymal patterns. In Proceedings of the 15th Medical Image Understanding and Analysis Meeting, Birmingham, UK, 2001; pp. 145–148. [Google Scholar]
  42. Petroudi, S.; Kadir, T.; Brady, M. Automatic classification of mammographic parenchymal patterns: A statistical approach. In Proceedings of the 25th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Cancun, Mexico, 17–21 September 2003; Volume 1, pp. 798–801. [Google Scholar]
  43. Subashini, T.; Ramalingam, V.; Palanivel, S. Automated assessment of breast tissue density in digital mammograms. Comput. Vis. Image Underst. 2010, 114, 33–43. [Google Scholar] [CrossRef]
  44. Oliver, A.; Lladó, X.; Pérez, E.; Pont, J.; Denton, E.R.; Freixenet, J.; Martí, J. A statistical approach for breast density segmentation. J. Digit. Imaging 2010, 23, 527–537. [Google Scholar] [CrossRef]
  45. Zwiggelaar, R. Local greylevel appearance histogram based texture segmentation. In Proceedings of the International Workshop on Digital Mammography, Girona, Spain, 16–18 June 2010; pp. 175–182. [Google Scholar]
  46. Liu, Q.; Liu, L.; Tan, Y.; Wang, J.; Ma, X.; Ni, H. Mammogram density estimation using sub-region classification. In Proceedings of the 4th International Conference on Biomedical Engineering and Informatics (BMEI), Shanghai, China, 15–17 October 2011; Volume 1, pp. 356–359. [Google Scholar]
  47. Tzikopoulos, S.D.; Mavroforakis, M.E.; Georgiou, H.V.; Dimitropoulos, N.; Theodoridis, S. A fully automated scheme for mammographic segmentation and classification based on breast density and asymmetry. Comput. Methods Programs Biomed. 2011, 102, 47–63. [Google Scholar] [CrossRef]
  48. Chen, Z.; Denton, E.; Zwiggelaar, R. Local feature based mammographic tissue pattern modelling and breast density classification. In Proceedings of the 2011 4th International Conference on Biomedical Engineering and Informatics (BMEI), Shanghai, China, 15–17 October 2011; Volume 1, pp. 351–355. [Google Scholar]
  49. Qu, Y.; Shang, C.; Wu, W.; Shen, Q. Evolutionary Fuzzy Extreme Learning Machine for Mammographic Risk Analysis. Int. J. Fuzzy Syst. 2011, 13, 4. [Google Scholar]
  50. Rampun, A.; Morrow, P.; Scotney, B.; Winder, J. Breast Density Classification Using Local Ternary Patterns in Mammograms. In Proceedings of the International Conference Image Analysis and Recognition, Montreal, QC, Canada, 5–7 July 2017; pp. 463–470. [Google Scholar]
  51. George, M.; Denton, E.; Zwiggelaar, R. Mammogram breast density classification using mean-elliptical local binary patterns. In Proceedings of the 14th International Workshop on Breast Imaging (IWBI 2018), Atlanta, GA, USA, 8–11 July 2018; p. 107180. [Google Scholar]
  52. Ojala, T.; Pietikäinen, M.; Mäenpää, T. Gray scale and rotation invariant texture classification with local binary patterns. In Proceedings of the European Conference on Computer Vision, Dublin, Ireland, 26 June–1 July 2000; pp. 404–420. [Google Scholar]
  53. Nanni, L.; Lumini, A.; Brahnam, S. Local binary patterns variants as texture descriptors for medical image analysis. Artif. Intell. Med. 2010, 49, 117–125. [Google Scholar] [CrossRef] [PubMed]
  54. Jabid, T.; Kabir, M.H.; Chae, O. Local directional pattern (LDP) for face recognition. In Proceedings of the International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 9–13 January 2010; pp. 329–330. [Google Scholar]
  55. George, M.; Zwiggelaar, R. Breast tissue classification using Local Binary Pattern variants: A comparative study. In Proceedings of the Annual Conference on Medical Image Understanding and Analysis, Southampton, UK, 9–11 July 2018; pp. 143–152. [Google Scholar]
  56. De Oliveira, J.E.; Machado, A.M.; Chavez, G.C.; Lopes, A.P.B.; Deserno, T.M.; Araújo, A.d.A. MammoSys: A content-based image retrieval system using breast density patterns. Comput. Methods Programs Biomed. 2010, 99, 289–297. [Google Scholar] [CrossRef] [PubMed]
  57. Epischina, J.; De Albuquerque, A.; Deserno, T. Content-based image retrieval applied to BI-RADS tissue classification in screening mammography. World J. Radiol. 2011, 28, 24–31. [Google Scholar]
  58. Li, H.; Giger, M.L.; Huo, Z.; Olopade, O.I.; Lan, L.; Weber, B.L.; Bonta, I. Computerized analysis of mammographic parenchymal patterns for assessing breast cancer risk: Effect of ROI size and location. Med. Phys. 2004, 31, 549–555. [Google Scholar] [CrossRef] [PubMed]
  59. Sharma, V. Comparative analysis of region of interest of different sizes for breast density classification. Int. J. Med. Res. Health Sci. 2017, 6, 76–84. [Google Scholar]
  60. Chen, Z.; Zwiggelaar, R. Segmentation of the breast region with pectoral muscle removal in mammograms. Med. Image Underst. Anal. 2010, 71–76. [Google Scholar]
  61. Suckling, J.; Parker, J.; Dance, D.; Astley, S.; Hutt, I.; Boggis, C.; Ricketts, I.; Stamatakis, E.; Cerneaz, N.; Kok, S.; et al. The Mammographic Image Analysis Society Digital Mammogram Database; International Congress Series; Exerpta Medica: Bridgewater, NJ, USA, 1994; Volume 1069, pp. 375–378. [Google Scholar]
  62. Liu, L.; Fieguth, P.; Guo, Y.; Wang, X.; Pietikäinen, M. Local binary features for texture classification: Taxonomy and experimental study. Pattern Recognit. 2017, 62, 135–160. [Google Scholar] [CrossRef]
  63. Rampun, A.; Scotney, B.W.; Morrow, P.J.; Wang, H.; Winder, J. Breast Density Classification Using Local Quinary Patterns with Various Neighbourhood Topologies. J. Imaging 2018, 4, 14. [Google Scholar] [CrossRef]
  64. Barkan, O.; Weill, J.; Wolf, L.; Aronowitz, H. Fast high dimensional vector multiplication face recognition. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Austrailia, 1–8 December 2013; pp. 1960–1967. [Google Scholar]
  65. Hall, M.A. Correlation-Based Feature Selection for Machine Learning. Ph.D. Thesis, University of Waikato, Hillcrest, New Zealand, 1999. [Google Scholar]
  66. Jabid, T.; Kabir, M.H.; Chae, O. Gender classification using local directional pattern (LDP). In Proceedings of the 2010 20th International Conference on Pattern Recognition (ICPR), Istanbul, Turkey, 23–26 August 2010; pp. 2162–2165. [Google Scholar]
  67. Tahmassebi, A.; Pinker-Domenig, K.; Wengert, G.; Helbich, T.; Bago-Horvath, Z.; Meyer-Baese, A. Determining the importance of parameters extracted from multi-parametric mri in the early prediction of the response to neo-adjuvant chemotherapy in breast cancer. Med. Imaging 2018. [Google Scholar] [CrossRef]
  68. Nguyen, H.T.; Caplier, A. Elliptical local binary patterns for face recognition. In Proceedings of the Asian Conference on Computer Vision, Daejeon, Korea, 5–6 November 2012; pp. 85–96. [Google Scholar]
  69. Stąpor, K. Evaluating and Comparing Classifiers: Review, Some Recommendations and Limitations. In Proceedings of the International Conference on Computer Recognition Systems, Polanica Zdroj, Poland, 22–24 May 2017; pp. 12–21. [Google Scholar]
  70. Tahmassebi, A.; Wengert, G.J.; Helbich, T.H.; Bago-Horvath, Z.; Alaei, S.; Bartsch, R.; Dubsky, P.; Baltzer, P.; Clauser, P.; Kapetas, P.; et al. Impact of machine learning with multiparametric magnetic resonance imaging of the breast for early prediction of response to neoadjuvant chemotherapy and survival outcomes in breast cancer patients. Investig. Radiol. 2019, 54, 110–117. [Google Scholar] [CrossRef]
  71. Duda, R.O.; Hart, P.E. Pattern Classification and Scene Analysis; A Wiley-Interscience Publication; Wiley: New York, NY, USA, 1973. [Google Scholar]
  72. Bishop, C.M. Neural Networks for Pattern Recognition; Oxford University Press: Oxford, UK, 1995. [Google Scholar]
  73. Bradley, A.P. The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognit. 1997, 30, 1145–1159. [Google Scholar] [CrossRef] [Green Version]
  74. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd Acm Sigkdd International Conference On Knowledge Discovery and Data Mining, San Francisco, CA, USA, 24–27 August 2016; pp. 785–794. [Google Scholar]
Figure 1. Example mammographic tissue types: (a) fatty, (b) glandular, and (c) dense tissue.
Figure 1. Example mammographic tissue types: (a) fatty, (b) glandular, and (c) dense tissue.
Jimaging 05 00024 g001
Figure 2. ROI extraction from the fibro-glandular disk region.
Figure 2. ROI extraction from the fibro-glandular disk region.
Jimaging 05 00024 g002
Figure 3. LBP pattern for (a) L B P P , R = 2 ( x c , y c ) , (b) E L B P P , R 1 = 3 , R 2 = 2 ( x c , y c ) , (c) M-ELBPP, R 1 = 3 , R 2 = 2 ( x c , y c ) where P = 8.
Figure 3. LBP pattern for (a) L B P P , R = 2 ( x c , y c ) , (b) E L B P P , R 1 = 3 , R 2 = 2 ( x c , y c ) , (c) M-ELBPP, R 1 = 3 , R 2 = 2 ( x c , y c ) where P = 8.
Jimaging 05 00024 g003
Figure 4. Effect of horizontal and vertical M-ELBP on mammographic RoI (a) Original Image (b) M-ELBP pattern for M-ELBP 8 , 7 , 4 at θ = 0 (c) M-ELBP pattern for M-ELBP 8 , 4 , 7 θ = 90 .M ELBP PATTERNS
Figure 4. Effect of horizontal and vertical M-ELBP on mammographic RoI (a) Original Image (b) M-ELBP pattern for M-ELBP 8 , 7 , 4 at θ = 0 (c) M-ELBP pattern for M-ELBP 8 , 4 , 7 θ = 90 .M ELBP PATTERNS
Jimaging 05 00024 g004
Figure 5. Summary of ROI selection, feature extraction and classification using ELBP variants.
Figure 5. Summary of ROI selection, feature extraction and classification using ELBP variants.
Jimaging 05 00024 g005
Figure 6. ROC curve for M-ELBP for a Bayesian Classifier.
Figure 6. ROC curve for M-ELBP for a Bayesian Classifier.
Jimaging 05 00024 g006
Figure 7. Classification accuracy as a function of descriptor size (N indicates a distinct run).
Figure 7. Classification accuracy as a function of descriptor size (N indicates a distinct run).
Jimaging 05 00024 g007
Figure 8. Area under ROC curve A z as a function of descriptor size (N indicates a distinct run).
Figure 8. Area under ROC curve A z as a function of descriptor size (N indicates a distinct run).
Jimaging 05 00024 g008
Figure 9. Classification accuracy as a function of ROI size (N indicates a distinct run).
Figure 9. Classification accuracy as a function of ROI size (N indicates a distinct run).
Jimaging 05 00024 g009
Figure 10. Area under ROC curve ( A z ) as a function of ROI size (N indicates a distinct run).
Figure 10. Area under ROC curve ( A z ) as a function of ROI size (N indicates a distinct run).
Jimaging 05 00024 g010
Table 1. A summary of existing approaches on breast density classification indicating authors, years of publication, techniques, experimental data for test, and classification.
Table 1. A summary of existing approaches on breast density classification indicating authors, years of publication, techniques, experimental data for test, and classification.
AuthorFeaturesClassifiersDatasetNumber of ImagesResultDensity ClassYear
Karssemeijer [40]Grey level histogramsKNNNijmegen61565.0%BIRADS I-IV1998
Blot and Zwiggelaar [41]Texture
GLCM
KNNMIAS26565.00%Fatty, Glandular and Dense2001
Bovis and Singh [32]SGLDANNDDSM37771.40%BIRADS I-IV2002
Petroudi et al. [42]Textons based on MR8
filters
KNNOxford13275.75%2 Class-Fatty and Dense2003
Zwiggelaar et al. [24]Gray level histogramPCA +KNNMIAS31271.5%Fatty, Glandular and Dense2005
Oliver et al. [29]Morphological
Features
KNN


Decision Tree
MIAS


DDSM
270


300 R-MLO
67.0%
73.0%

47.0%
BIRADS I-IV2005
Muhimmah and Zwiggelaar [23]Multiresolution histogram
Features
DAG-SVMMIAS32177.57%Fatty, Glandular and Dense2006
Oliver et al. [6]TextureKNN+SFS (sequential
forward selection (SFS))
MIAS

DDSM
322

831
86.0%

77.0%
BIRADS I-IV2008
Subashini et al. [43]Statistical featuresSVMMIAS4395.55%Fatty, Glandular and Dense2010
Oliver et al. [44]Connected density clusters
taking the spatial information
LDA-PCADDSM83179.0%Fatty and dense2010
Zwiggelaar [45]LGAKNNMIAS32264.0%BIRADS I-IV2010
Liu et al. [46]StatisticalSVMTianjin Tumor
Hospital
8886.40%BIRADS I-IV2011
Tzikopoulos et al. [47]Intensity basedSVMMIAS32285.70%Fatty, Glandular and Dense2011
Chen et al. [48]LBP
LGA
BIF
Textons
Topographic
KNN,
Bayesian
MIAS32259.0%
72.0%
70.0%
75.0%
76.0%
BIRADS I-IV2011
Qu et al. [49]UnknownE-FELM
(Evolutionary Fuzzy
Extreme Learning Machine)
MIAS32272.6%BIRADS I-IV2011
Muštra et al. [30]GLCMKNN+Naïve BayesianKBD-FER,
University Hospital Dubrava,
Zagreb
14479.30%2 Class-Fatty and Dense2012
Muštra et al. [30]GLCMKNN+Naive BayesianMIAS14482.0%2 class-Fatty, Dense2012
He et al. [28]TextureBinary model based Bayes ClassifierMIAS32278.0%BIRADS I-IV2012
Rampun et al. [50]LTPSVMMIAS32282.3%BIRADS I-IV2017
George et al. [51]M-ELBP
LBP
BayesianMIAS32174.0%
66.5%
BIRADS I-IV2018
Table 2. Confusion matrix for automatic tissue classification using Local Binary Pattern (LBP) descriptor for ROI size 256 × 256 .
Table 2. Confusion matrix for automatic tissue classification using Local Binary Pattern (LBP) descriptor for ROI size 256 × 256 .
Automatic Classification
Truth DataFattyGlandularDense
Fatty86191
Glandular167315
Dense73866
Table 3. Confusion matrix for automatic tissue classification using the Elliptical Local Binary Pattern (ELBP) descriptor for ROI size 256 × 256 and R = { 4 , 7 } and P = 8 .
Table 3. Confusion matrix for automatic tissue classification using the Elliptical Local Binary Pattern (ELBP) descriptor for ROI size 256 × 256 and R = { 4 , 7 } and P = 8 .
Automatic Classification
Truth DataFattyGlandularDense
Fatty86200
Glandular116825
Dense32286
Table 4. Confusion matrix for automatic tissue classification using the Uniform-Elliptical Local Binary Pattern (uniform-ELBP) descriptor for ROI size 256 × 256 and R = { 4 , 7 } and P = 8 .
Table 4. Confusion matrix for automatic tissue classification using the Uniform-Elliptical Local Binary Pattern (uniform-ELBP) descriptor for ROI size 256 × 256 and R = { 4 , 7 } and P = 8 .
Automatic Classification
Truth DataFattyGlandularDense
Fatty86191
Glandular116924
Dense22782
Table 5. Confusion matrix for automatic tissue classification using the Mean Elliptical Local Binary Pattern (M-ELBP) descriptor for ROI size 256 × 256 and R = { 4 , 7 } and P = 8 .
Table 5. Confusion matrix for automatic tissue classification using the Mean Elliptical Local Binary Pattern (M-ELBP) descriptor for ROI size 256 × 256 and R = { 4 , 7 } and P = 8 .
Automatic Classification
Truth DataFattyGlandularDense
Fatty92140
Glandular137615
Dense02388
Table 6. Confusion matrix for automatic tissue classification using the Local Directional Pattern (LDP) descriptor for ROI size 256 × 256 .
Table 6. Confusion matrix for automatic tissue classification using the Local Directional Pattern (LDP) descriptor for ROI size 256 × 256 .
Automatic Classification
Truth DataFattyGlandularDense
Fatty92131
Glandular107420
Dense33276
Table 7. Classification accuracy results by various classifiers for LBP variants for ROI size 256 × 256  pixels for 10-run 10-fold cross validation. (* indicates that the CA% are not significantly different from the best result using an unpaired t test ( p = 0.05 ) level).
Table 7. Classification accuracy results by various classifiers for LBP variants for ROI size 256 × 256  pixels for 10-run 10-fold cross validation. (* indicates that the CA% are not significantly different from the best result using an unpaired t test ( p = 0.05 ) level).
ClassifierLBPELBPUniform ELBPM-ELBPLDP
Bayesian Network 69.44 ± 0.92 75.41 ± 1.05 72.86 ± 1.06 77.38 ± 1.06 74.92 ± 0.67
KNN 69.34 ± 1.09 70.43 ± 1.03 71.46 ± 1.41 75.46 ± 0.78 75.95 ± 0.96
SVM 67.07 ± 0.75 73.11 ± 0.76 73.29 ± 0.64 74.42 ± 0.92 74.17 ± 0.52
Random Forest 69.19 ± 1.31 73.26 ± 1.08 73.04 ± 0.97 73.20 ± 1.36 75.91 ± 1.01
Table 8. Area under ROC ( A z ) classification results for LBP variants using ROI size 256 × 256 pixels for a 10 run 10 fold cross validation. * indicates that the CA% are not significantly different from the best result using an unpaired t test ( p = 0.05 ) level).
Table 8. Area under ROC ( A z ) classification results for LBP variants using ROI size 256 × 256 pixels for a 10 run 10 fold cross validation. * indicates that the CA% are not significantly different from the best result using an unpaired t test ( p = 0.05 ) level).
ClassifierLBPELBPUniform ELBPM-ELBPLDP
Bayesian Network 0.89 ± 0.006 0.90 ± 0.001 0.89 ± 0.003 0.92 ± 0.000 0.89 ± 0.004
KNN 0.86 ± 0.005 0.85 ± 0.003 0.87 ± 0.005 0.89 ± 0.005 0.89 ± 0.002
SVM 0.78 ± 0.004 0.83 ± 0.005 0.84 ± 0.003 0.85 ± 0.005 0.84 ± 0.002
Random Forest 0.86 ± 0.004 0.89 ± 0.004 0.89 ± 0.004 0.89 ± 0.004 0.90 ± 0.003

Share and Cite

MDPI and ACS Style

George, M.; Zwiggelaar, R. Comparative Study on Local Binary Patterns for Mammographic Density and Risk Scoring. J. Imaging 2019, 5, 24. https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging5020024

AMA Style

George M, Zwiggelaar R. Comparative Study on Local Binary Patterns for Mammographic Density and Risk Scoring. Journal of Imaging. 2019; 5(2):24. https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging5020024

Chicago/Turabian Style

George, Minu, and Reyer Zwiggelaar. 2019. "Comparative Study on Local Binary Patterns for Mammographic Density and Risk Scoring" Journal of Imaging 5, no. 2: 24. https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging5020024

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop