NguyenHai Thanh
NguyenCham Ngoc Thi
PhanThao Minh Nguyen
DaoTinh Cong
-
(College of Information and Communication Technology (CICT), Can Tho University, Vietnam
nthai@cit.ctu.edu.vn, {ntncham0109, pnmthaoct, dcongtinh}@gmail.com
)
Copyright © The Institute of Electronics and Information Engineers(IEIE)
Keywords
Disease prediction, Model explanation, Chest X-ray (CXR) images, Artificial intelligence, Computer-aided diagnosis, Pleural effusion
1. Introduction
Pleural effusion is an increase in fluid between the two membranes that cover the
lungs the chest wall, indicating an imbalance between pleural fluid formation and
removal. The amount of fluid is regularly spread thinly over the visceral and parietal
pleura that acts as a lubricant between the two membranes. Pleural effusion is a significant
increase in the amount of pleural fluid. Determining the cause is important for treating
to treat pleural effusion suitably. Some of the most ordinary symptoms of pleural
effusion are chest pain and pleurisy. Pleural effusion has many causes, such as heart
and kidney failure, low albumin levels in the blood, infections, pulmonary embolism,
and malignancies. Many cases show no symptoms but are discovered during a physical
examination or are detected by Chest X-ray (CXR) imaging, which is the most advantageous
method for making a diagnosis.
Coronavirus disease (COVID-19) is an infectious lung disease that has serious consequences
for human health globally. According to the WHO [1], more than 37 million coronavirus infections have been detected worldwide, with approximately
one million deaths. This highlights the stage of emergency caused by the coronavirus
and the impact on the body of the infected patients. The lungs can be damaged directly
after a coronavirus infection. Radiologists can diagnose pleural effusion based on
the CXR images. Pleural effusion is one of the most prominent manifestations of COVID-19.
The Guardian reports that once the lungs become infected, the small air sacs can fill
with cells and fluid, blocking oxygen flow.
The symptoms of pleural effusion manifest as shortness of breath while lying down.
The degree of dyspnea depends on the amount of fluid in the pleural space. A rapid
increase in the amount of pleural fluid causes acute difficulty in breathing. The
clinical examination reveals syndromes, such as decreased vocal fibrillation, decreased
alveolar fibrillation, and percussion. The manifestation of the disease will differ
according to the cause of the disease. Pleural effusion has many causes and can be
benign, but there are many cases of malignant pleural effusion that can cause death.
The typical causes of hydrocephalus include pleural effusion, pleurisy, cancer, previous
systemic history of heart failure, malnutrition, cirrhosis, and other chronic diseases.
Pleural effusion can often be diagnosed based on the typical symptoms in a physical
examination and the patient's history. The doctor listens carefully to the lungs using
a stethoscope and measures the pulse and blood pressure. An x-ray of the lungs can
also be taken. An X-ray can identify the position and detect the quantity of pleural
effusion because they normally occur as whitish areas at the lung base and appear
unilaterally or bilaterally. If a patient lies on their side for a few minutes, most
pleural effusions will move and layer out along that side of the chest space, which
was positioned downward because of the effects of gravity. This movement of the pleural
effusion can be seen on a CXR image taken with the patient lying on their side (a
lateral decubitus X-ray).
Currently, radiologists perform diagnostic imaging in hundreds of cases. The diagnosis
process takes significant time, but errors may also increase, which affects the diagnosis
quality. In addition, some diseases, such as pleural effusion, pleural effusion, and
lung cancer, have similar symptoms. Therefore, it is easy to confuse and difficult
to diagnose accurately using traditional methods. As a result, there might be many
more false negatives that may lead to a deterioration of the patient's health condition,
possibly even death. Therefore, the deep learning model helps doctors make more accurate
decisions quickly in the pleural effusion diagnosis process, which can allow the development
of the earliest treatment plan for the patient and reduce the risk of death.
Deep learning is a promising approach in the medicine domain. Considerable research
has been carried out to examine chest and lung diseases using machine learning. Deep
learning innovations have brought promising improvements in medical technology, such
as discovering antibiotics, analyzing electronic health records, or predicting medical
events. Several deep learning approaches in medical imaging have been proposed for
disease detection and diagnosis. On the other hand, this technology does not mean
the ultimate replacement of physicians, particularly radiologists. Instead, it helps
radiologists diagnose patients more accurately. Because belief is fundamental, people
can decide to act on a prediction or whether to implement a new model. Providing more
model details is important to turn an unreliable model or prediction into a reliable
one. Although machine learning is at the core of many recent advances in science and
technology, the important human role is an overlooked aspect of this field. Whether
people are directly using machine learning classifiers as tools or deploying models
into the products, an important concern remains if they believe in a model or prediction
or whether they will use it with greater confidence.
The performance of intelligent systems depends on the size of the training data. One
of the greatest drawbacks is that deep learning models are not generalizing well,
and the amount of public data for model training could be limited for some diseases.
Recently, several approaches have been made to solve this problem. A previous study
[2] proposed the deep k-nearest neighbor (kNN) [3]. The research improved the learning strategy to unify the kNN classification and
feature extraction on multiple small-class and class-imbalanced medical image datasets.
Therefore, this method has solved the non-parametric characteristics of kNN while
associated with a feature extractor. In addition, most medical images lack annotated
data public using the supervised methods in deep learning models. Pooch et al. [4] examined some novel semi-supervised classification methods using the ChestX-ray14
dataset. They were based on pseudo-labeling and consistency regularization to produce
multi-label classification and use a modern model architecture in CXR image classification.
In recent years, deep learning techniques have emerged as state-of-the-art methods
to classify medical images. Tuning the hyperparameters of the networks trained and
using deep learning libraries is difficult. Inés et al. [5] investigated an Automated Machine Learning (AutoML) approach to solve these issues.
Their approach associated transfer learning with a new semi-supervised learning method
when several annotated images were available.
This paper proposes a Convolutional Neural Network architecture that compares with
some pre-trained models to assist in the control and early detection of pleural effusion
from CXR images. Moreover, applying the model explanation algorithm to assist in medical
diagnosis is a promising method. Because deep learning models still function as a
black box, interpreting the generated output is a challenge.
· A CNN model that can distinguish between normal samples and pleural effusion patients
efficiently was proposed.
· This study compared the pleural effusion classifi-cation performance of various
scenarios, such as using the architecture with only a Fully Connected Layer (FC) and
the changes in the quantity of convolutional layers that were explored from one to
three convolutional layers with 32 to 64 filters for each convolutional layer.
· The use of two popular activate functions, such as Sigmoid and Softmax, was investigated
to improve accuracy.
· Local Interpretable Model-Agnostic Explanations (LIME) [6] for visual explanations and interpretation purposes in the pleural effusion diagnosing
process were examined.
The remainder of this paper is as follows. Section 2 presents related work. Section
3 reports methods to solve the problem, including to support pleural effusion with
a combination of CNNs and algorithms to provide explanations from the output of the
CNNs. Section 4 presents the experimental results of the proposed method. The conclusions
are reported in Section 5.
2. Related Work
One of the most challenges for doctors is to make an exact decision and reduce the
time to detect diseases. Therefore, the use of deep learning algorithms in the analysis
and processing of biomedical images has provided more satisfactory results. Advances
in deep learning and big data have enabled artificial intelligence to replace humans
gradually. Several approaches have been introduced to support the detection of pleural
effusion using CXR images. Some of these methods use feature extraction techniques
along with a machine learning algorithm as a classification technique, whereas others
use deep learning techniques for feature extraction and classification.
Maduskar et al. [7] presented pre-trained models, namely DenseNet121 and MobileNetV2, using the CheXpert
[8] to support the successful diagnosis of pleural effusion. In tuberculous pleural effusion
(TPE) diagnostic models, a previous study [9] proposed four machine learning algorithms (MLAs), such as logistic regression, k-nearest
neighbors (KNN) [3], support vector machine (SVM) [10], and random forest (RF) [11]. The random forest reached the best result with a sensitivity and specificity of
89.1% and 93.6%, respectively. TPE using RF was produced improved the results, lowered
the computation cost, and resulted in a diagnosis process. Shu et al. [12] constructed a logistic regression model for the diagnosis of TPE with a sensitivity
and specificity of 82.9% and 86.7%, respectively. CheXNeXt is a complex neural network
that detects 14 different diseases simultaneously, including pleural effusion [13]. Cicero et al. [14] examined the diagnosis of pleural effusion using two popular CNN architectures that
have shown excellent results on the ImageNet dataset: AlexNet [15] and GoogLeNet [16]. Bustomi et al. [17] combined image processing analysis with a Na\"{i}ve Bayes Classifier (NBC) to overcome
the subjective problems when radiologists analyzed the medical image. The five image
features were used as predictors to determine the lung class, i.e., normal lungs class,
pleural effusion class, and lung cancer class. The NBC method showed an accuracy of
70%.
Sunny et al. [18] used Chexpert datasets to compare the performance of VGG16 and InceptionV3 to predict
pleural effusion. The best model was the InceptionV3 multi-class model, in which the
additional labels were provided to help detect pleural effusion using CXR images.
Behzadi-khormouji et al. [19] proposed two types of deep convolutional neural networks to detect lung abnormalities
through CXR images with certain accuracy. In the first type, the state-of-the-art
Deep Convolutional Neural Networks (DCNNs), such as VGG16 [20], DenseNet121 [21], and pyramid convolutional structure [22], were applied to the Pediatric Chest X-ray dataset. The VGG16 and DenseNet121 were
pre-trained on the ImageNet dataset, including 1.2 million color images [23]. In the second type, an efficient architecture design of deep convolutional neural
networks called ChestNet was proposed. This model was constructed based on the context
in the datasets under study.
3. The Proposed Scheme
This study examined the CheXpert dataset [24] (the smaller data set size was 11GB) to propose an automated CXR interpretation framework
(Table 1). The dataset was collected from Stanford Hospital, which was performed between October
2002 and July 2017 in inpatient and outpatient centers, along with their associated
radiology reports. Many board-certified radiologists annotated the dataset manually.
Several samples from the original dataset were modified to improve the validation
accuracy of the study. All images with the same resolution as the original image size
were downsized to 64 x 64 for the prediction task.
Table 1. Experiment Parameters with the size of the training and test sets.
Class
|
Training
|
Testing
|
Total
|
Positive
|
68471
|
29344
|
97815
|
Negative
|
24777
|
10619
|
35396
|
4. Method
The CNN model was designed from the beginning. The Keras library with TensorFlow was
applied to use the models. The LIME method was used to explain the predictions. The
trust of the model was measured using AUC and accuracy. The proposed CNN model with
various models reported elsewhere. The research methodology is presented in Fig. 1. Three Chest X-ray images were fetched into the proposed framework. In the training
phase, the Chest X-ray image data will go through pre-processing. This paper proposes
the CNN model and pre-trained models, such as ResNet, InceptionResnetV2, Xception.
Finally, the CNN model, which was designed from scratch, was selected, and LIME was
applied to visualize the explanations that helped highlight the effusion in the lungs.
The output showed the results and probability of three Chest X-ray images after diagnosis.
Fig. 1. Proposed Architecture for the Pleural Effusion Diagnosis Support System.
4.1 Convolutional Neural Network Architectures
Several shallow CNN architectures were investigated, and the efficiency of each architecture
on the CXR images for Pleural Effusion classification was evaluated. The number of
convolutional layers increased from 1 to 3, while the number of filters per convolutional
layer was 32 and 64. Two popular activation functions consisting of Softmax and Sigmoid
were examined in such a configuration.
From the experimental results, the architecture shown in Fig. 2 was used to apply model explanation in the next steps for Pleural Effusion diagnosis.
The input image size was 64 $\times ~ $64 with 32 filters, and ReLU activates the
function for training. Three convolution layers with up to 256 features per class
were added to extract the object better and improve the result. In addition, the max-pooling
layer was used after the convolutional layer with 2 $\times $ 2 dimensions. The learning
rate was 0.000001 to decrease the overfitting problem. The cost was compared using
the Adam optimizer with Categorical Crossentropy. The softmax function with 20 epochs
was suggested. The proposed algorithm focused primarily on the classification of images,
which presents two outputs, including pleural effusion and normal.
Fig. 2. Architecture of shallow Convolutional Neural Network for classifying the Pleural Effusion.
4.2 Explainable Deep Learning using LIME
Ribeiro et al. proposed Local Interpretable Model-Agnostic Explanations (LIME) [6] in 2016. The primary idea was that it is easier to interpret a locally approximated
black-box model using a simpler glass-box model. LIME is a popular approach to enhance
the interpretability and visualize the explainability of black-box Deep Learning (DL)
algorithms. LIME frequently produces an explanation for a single prediction by any
model by training a simpler interpretable model, e.g., linear classifier around the
prediction by producing simulated data around the instance by random perturbation
and obtaining the feature importance by applying some form of feature selection [25]. Therefore, LIME provides a patient-specific explanation for a given classification
that enhances the possibility of any classifier to supply efficient usage in clinical
conditions. LIME is clinically relevant with a very high concordance with explanations
provided by physicians, according to the evaluation by field experts.
LIME can be explained with a local model that shows the axes of the values of two
continuous explanatory variables. The colored regions correspond to the decision areas
for a binary classifier, i.e., combinations of the values of the two variables. The
model classifies the observation to one of the two classes. The black crosses present
the observation of interest. The dots show the artificial data generated to understand
the local behavior of the complex model around the point of interest. The dashed line
indicates a simple linear model fitted to construct the local approximation in Fig. 3. The local behavior of the black-box model can be explained by a simple model around
the observation of interest. The simpler model serves as a local explainer for the
more complex model.
The features and interpretable data representations, which use an understandable representation
to humans, can be separated. For example, the text classifier is a binary vector corresponding
to the presence or absence of a word, or image classification has an interpretable
representation that can be a binary vector that can present the presence or absence
of a contiguous patch of similar pixels.
Formula 1 shows the LIME method that can produce an explanation by the following:
· $x~ \in R$ is designated as a prior representation of an explained instance.
· $g~ \in G,$where $G$ is a class of capability interpretable models as an explanation.
[26]
· $\Omega(g)$ is a measure of the complexity of the explanation $g~ \in G$.
· $f\left(x\right)$ is the probability that $x$ belongs to a certain class.
· $\pi _{x}\left(z\right)$ is a proximity measure to identify locality around x between
$z$ to $x$.
· $L\left(f,g,\pi _{x}\right)$ measure $g$ approximating $f$ in the locality defined
by $\pi _{x}$. $L\left(f,g,\pi _{x}\right)$ is needed to enhance the interpretability
and local faith, and $\Omega(g)$ should be low to be explainable by humans.
Fig. 4 gives an example of the explanations of an image by LIME. The original image is shown
in Fig. 4(a), and 100 \textit{superpixels} are identified in Fig. 4(b). Figs. 4(c) and (d) present an explanation for the poodle class, goose class, respectively.
In medical systems, the basis for providing analytical results also needs to be provided
to implement processes correctly based on AI analysis. For example, an AI system can
analyze whether a patient has cancer and show why it came to such conclusions. Furthermore,
it is almost impossible to design a safety model without understanding how it works.
Increasing the interpretability of the AI systems helps analyze their logical vulnerabilities
or data blind spots, thereby improving the security of the AI systems. Malhi et al.
[28] proposed an explainable machine-learning tool that can be used potentially for decision
support in medical image analysis cases, particularly on in-vivo gastral images obtained
from capsule endoscopy. They applied LIME to visualize the explanations to help health
experts trust the black box predictions. The method presented elsewhere [29] highlights the need for explainable AI in the medical field to help medical experts
make interpretable, transparent, and easy to understand decisions that are easy to
deploy, have a fast computation time, and provide short, selective explanations that
suit busy physicians.
LIME is widespread across different fields, especially the medical domain. The interpretation
is generated by initiating the original model (black box) closer to the explained
model. The explainable model was produced based on the perturbation of the original
example from which the chosen components were excluded, e.g., removing words or hiding
a part of an image dividing the image into interpretable parts. The image was altered
in several ways to enable or disable some of the interpretable components. For each
of the generated images, the probability on it was obtained. To learn a linear model
weighed locally on the dataset, it is important to understand how it conducts. The
original model, in which the components with the highest weights are presented, are
explained. Because the true positive predictions showed the visual explanations with
the largest weights, the LIME identifies the relevant superpixels, and more true positive
explanations are produced. LIME provides visual explanations that can interpret examples
with fewer essential features. The selected feature in a good explanation meets the
need because people prefer sparse explanations. LIME can predict the change in probability
of a label based on the change in an input property, which provides a good contrast
explanation.
Fig. 3. Idea behind the LIME approximation with a local glass-box model.
Fig. 4. Example of the Explanations of an Image Obtained by the VGG16 network with ImageNet weights for the half-duck and half-horse image.
5. Results
5.1 Settings
This study used Ubuntu 18.04 server with a CPU of 20 cores and 64GB of RAM that is
a robust hardware environment to test the proposed algorithm. The experiment code
for all the models is available through python that supports several libraries, such
as Keras, TensorFlow, Matplotlib, and Pandas.
5.2 Experimental Results
Table 2 lists the experimental results using the proposed CNN architecture with various scenarios.
The architecture with three layers containing 32 filters and optimized Adam function
showed the highest accuracy and AUC of $\textbf{85.96%}$ and $\textbf{93.19%,}$ respectively.
As observed, the performance with Softmax used for activation can produce better results
compared to Sigmoid. With the image dataset investigated, increasing the number of
convolutional layers may be a better solution for improving the performance rather
than increasing the filters of each convolutional layer. Table 3 compares other conventional methods with the proposed method.
The proposed method produced a better result than the existing methods. Table 4 lists the detailed studies mentioned by the authors.
Table 2. Brief synthetic of our experimental results.
Method
|
Activation
|
ACC
(%)
|
AUC (%)
|
Time
(s)
|
CNN-L1F32
|
Softmax
|
82.89
|
90.56
|
582
|
CNN-L2F32
|
84.18
|
91.83
|
1232
|
CNN-L3F32
|
85.96
|
93.19
|
1840
|
CNN-L1F64
|
82.49
|
90.38
|
591
|
CNN-L2F64
|
84.96
|
92.37
|
1243
|
CNN-L3F64
|
85.59
|
92.87
|
2358
|
CNN-L1F32
|
Sigmoid
|
82.92
|
86.83
|
577
|
CNN-L2F32
|
82.84
|
86.56
|
576
|
CNN-L3F32
|
85.12
|
89.75
|
1255
|
CNN-L1F64
|
84.87
|
89.31
|
2114
|
CNN-L2F64
|
85.96
|
90.61
|
1853
|
CNN-L3F64
|
86.10
|
90.86
|
3544
|
Table 3. Comparison of other conventional methods with the proposed method.
Method
|
ACC (%)
|
AUC (%)
|
Time
(s)
|
Resnet
|
76.37
|
82.76
|
13224
|
Xception
|
81.92
|
89.78
|
3831
|
InceptionResnetV2
|
83.00
|
90.74
|
4235
|
CNN-L3F32
(Proposed method)
|
85.96
|
93.19
|
1840
|
Table 4. Comparison of other conventional methods with the proposed method.
Method
|
ACC (%)
|
AUC (%)
|
Time
(s)
|
Wall et al. [27]
|
-
|
92.00
|
-
|
Singh [30]
|
-
|
78.20
|
-
|
CNN-L3F32
(Proposed method)
|
85.96
|
93.19
|
1840
|
5.3 Comparative of Explanation of the Results using LIME
The proposed method was evaluated using LIME technology (Fig. 5). Fig. 5(a) presents the original CXR image. Fig. 5(b) explains LIME regarding identifying pleural effusion. The yellow highlight in Fig. 5(b) indicates the region of fluid causing the disease. Fig. 5(c) is a visualization of Fig. 5(b), which distinguishes the normal (green area) and abnormal (red area).
5.4 Performance Analysis
These measurements require True Positive (TP), False Positive (FP), True Negative
(TN), and False Negative (FN) values. A TP presents the positive results of pleural
effusion that were classified correctly by the system. TN is the amount of negative
or normal data that were classified correctly by the system. FN is the amount of negative
data that were classified incorrectly by the system. FP is the amount of positive
data that were classified incorrectly by the system. Fig. 6 presents a confusion matrix of the TP, FP, TN, and FN measurements.
Fig. 7 presents the statistics of the training process with different metrics, such as Accuracy
and AUC after 20 epochs. Both line graphs provide information on the proposed method
throughout two metrics: Accuracy and AUC. The Training and Validation Accuracy showed
a rapid increase in the performance until 20 epochs. On the other hand, at some epoch,
test accuracy reached the saturated value. Similarly, the other chart also shows the
proportion of the AUC that rises significantly at the initial stage and gains a stability
value of more than 93%.
Fig. 5. Explaining the prediction of pleural effusion in positive patients.
Fig. 6. Confusion matrix for our proposed method.
Fig. 7. Experimental results of the proposed architecture in various evaluation metrics, including Accuracy, AUC.
6. Conclusion
Pleural effusion is generally diagnosed with CXR images. A CXR image contains abnormalities
as consolidation resulting from pleural effusion. In this study, a method was proposed
to detect imaging features of pleural effusion efficiently using convolutional neural
networks. The proposed approach is expected to support doctors for pleural effusion
diagnosis.
A CheXpert dataset of chest radiographs, which have uncertainty labels and radiologist-labeled
reference standard evaluation sets, was evaluated. A variety of CNN architectures
were examined to discriminate between normal samples and pleural effusion images.
LIME was leveraged to provide explanations from the trained model on the CXR images.
This promising technology can help explain many tasks related to assessing the trust
for effective human interactions with machine learning systems and obtaining more
informative predictions. The proposed method will provide growth and acknowledgment
of CXR interpretation models to improve healthcare in the future.
REFERENCES
World Health Organization. , Coronavirus disease Situation Report-150., Retrieved
from https://www.who.int/emergencies/diseases/novel-coronavirus-2019/situation-reports
Zhuang J., Cai J., Wang R., Zhang J., Zheng WS., 2020, Deep kNN for Medical Image
Classification., In: Martel A.L. et al. (eds) Medical Image Computing and Computer
Assisted Intervention - MICCAI 2020. MICCAI 2020. Lecture Notes in Computer Science.
Springer, Cham., Vol. 12261
Xinran Li, Chenhui Xiang., 2012, Correlation-based K-nearest neighbor algorithm.,
2012 IEEE International Conference on Computer Science and Automation Engineering,
Beijing, pp. 185-187
Pooch E.H.P., Ballester P., Barros R.C., MIL3ID 2020, Semi-supervised Classification
of Chest Radiographs., In: Cardoso J. et al. (eds) Interpretable and Annotation-Efficient
Learning for Medical Image Computing. IMIMIC 2020, MIL3ID 2020, LABELS 2020. Lecture
Notes in Computer Science Springer, Cham., Vol. 12446
Inés A., Domínguez C., Heras J., Mata E., Pascual. V., Biomedical image classification
made easier thanks to transfer and semi-supervised learning.
Marco Tulio Ribeiro , Sameer Singh , Carlos Guestrin. , 2016, Why should i trust you?:
Explaining the predictions of any classifier., In:Proceedings of the 22nd ACM SIGKDD
International Conference on Knowledge Discovery and Data Mining. ACM. 2016, pp. 1135-1144.
Maduskar P., Philipsen RH., Melendez J., Scholten E., Chanda D., Ayles H., Sánchez
CI., van Ginneken B., 2016 Feb., Automatic detection of pleural effusion in chest
radiographs., Med Image Anal., Vol. 28, pp. 22-32
Jeremy Irvin , Pranav Rajpurkar , Michael Ko , Yifan Yu , Silviana Ciurea-Ilcus ,
Chris Chute , Henrik Marklund , Behzad Haghgoo , Robyn Ball , Katie Shpanskaya , et
al. , 2019, Chexpert: A large chest radiograph dataset with uncertainty labels and
expert comparison., In Thirty-Third AAAI Conference on Articial Intelligence
Ren Z., Hu Y., 2019, , Identifying tuberculous pleural effusion using artificial intelligence
machine learning algorithms., Respir Res 20, Vol. 220, pp. -
Yujun Yang , Jianping Li , Yimei Yang , 2015, The research of the fast SVM classifier
method, 2015 12th International Computer Conference on Wavelet Active Media Technology
and Information Processing (ICCWAMTIP), Chengdu, pp. 121-124
Kern C., Klausch T., Kreuter F., 2019, . Tree-based Machine Learning Methods for Survey
Research., Survey research methods, Vol. 13, No. 1, pp. 73-93
Shu CC, Wang JY, Hsu CL, et al. , 2015, Diagnostic role of inflammatory and anti-inflammatory
cytokines and effector molecules of cytotoxic T lymphocytes in tuberculous pleural
effusion., Respirology., Vol. 20, No. 1, pp. 147-54
Rajpurkar P, Irvin J, Ball RL, 2018 Nov. 20, , Deep learning for chest radiograph
diagnosis: A retrospective comparison of the CheXNeXt algorithm to practicing radiologists.,
PLoS Med., Vol. 15, No. 11, pp. e1002686.-
Cicero M, Bilbily A, Colak E, Dowdell T, Gray B, Perampaladas K, 2017 May., Training
and Validating a Deep Convolutional Neural Network for Computer-Aided Detection and
Classification of Abnormalities on Frontal Chest Radiographs., Invest Radiol., Vol.
52, No. 5, pp. 281-287
Alex Krizhevsky , Ilya Sutskever , Geoffrey E. Hinton. , 2012, ImageNet classification
with deep convolutional neural networks., In Proceedings of the 25th International
Conference on Neural Information Processing Systems - Volume 1 (NIPS’12). Curran Associates
Inc. Red Hook, NY, USA, pp. 1097-1105
Szegedy C, Liu W, Jia Y, et al. , 2015, Going deeper with convolutions., In Proceedings
of the IEEE Conference on Computer Vision and Pattern Recognition., pp. 1-9
M. Arief Bustomi , Anifatul Faricha , Alfiana Ramdhan , Faridawati1 , 2018, Integrated
image processing analysis and Naïve Bayes Classifier method for lungs X-ray image
classification, ARPN Journal of Engineering and Applied Sciences. 2018, pp. 718-724
Sunny Mahesh , Eliana Marostica , ,Automated Classification of Pleural Effusion in
Chest X-rays., Harvard Medical School Department of Biomedical Informatics.
Hamed Behzadi-khormouji , Habib Rostami , Sana Salehi , Touba Derakhshande-Rishehri
, Marzieh Masoumi , Siavash Salemi , Ahmad Keshavarz , Ali Gholamrezanezhad , Majid
Assadi , Ali Batouli , 2019, Deep Learning, Reusable and Problem Based Architectures
for Detection of Consolidation on Chest X-Ray Images, Computer Methods and Programs
in Biomedicine
C.C. Ting , B.F. Wu , M.L. Chung , C.C. Chiu , Y.C. Wu , 2015, Visual contrast enhancement
algorithm based on histogram equalization, Sensors (Switzerland)., Vol. 15, pp. 16981-16999
Singh K., Vishwakarma D.K., Walia G.S., Kapoor R., 2016, Contrast enhancement via
texture region based histogram equalization, J. Mod. Opt., Vol. 63, pp. 1444-1450
Shen D., Wu G., Suk. H.-I., 2017, Deep Learning in Medical Image Analysis, Annu. Rev.
Biomed. Eng., Vol. 19, pp. 221-248
Shen D., Wu G., Suk. H.-I., 2017, Deep Learning in Medical Image Analysis, Annu. Rev.
Biomed. Eng., Vol. 19, pp. 221-248
JeremyIrvin , PranavRajpurkar , MichaelKo , YifanYu , SilvianaCiurea-Ilcus , ChrisChute
, HenrikMarklund , BehzadHaghgoo , RobynBall , KatieShpanskaya , JayneSeekins , DavidA.Mong
, SafwanS.Halabi , JesseK.Sandberg , RickyJones , DavidB.Larson , CurtisP.Langlotz
, BhavikN.Patel , MatthewP.Lungren , CheXpert: A Large Chest Radiograph Dataset with
Uncertainty Labels and Expert Comparison.
Muhammad Rehman Zafar , Naimul Mefraz Khan. , ,
DLIME: A Deterministic Local Interpretable Model-Agnostic Explanations Approach
for Computer-Aided Diagnosis Systems.
Wang F., Rudin. C., 2015, Falling rule lists., In Artificial Intelligence and Statistics
(AISTATS)
Wall Nathan , Palanisamy Muthu , Santerre John , 2019, Automated Pleural Effusion
Detection on Chest X-Rays, SMU Data Science Review:, Vol. 2, No. 2, pp. article 15
Malhi A., Kampik T., Pannu H. S., Madhikermi M., Främling K., 2019, Explaining Machine
Learning-based Classifications of in-vivo Gastral Images., In 2019 Digital Image Computing:
Techniques and Applications (DICTA) [8945986] IEEE.
Holzinger A., Biemann C., Pattichis C. S., Kell D. B., What do we need to build explainable
ai systems for the medical domain?
Singh R, Kalra MK, Nitiwarangkul C, Patti JA, Homayounieh F, Padole A, et al., 2018,
Deep learning in chest radiography: Detection of findings and presence of change.,
PLoS ONE 13(10): e0204155.
Author
Hai Thanh Nguyen (Email: nthai@cit.ctu.edu.vn) is a lecturer of College of Information
and Communi-cation Technology, Can Tho Univer-sity, Vietnam. He received his Engineering
degree in Informatics from Can Tho University, the master degree in Computer Science
and Engineering from National Chiao Tung University, Taiwan, and obtained the PhD
degree in Computer Science from Sorbonne University, France. His PhD thesis studied
approaches for disease prediction using metagenomic data. His current research includes
Bioinformatics, Health care system, Computer Vision, modeling of decisions, and recommender
system.
Cham Ngoc Thi Nguyen is a final year student in Information systems, the College
of Information and Communication Technology, Can Tho University. In 2018, she was
a short-term exchange student at King Mongkut's University of Technology North Bangkok.
Thao Minh Nguyen Phan is currently a final-year undergraduate student pursuing
Information System from Can Tho University, Can Tho, Vietnam. Her research interests
include deep learning, data science, recommender system and web programming.
Tinh Cong Dao is a senior pursuing Computer Science from Can Tho University, Can
Tho, Vietnam. His research interests include machine learning, deep learning, computer
vision, data mining and recommender system.