KRAS is a pathogenic gene frequently implicated in non-small cell lung cancer (NSCLC). However, biopsy as a diagnostic method has practical limitations. Therefore, it is important to accurately determine the mutation status of the KRAS gene non-invasively by combining NSCLC CT images and genetic data for early diagnosis and subsequent targeted therapy of patients. This paper proposes a Semi-supervised Multimodal Multiscale Attention Model (S2MMAM). S2MMAM comprises a Supervised Multilevel Fusion Segmentation Network (SMF-SN) and a Semi-supervised Multimodal Fusion Classification Network (S2MF-CN). S2MMAM facilitates the execution of the classification task by transferring the useful information captured in SMF-SN to the S2MF-CN to improve the model prediction accuracy. In SMF-SN, we propose a Triple Attention-guided Feature Aggregation module for obtaining segmentation features that incorporate high-level semantic abstract features and low-level semantic detail features. Segmentation features provide pre-guidance and key information expansion for S2MF-CN. S2MF-CN shares the encoder and decoder parameters of SMF-SN, which enables S2MF-CN to obtain rich classification features. S2MF-CN uses the proposed Intra and Inter Mutual Guidance Attention Fusion (I2MGAF) module to first guide segmentation and classification feature fusion to extract hidden multi-scale contextual information. I2MGAF then guides the multidimensional fusion of genetic data and CT image data to compensate for the lack of information in single modality data. S2MMAM achieved 83.27% AUC and 81.67% accuracy in predicting KRAS gene mutation status in NSCLC. This method uses medical image CT and genetic data to effectively improve the accuracy of predicting KRAS gene mutation status in NSCLC.
Lung cancer is specifically divided into non-small cell lung cancer (NSCLC) and small cell lung cancer. NSCLC accounts for approximately 85% of newly diagnosed lung cancers yearly [[
In recent years, researchers have used CT images to predict gene mutations based on traditional radiomics and machine learning. Song et al. [[
The radiomics and machine learning methods mentioned above have successfully predicted gene mutations. However, most of these methods rely on hand-crafted features. In recent years, deep learning based on convolutional neural networks has attracted much attention in the field of medical image computing. This data-driven approach can automatically extract complex image features [[
Although the above model achieved considerable performance, there are still some challenges in the study of deep learning methods based on image and genetic data for predicting KRAS mutation status in NSCLC: 1) Majority of deep learning methods [[
To overcome these difficulties and achieve non-invasive and accurate prediction of KRAS gene mutations in NSCLC. We propose a Semi-supervised Multimodal Multiscale Attention Model (S
In contrast to conventional radiomics and machine learning [[
The contributions of this paper are as follows:
- A Semi-supervised Multimodal Multiscale Attention Model (S
2 MMAM) based on imaging genomics is proposed, which effectively solves the problem of difficult intermediate fusion of multimodal heterogeneous data. S2 MMAM exploits the facilitation of supervised segmentation features for semi-supervised classification tasks to improve the model performance for predicting KRAS gene mutation status in NSCLC. - A new Triple Attention-guided Feature Aggregation (TAFA) module is designed. It is based on the attention module to adaptively fuse high-level semantic features with low-level semantic features. TAFA can suppress low-level background noise and retain detailed local semantic information.
- We use the Intra and Inter Mutual Guidance Attention Fusion (I
2 MGAF) module to guide segmentation and classification feature fusion, as well as CT image and genetic data fusion, respectively. It can achieve multi-scale multimodal information fusion and improve classification performance.
Semi-supervised learning has been studied in the medical imaging community for a long time [[
Graph
Table 1 Comparison of three commonly used consistency-based semi-supervised methods.
Methods Purpose Limitations Π-Model Based on the consistency principle and perturbs the input data High complexity and nosiy-prone results Temporal Ensembling (TE) Employs an exponential moving average (EMA) prediction for each unlabeled data as the consistency target. Maintain a huge prediction matrix during the prediction process, and the training time complexity is high for large data sets. Mean Teacher Improves the problem of high time complexity caused by the TE method. Constructs a teacher model using the EMA weights of the student model.
In recent years, Mean Teacher has achieved good results as a basic framework in semi-supervised classification tasks. Wang et al. [[
Using segmentation tasks to facilitate classification network tasks is a basic form of multitask learning [[
According to Table 2, the above works demonstrate that segmentation has a facilitating effect on classification. However, there is a common problem: they are all studied for supervised models. Supervised models have high requirements for data labeling costs. We believe that the combination of segmentation and classification tasks can make the network more informative. Therefore, our research aims to combine the idea of segmentation facilitating classification with semi-supervised models. We combined two related tasks of NSCLC lesion segmentation and KRAS gene mutation status prediction. S
Graph
Table 2 Comparison of three commonly used consistency-based semi-supervised methods.
Methods Contributions Limitations Xie et al. [ Proposed the Mutual Bootstrapping Deep Convolutional Neural Networks (MB-DCNN) model for simultaneous segmentation and classification of skin lesions. The rough lesion masks generated by the segmentation network in MB-DCNN help the classification network for training. The segmentation and classification networks transfer knowledge to each other in a bootstrap manner and facilitate each other. Non-end-to-end model Professional doctors are needed to manually label each image Zhao et al. [ Proposed a Segmentation-based Sequence Residual Attention Model (SSRAM) for the dual task of colorectal cancer lesion segmentation and KRAS gene mutation status prediction. The SSRAM utilizes the information provided by the segmentation network and the mask to successfully improve the accuracy of the classification task. 1. Data pre-processing is more complex 2、Professional doctors are needed to manually label each image Song et al. [ Utilized the lung nodule segmentation task to assist the lung nodule malignant development prediction task. Professional doctors are needed to manually label each image
Traditional convolution operations mostly focus on extracting local features. However, due to the limited information contained in local features, the model cannot learn the full range of region of interest contents well. Multi-scale features contain local features of multiple regions of interest. The extracted local features are fused with other operations to obtain comprehensive information about the target, which helps the network model to learn. To extract multi-scale features, The Atrous Spatial Pyramid Pooling (ASPP) module [[
The Convolutional Block Attention Module (CBAM) [[
Currently, it is widely believed that both multi-scale features and attention mechanisms can help models enhance the recognition of feature maps from different dimensions. However, the above papers have a common problem: they do not combine the ideas of multi-scale and attention mechanism. Therefore, we combine these two techniques and design the TAFA module. On the one hand, fuse high and low dimensional segmentation features to obtain abstract and detailed information. On the other hand, we fuse segmentation and classification features of different levels to guide the features to learn key factors adaptively and enhance the ability of the network to capture lesions. Thus, the predictive capability of the model is improved.
In this paper, we propose a Semi-supervised Multimodal Multiscale Attention Model (S
Graph: Overview of our S2MMAM, including: (a) Supervised Multilevel Fusion Segmentation Network (SMF-SN). The inputs are CT images and pixel-level mask images, and the outputs are segmented lesion images, (b) Semi-supervised Multimodal Fusion Classification Network (S2MF-CN), and (c) processing of gene data. In the S2MMAM, the useful information of CT images is captured by SMF-SN and transferred to S2MF-CN to facilitate the execution of image prediction tasks. The S2MMAM utilizes the fusion of CT images and genetic data to accurately predict whether KRAS is mutated in NSCLC.
In the NSCLC dataset, each patient corresponds to a set of CT images and gene data (Section Dataset). Specifically, in our problem setting, we are given a training set containing N labeled data and M unlabeled data where N<
Graph
Graph
Graph
Graph
Graph
Graph
Graph
Graph
Graph
This section introduces a supervised segmentation network based on multidimensional feature fusion. SMF-SN can precisely localize lesion edges and internal regions and greatly reduce the impact of image background noise on network performance. SMF-SN mainly utilizes our proposed SE-ResNeXt and TAFA modules.
We use the enhanced segmentation training dataset S
As shown in Fig 2, SMF-SN includes a stem block, three encoder blocks, three TAFA blocks, a bridge block, three decoder blocks, and an output block.
Graph: We adjust the dilation rates in ASPP in the bridge from 6,12,18 to 3,6,9 to better adapt SMF-SN to our segmentation task.
In the encoder, each encoder is composed of a SE-ResNeXt and a max-pooling layer with step size 2. As shown in Fig 3, SE-ResNeXt is improved from ResNeXt with SENet. ResNext achieves aggregating a set of transitions with the same topology by repeating multiple blocks. SENet can perform feature learning on the aggregated features in the channel dimension to form the importance of each channel. SE-ResNeXt can enhance the network in both the channel and spatial dimensions to capture richer segmentation features. Applying the MaxPooling layer can reduce the spatial dimension of the feature map by half to reduce the computational cost. The output of the encoder is passed through a bridge consisting of SE-ResNeXt and Atrous Spatial Pyramid Pooling (ASPP). It provides the largest receptive domain for TAFA to include a wider range of contextual information, facilitating more efficient integration between multiple levels. Between high-level and low-level semantics, we use the proposed TAFA module. This module utilizes multi-scale and attention fusion mechanisms. The module both suppresses low-level irrelevant background noise and complements each other with contextual difference information, preserving more detailed local semantic information and better learning of focal information. TAFA module is depicted in detail in Section Triple Attention-guided Feature Aggregation.
Graph: SE-ResNeXt is improved from ResNeXt with SENet.
Since CT images of lung nodules may contain a large amount of noise, for example, there are problems of grayscale overlap between lung tissues, blurred boundaries, and challenging to distinguish. High-level features of the decoder and low-level features of the encoder are crucial for capturing lesion features. However, most of the existing UNet-based connection methods directly connect shallow and deep semantic features of different scales. This behavior ignores that high-level features contain rich semantic information that can help low-level features identify semantically important locations. Likewise, low-level features contain rich spatial information that can help high-level features reconstruct accurate details.
Considering the above factors, we design a Triple attention-guided feature aggregation (TAFA) module to guide the fusion between high and low-dimensional features. TAFA can guide different layers to extract key feature information individually and then fuse after retaining the domain invariant key information, as shown in Fig 4. In the TAFA module, we first upsample the high-dimensional feature
Graph
Graph
Graph
Graph
Where Concat represents the concatenation operation, f
Graph
Graph
Graph
Graph
Graph
Graph
Where f
Graph
Graph
Graph
Graph
Graph
Graph
Graph: Fig 4 info:doi/10.1371/journal.pone.0297331.g004
Finally, the weighted features are concatenated. The concatenated feature maps are multiplied with W
Graph
Where Conv represents 1×1 convolution operation, ⊕ represents element-wise sum and ⊗ represents element-wise multiplication.
Our proposed TAFA transfers features from shallower convolutional layers to deeper convolutional layers. Performing the shallow features in the deeper convolutional layers prevents the shallow features from being forgotten. It makes the obtained features have more vital characterization ability. By gradually guiding the fusion between high and low features, SMF-SN can be guided to adaptively combine high and low-dimensional semantic information to reassign feature weights and better capture critical domain invariant information. Thus, lung nodules can be separated from the noise.
The proposed S
Graph: The overview of the Student Module, including (a) the specific implementation details of the Student Model, (b) Intra fusion component (IntraFC) aims to fuse classification and segmentation features at different levels, and (c) Inter fusion component (InterFC) aims to fuse CT image features and genetic features.
In the S
- Intra Fusion Component (IntraFC)
- We propose the IntraFC based on the MultiRes Block, which can capture multi-scale information [[
26 ]]. We adopted a strategy of fusing classification features with segmentation features at each level. The information favoring the prediction of KRAS gene mutation status is jointly retained. - The specific structure of the IntraFC component is shown in Fig 5(B). The final level segmentation features
Graph
Graph
Graph
Graph
- Inter Fusion Component (InterFC)
- We propose the InterFC to find the bidirectional mapping relationship between lung cancer image features and causative genes from the sagittal view (x-axis), coronal view (y-axis), and axial view (z-axis), respectively. InterFC can adaptively enhance the necessary information in different modal features, allowing a more adequate fusion of multimodal features.
- The specific structure of the InterFC component is shown in Fig 5(C). The initial classification feature F
C , the fusion result
Graph
•
Graph
- Where Concat denotes the concatenation operation. Then the concatenated multimodal data features are fed to three convolutional layers with BN and ReLU. The size of the convolution kernel is 1×3×1, 3×1×1 and 1×1×3 respectively, to produce three feature maps Query∈R
C×H×W , Key∈RC×H×W and Value∈RC×H×W (where C,H,W indicate the channel, height, width of the input features F respectively). We first transpose the Query feature. Then, we perform a softmax layer on the matrix multiplication of QueryT and Key to encode the feature relationships in sagittal and coronal views. Finally, matrix multiplication is multiplied with Value to obtain the voxel-level attention enhanced fusion features FInter , which are then reshaped to be in RC×H×W .
•
Graph
- Where ⊕ denotes element-wise sum, ⊗ denotes element-wise multiplication.
In this study, we applied NSCLC-Radiogenomics [[
Graph
Table 3 Patients' medical record information in the dataset.
Category Total Mutation Wildtype Amount 124 30 94 Gender Male 93 24 69 Female 31 6 25 Smoking History Smoking 107 30 77 Non-smoking 17 0 17 Pathological type Adenocarcinoma 105 29 76 Squamous Carcinoma 17 0 17 Other 2 1 1
In our experiments, for 124 sets of CT images inspired by Cubuk et al. [[
The gene expression data used in this study is RNA-seq data. Since the vast gene dataset contains more than 20,000 gene expression data per patient, the huge amount of gene expression data can significantly increase the computational cost and decrease the prediction accuracy. Therefore, before training the model, we screened the gene expression data from RNA-seq sequencing by the feature selection algorithm [[
Our model S
Graph
Table 4 The initialization network configurations of model.
Network Configurations Setting Epochs per fold 20 Optimizer Adams Initial learning rate 0.001 Batch size 16
To quantitatively analyze the experimental results, we used six performance metrics to evaluate the classification results obtained, including Accuracy (AC), Recall, Precision, Specificity (SP), Area Under the receiver operating Curve (AUC) and F1 score (F1). They are defined as follows:
Graph
Graph
Graph
Graph
Graph
Graph
Where TP is true positive, TN is true negative, FP is false positive, FN is false negative, t
In this section, we evaluate the impact of the SE-ResNeXt, TAFA module, and the I
Using SE-ResNeXt as the backbone of the network can not only enhance the network to extract focal features. It can also take advantage of the lightweight feature of ResNeXt to reduce the computational burden of the network and improve the network's efficiency. To verify the performance of our proposed SE-ResNeXt, we replace the backbone network with S
Graph
Table 5 Comparison of classification performance of UNet, ResNet, ResNeXt, Inception-v3 and SE-ResNeXt on S2MMAM. SE-ResNeXt(Ours) achieved the best results in all six comparative metrics.
Methods AC(%) Recall(%) Precision(%) SP(%) AUC(%) F1(%) UNet [ 71.29±0.53 71.35±0.11 73.24±0.64 70.09±0.36 70.33±0.38 72.28±0.37 ResNet [ 73.95±1.05 74.01±2.19 74.15±0.41 76.32±2.46 76.48±0.47 74.07±1.29 ResNeXt [ 77.99±3.16 76.92±1.37 78.14±3.21 75.28±3.14 77.31±2.22 77.53±2.27 Inception-v3 [ 75.29±2.86 77.39±3.15 75.14±2.21 77.20±2.18 76.84±1.43 76.25±2.66 SE-ResNeXt(Ours) 81.67±2.67 82.31±2.51 83.15±1.21 82.66±2.07 83.27±1.49 82.73±1.86
As shown in Table 5, it is evident from the results that our S
Using TAFA as the basic module to build S
Graph
Table 6 Comparison of classification performance of TAFA on S2MMAM and four models with different fusion blocks. TAFA(Ours) achieved the best results in all six comparative metrics.
Methods AC(%) Recall(%) Precision(%) SP(%) AUC(%) F1(%) Addition 72.56±1.02 72.28±1.69 71.43±2.14 72.47±1.79 72.49±1.23 71.85±1.97 Concatenation 73.03±1.52 73.63±1.46 75.24±1.22 73.13±1.86 72.15±0.87 74.43±1.34 AEAF [ 77.25±1.67 78.22±1.62 77.73±2.44 76.56±2.77 78.88±2.45 78.27±2.72 ASSCM [ 78.39±1.42 78.37±1.19 78.49±0.76 78.57±2.06 77.89±1.06 78.43±0.97 TAFA(Ours) 81.67±2.67 82.31±2.51 83.15±1.21 82.66±2.07 83.27±1.49 82.73±1.86
The results show that the highest performance metrics were achieved on the classification task using our proposed S
The I
Graph: Fig 6 info:doi/10.1371/journal.pone.0297331.g006
Graph: Fig 7 info:doi/10.1371/journal.pone.0297331.g007
Fig 6 shows a visual comparison of the six classification performance metrics after replacing the IntraFC module in I
Fig 7 shows the comparison of the six classification performance metrics after replacing the InterFC module in I
We compare the proposed S
Table 7 shows that the key evaluation metrics of S
Graph: Fig 8 info:doi/10.1371/journal.pone.0297331.g008
Graph
Table 7 Comparison of the classification performance of S2MMAM and five other semi-supervised medical image classification models.
Methods Labeled Unlabeled Data Result(%) CT Gene AC Recall Precision SP AUC F1 Π-Model [ 100% 0 √ 76.35±2.32 78.21±2.36 79.32±2.68 77.32±2.65 76.23±2.31 78.76±2.51 Mean Teacher [ √ 81.24±2.18 80.15±1.81 82.34±1.79 81.86±2.43 80.04±2.78 81.23±1.8 RSM [ √ 84.21±1.26 81.93±2.15 84.21±2.03 84.72±2.07 83.41±2.65 83.05±2.09 SS-TBN [ √ 80.25±1.71 79.38±2.16 79.88±2.71 78.62±3.89 81.23±2.44 80.35±2.66 DAB [ √ 81.79±2.3 80.42±1.51 82.11±2.23 82.8±2.43 83.56±2.36 82.37±1.97 S2MMAM(Ours) √ √ 86.94±3.12 85.97±2.19 84.28±1.73 86.11±2.54 87.92±1.69 85.12±1.96 Π-Model [ 30% 70% √ 71.23±2.49 72.11±1.65 72.56±1.24 71.16±2.77 70.15±2.17 72.33±1.45 Mean Teacher [ √ 74.28±2.53 74.16±2.73 75.29±2.94 74.62±2.82 74.21±3.48 74.72±2.83 RSM [ √ 75.91±2.37 75.13±3.21 76.37±2.86 75.49±3.53 75.94±2.34 75.74±3.04 SS-TBN [ √ 76.01±1.54 75.17±1.89 77.22±1.47 77.01±2.15 76.37±2.22 77.74±1.98 DAB [ √ 75.73±2.46 77.22±2.49 76.16±2.73 76.49±2.87 76.06±1.87 76.58±2.12 S2MMAM(Ours) √ √ 81.67±2.67 82.31±2.51 83.15±1.21 82.66±2.07 83.27±1.49 82.73±1.86
Although ablation studies and comparison experiments have demonstrated the merits of our proposed method, further discussions are needed on 1) the positive effects of segmentation features for the classification task, 2) the superiority of multimodal data over single modal data, and 3) the selection of the proportion of labeled images within the training dataset.
We designed three sets of experiments and empirically used data with the proportion of labeled data of 100%, 40%, and 30% as the training dataset. Baseline is used as our base architecture, where Baseline is only constructed by S
Graph
Table 8 Six metrics were achieved on the test set by Baseline, Baseline+SMF-SN, Baseline+Gene, and our S2MMAM when using 30%, 40%, and 100% labeled training images.
Methods Labeled Unlabeled Data Result(%) CT Gene AC Recall Precision SP AUC F1 Baseline 100% 0 √ 76.29±5.22 75.39±2.31 74.92±2.64 77.57±2.84 78.26±3.14 75.15±2.39 Baseline+SMF-SN √ 83.19±2.46 79.38±1.37 81.31±1.67 82.47±3.64 84.29±4.36 80.33±1.52 Baseline+Gene √ √ 82.37±1.67 80.06±3.49 78.15±2.17 81.66±3.58 82.2±2.61 79.09±2.81 S2MMAM(Ours) √ √ 86.94±3.12 85.97±2.19 84.28±1.73 86.11±2.54 87.92±1.69 85.12±1.96 Baseline 40% 60% √ 74.04±1.21 73.72±3.11 74.3±2.16 73.74±2.94 75.61±2.47 73.28+2.76 Baseline+SMF-SN √ 78.37±2.14 77.49±2.76 78.06±1.98 78.34±2.48 79.23±3.16 79.35+2.54 Baseline+Gene √ √ 78.41±1.43 78.16±1.32 79.64±2.34 78.14±1.79 78.02±1.99 79.87+1.29 S2MMAM(Ours) √ √ 82.35±1.72 83.14±1.48 83.78±1.77 81.87±1.76 83.98±1.01 83.62+1.52 Baseline 30% 70% √ 73.65±2.18 72.91±1.03 72.11±2.36 73.67±2.72 73.33±2.31 72.51±1.7 Baseline+SMF-SN √ 77.29±5.22 76.43±4.21 77.24±3.21 77.17±1.2 77.44±5.32 78.8±3.69 Baseline+Gene √ √ 75.11±2.3 78.81±2.92 79.35±2.16 75.39±3.44 75.14±3.26 79.08±2.54 S2MMAM(Ours) √ √ 81.67±2.67 82.31±2.51 83.15±1.21 82.66±2.07 83.27±1.49 82.73±1.86
- 1) The positive effects of segmentation features for the classification task
As shown in Table 8, better classification results are obtained when the model utilizes the idea of segmentation to facilitate classification. Compared to Baseline, Baseline+SMF-SN improves the AUC values by 6.03%, 3.62%, and 4.11% in 30%, 40%, and 100% labeled datasets, respectively. We also visualize some of our Baseline and Baseline+SMF-SN segmentation results in Fig 9. The results are output in the form of a segmentation graph, which visualizes the ability of the network to localize the lesion area. As can be seen from Fig 9, the model with segmentation task can better localize the lesion area. It can avoid mixing impurities that can easily interfere with the judgment to improve the accuracy of diagnosis.
Graph: Baseline+SMF-SN: classification task and segmentation task. (a) and (b) are the wild type of NSCLC. (c) and (d) are the mutation of NSCLC. The region surrounded by the red line is the ground truth, and the region surrounded by the green line is the segmentation results.
- 2) The superiority of multimodal data over single modal data
As shown in Table 8, when we used genetic data, the AUC improved by 3.94%, 2.41%, and 2.81%, respectively, compared with Baseline. This indicates that image data can also extract genotypic features from biological data that can express individual differences and reflect disease characteristics at the micro level. Further, enhances the network information richness and promotes the classification performance.
- 3) The selection of the proportion of labeled images within the training dataset
As shown in Table 8, when the proportion of labeled data was 30% and 40%, respectively, the difference in the values of the four metrics was small, with a 0.71% difference in AUC and a 0.83% difference in Recall. Compared with the cost of physician labeling, this result indicates that the guidance information contained in 30% labeled training images is sufficient for the network to learn the key information of the lesion. Therefore, we used 30% labeled images and 70% unlabeled images as the training ratio of the model.
To show the classification performance of our S
Graph: Fig 10 info:doi/10.1371/journal.pone.0297331.g010
Graph: Fig 11 info:doi/10.1371/journal.pone.0297331.g011
In summary, the strategy of sharing segmentation network parameters by the classification network can assist the network to better localize the lesion region. The complementary nature of multimodal data allows the network to learn more abstract features besides addressing the challenge of less information in semi-supervised strategies. Therefore, our S
In order to demonstrate the scalability of our model, our application scenarios will not be limited to semi-supervised learning but will be extended to supervised learning. We compare our S
As shown in Table 9, our S
Graph
Table 9 Comparison of the classification performance of S2MMAM and two other supervised medical image classification models.
Methods AC(%) Recall(%) Precision(%) SP(%) AUC(%) F1(%) MFFDM [ 84.15±1.45 84.22±2.04 83.98±2.77 84.02±1.97 84.17±1.03 84.16±1.17 PLNM [ 86.34±2.11 86.21±2.61 85.24±1.69 85.73±2.65 86.32±1.87 85.23±2.31 S2MMAM(Ours) 86.94±3.12 85.97±2.19 84.28±1.73 86.11±2.54 87.92±1.69 85.12±1.96
In this paper, we propose an integrating Image and Gene Data with a Semi-Supervised Attention Model for the Prediction of KRAS Gene Mutation Status in Non-Small Cell Lung. The model consists of two components: supervised multilevel fusion segmentation network (SMF-SN) and semi-supervised multimodal fusion classification network (S
However, our S
Gwak Jeonghwan Academic Editor
22 Aug 2023
PONE-D-23-16921Integrating Image and Gene-Data with a Semi-Supervised Attention Model for Prediction of KRAS Gene Mutation Status in Non-Small Cell Lung CancerPLOS ONE
Dear Dr. Zhao,
Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE's publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.
Please submit your revised manuscript by Oct 06 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://
Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.
If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.
We look forward to receiving your revised manuscript.
Kind regards,
Jeonghwan Gwak, PhD
Academic Editor
PLOS ONE
Journal requirements:
When submitting your revision, we need you to address these additional requirements.
1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at
https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne%5fformatting%5fsample%5fmain%5fbody.pdf and
https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf
- 2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.
- 3. Thank you for stating the following financial disclosure:
"This work was supported by the National Natural Science Foundation of China (Grant No. U21A20469); the National Natural Science Foundation of China (Grant No. 61972274); the Central Government Guides Local Science and Technology Development Fund Project (Grant No. YDZJSX2022C004); the Natural Science Foundation of Shanxi Province (Grant No. 202103021224066); and NHC Key Laboratory of Pneumoconiosis Shanxi China Project, (Grant No.2020-PT320-005), the Non-profit Central Research Institute Fund of Chinese Academy of Medical Science."
Please state what role the funders took in the study. If the funders had no role, please state: "The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript."
If this statement is not correct you must amend it as needed.
Please include this amended Role of Funder statement in your cover letter; we will change the online submission form on your behalf.
4. Thank you for stating the following in the Acknowledgments Section of your manuscript:
"This work was supported by the National Natural Science Foundation of China (Grant No. U21A20469); the National Natural Science Foundation of China (Grant No. 61972274); the Central Government Guides Local Science and Technology Development Fund Project (Grant No. YDZJSX2022C004); the Natural Science Foundation of Shanxi Province (Grant No. 202103021224066); and NHC Key Laboratory of Pneumoconiosis Shanxi China Project, (Grant No.2020-PT320-005), the Non-profit Central Research Institute Fund of Chinese Academy of Medical Science."
We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form.
Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows:
"This work was supported by the National Natural Science Foundation of China (Grant No. U21A20469); the National Natural Science Foundation of China (Grant No. 61972274); the Central Government Guides Local Science and Technology Development Fund Project (Grant No. YDZJSX2022C004); the Natural Science Foundation of Shanxi Province (Grant No. 202103021224066); and NHC Key Laboratory of Pneumoconiosis Shanxi China Project, (Grant No.2020-PT320-005), the Non-profit Central Research Institute Fund of Chinese Academy of Medical Science."
Please include your amended statements within your cover letter; we will change the online submission form on your behalf.
Additional Editor Comments:
AE Comments: Thank you for submitting your manuscript. I appreciate the efforts you have put into this research. I have received feedback from the reviewers, and I would like to share their comments and suggestions with you.
1) Clarity and Comprehension: Reviewer 1 points out a lack of clarity in the explanation of your proposed method. The reviewer found it difficult to understand, making it challenging to reproduce the experiments. Specific feedback has been given regarding figures and their captions (Fig. 1, Fig. 2, and Fig. 5), as well as the use of equations (e.g., Equations 8 and 9).
- 2) Novelty and Originality: Reviewer 2 has raised concerns about the originality of the work. It's essential to clarify the unique contributions of your research compared to existing literature.
- 3) Related Work: Both reviewers emphasize the need to improve the section on related works. The current version lists existing works without analyzing their limitations. Consider adding a more detailed analysis and perhaps summarizing existing studies in a tabular form to improve readability.
- 4) Methodology and Experimental Details: Both reviewers have made suggestions to provide more information on the methodology, hyperparameters, network configurations, and a thorough description of the experimental phases.
- 5) Source Code: Reviewer 2 suggests providing a GitHub link for the source code to enhance repeatability and verification of the study.
- 6) Grammar and Typos: Both reviewers have found grammatical errors and typos in the manuscript. It is advised to run the manuscript through a grammar checker and proofread it carefully.
- 7) Additional Feedback: Reviewer 2 has provided an extensive list of recommendations to enhance the quality and clarity of the manuscript. These include improving the introduction, elaborating on tables, addressing overfitting, revisiting results, and ensuring that the references are up-to-date.
In light of the feedback, I recommend revising your manuscript, addressing the concerns raised by the reviewers. This will not only enhance the clarity and quality of your research but also strengthen its contribution to the field.
I hope this feedback is constructive and assists you in enhancing your manuscript. I look forward to receiving your revised submission.
[Note: HTML markup is below. Please do not edit.]
Reviewers' comments:
Reviewer's Responses to Questions
1. Is the manuscript technically sound, and do the data support the conclusions?
The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.
Reviewer #1: Yes
Reviewer #2: Partly
***
2. Has the statistical analysis been performed appropriately and rigorously?
Reviewer #1: Yes
Reviewer #2: I Don't Know
***
3. Have the authors made all data underlying the findings in their manuscript fully available?
The
Reviewer #1: Yes
Reviewer #2: Yes
***
4. Is the manuscript presented in an intelligible fashion and written in standard English?
PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.
Reviewer #1: Yes
Reviewer #2: No
***
5. Review Comments to the Author
Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)
Reviewer #1: This study proposes a deep learning-based methodology for classifying the oncogenic gene KRAS, which frequently involves non-small cell lung cancer (NSCLC), using CT images and genetic information. Their main contributions are the development of the 'Semi-supervised Multimodal Multiscale Attention' mechanism and the novel 'Attention-guided Feature Aggregation' module. The proposed method appears to be novel and innovative, and the data and analysis seem to fully support their claims. However, due to the lack of clarity in the explanation of the proposed method, it is difficult to understand, making it seem impossible to reproduce the experiments. Therefore, 'minor revisions' are suggested to improve the paper.
In Section 3.1, while it seems that the input of the Supervised Multilevel Fusion Segmentation Network (SMF-SN) is X_L and the output is Y_L, in Fig. 1. (a), it is not clearly introduced what the input and output of SMF-SN are, as both X_L and Y_L are shown. This requires modification.
Overall, the introduction of the proposed system is difficult to comprehend. For example, in Fig. 2, there is only one ASPP block, but the caption suggests the presence of multiple ASPPs. Additionally, in Fig. 1 (b), is the input of the Student Model S_L and the input of the Teacher Model C_U? How are the predictions of the Student and Teacher integrated?
In Fig. 5, an explanation is needed for 'Genes Selection'.
The paper requires revisions for grammatical errors.
Equations must be used in the right way (e.g. Equations 8 and 9 should have the first two letters of Recall and Precision in italics.)
Reviewer #2: The experimental study is interesting information in this paper. However, the main weakness of the paper lies in its lack of originality and novelty. The following suggestions may be considered to enhance the quality and clarity of the manuscript
1- The motivation is not clear. Why did this work? Is any problem does it address that the previous methods cannot?
- 2- The introduction section could be improved by clarifying the similarities and differences between the related work and the proposed method are not clearly described. It is recommended to add a separate subsection and clear description in this regard.
- 3- Related work: The paper only lists existing works in the research community without any analysis of existing work's limitations. Therefore, I suggest that the authors mention more summary and limitation analysis so that readers can easily appreciate the contributions made by this paper.
- 4- In the related works, existing studies can also be summarized in a tabular form to improve readability
- 5- Elaborate all tables briefly.
- 6- How to deal with overfitting in your model?
- 7- Results and illustrations need to be revisited.
- 8- Background information of this work can be provided more systematically and comprehensively, i.e. logic of this paper should be further enhanced.
- 9- - Hyperparameters of the model:
- The initialization method is not mentioned.
- 10- Similarly, the network configurations can be summarized in a table e.g. input size, # of layers, learning rate, optimizers etc.
- 11- Furthermore, the study's application is not explained in an intelligible manner. You should include an experimentation section to provide readers with a thorough description of all the experimental phases in a straightforward and accessible manner.
- 12- The theoretical and practical sections of the study are not adequately convincing, and the writing style is absolutely insufficient to highlight the subjective contribution to your research when compared to past research findings.
- 13- Another important aspect of scientific research is the capacity to repeat the experiment or study in a different setting and reuse or adapt the findings. This is an important point, and you could elaborate on it further in the discussion area to give additional scientific value to this critical study.
- 14- Please include a link in the research article that allows the complete applied side of this study to be downloaded for verification, validation, and inspection, as well as so that it may be used as a scientific reference.
The code source of this work must be added as a comment to the paper and must be uploaded as a GitHub link to be visible and referenceable.
- 15- In addition to these specific recommendations, the authors should also run the manuscript through a grammar checker like Grammarly to address any language or grammatical errors. Finally, the authors should ensure that all references cited in the manuscript are up-to-date and relevant to the research topic.
- 16- Typos/Grammatical Errors:
Subsection Segmentation facilitates classification
Deep Convolutional Nneural Networks --> N should be removed from neural
Section Conclusion:
Mutation Status in Non-Small Cell Lung The model --> period (.) is missing
network (S2 MF-CN). fusion. --> the extra period (.) should be removed
***
6. PLOS authors have the option to publish the peer review history of their article (https://journals.plos.org/plosone/s/editorial-and-peer-review-process#loc-peer-review-history). If published, this will include your full peer review and any attached files.
If you choose "no", your identity will remain anonymous but your review may still be made public.
Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our https://
Reviewer #1: No
Reviewer #2: No
***
[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]
While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.
13 Oct 2023
Responses to Reviewers'
Dear Editors and Reviewers,
Thank you for your letter and comments on our manuscript entitled "Integrating Image and Gene-Data with a Semi-Supervised Attention Model for Prediction of KRAS Gene Mutation Status in Non-Small Cell Lung Cancer (ID: PONE-D-23-16921)". We sincerely thank all reviewers for their time and effort. According to the constructive comments of the editors and reviewers on improving the quality of the revised version of this paper, we have revised the whole manuscript carefully and tried to avoid any grammar or syntax errors. In addition, we have asked several colleagues who are skilled in English papers to help us thoroughly check the organization and language of the paper. We hope for accepting our further improved submitted paper for possible publication in the PLOS ONE distinguished journal.We have revised the manuscript point by point. We apologize for not using "the Tracked Changes function in Word." The reason is that we revised all grammar or syntax errors and made a lot of changes, which might have interfered with the editors and reviewers reviewing the paper. We have highlighted the important changes in red. We hope you will be satisfied with our revised manuscript. Responses to comments, as well as details of revisions, are given below.
Sincerely yours,
Juanjuan Zhao (on behalf of all the co-authors)
Reviewer #1:
Comment1:
In Section 3.1, while it seems that the input of the Supervised Multilevel Fusion Segmentation Network (SMF-SN) is X_L and the output is Y_L, in Fig. 1. (a), it is not clearly introduced what the input and output of SMF-SN are, as both X_L and Y_L are shown. This requires modification.
Response1:
We sincerely thank the reviewers for asking rigorous questions.
We feel very sorry that we lacked some explanations here. The input of SMF-SN is CT images, which is , and pixel-level mask images annotated by physicians, which is. The output of SMF-SN is the segmented lesion map. Since segmentation performance is not a concern in this study, it is not highlighted in the figure. We have added explanations for input and output in the caption of Fig. 1(a) in new manuscript to improve readability.
Comment2:
Overall, the introduction of the proposed system is difficult to comprehend. For example, in Fig. 2, there is only one ASPP block, but the caption suggests the presence of multiple ASPPs. Additionally, in Fig. 1 (b), is the input of the Student Model S_L and the input of the Teacher Model C_U? How are the predictions of the Student and Teacher integrated?
Response2:
Sorry we sincerely appreciate these insightful questions and apologize for our lack of rigor.
(
(
Fig.1 The Mean Teacher method. The figure depicts a training batch with a single labeled example. Both the student and the teacher model evaluate the input applying noise (η, η') within their computation. The softmax output of the student model is compared with the one-hot label using classification cost and with the teacher output using consistency cost. After the weights of the student model have been updated with gradient descent, the teacher model weights are updated as an exponential moving average of the student weights. Both model outputs can be used for prediction.
We very much apologize for not articulating this clearly. Fig.1 comes from the paper Mean Teacher Model [
Based on your suggestion, We redraw Fig. 1 to clearly show the inputs to the model. means labled dataset for segmentation. means labled dataset for classification. means unlabled dataset for classification. is the input of SMF-SN but not for classification network.
Comment3:
In Fig. 5, an explanation is needed for 'Genes Selection'.
Response3:
We sincerely thank the reviewers for the detailed comments.
We explained a detailed description of 'Genes Selection' in Section Data preprocessing. However, in Section Data preprocessing, our caption was set to 'Gene Data' in the previous version of the manuscript, causing ambiguity. We feel very sorry for this and have changed 'Gene Data' to 'Genes Selection' in the new version of the manuscript to prevent ambiguity.
Comment4:
The paper requires revisions for grammatical errors.
Response4:
We sincerely thank the reviewers for the detailed comments.
According to your suggestion, We have checked our manuscript carefully and corrected the grammatical, styling, and typos found in our new manuscript. Moreover, we have asked several colleagues who are skilled in English papers to help us thoroughly check the organization and language of the paper.
Comment5:
Equations must be used in the right way (e.g. Equations 8 and 9 should have the first two letters of Recall and Precision in italics.)
Response5:
We sincerely thank the reviewers for this insightful question.
We have checked all the Equations carefully for formatting issues and made corrections.
Reviewer #2:
Comment1:
The motivation is not clear. Why did this work? Is any problem does it address that the previous methods cannot?
Response1:
We sincerely thank the reviewers for asking rigorous questions.
We have supplemented a detailed description of the motivation for the research in the new manuscript (Section Introduction, P1, L3-L4). The emergence of targeted therapy has substantially increased the survival rate of NSCLC patients. Mutations of essential pathogenic genes should be identified before targeted therapy. KRAS is a gene type with a high probability of mutation. It is necessary for diagnosing whether a patient has a KRAS gene mutation.
We listed the limitations of the previous methodology in the fourth paragraph of Section Introduction. The main solution of this study is the three problems listed. 1)Majority of deep learning methods that study classification tasks focus only on methods for classification. However, these studies did not use the segmentation features generated by the segmentation task to facilitate the classification task to improve the performance and effectiveness of the classification task. 2) Most of the studied fusion methods used simple fusion means of direct concatenation. But, they ignore the correlation and difference between medical images and genetic data. It not only leads to ineffective mining of useful semantic features between multi-scale image features and gene features, but also fails to make full use of the complementarity of multimodal information. 3) Many studies used models that overemphasized the deep features of lesion abstraction. Nonetheless, they did not pay sufficient attention to the importance of detailed shallow features in prediction results. This leads to limitations in improving accuracy.
Comment2:
The introduction section could be improved by clarifying the similarities and differences between the related work and the proposed method are not clearly described. It is recommended to add a separate subsection and clear description in this regard.
Response2:
We feel great thanks for your professional review work on our paper.
We have added a new subsection according to the your suggestion in sixth paragraph of Introduction section to further compare the similarities and differences between previous work and ours.
Comment3:
Related work: The paper only lists existing works in the research community without any analysis of existing work's limitations. Therefore, I suggest that the authors mention more summary and limitation analysis so that readers can easily appreciate the contributions made by this paper.
Response3:
Thank you very much for the professional review work you have done on our papers.
Following your third and fourth comments, we have summarized and compared past work in a tabular format. The reason for not tabulating the comparison in the section Multiscale features and attention learning is that we consider both approaches classic and valid. We are concerned that the papers listed above all focus on only one aspect. Our contribution is to combine both methods to obtain better performance in extracting lesion information.
Comment4:
In the related works, existing studies can also be summarized in a tabular form to improve readability.
Response4:
Thank you very much for the professional review work you have done on our papers.
Following your third and fourth comments, we have summarized and compared past work in a tabular format to improve readability.
Comment5:
Elaborate all tables briefly.
Response5:
We feel great thanks for your professional review work on our paper.
Based on your suggestions, we have further elaborated all the tables for a better understanding of the readers.
Comment6:
How to deal with overfitting in your model?
Response6:
We feel great thanks for your professional review work on our paper.
We mainly used a cross-validation approach to prevent overfitting problems. We set the cross-validation to 5-fold, 10-fold and 15-fold. Table 1 records the AUC values of the test dataset at different parameter settings and with different scales of labeled data. The results show that the test dataset has the highest accuracy when the cross-validation is equal to 10-fold.
Table 1. AUC values of the test dataset under different parameter settings for different scales of labeled data.
Setting 5-fold 10-fold 15-fold
- 100% Labeled 84.16 87.92 88.76
- 40% Labeled 78.64 83.98 82.44
- 30% Labeled 70.02 83.27 80.69
Comment7:
Results and illustrations need to be revisited.
Response7:
Thank you very much for your professional review of our paper.
We feel sorry for our lack of rigor. We have re-examined the results and illustrations. We found that some of the illustrations' descriptions did not match the illustrations' content, as in Fig 8. We have corrected the incorrect parts and confirmed that the results and illustrations are correct. We will pay more attention to the uploading requirements to ensure readers see the correct results and illustrations.
Comment8:
Background information of this work can be provided more systematically and comprehensively, i.e. logic of this paper should be further enhanced.
Response8:
We feel great thanks for your professional review work on our paper.
Based on your suggestions, we have reorganized the logic of the article and partially rewritten it to make it present the objectives of the study more clearly.
Comment9:
Hyper-parameters of the model:The initialization method is not mentioned.
Response9:
We feel great thanks for your professional review work on our paper.
The initialization of hyper-parameters is mentioned in the Implementation details of Section Implementation details. The following is the initialization information for the hyper-parameters: All models in the experiments are trained using 10-fold cross-validation, with the number of epochs per fold set to 20. Adams was used as our optimizer. The initial learning rate is set to 0.001 empirically, and the batch size is set to 16.
Comment10:
Similarly, the network configurations can be summarized in a table e.g. input size, # of layers, learning rate, optimizers etc.
Response10:
We sincerely thank the reviewers for asking rigorous questions.
Based on your suggestion, we have redrawn Fig. 2. In Fig. 2, we have added the basic parameters of the network, such as input size. Since SMF-SN and S2MF-CN have the same structure, there is no separate structural diagram for S2MF-CN.
According to your suggestion, we design the hyper-parameter contents to be summarized in the form of Table 4 to improve readability.
Comment11:
Furthermore, the study's application is not explained in an intelligible manner. You should include an experimentation section to provide readers with a thorough description of all the experimental phases in a straightforward and accessible manner.
Response11:
Thank you very much for your advice.
We have added the visualization of the results of the experimental procedure in the Section Superiority of the model. The visualization shows the focus and significance of our study in an intuitive way. We believe that readers can understand the advantages of our method in this way.
Comment12:
The theoretical and practical sections of the study are not adequately convincing, and the writing style is absolutely insufficient to highlight the subjective contribution to your research when compared to past research findings.
Response12:
We feel great thanks for your professional review work on our paper.
In the theoretical part, we have comprehensively revised the logic of the paper with the above comments to highlight the superiority of our method.
In the experimental part, we have further analyzed the experimental results according to your suggestions. Detailed descriptions are provided in Sections Ablation studies and Sections Comparison experiment, to better demonstrate the good classification performance of our model. In the Discussion section, we have reorganized the logic to highlight the advantages of the model in a more logical form. In addition, according to #comment13, we have also supplemented extended experiments in the Discussion section to demonstrate the expandability and reusability of the experiments.
Through the above methods, we hope to make the paper more convincing.
Comment13:
Another important aspect of scientific research is the capacity to repeat the experiment or study in a different setting and reuse or adapt the findings. This is an important point, and you could elaborate on it further in the discussion area to give additional scientific value to this critical study.
Response13:
We feel great thanks for your professional review work on our paper.
Following your suggestion, we have added 'Performance in Supervised Learning' in Section Performance in Supervised Learning to demonstrate the extensibility of our model. The experimental proof demonstrates that our model is not only applicable in semi-supervised learning but can also be used in supervised learning. This shows that our research can be realized in various application scenarios.
Comment14:
Please include a link in the research article that allows the complete applied side of this study to be downloaded for verification, validation, and inspection, as well as so that it may be used as a scientific reference.The code source of this work must be added as a comment to the paper and must be uploaded as a GitHub link to be visible and referenceable.
Response14:
We feel great thanks for your professional review work on our paper.
We've included a link to the code in the Data availability.
The specific modifications are as follows:
The data are available from the website
https://wiki.cancerimagingarchive.net/display/Public/NSCLC+Radiogenomics. The code for S2MMAM is available on a GitHub repository at https://github.com/xyttttboom/SSMMAM.
Comment15:
In addition to these specific recommendations, the authors should also run the manuscript through a grammar checker like Grammarly to address any language or grammatical errors. Finally, the authors should ensure that all references cited in the manuscript are up-to-date and relevant to the research topic.
Response15:
We sincerely appreciate these insightful questions and apologize for our lack of rigor.
According to your suggestion, we have used Grammarly to address all language and grammatical errors.Moreover, we have asked several colleagues who are skilled in English papers to help us thoroughly check the organization and language of the paper.
We rechecked the references, deleting papers with little relevance to the topic and adding new papers with vital relevance. In Section Comparison experiment, we also rechecked the literature and compared it with recently published papers with better results.
e.g.
34.Cai M, Zhao L, Zhang Y, Wu W, Jia L, Zhao J, Yang Q, Qiang Y. A progressive phased attention model fused histopathology image features and gene features for lung cancer staging prediction. Int J Comput Assist Radiol Surg. 2023 Oct;18(
39.Zeng LL, Gao K, Hu D, Feng Z, Hou C, Rong P, Wang W. SS-TBN: A Semi-Supervised Tri-Branch Network for COVID-19 Screening and Lesion Segmentation. IEEE Trans Pattern Anal Mach Intell. 2023 Aug;45(
40.Chen X, Bai Y, Wang P, Luo J. Data augmentation based semi-supervised method to improve COVID-19 CT classification. Math Biosci Eng. 2023 Feb 6;20(
41.Tu Y, Lin S, Qiao J, et al. Alzheimer's disease diagnosis via multimodal feature fusion. Computers in Biology and Medicine, 2022, 148: 105901.
Comment16:
Typos/Grammatical Errors:
Subsection Segmentation facilitates classification
Deep Convolutional Nneural Networks --> N should be removed from neural
Section Conclusion:
Mutation Status in Non-Small Cell Lung The model --> period (.) is missing
network (S2MF-CN). fusion. --> the extra period (.) should be removed.
Response16:
We sincerely appreciate these insightful questions and apologize for our lack of rigor.
Based on your comments, we have made the corrections to revise the typos and grammatical errors throughout the paper.
Attachment
Submitted filename: Response to Reviewers.docx
Gwak Jeonghwan Academic Editor
3 Jan 2024
Integrating Image and Gene-Data with a Semi-Supervised Attention Model for Prediction of KRAS Gene Mutation Status in Non-Small Cell Lung Cancer
PONE-D-23-16921R1
Dear Dr. Zhao,
We're pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.
Within one week, you'll receive an e-mail detailing the required amendments. When these have been addressed, you'll receive a formal acceptance letter and your manuscript will be scheduled for publication.
An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at
If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they'll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.
Kind regards,
Jeonghwan Gwak, PhD
Academic Editor
PLOS ONE
Additional Editor Comments (optional):
AE: After careful consideration and based on the insightful feedback from our reviewers, I am delighted to announce that your paper is now deemed publishable.
Reviewers' comments:
Reviewer's Responses to Questions
1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the "Comments to the Author" section, enter your conflict of interest statement in the "Confidential to Editor" section, and submit your "Accept" recommendation.
Reviewer #2: All comments have been addressed
***
2. Is the manuscript technically sound, and do the data support the conclusions?
The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.
Reviewer #2: Yes
***
3. Has the statistical analysis been performed appropriately and rigorously?
Reviewer #2: Yes
***
4. Have the authors made all data underlying the findings in their manuscript fully available?
The
Reviewer #2: Yes
***
5. Is the manuscript presented in an intelligible fashion and written in standard English?
PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.
Reviewer #2: Yes
***
6. Review Comments to the Author
Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)
Reviewer #2: (No Response)
***
7. PLOS authors have the option to publish the peer review history of their article (https://journals.plos.org/plosone/s/editorial-and-peer-review-process#loc-peer-review-history). If published, this will include your full peer review and any attached files.
If you choose "no", your identity will remain anonymous but your review may still be made public.
Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our https://
Reviewer #2: Yes: Zahid Ullah
***
Gwak Jeonghwan Academic Editor
1 Mar 2024
PONE-D-23-16921R1
PLOS ONE
Dear Dr. Zhao,
I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.
At this stage, our production department will prepare your paper for publication. This includes ensuring the following:
* All references, tables, and figures are properly cited
* All relevant supporting information is included in the manuscript submission,
* There are no issues that prevent the paper from being properly typeset
If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.
Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.
If we can help with anything else, please email us at customercare@plos.org.
Thank you for submitting your work to PLOS ONE and supporting open access.
Kind regards,
PLOS ONE Editorial Office Staff
on behalf of
Dr. Jeonghwan Gwak
Academic Editor
PLOS ONE
By Yuting Xue; Dongxu Zhang; Liye Jia; Wanting Yang; Juanjuan Zhao; Yan Qiang; Long Wang; Ying Qiao and Huajie Yue
Reported by Author; Author; Author; Author; Author; Author; Author; Author; Author