Anna Luíza Damaceno Araújo, Viviane Mariano da Silva, Matheus Cardoso Moraes, Henrique Alves de Amorim, Felipe Paiva Fonseca, Maria Sissa Pereira Sant'Ana, Ricardo Alves Mesquita, Bruno Augusto Linhares Almeida Mariz, Hélder Antônio Rebelo Pontes, Lucas Lacerda de Souza, Cristina Saldivia-Siracusa, Syed Ali Khurram, Alexander T. Pearson, Manoela Domingues Martins, Marcio Ajudarte Lopes, Pablo Agustin Vargas, Luiz Paulo Kowalski, Alan Roger Santos-Silva
{"title":"使用最先进的深度学习架构进行口腔上皮发育不良分级:比较评价","authors":"Anna Luíza Damaceno Araújo, Viviane Mariano da Silva, Matheus Cardoso Moraes, Henrique Alves de Amorim, Felipe Paiva Fonseca, Maria Sissa Pereira Sant'Ana, Ricardo Alves Mesquita, Bruno Augusto Linhares Almeida Mariz, Hélder Antônio Rebelo Pontes, Lucas Lacerda de Souza, Cristina Saldivia-Siracusa, Syed Ali Khurram, Alexander T. Pearson, Manoela Domingues Martins, Marcio Ajudarte Lopes, Pablo Agustin Vargas, Luiz Paulo Kowalski, Alan Roger Santos-Silva","doi":"10.1111/jop.13477","DOIUrl":null,"url":null,"abstract":"<div>\n \n \n <section>\n \n <h3> Background</h3>\n \n <p>Dysplasia grading systems for oral epithelial dysplasia are a source of disagreement among pathologists. Therefore, machine learning approaches are being developed to mitigate this issue.</p>\n </section>\n \n <section>\n \n <h3> Methods</h3>\n \n <p>This cross-sectional study included a cohort of 82 patients with oral potentially malignant disorders and correspondent 98 hematoxylin and eosin-stained whole slide images with biopsied-proven dysplasia. All whole-slide images were manually annotated based on the binary system for oral epithelial dysplasia. The annotated regions of interest were segmented and fragmented into small patches and non-randomly sampled into training/validation and test subsets. The training/validation data were color augmented, resulting in a total of 81,786 patches for training. The held-out independent test set enrolled a total of 4,486 patches. Seven state-of-the-art convolutional neural networks were trained, validated, and tested with the same dataset.</p>\n </section>\n \n <section>\n \n <h3> Results</h3>\n \n <p>The models presented a high learning rate, yet very low generalization potential. At the model development, VGG16 performed the best, but with massive overfitting. In the test set, VGG16 presented the best accuracy, sensitivity, specificity, and area under the curve (62%, 62%, 66%, and 65%, respectively), associated with the higher loss among all Convolutional Neural Networks (CNNs) tested. EfficientB0 has comparable metrics and the lowest loss among all convolutional neural networks, being a great candidate for further studies.</p>\n </section>\n \n <section>\n \n <h3> Conclusion</h3>\n \n <p>The models were not able to generalize enough to be applied in real-life datasets due to an overlapping of features between the two classes (i.e., high risk and low risk of malignization).</p>\n </section>\n </div>","PeriodicalId":2,"journal":{"name":"ACS Applied Bio Materials","volume":null,"pages":null},"PeriodicalIF":4.6000,"publicationDate":"2023-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"The use of deep learning state-of-the-art architectures for oral epithelial dysplasia grading: A comparative appraisal\",\"authors\":\"Anna Luíza Damaceno Araújo, Viviane Mariano da Silva, Matheus Cardoso Moraes, Henrique Alves de Amorim, Felipe Paiva Fonseca, Maria Sissa Pereira Sant'Ana, Ricardo Alves Mesquita, Bruno Augusto Linhares Almeida Mariz, Hélder Antônio Rebelo Pontes, Lucas Lacerda de Souza, Cristina Saldivia-Siracusa, Syed Ali Khurram, Alexander T. Pearson, Manoela Domingues Martins, Marcio Ajudarte Lopes, Pablo Agustin Vargas, Luiz Paulo Kowalski, Alan Roger Santos-Silva\",\"doi\":\"10.1111/jop.13477\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div>\\n \\n \\n <section>\\n \\n <h3> Background</h3>\\n \\n <p>Dysplasia grading systems for oral epithelial dysplasia are a source of disagreement among pathologists. Therefore, machine learning approaches are being developed to mitigate this issue.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Methods</h3>\\n \\n <p>This cross-sectional study included a cohort of 82 patients with oral potentially malignant disorders and correspondent 98 hematoxylin and eosin-stained whole slide images with biopsied-proven dysplasia. All whole-slide images were manually annotated based on the binary system for oral epithelial dysplasia. The annotated regions of interest were segmented and fragmented into small patches and non-randomly sampled into training/validation and test subsets. The training/validation data were color augmented, resulting in a total of 81,786 patches for training. The held-out independent test set enrolled a total of 4,486 patches. Seven state-of-the-art convolutional neural networks were trained, validated, and tested with the same dataset.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Results</h3>\\n \\n <p>The models presented a high learning rate, yet very low generalization potential. At the model development, VGG16 performed the best, but with massive overfitting. In the test set, VGG16 presented the best accuracy, sensitivity, specificity, and area under the curve (62%, 62%, 66%, and 65%, respectively), associated with the higher loss among all Convolutional Neural Networks (CNNs) tested. EfficientB0 has comparable metrics and the lowest loss among all convolutional neural networks, being a great candidate for further studies.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Conclusion</h3>\\n \\n <p>The models were not able to generalize enough to be applied in real-life datasets due to an overlapping of features between the two classes (i.e., high risk and low risk of malignization).</p>\\n </section>\\n </div>\",\"PeriodicalId\":2,\"journal\":{\"name\":\"ACS Applied Bio Materials\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":4.6000,\"publicationDate\":\"2023-09-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ACS Applied Bio Materials\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1111/jop.13477\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"MATERIALS SCIENCE, BIOMATERIALS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACS Applied Bio Materials","FirstCategoryId":"3","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1111/jop.13477","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"MATERIALS SCIENCE, BIOMATERIALS","Score":null,"Total":0}
The use of deep learning state-of-the-art architectures for oral epithelial dysplasia grading: A comparative appraisal
Background
Dysplasia grading systems for oral epithelial dysplasia are a source of disagreement among pathologists. Therefore, machine learning approaches are being developed to mitigate this issue.
Methods
This cross-sectional study included a cohort of 82 patients with oral potentially malignant disorders and correspondent 98 hematoxylin and eosin-stained whole slide images with biopsied-proven dysplasia. All whole-slide images were manually annotated based on the binary system for oral epithelial dysplasia. The annotated regions of interest were segmented and fragmented into small patches and non-randomly sampled into training/validation and test subsets. The training/validation data were color augmented, resulting in a total of 81,786 patches for training. The held-out independent test set enrolled a total of 4,486 patches. Seven state-of-the-art convolutional neural networks were trained, validated, and tested with the same dataset.
Results
The models presented a high learning rate, yet very low generalization potential. At the model development, VGG16 performed the best, but with massive overfitting. In the test set, VGG16 presented the best accuracy, sensitivity, specificity, and area under the curve (62%, 62%, 66%, and 65%, respectively), associated with the higher loss among all Convolutional Neural Networks (CNNs) tested. EfficientB0 has comparable metrics and the lowest loss among all convolutional neural networks, being a great candidate for further studies.
Conclusion
The models were not able to generalize enough to be applied in real-life datasets due to an overlapping of features between the two classes (i.e., high risk and low risk of malignization).