Abstract
Tubular injury is the most common cause of acute kidney injury. Histopathological diagnosis may help distinguish between the different types of acute kidney injury and aid in treatment. To date, a limited number of study has used deep-learning models to assist in the histopathological diagnosis of acute kidney injury. This study aimed to perform histopathological segmentation to identify the four structures of acute renal tubular injury using deep-learning models. A segmentation model was used to classify tubule-specific injuries following cisplatin treatment. A total of 45 whole-slide images with 400 generated patches were used in the segmentation model, and 27,478 annotations were created for four classes: glomerulus, healthy tubules, necrotic tubules, and tubules with casts. A segmentation model was developed using the DeepLabV3 architecture with a MobileNetv3-Large backbone to accurately identify the four histopathological structures associated with acute renal tubular injury in PAS-stained mouse samples. In the segmentation model for four structures, the highest Intersection over Union and the Dice coefficient were obtained for the segmentation of the “glomerulus” class, followed by “necrotic tubules,” “healthy tubules,” and “tubules with cast” classes. The overall performance of the segmentation algorithm for all classes in the test set included an Intersection over Union of 0.7968 and a Dice coefficient of 0.8772. The Dice scores for the glomerulus, healthy tubules, necrotic tubules, and tubules with cast are 91.78 ± 11.09, 87.37 ± 4.02, 88.08 ± 6.83, and 83.64 ± 20.39%, respectively. The utilization of deep learning in a predictive model has demonstrated promising performance in accurately identifying the degree of injured renal tubules. These results may provide new opportunities for the application of the proposed methods to evaluate renal pathology more effectively.
Similar content being viewed by others
Introduction
Acute kidney injury (AKI) is characterized by sudden decrease in renal function. Pathologists use acute tubular injury (ATI) to describe the histopathological findings of AKI caused by damage to the tubules due to ischemia or toxin-induced toxicity. In practice, rather than using the term acute tubular necrosis (ATN), which has been traditionally employed despite the lack of necrosis in several cases, semiquantitative histopathological assessment of ATI is classified into three levels: mild, moderate, or severe1. Although the histopathology of ATI may differ between distinct pathologies, it is generally characterized by focal or diffuse tubular dilatation, thinning of the lining epithelium, vacuolation, loss of the brush border in proximal tubules, loss of nuclei, rupture of the basement membrane, and tubular cast formation in toxic acute tubular injury2,3. Kidney Disease: Improving Global Outcomes (KDIGO) urges the discovery of the etiology of AKI whenever possible4,5. Histopathological assessment may help distinguish different types of AKI and aid in patient care1.
Deep learning is the most recent machine-learning innovation and provides an unrivaled capacity to efficiently manage patients, render diagnostic support, and guide therapies6,7,8. Recent breakthroughs in deep learning, particularly convolutional neural networks (CNNs), have provided new techniques for developing systems that can assist pathologists in clinical diagnoses. Advances in whole-slide imaging technology have promoted new deep learning applications in renal histopathology9,10. Pathologists' common tasks of recognizing and identifying tissue components can be decomposed into computer vision tasks such as segmentation and detection.
Various deep learning algorithms have recently been developed for the multiclass segmentation of whole renal slide images from human and mouse kidney diseases. Most studies have focused on glomerular segmentation11. Recently, Massimo Salvi et al.12 demonstrated that an automated method using the RENTAG algorithm may be effective in quantifying glomerulosclerosis and tubular atrophy. However, few studies have used deep-learning models for the histopathological assessment of renal tubular injury after AKI. Therefore, this study was conducted to apply deep-learning models to the histopathological segmentation of the four structures in acute renal tubular injury.
In summary, our contributions are as follows. A segmentation model was developed using the DeepLabV3 architecture to accurately identify the four histopathological structures associated with acute renal tubular injury: glomerulus, necrotic tubules, healthy tubules, and tubules with cast. Our approach achieves promising performance in accurately identifying the degree of injured renal tubules.
Material and method
Kidney sample and criteria of acute kidney injury
This study was performed with the approval of the Ethical Committee of Jeonbuk National University Hospital. All methods were performed in accordance with the relevant guidelines and regulations. In a previous study, kidney samples were collected from a mouse model of cisplatin-induced acute tubular injury13. We re-analyzed kidney samples from male C57BL/6 mice (age: 8–9 weeks; weight: 20–25 g). The mice were divided into two groups: control buffer-treated and cisplatin-treated. Mice in the cisplatin group were intraperitoneally administered a single dose of cisplatin (Cis; 20 mg/kg; Sigma Chemical Co., St. Louis, MO, USA), whereas mice in the control group were intraperitoneally administered saline. Histological measurements were performed 72 h after treatment with cisplatin or the control buffer. To evaluate the function of the injured kidney, blood samples were collected three days after cisplatin administration to measure serum creatinine levels. When serum creatinine was above 0.5 mg/dL, acute kidney injury caused by cisplatin was determined.
Histopathology and assessment of tubular injury
Kidney tissue was fixed in formalin and embedded in paraffin blocks. Hematoxylin and eosin (HE) staining was performed to assess renal tubular injury. Sections of 3-µm thickness were stained using the Periodic acid-Schiff (PAS) Stain Kit (Abcam, Cambridge, MA, USA; catalog no. 150680) in accordance with the manufacturer’s instructions12,14. Tubular injury was evaluated by three blinded observers who examined at least 20 cortical fields (× 200 magnification) of the PAS-stained kidney sections. Tubular injury (necrotic tubules) was defined as tubular dilation, tubular atrophy, tubular cast formation, brush border loss, or thickening of the tubular basement membrane. Finally, the slides were digitized using a Motic Easy ScanPRO slide scanner (Motic Asia Corp., Kowloon, Hong Kong) at 40× magnification.
Datasets
Forty-five whole-slice images (WSIs) with 400 generated patches were used for the segmentation model devolopment. Ground-truth annotations were created using the SUPERVISELY polygon tool (supervisely.com). Polygons mark segment annotations by placing waypoints along the boundaries of the objects that the model must segment. All annotations were reviewed by three nephrologists with extensive experience in nephropathology. The pathologists engaged in discussions to resolve disagreements. Four predefined classes were annotated: (1) glomerulus, (2) healthy tubules, (3) necrotic tubules, and (4) tubules with casts. Figure 1A, B and C show examples of the whole-slide images of H&E and PAS-stained kidney section obtained using a slide scanner and a randomly generated patch without annotations, respectively. The annotations consisting of four different structures, ‘glomerulus,’ ‘healthy tubules,’ ‘necrotic tubules,’ and ‘tubules with cast’ are shown in Fig. 1D. In total, 27,478 annotations, along with their corresponding patches, were partitioned into two distinct proportions: a training subset comprising 80% of the data and a testing subset constituting the remaining 20%. Patches that belonged to the same WSI did not appear in either the training or testing proportions to ensure robust generalization of the segmentation models. Subsequently, to fine-tune the model hyperparameters, the training subset underwent further random splitting into training (80%) and validation (20%) subsets. This approach aimed to facilitate the refinement of model performance by iteratively adjusting the hyperparameters based on the validation set, while preserving the independence of the testing set for the final evaluation of model generalization (Table 1 and Figs. 2, 3 and 4).
Preprocessing
Because the pathology images were represented in an RGB data structure, the pixel values of the images ranged from 0 to 255. The pixels were scaled to a range between zero and one to avoid gradient explosions during the training phase. The patch images were resized to 512 × 512 pixels before being fed into the deep-learning model for segmentation. Three different augmentation methods were used to address overfitting resulting from a limited number of samples: horizontal flipping, rotation, and brightness adjustment. The third augmentation method was used because of varying degrees of slide brightness. Although we performed PAS staining for all histological slides using the same protocol, the degree of staining and, consequently, the overall brightness of the specimen may have differed among the different slides because the tissue embedded in paraffin was collected at various times. Thus, a random adjustment of the contrast of patches can improve model performance. The augmentation methods were applied only to the training and validation datasets, and not to the test set. All augmentation protocols were implemented using the Python Albumentation library15. We applied 3 augmentation methods to the 50% of the training images: (1) horizontal flip, (2) rotate images with random angles from − 90 to 90°, and (3) contrast change.
Proposed model framework
In this study, we proposed to use DeepLabV316, which is a two-stage segmentation framework for the segmentation task. The architecture of the DeepLabV3 encoder consists of Atrous Spatial Pyramid Pooling (ASPP) blocks that allow it to maintain the Field-of-View (FOV) of the network layers and effectively capture contextual information at different scales. Moreover, DeepLabV3 uses dilated or “-atrous” convolution layers to maintain high-precision predictions while maintaining a wide FOV. This is particularly critical for histopathological imaging because of the fine-grained structures and textures. In addition, the dense structure of the images leads to an extreme foreground–background class-imbalance phenomenon. To overcome this challenge, we integrated an objective function, which is the summation of the Dice Loss17 and Focal Loss18 functions. Unlike classification tasks, the outputs of segmentation problems are continuous, rather than categorical. Thus, Dice Loss is particularly suitable for continuous maps because it measures the overlap between a prediction and target. Furthermore, Dice Loss is independent of the statistical distribution of labels and penalizes misclassifications based on the overlap between the predicted regions and ground truths. The last part of our object function is the Focal Loss function, which was used in the RetinaNet18 deep-learning model to mitigate the class-imbalance problem in dense object detection. Furthermore, we integrated DeepLabV3 with a MobileNet backbone designed for mobile and embedded devices such that the developed model can be applied to devices that might have limited computational resources in clinical environments.
As presented in Table 2, our datasets were imbalanced, with the number of annotations for the Glomerulus class being relatively small compared to the other classes. To address this issue, the objective function assigns a higher weight to examples in the minority class, and a lower weight to those in the majority class. Mathematically, the objective function can be described by the following equation:
where \(y\), \(\overline{p }\), and \(\gamma\) correspond to the ground truth, model prediction, and the parameter that controls the degree of focus on the difficulty of the examples, respectively. If \(\gamma\) is set to 0, the Focal Loss is reduced to the standard cross-entropy loss. The proposed model was implemented using PyTorch19, and the loss function was obtained from the MONAI library19,20. The training procedure took approximately 4 h on a graphics processing unit (GPU) RTX 3090 24 GB.
Data analyses
Network performance was quantitatively assessed using instance-level DICE and IoU scores. In image segmentation, the DICE and IoU are commonly used to evaluate the performance of segmentation algorithms. They measured the similarity between the predicted segmentation mask and ground-truth mask. While DICE measures the ratio of the intersection of the two masks to the sum of their areas, the IoU metric calculates the overlap between the predictions and human masks by taking the ratio of their intersection to their union. In addition, sensitivity, specificity, and accuracy were calculated. In this study, we used these metrics to evaluate the performance of the proposed system comprehensively.
Comparison with other model
In our comprehensive comparative analysis, we used U-Net21 and SegFormer22, two widely used neural network architectures. U-Net, a widely used convolutional neural network architecture for semantic segmentation, features a distinctive U-shaped design comprising the contracting, bottleneck, and expansive paths. It excels at capturing intricate spatial features and is known for its success in medical image segmentation tasks. SegFormer, a state-of-the-art algorithm for segmentation, adopts a transformer-based architecture23 with lightweight multilayer perception. It demonstrates an extremely high level of performance on the Cityscapes24 dataset, highlighting its effectiveness in diverse computer vision applications. We applied the standard architectures of U-Net and SegFormer without modification and used the same training, validation, and test subsets as in our model. The DICE and IoU values of U-Net and SegFormer were measured for comparison.
Statistical analyses
We used One-way ANOVA (or t-tests) for comparison between deepLabV3, UNet and Segformer by comparing respective Dice and IoU coeffecienct. P < 0.05 was considered statistically significant.
Results
Model parameter optimization
We trained the model using the following hyperparameters: a learning rate of 0.5, batch size of 32, 60 epochs, and \(\gamma\) of 2. We evaluated the performance of each combination of hyperparameters using a held-out validation dataset. We found that the learning rate had a significant impact on model performance, with higher learning rates leading to faster convergence but a lower Dice coefficient (DICE) and Intersection over Union (IoU). In contrast, a lower learning rate results in overfitting. The batch size had a less pronounced effect, with a larger batch size generally resulting in faster convergence and improved validation performance. In addition to learning rate and batch size, we discovered that \(\gamma\) of Focal Loss was very sensitive to the performance of the model. A small value led to overfitting of the majority classes, whereas a large value resulted in poor performance in the training dataset.
Performance of segmentation model
The effectiveness of the proposed segmentation model for each class is summarized in Table 2. The average (± standard deviation) DICE scores for the glomerulus, healthy tubules, necrotic tubules, and tubules with cast were 91.78 ± 11.09, 87.37 ± 4.02, 88.08 ± 6.83, and 83.64 ± 20.39%, respectively. These results suggest that the proposed segmentation model is highly accurate in identifying different classes of objects, with the glomerulus class achieving the highest DICE score. Analysis of the IoU scores yielded similar results. The average (± standard deviation) IoU for the glomerulus, healthy tubules, necrotic tubules, and tubules with cast were 86.09 ± 12.87, 77.79 ± 6.11, 79.36 ± 10.89, and 75.49 ± 21.21%, respectively, thus demonstrating the accuracy of the proposed segmentation model across all classes with the glomerulus class achieving the highest IoU score.
In addition, the sensitivity, specificity, and accuracy of the proposed model were evaluated. The sensitivity values for the glomerulus, healthy tubules, necrotic tubules, and tubules with cast were 84.84 ± 27.11, 86.72 ± 6.49, 75.96 ± 32.59, and 69.44 ± 35.87%, respectively. The specificity values for the glomerulus, healthy tubules, necrotic tubules, and tubules with cast were 99.69 ± 2.65, 90.25 ± 10.04, 88.54 ± 7.32, and 98.74 ± 1.29%, respectively. The accuracy values for the glomerulus, healthy tubules, necrotic tubules, and tubules with cast were 99.43 ± 0.37, 90.95 ± 3.27, 90.07 ± 5.13, and 97.98 ± 1.85%, respectively.
Comparison with other studies
We compared our model with existing state-of-the-art methods (U-Net and SegFormer) for histopathological assessment of renal tubular injury. Table 3 presents a comparison between the performances of the three models for the testing subset. Our model (DeepLabV3) exhibited a comparable or slightly better performance than SegFormer. The performance of the proposed model was better than that of U-Net, particularly in segmenting necrotic tubules and tubules with cast.
Discussion
Over the last decade, numerous studies have focused on the development of deep-learning models for nephropathology. In several previous studies, neural networks have been trained and successfully applied to specific glomerular segmentation tasks, such as distinguishing between glomerular and non-glomerular regions and classifying healthy and injured glomeruli in WSIs of both human disease and animal models25,26,27. In 2020, Uchino et al. developed a comprehensive deep-learning model to classify multiple glomerular images and suggested its potential use in enhancing the diagnostic accuracy for clinicians28.
The initial results of the multiclass segmentation task for kidneys were reported in 201829. They proposed a method for renal segmentation of PAS-stained digital slides of renal allograft resections using CNNs for nine classes, including five healthy structures (glomerulus, distal tubules, proximal tubules, arterioles, and capillaries) and four pathological structures (atrophic tubules, sclerotic glomeruli, fibrotic tissue, and inflammatory infiltrates). Three different network architectures were used to perform this task: a fully convolutional network, U-net, and a multiscale fully convolutional network.
Another CNN for the multiclass segmentation of kidney sections with PAS staining was developed by Hermsen et al.30. Dice coefficients were used to assess the segmentation performance for ten classes (glomerulus, sclerotic glomerulus, empty Bowman's capsules, proximal tubules, distal tubules, atrophic tubules, undefined tubules, arteries, interstitium, and capsule) of nephrectomy and transplant biopsy specimens. In both datasets, the glomerulus was the best-segmented class (Dice coefficients of 0.95 and 0.94)30. Recently, Bouteldja et al. published high-performance deep-learning algorithms for the multiclass segmentation of kidney histology for various diseases in mouse models and other species. In this study, six annotated structures were used: tubules, full glomerulus, glomerular tuft, artery, arterial lumen, and vein31. Although previous studies have focused on developing models for segmenting renal tubular structures, the predefined classes of tubules included only normal tubular types, such as proximal and distal tubules, or abnormal tubular types, such as atrophic tubular structures, in a renal fibrosis model32.
To the best of our knowledge, there have been a limited number of reports on segmentation models for identifying injured tubules in patients with acute kidney injury. Our study presents a deep learning-based segmentation model for evaluating acute renal tubular injury in digitized PAS-stained images. We applied deep-learning models to identify the typical structural types of toxicity-induced acute tubular injuries, including glomeruli, healthy tubules, necrotic tubules, and tubules with casts. The DICE scores and IoU showed high and consistent performances in the segmentation of these regions. Notably, the performance of the proposed model was the highest for the glomerulus despite the glomerulus class having the smallest number of annotations. This suggests that the performance of the model can be improved further by adding more training data, particularly for the glomerulus class. Overall, the results suggest that the proposed segmentation model has the potential to be used in clinical applications for the accurate identification and segmentation of different kidney structures, particularly injured tubules. In future, we intend to translate the technique developed in this study to a human biopsy dataset. As a dissociation exists between histopathological findings and the clinical symptoms of AKI in some cases (such as volume depletion-induced AKI in allergic, cardiogenic, or hemorrhagic shock), renal biopsy may assist in assessing structural injury, differentiating the cause of AKI, and aiding in treatment1.
The proposed approach exhibited a similar or slightly higher performance than the state-of-the-art models. The mean DICE values for SegFormer and U-Net were 81.49% (ranging from 75.69 to 86.69%) and 70.27% (ranging from 53.66 to 82.18%), respectively, across the four classes, whereas our model yielded a mean DICE of 87.71% (ranging from 83.64 to 91.78%). The mean IoUs for SegFormer and U-Net were 69.97% (ranging from 61.55 to 76.77%) and 62.48% (ranging from 53.41 to 72.78%) across the four classes, respectively, whereas our model had a mean IoU of 79.68% (ranging from 75.49 to 86.09%). Therefore, compared with previously used methods for assessing renal tubular injury, the method proposed in this study may be effective for identifying injured renal tubules in acute kidney injury in terms of segmentation performance and computational complexity. It is noteworthy that our model exhibited a comparable or slightly better performance than Segformer, with significantly simpler computational complexity. SegFormer produced results with a high degree of parameter counts of 64 million, whereas our model, DeepLabV3, based on Mobile-net, presented relatively high efficiency with only 11 million parameter counts. This efficiency underscores the potential practical advantages of our model in terms of computational resources and model complexity.
Our study has some limitations. First, a deep-learning model was developed to evaluate the histological images of murine cisplatin-induced acute tubular injury. Although the histological structures of the mouse and human kidneys are similar, the distance or connective tissue area among the structures in the mouse kidney tissue is relatively small compared to that in humans. These closely located structures make it more difficult to distinguish the boundaries between them, particularly in necrotic areas where the basement membranes are occasionally not intact. Second, the number of WSIs and patches generated in this study was limited. A study that includes a larger number of annotations is underway and is expected to achieve higher performance in training the model. Third, when substances such as casts are present in the injured tubular lumen, the effectiveness of measuring the degree of tubular injury decreases.
Conclusion
The deep-learning segmentation model developed in this study can accurately identify the histopathological structures of injured renal tubules. The results serve as the basis for future studies with larger datasets, including mouse and human biopsy samples, which can provide new opportunities for applying the proposed methods to renal pathology.
Data availability
The datasets used and/or analysed during the current study available from the corresponding author on reasonable request.
References
Gaut, J. P. & Liapis, H. Acute kidney injury pathology and pathophysiology: a retrospective review. Clin. Kidney J. 14, 526–536. https://doi.org/10.1093/ckj/sfaa142 (2021).
Tavares, M. B. et al. Acute tubular necrosis and renal failure in patients with glomerular disease. Ren. Fail. 34, 1252–1257. https://doi.org/10.3109/0886022x.2012.723582 (2012).
Racusen, L. C. The histopathology of acute renal failure. New Horiz. 3, 662–668 (1995).
Khwaja, A. KDIGO clinical practice guidelines for acute kidney injury. Nephron Clin. Pract. 120, c179-184. https://doi.org/10.1159/000339789 (2012).
Okusa, M. D. & Davenport, A. Reading between the (guide)lines–the KDIGO practice guideline on acute kidney injury in the individual patient. Kidney Int. 85, 39–48. https://doi.org/10.1038/ki.2013.378 (2014).
Ramesh, A. N., Kambhampati, C., Monson, J. R. & Drew, P. J. Artificial intelligence in medicine. Ann. R. Coll. Surg. Engl. 86, 334–338. https://doi.org/10.1308/147870804290 (2004).
Miotto, R., Wang, F., Wang, S., Jiang, X. & Dudley, J. T. Deep learning for healthcare: Review, opportunities and challenges. Brief Bioinform. 19, 1236–1246. https://doi.org/10.1093/bib/bbx044 (2018).
Park, K. et al. Deep learning predicts the differentiation of kidney organoids derived from human induced pluripotent stem cells. Kidney Res. Clin. Pract. 42, 75–85. https://doi.org/10.23876/j.krcp.22.017 (2023).
Becker, J. U. et al. Artificial intelligence and machine learning in nephropathology. Kidney Int. 98, 65–75. https://doi.org/10.1016/j.kint.2020.02.027 (2020).
Ghaznavi, F., Evans, A., Madabhushi, A. & Feldman, M. Digital imaging in pathology: whole-slide imaging and beyond. Annu. Rev. Pathol. 8, 331–359. https://doi.org/10.1146/annurev-pathol-011811-120902 (2013).
Huo, Y., Deng, R., Liu, Q., Fogo, A. B. & Yang, H. AI applications in renal pathology. Kidney Int. 99, 1309–1320. https://doi.org/10.1016/j.kint.2021.01.015 (2021).
Salvi, M. et al. Automated assessment of glomerulosclerosis and tubular atrophy using deep learning. Comput. Med. Imaging Graph. 90, 101930. https://doi.org/10.1016/j.compmedimag.2021.101930 (2021).
Jung, Y. J., Park, W., Kang, K. P. & Kim, W. SIRT2 is involved in cisplatin-induced acute kidney injury through regulation of mitogen-activated protein kinase phosphatase-1. Nephrol. Dial. Transplant. 35, 1145–1156. https://doi.org/10.1093/ndt/gfaa042 (2020).
Jiang, L. et al. A deep learning-based approach for glomeruli instance segmentation from multistained renal biopsy pathologic images. Am. J. Pathol. 191, 1431–1441. https://doi.org/10.1016/j.ajpath.2021.05.004 (2021).
Buslaev, A. et al. Albumentations: fast and flexible image augmentations. arXiv:1809.06839 (2018).
Chen, L. C., Papandreou, G., Schroff, F. & Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv:1706.05587 (2017).
Sudre, C. H., Li, W., Vercauteren, T., Ourselin, S. & Jorge Cardoso, M. Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support 240–248 (Springer, 2017).
Lin, T. Y., Goyal, P., Girshick, R., He, K. & Dollar, P. Focal loss for dense object detection. IEEE Trans. Pattern Anal. Mach. Intell. 42, 318–327. https://doi.org/10.1109/tpami.2018.2858826 (2020).
Paszke, A. et al. Pytorch: An imperative style, high-performance deep learning library. arXiv:1912.01703 (2019)
Cardoso, M. J. et al. MONAI: An open-source framework for deep learning in healthcare (2022).
Ronneberger, O., Fischer, P. & Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. CoRR https://arxiv.org/abs/1505.04597 (2015).
Xie, E., Wang, W., Yu, Z., Anandkumar, A., Alvarez J. M. & Luo, P. SegFormer: Simple and efficient design for semantic segmentation with transformers. In Advances in Neural Information Processing Systems. https://doi.org/10.48550/arXiv.2105.15203 (2021).
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, Ł. & Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems (eds. Guyon, I., Von Luxburg, U., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S. & Garnett, R.) Vol. 30 (Curran Associates, Inc., 2017).
Cordts, M., Omran, M., Ramos, S., Rehfeld, T., Enzweiler, M., Benenson, R., Franke, U., Roth, S. & Schiele, B. The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016).
Kannan, S. et al. Segmentation of glomeruli within trichrome images using deep learning. Kidney Int. Rep. 4, 955–962. https://doi.org/10.1016/j.ekir.2019.04.008 (2019).
Gadermayr, M. et al. Segmenting renal whole slide images virtually without training data. Comput. Biol. Med. 90, 88–97. https://doi.org/10.1016/j.compbiomed.2017.09.014 (2017).
Bueno, G., Fernandez-Carrobles, M. M., Gonzalez-Lopez, L. & Deniz, O. Glomerulosclerosis identification in whole slide images using semantic segmentation. Comput. Methods Programs Biomed. 184, 105273. https://doi.org/10.1016/j.cmpb.2019.105273 (2020).
Uchino, E. et al. Classification of glomerular pathological findings using deep learning and nephrologist-AI collective intelligence approach. Int. J. Med. Inform. 141, 104231. https://doi.org/10.1016/j.ijmedinf.2020.104231 (2020).
de Bel, T. et al. Automatic segmentation of histopathological slides of renal tissue using deep learning. In SPIE. https://doi.org/10.1117/12.2293717 (2018).
Hermsen, M. et al. Deep learning-based histopathologic assessment of kidney tissue. J. Am. Soc. Nephrol. 30, 1968–1979. https://doi.org/10.1681/asn.2019020144 (2019).
Bouteldja, N. et al. Deep learning-based segmentation and quantification in experimental kidney histopathology. J. Am. Soc. Nephrol. 32, 52–68. https://doi.org/10.1681/asn.2020050597 (2021).
Ginley, B. et al. Computational segmentation and classification of diabetic glomerulosclerosis. J. Am. Soc. Nephrol. 30, 1953–1967. https://doi.org/10.1681/asn.2018121259 (2019).
Acknowledgements
We would like to thank Editage (www.editage.co.kr) for English language editing.
Funding
This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education, Science, and Technology (NRF-2023R1A2C2006929, K.W., NRF-2020R1I1A3055652, J.Y.J., 2022R1I1A1A01056360, P.W., NRF-2022R1I1A3072856, I. P. and RS-2023-00242528, I. P.), Institute of Information & communications Technology Planning & Evaluation (IITP) under the Artificial Intelligence Convergence Innovation Human Resources Development (IITP-2023-RS-2023-00256629, I. P.) funded by the Korea government (MSIT) and the Fund of Biomedical Research Institute, Jeonbuk National University Hospital (K.W.).
Author information
Authors and Affiliations
Contributions
T.T.U.N., W.K., H.K., Y.J.J., I.P. and W.P. performed study concept and design; T.T.U.N., A-T.N., W.K., and I.P. performed development of methodology and writing, review and revision of the paper; T.T.U.N., A-T.N., I.P. and W.K. provided acquisition, analysis and interpretation of data, and statistical analysis; K.M.K. provided technical and material support. All authors read and approved the final paper.
Corresponding authors
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Nguyen, T.T.U., Nguyen, AT., Kim, H. et al. Deep-learning model for evaluating histopathology of acute renal tubular injury. Sci Rep 14, 9010 (2024). https://doi.org/10.1038/s41598-024-58506-9
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-024-58506-9
Keywords
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.