Skip to main content

Automation of Wilms’ tumor segmentation by artificial intelligence

Abstract

Background

3D reconstruction of Wilms’ tumor provides several advantages but are not systematically performed because manual segmentation is extremely time-consuming. The objective of our study was to develop an artificial intelligence tool to automate the segmentation of tumors and kidneys in children.

Methods

A manual segmentation was carried out by two experts on 14 CT scans. Then, the segmentation of Wilms’ tumor and neoplastic kidney was automatically performed using the CNN U-Net and the same CNN U-Net trained according to the OV2ASSION method. The time saving for the expert was estimated depending on the number of sections automatically segmented.

Results

When segmentations were performed manually by two experts, the inter-individual variability resulted in a Dice index of 0.95 for tumor and 0.87 for kidney. Fully automatic segmentation with the CNN U-Net yielded a poor Dice index of 0.69 for Wilms’ tumor and 0.27 for kidney. With the OV2ASSION method, the Dice index varied depending on the number of manually segmented sections. For the segmentation of the Wilms’ tumor and neoplastic kidney, it varied respectively from 0.97 to 0.94 for a gap of 1 (2 out of 3 sections performed manually) to 0.94 and 0.86 for a gap of 10 (1 section out of 6 performed manually).

Conclusion

Fully automated segmentation remains a challenge in the field of medical image processing. Although it is possible to use already developed neural networks, such as U-Net, we found that the results obtained were not satisfactory for segmentation of neoplastic kidneys or Wilms’ tumors in children. We developed an innovative CNN U-Net training method that makes it possible to segment the kidney and its tumor with the same precision as an expert while reducing their intervention time by 80%.

Backgroung

Wilms’ tumor, or nephroblastoma, is one of the most common malignant tumors in children, affecting 1 in 10,000 children [1, 2]. Irrespective of the therapeutic protocol followed (International Society of Paediatric Oncology or Children’s Oncology Group), surgery retains an essential role in the care of such children. 3D reconstruction of the neoplastic kidney, based on patient imaging data, provides several advantages: pre-operative surgical planning, anticipation of operative risks (especially vascular), help in the selection of patients who can benefit from nephron-sparing surgery, and improvement of the information provided to families [3]. Currently, these 3D reconstructions are not systematically performed in clinical practice. Indeed, the construction of 3D models requires a preliminary segmentation phase (assigning a label to each pixel of the image), which can be extremely time-consuming and a source of human error when performed manually [3, 4].

Artificial intelligence (AI) encompasses a set of concepts and technologies that enable machines to simulate human intelligence. It uses artificial neural networks, mathematical logic, and computer science. Thus, AI can be defined as a set of algorithms that gives machines the ability to reason or perform certain cognitive functions such as problem-solving, object or word recognition, and decision-making [5]. Deep Learning refers to a subset of machine learning techniques that involve training and using artificial neural networks with multiple layers to learn and extract complex patterns and representations from data. There are different kinds of deep learning architectures that can be used for medical image analysis, and most of them are built from convolutional neural networks (CNN) [4]. Thus, CNN are a type of deep learning model specifically designed for processing and analyzing structured data, such as images or time-series data. They utilize convolutional layers to automatically extract hierarchical features from the input data, making them highly effective for tasks like image recognition, classification or segmentation [6]. In the context of tumor pathology, these tools can be used for tumor segmentation, differential diagnosis, tumor staging and grading [4].

The objective of our study was to develop an artificial intelligence tool to, as much as possible, automate the segmentation of the neoplastic kidney in order to limit the need for intervention by a human expert and thus allow it to be performed in routine clinical practice.

Materials and methods

We built a database from 14 CT scans of 12 patients who had been treated for Wilms’ tumor in our department at the University Hospital of Besançon, France. All scanners had an arterial contrast phase, and 5 scanners had a late acquisition time allowing the urinary tract to be assessed. For each CT scan, we performed manual or semi-automatic segmentation of healthy kidneys, neoplastic kidneys (preserved renal parenchyma around the tumor), Wilms’ tumors, arterial vascularization, venous vascularization, and urinary cavities, using 3D Slicer software version 4.8.1 (https://www.slicer.org/). This manual segmentation was carried out by two different experts in order to calculate the inter-individual variability (one pediatric surgeon and one pediatric radiologist with more than 5 years’ experience in pediatric oncology). The patient demographics and segmentation method were described in a previous paper [3]. These established data were then available for the development and training of AI tools.

Convolutional neural networks (CNN)

Segmentation of the neoplastic kidney and the tumor was initially performed with U-Net [7], which is the most commonly used CNN for medical image segmentation. U-Net has two stages: a down-sampling stage (an encoder process uses the max-pooling strategy to compress image features) and an up-sampling stage (a decoder process uses the unpooling strategy to outpout the results) [8]. For each patient p, the U-Net has been trained over all others the patients except p, during 200 epochs (the number of epochs is a hyperparameter of CNN that controls the number of complete passes through the training dataset), with a batch size of 16 (the batch size is a hyperparameter of CNN that controls the number of training samples to work through before the model’s internal parameters are updated).

The segmentation was then performed with the same CNN U-Net trained according to the OV2ASSION (overlearning vector for valid sparse segmentations) training method, described in a previous article [9, 10]. In this method, CNN training is performed with a variable number of patient CT scan sections that have been manually segmented by an expert. The objective of the CNN is then to automatically complete the segmentation of the missing sections. The results obtained are then used to define the minimum number of manually segmented sections so that the CNN completes the segmentation automatically with good accuracy. The notion of “gap”, therefore, defines the interval between two manually segmented sections (Fig. 1). The larger the gap, the greater the interval between two sections and, therefore, the more the need for intervention by an expert is restricted. The time saving for the human expert is estimated depending on the number of sections automatically segmented. Computations have been performed on the supercomputer facilities of the Franche-Comté Computation Mesocenter.

Fig. 1
figure 1

Representation of manually segmented sections and the notion of a gap with the OV2ASSION method

Evaluation of the results

The segmentations obtained by AI were compared to the manual segmentations performed by an expert using the Dice similarity index according to the following formula:

$${\rm{Dice}}\,{\rm{similarity}}\,{\rm{index}}\,{\rm{ = }}\,\frac{{2\left| {X \cap Y} \right|}}{{\left| {X \cup Y} \right|}}\, = \,\frac{{2A}}{{2A + B + C}}$$

where:

X is the segmentation obtained by AI.

Y is the segmentation performed manually by an expert.

A is the number of common pixels, present in class X and in class Y.

B is the number of pixels present in class Y and absent in class X.

C is the number of pixels present in class X and absent in class Y.

Results

When segmentations were performed manually by two experts, the inter-individual variability resulted in an average Dice index of 0.95 [0.91–0.97] for Wilms’ tumor (Table 1) and 0.87 [0.69–0.96] for neoplastic kidney (Table 2). The Dice index was greater than 0.90 for all patients concerning the renal tumor and it was less than 0.80 for two patients concerning the neoplastic kidney.

Table 1 Dice indices and estimated time saving for the segmentation of Wilms’ tumors with manual segmentation (inter-individual variability), automatic segmentation (CNN U-Net) and semi-automatic segmentation (CNN U-Net trained with the OV2ASSION method)
Table 2 Dice indices and estimated time saving for the segmentation of neoplastic kidneys with manual segmentation (inter-individual variability), automatic segmentation (CNN U-Net) and semi-automatic segmentation (CNN U-Net trained with the OV2ASSION method)

Of the 14 patients tested, automatic segmentation with the CNN U-Net yielded an average Dice index of 0.69 [0.01–0.93] for Wilms’ tumor segmentation (Table 1) and 0.27 [0.02–0.56] for neoplastic kidney segmentation (Table 2). The Dice index was greater than 0.80 for 9 out of 14 patients regarding renal tumor segmentation. In contrast, the Dice index was less than 0.80 for all patients regarding neoplastic kidney segmentation, with a maximum Dice value of 0.56.

With the OV2ASSION drive method, the average Dice index varied depending on the number of manually segmented sections. Logically, the Dice index tends to decrease when the gap increases because it means that the CNN has less established data for training. For the segmentation of the Wilms’ tumors (Table 1), it varied from 0.97 for a gap of 1 (2 out of 3 sections performed manually) to 0.94 for a gap of 10 (1 section out of 6 performed manually). The Dice index was greater than 0.80 for all patients, regardless of the gap used. For neoplastic kidney segmentation (Table 2), the average Dice index ranged from 0.94 for a gap of 1 to 0.86 for a gap of 10. Twelve patients had a Dice index greater than 0.80 for all gaps tested. On the other hand, two patients (patients 9 and 10) had a Dice index below 0.80, with a poor result for patient 9 since the Dice index decreased to less than 0.70 from gap 7 and fell below 0.50 from gap 9.

The Fig. 2 shows an example of renal and tumor segmentations achieved by CNN U-Net with and without OV2ASSION method. Thus, the CNN driven by the OV2ASSION method obtained identical results to the inter-individual variability for a gap of 7 concerning the Wilms’ tumor and for a gap of 8 concerning the neoplastic kidney. In other words, the segmentation with the AI tool was as precise as the segmentation by an expert when provided with 1 out of 5 manually segmented sections. The time saving for the human expert varied depending on the gap, from 33% for a gap of 1 (1 section out of 3 was automatically segmented) to 83% for a gap of 10 (10 sections out of 12 were automatically segmented).

Fig. 2
figure 2

Results of renal (first line) and tumor (second line) segmentations. Column A: CT cross-sections. Column B: manual segmentations performed by a human expert. Column C: automatic segmentations obtained with CNN U-Net. Column D: segmentations obtained with CNN U-Net + OV2ASSION training method

Discussion

AI tools are increasingly used for analysis and processing of medical images, allowing different tasks to be performed such as classification, detection, and segmentation [6]. In our study, we used convolutional neural networks for segmentation of Wilms’ tumors in children. Although it is possible to use already developed neural networks, such as U-Net [7] or FCN [11], we found that the results obtained were not satisfactory for segmentation of neoplastic kidneys or Wilms’ tumors in children. Indeed, use of the CNN U-Net on our sample yielded an average Dice index for Wilms’ tumor segmentation (0.69) and a poor index for neoplastic kidney segmentation (0.27). Performing a fully automated segmentation remains a challenge in the field of medical image processing, especially when pathological situations in children are being investigated. Several authors have proposed AI tools to achieve good results with fully automated segmentation of healthy kidneys [12, 13]. However, the situation is complicated when pathological kidneys are examined, whether or not they are malformed [14] or neoplastic [15]. To overcome this problem, a competition was created in adults, called the 2019 Kidney and Kidney Tumor Segmentation Challenge (KiTS19) at the International Conference on Medical Image Computing and Computer Assisted Intervention [16]. This competition had two objectives: (1) to allow a fair and objective comparison of the various methods (since all teams had the same training set and were evaluated by the same metrics on the same test set) and (2) to stimulate research on the challenge of automatic segmentation by making a quantity of established data available for the entire international research community. The results proved to be very good, as the winning team had developed an AI tool capable of automatically segmenting the pathological kidney and the renal tumor with average Dice indices of 0.97 and 0.85, respectively [16]. Other studies have demonstrated good results in renal tumor segmentation in adults using deep learning methods [8, 17, 18]. However, these results are not transferable to Wilms’ tumors in children. Indeed, this tumor type is very different from kidney tumors in adults. They are often much larger, thus compressing the renal parenchyma, which can be very thinned or even fragmented and clearly modify the usual anatomical ratios. In addition, Wilms’ tumor can have a very heterogeneous appearance from one patient to another (regarding its location, size, spatial complexity, intensity, contrast, or relationships with neighboring organs), thus complicating the learning of AI tools.

Another obstacle concerns the quality of medical images, especially CT scans, which is often worse in children than in adults. This can be explained by the use of low-dose acquisition protocols to limit the irradiation of children, by greater movement artifacts during image acquisition, and by a lower amount of fat in children decreasing contrast within the image [3]. It is also possible to use magnetic resonance imaging (MRI). However, the time taken to acquire such images is relatively long, so these examinations may need to be performed under general anesthesia. The progress made in this field with the emergence of more efficient MRI units has reduced the image acquisition time and may hence have reduced this constraint.

The amount of information available (i.e., images segmented and labeled by an expert) is still very limited, as this pathology remains relatively rare. One of the main challenges in medical imaging-based deep learning in kidney diseases is the lack of large, diverse datasets [4]. This is undeniably a major obstacle to the training of AI tools, largely explaining the disappointing results obtained in our study by CNN U-Net.

This is why we have developed a special training method called OV2ASSION. This method allows the AI tool to perform learning from a few patient sections manually segmented by an expert. This method is certainly not completely automated, but it allows for a quite significant reduction in the duration of the expert’s intervention. In our study, the CNN managed to automatically complete the segmentation of the neoplastic kidney and Wilms’ tumor with the same accuracy as an expert when it was trained based on one manually segmented section out of 5. If the segmentation time is assumed to be identical for each section, this reduces the intervention time of the expert by 80%.

The original objective has not been fully met as the method presented here is not entirely automated. However, several perspectives are worth further consideration to achieve this. To improve the performance of AI tools, it is essential to increase the amount of information (i.e., source images with their labeled segmentations) for the learning phase. One can then envision creating a database specific to Wilms’ tumor in children that can be used by the entire international scientific community (similar to what was done with the Kidney and Kidney Tumor Segmentation Challenge in adults). It is also possible to artificially increase this amount of information using only computer processes with data augmentation (where the number of images can be artificially increased using rotation, translation, or reversal techniques on the available images).

It would also be worthwhile to optimize the performance of AI tools by providing them with anatomical knowledge [19] so they can reason like an expert when performing manual segmentation. This then raises the question of the organization and prioritization of this knowledge so that it can be read and exploited by computer tools. Several ways can be envisioned to achieve this, such as the use of ontologies (for pathological anatomy) or atlases (for normal anatomy).

The development of a fully automated method remains a complex challenge, especially in regard to the tumor pathology of children. An alternative would, therefore, be to move towards the development of a semi-automated method during which intervention by the expert would be very limited (such as manually positioning the germinating pixel on a section with a single click or marking the boundaries of the tumor by a few reference points on a limited number of sections). Although this method is not fully automated, it would still be readily usable by the operator in daily clinical practice.

Finally, it will be essential to extend these segmentation processes to the arterial vascularization, venous vascularization, and urinary tract since it is paramount that these anatomical elements are also analyzed for decision-making [20]. This will lead to a new challenge to be solved, that of registration, because these structures can only be segmented on images acquired at different times.

Conclusions

Fully automated segmentation remains a challenge in the field of medical image processing, especially when the tumor pathology in children is assessed. We developed a process for segmentation of Wilms’ tumors and neoplastic kidneys using the CNN U-Net trained according to the OV2ASSION method. This technique, which is not fully automated, makes it possible to carry out segmentation of the kidney and its tumor with the same precision as an expert while reducing their intervention time by 80%.

Data availability

The data sets generated during and/or analyzed during the current study are not publicly available due to intellectual property/privacy but are available from the corresponding author on reasonable request.

Abbreviations

AI:

Artificial Intelligence

CNN:

Convolutional Neural Networks

CT:

Computed Tomography

OV2ASSION:

Overlearning vector for valid sparse segmentations

References

  1. Owens CM, Brisse HJ, Olsen OE, Begent J, Smets AM. Bilateral disease and new trends in Wilms tumour. Pediatr Radiol. 2008;38(1):30–9. https://doi.org/10.1007/s00247-007-0681-0

    Article  PubMed  Google Scholar 

  2. Han Q, Li K, Dong K, Xiao X, Yao W, Liu G. Clinical features, treatment, and outcomes of bilateral Wilms’ tumor: a systematic review and meta-analysis. J Pediatr Surg. 2018;53(12):2465–9. https://doi.org/10.1016/j.jpedsurg.2018.08.022

    Article  PubMed  Google Scholar 

  3. Chaussy Y, Vieille L, Lacroix E, Lenoir M, Marie F, Corbat L, et al. 3D reconstruction of Wilms’ tumor and kidneys in children: variability, usefulness and constraints. J Pediatr Urol. 2020;16(6):830.e1-830.e8. https://doi.org/10.1016/j.jpurol.2020.08.023

    Article  Google Scholar 

  4. Zhang M, Ye Z, Yuan E, Lv X, Zhang Y, Tan Y, et al. Imaging-based deep learning in kidney diseases: recent progress and future prospects. Insights Imaging. 2024;15(1):50. https://doi.org/10.1186/s13244-024-01636-5

    Article  PubMed  PubMed Central  Google Scholar 

  5. Wall J, Krummel T. The digital surgeon: how big data, automation, and artificial intelligence will change surgical practice. J Pediatr Surg. 2020;55S:47–50. https://doi.org/10.1016/j.jpedsurg.2019.09.008

    Article  PubMed  Google Scholar 

  6. Litjens G, Kooi T, Bejnordi BE, Setio AAA, Ciompi F, Ghafoorian M, et al. A survey on deep learning in medical image analysis. Med Image Anal. 2017;42:60–88. https://doi.org/10.1016/j.media.2017.07.005

    Article  PubMed  Google Scholar 

  7. Ronneberger O, Fischer P, Brox T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In: Navab N editors. Medical Image Computing and Computer-assisted intervention – MICCAI 2015. Lecture Notes in Computer Science, vol 9381. Springer; Cham. p 234–41. https://doi.org/10.1007/978-3-319-24574-4_28

  8. Yang Y, Chen F, Liang H, Bai Y, Wang Z, Zhao L, et al. CNN-based automatic segmentations and radiomics feature reliability on contrast-enhanced ultrasound images for renal tumors. Front Oncol. 2023;13:1166988. https://doi.org/10.3389/fonc.2023.1166988

    Article  PubMed  PubMed Central  Google Scholar 

  9. Marie F, Corbat L, Chaussy Y, Delavelle T, Henriet J, Lapayre J-C. Segmentation of deformed kidneys and nephroblastoma using case-based reasoning and convolutional neural network. Expert Syst Appl. 2019;127:282–94. https://doi.org/10.1016/j.eswa.2019.03.010

    Article  Google Scholar 

  10. Corbat L, Henriet J, Chaussy Y, Lapayre J-C. Fusion of multiple segmentations of medical images using OV2ASSION and deep learning methods: application to CT-Scans for tumoral kidney. Comput Biol Med. 2020;124:103928. https://doi.org/10.1016/j.compbiomed.2020.103928

    Article  PubMed  Google Scholar 

  11. Long J, Shelhamer E, Darrell T. Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2015:3431–3440. https://openaccess.thecvf.com/content_cvpr_2015/papers/Long_Fully_Convolutional_Networks_2015_CVPR_paper.pdf

  12. Kim T, Lee KH, Ham S, Park B, Lee S, Hong D, et al. Active learning for accuracy enhancement of semantic segmentation with CNN-corrected label curations: evaluation on kidney segmentation in abdominal CT. Sci Rep. 2020;10(1):366. https://doi.org/10.1038/s41598-019-57242-9

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  13. Müller D, Kramer F. MIScnn: a framework for medical image segmentation with convolutional neural networks and deep learning. BMC Med Imaging. 2021;21(1):12. https://doi.org/10.1186/s12880-020-00543-7

    Article  PubMed  PubMed Central  Google Scholar 

  14. Sharma K, Rupprecht C, Caroli A, Aparicio MC, Remuzzi A, Baust M, et al. Automatic segmentation of kidneys using deep learning for total kidney volume quantification in autosomal Dominant polycystic kidney disease. Sci Rep. 2017;7(1):2049. https://doi.org/10.1038/s41598-017-01779-0

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  15. Qayyum A, Lalande A, Meriaudeau F. Automatic segmentation of tumors and affected organs in the abdomen using a 3D hybrid model for computed tomography imaging. Comput Biol Med. 2020;127:104097. https://doi.org/10.1016/j.compbiomed.2020.104097

    Article  PubMed  Google Scholar 

  16. Heller N, Isensee F, Maier-Hein KH, Hou X, Xie C, Li F, et al. The state of the art in kidney and kidney tumor segmentation in contrast-enhanced CT imaging: results of the KiTS19 challenge. Med Image Anal. 2021;67:101821. https://doi.org/10.1016/j.media.2020.101821

    Article  PubMed  Google Scholar 

  17. Feng H, Kou X, Tang Z, Li L, FYU-Net. A Cascading Segmentation Network for Kidney Tumor Medical Imaging. Comput Math Methods Med. 2022;2022:4792532. https://doi.org/10.1155/2022/4792532

    Article  PubMed  PubMed Central  Google Scholar 

  18. Chen S, Song D, Chen L, Guo T, Jiang B, Liu A, et al. Artificial intelligence-based non-invasive tumor segmentation, grade stratification and prognosis prediction for clear-cell renal-cell carcinoma. Precis Clin Med. 2023;6(3):pbad019. https://doi.org/10.1093/pcmedi/pbad019

    Article  PubMed  PubMed Central  Google Scholar 

  19. Liu L, Wolterink JM, Brune C, Veldhuis RNJ. Anatomy-aided deep learning for medical image segmentation: a review. Phys Med Biol. 2021;66(11):11TR01. https://doi.org/10.1088/1361-6560/abfbf4

    Article  Google Scholar 

  20. He Y, Yang G, Yang J, Ge R, Kong Y, Zhu X, et al. Meta grayscale adaptive network for 3D integrated renal structures segmentation. Med Image Anal. 2021;71:102055. https://doi.org/10.1016/j.media.2021.102055

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

The authors would like to thank the European Union, the Swiss Confederation and the Swiss cantons of Vaud and Neuchatel for financing this project as part of the SAIAD and SAIAD 2 INTERREG V France-Switzerland programs and also the SAIAD and SAIAD 2 consortium partners. Computations have been performed on the supercomputer facilities of the Franche-Comté Computation Mesocenter.

We thank Sophie Domingues for her assistance for the preparation of this manuscript (language help).

Funding

This work was supported by the European Regional Development Fund (European Program Interreg V France-Suisse 2014–2020).

Author information

Authors and Affiliations

Authors

Contributions

Conception and design of the study: Olivier Hild, Pierre Berriet, Yann Chaussy.

Acquisition of data: Pierre Berriet, Lor?dane Salvi, Marion Lenoir, Fr?d?ric Auber, Yann Chaussy.

Analysis and interpretation of data: Olivier Hild, J?r?mie Nallet, Julien Henriet, Jean-Philippe Thiran, Yann Chaussy.

Drafting the manuscript: Olivier Hild, Pierre Berriet, Lor?dane Salvi, Yann Chaussy.

Critically revising the manuscript: J?r?mie Nallet, Marion Lenoir, Julien Henriet, Jean-Philippe Thiran, Fr?d?ric Auber.

Final approval: All authors.

Corresponding author

Correspondence to Yann Chaussy.

Ethics declarations

Ethics approval and consent to participate

This work was approved by our internal Institutional Review Board at CHU Besançon.

Consent for publication

Patient consent forms are not required due to the type of this research with non-identifying data.

Competing interests

The authors declare that they have no competing interest.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Hild, O., Berriet, P., Nallet, J. et al. Automation of Wilms’ tumor segmentation by artificial intelligence. Cancer Imaging 24, 83 (2024). https://doi.org/10.1186/s40644-024-00729-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s40644-024-00729-0

Keywords