Abstract
During the process of labor, the intrapartum transperineal ultrasound examination serves as a valuable tool, allowing direct observation of the relative positional relationship between the pubic symphysis and fetal head (PSFH). Accurate assessment of fetal head descent and the prediction of the most suitable mode of delivery heavily rely on this relationship. However, achieving an objective and quantitative interpretation of the ultrasound images necessitates precise PSFH segmentation (PSFHS), a task that is both time-consuming and demanding. Integrating the potential of artificial intelligence (AI) in the field of medical ultrasound image segmentation, the development and evaluation of AI-based models rely significantly on access to comprehensive and meticulously annotated datasets. Unfortunately, publicly accessible datasets tailored for PSFHS are notably scarce. Bridging this critical gap, we introduce a PSFHS dataset comprising 1358 images, meticulously annotated at the pixel level. The annotation process adhered to standardized protocols and involved collaboration among medical experts. Remarkably, this dataset stands as the most expansive and comprehensive resource for PSFHS to date.
Similar content being viewed by others
Background & Summary
Detecting maternities at risk of requiring a cesarean section is paramount in enhancing perinatal outcomes and maternal satisfaction during childbirth. Prolonged labor or failure to progress is one of the common indications that causes approximately one-third of all cesarean deliveries, underscoring the vital need for precise prediction of prolonged labor to mitigate the occurrence of unplanned emergency cesarean procedures. Notably, the prevalence of cesarean section rates has witnessed a recent increase, often attributed to indications concerning the position of the fetal head (FH) and the progression of labor1.
Traditional methods involving subjective digital vaginal examinations for ascertaining FH position, rotation, and descent during delivery have demonstrated a lack of accuracy at times2. In this context, intrapartum transperineal ultrasound has emerged as an efficacious approach for monitoring FH descent. A critical advancement offered by this technique is the angle of progress (AOP), which serves as an objective, accurate, and reproducible indicator. Notably surpassing the limitations of digital vaginal examination3, the AOP offers insight into the relationship between the pubic symphysis (PS) and FH (PSFH). According to the practice guideline of the International Society of Ultrasound in Obstetrics and Gynecology (ISUOG), AOP is measured on a static 2D ultrasound image and is defined as the angle the angle between the long axis of the pubic bone and a line from the lowest edge of the pubic symphysis that tangentially touches the deepest bony part of the fetal skull. Research indicates that an AoP greater than or equal to 120 degrees is closely linked with a high chance of spontaneous vaginal delivery. Therefore, Therefore, AoP measured based on a single ultrasound image can be used as a predictive indicator of the mode of delivery.
The first step in interpreting the morphometrics is performing PSFH segmentation (PSFHS) - extracting visible PSFH contours from transperineal ultrasound images. However, PSFHS is a challenging task, involving accurate identification and delineation of the PSFH boundaries. FHs can vary widely in shape and orientation during different stages of labor, and surrounding structures like amniotic fluid and placenta can overlap with or obstruct parts of the head, introducing segmentation ambiguity. Additionally, the size and position of PSFH can vary significantly among individuals, making it difficult to develop a single generalized segmentation model. The inherent characteristics of ultrasound images, such as poor resolution, noise, and artifacts, further complicate the PSFHS process, especially during the dynamic changes in relative positions of PS and FH during the second stage of labor.
To address these challenges, automatic segmentation with Artificial Intelligence (AI) offers a promising approach4. Modifications to UNet structures have been introduced, incorporating attention mechanisms for improved detail capture5,6. Notably, Bai et al. devised a dual decoder strategy integrating traditional and deformable convolutions to concurrently extract morphological and global features7,8. Beyond architectural enhancements, Lu et al. explored the use of a shape-constrained loss function, reinforcing the UNet variant’s resilience to noise through the integration of a convex shape prior9. Despite the commendable performance of these UNet variations on proprietary datasets, their comparative evaluation remains challenging due to dataset diversity and size constraints. For example, Lu et al. released the JNU-IFM dataset10. Images of the JNU-IFM were acquired with the ObEye system from 51 pregnant women and collected from NanFang Hospital of Southern Medical University10. However, the development of AI models relies heavily on datasets from multiple centers, involving different patients, and sourced from various ultrasound devices. These factors significantly enhance the model’s performance, generalizability, and practical applicability in real-world clinical settings. Here are the key reasons why such diverse datasets are important: (1) Different centers may have patient populations with varying demographics, disease prevalence, and comorbidities; (2) Ultrasound devices from different manufacturers or even different models from the same manufacturer can produce images with varying qualities, resolutions, and characteristics; (3) Different centers might have varying protocols for image acquisition, patient preparation, and even image annotation standards; and (4) Images from different sources can have various types of noise, artifacts, or quality issues. A model trained on a diverse dataset is likely to generalize better to unseen data, reducing the risk of overfitting to specific characteristics of the training data. This is crucial for medical applications where the cost of errors can be very high. Therefore, larger and more comprehensive datasets are essential8.
To address these challenges and accelerate progress in AI research, it is imperative to promote data sharing and establish more comprehensive and representative datasets. In line with this objective, a proposed PSFHS dataset has been introduced, encompassing intrapartum transperineal ultrasound images that have been meticulously annotated at the pixel level through standard crowdsourcing among medical professionals. 1358 images from 1124 patients were gathered to form the PSFHS dataset. The accuracy of the pixelwise annotation, carried out by a group of trained students, was verified by expert physicians. This dataset is expected to facilitate the monitoring of labor progression through computer-aided systems and promote the practical implementation of technology in clinical settings, ultimately contributing to enhanced childbirth outcomes and improved care for both mother and fetus.
Methods
Subject characteristics
This retrospective image collection included 1124 pregnant women from two different medical institutions. Inclusion criterias were defined as: singleton pregnancy at term gestation (37 weeks or more), fetus in cephalic presentation, absence of documented fetal malformations. There are two parts in the PSFHS dataset: 1045 images from 1040 patients of Zhujiang Hospital of Southern Medical University, 313 images from 84 pregnant women from the Department of Obstetrics and Gynecology of the First Affiliated Hospital of Jinan University (Fig. 1a). This study received approval from the institutional review boards of Zhujiang Hospital of Southern Medical University (No. 2023-SYJS-023) and the First Affiliated hospital of Jinan University (No. JNUKY-2022-019). Informed consent was waived because of the retrospective nature of the study and the analysis used anonymous medical image data.
Image acquisition
Ultrasound acquisitions were performed using a portable machine equipped with a 3.5 MHz probe. The ‘ObEye’ system (Guangzhou, China; http://lian-med.com) and the Esaote My Lab were used in the First Affiliated Hospital of Jinan University and Zhujiang Hospital of Southern Medical University, respectively. During each acquisition, the operator positioned the probe longitudinally in the translabial area to visualize both the PS horizontally in the upper central part of the image and the edges of the FH in the lower part11. These original images were cropped to remove sensitive information about patients.
Image annotation
The team responsible for annotations included 2 proficient physicians and 18 students specializing in biomedical studies (refer to Fig. 1b). Before commencing their tasks, annotators received comprehensive training that involved familiarizing them with the structures of PSFH and the key aspects of ultrasound images. This training was facilitated through a combination of online sessions and in-person guidance by the physicians. Each annotator was assigned 15 test images, which were subsequently assessed by the physicians. If the annotations were deemed inadequate, the images were returned to the respective student for refinement. Annotators were instructed to utilize the pencil tool in Pair (https://www.aipair.com.cn/) for precise pixel-wise segmentation. They used red color for the pixel outlines of PS and green for the contour pixels of FH. In instances where the contours appeared fragmented or discontinuous, annotators were instructed to ensure that the contours maintained a complete elliptical shape. This instruction was essential considering the ultimate clinical application’s requirement to calculate AOP based on the segmented PSFH contours. The final segmentation ground truth was represented by a three-color image, where red pixels denoted PS, green pixels represented FH, and black pixels indicated the background. During the official annotation phase, each image was annotated by two annotators. Any overlapping pixels annotated by both annotators were further reviewed and adjusted by a highly experienced physician with a decade of expertise (refer to Fig. 1c). Note: Dropped artifacts in ultrasound images can significantly impact the quality and accuracy of the images, leading to potential errors in annotation. If artifact annotation may help to develop application-oriented robust algorithms, such as uncertain segmentation algorithms.
Morphological parameters
Class imbalance is a common issue in image segmentation tasks that significantly affects the performance of deep learning models. When there is a class imbalance, it means that the number of pixels belonging to one class significantly outnumbers the pixels belonging to other classes. The pixel ratio of background to target (or among various targets) is a critical metric for understanding class imbalance. In the PSFHS dataset, the pixel proportions of PS, FH and background are, 1.78% ± 0.66%, 14.55 ± 5.73%, and 83.66 ± 6.27%, respectively. Based on the ground truth of PSFH, ellipse fitting is performed and thereby AOP is measured according to its definition—the angle between the longitudinal axis of the pubic symphysis and a line originating from its inferior edge to the leading edge of the fetal cranium tangentially (Fig. 1d)7,9. AOP is a predictor of the mode of delivery and the average value of AOP in the PSFHS dataset is 98.33° ± 21.11°.
Data Records
All data records12 are available as files on the web page https://doi.org/10.5281/zenodo.10969427. The unzipped file folder of this dataset contains the original transperineal ultrasound images and annotation ground truth images. The unzipped file is organized into 2 folders, named “image_mha” and “label_mha”, that contain original transperineal ultrasound images and corresponding ground truth images, respectively. The images in these 2 folders are stored, named and arranged according to the same rule, where a specific image in the “label_mha” folder is the ground truth of the image with the same name in the “image_mha” folder. Images are named as “n.mha”, where “n” means the number of images. In the dataset, there are 1358 images (“n” from 03744 to 05101) of 1124 pregnant women. The images in the “image_mha” folder contain pixels labelled as 0, 1, or 2, where 0 represents the background, 1 represents the PS, and 2 represents the FH. These data can be accessed using the software “Insight Segmentation and Registration Toolkit”, available at https://itk.org/.
Technical Validation
In this research, each ultrasound image underwent double annotation, followed by refinement by a medical professional. This process prompted the investigation of three distinct types of consistencies: intra-annotator consistency, referring to the same annotator at different time points; inter-annotator consistency among annotators at the same level; and inter-annotator consistency among annotators at different levels13.
To assess the intra-annotator consistency of the different annotators across various instances, a set of 40 images was selected from the complete dataset. These 40 images were annotated twice by three different annotators, including one physician and two others, on separate occasions. The Dice coefficient was then calculated between the annotations from the first and second rounds. The mean Dice coefficient for all 40 images and three annotators was 0.8817, with a confidence interval of 0.8502–0.9002.
To evaluate the inter-annotator consistency among annotators at the same level, discrepancies were assessed by computing the Dice coefficient between the annotations produced by the two annotators. The resulting mean Dice coefficient was 0.8750 (0.8520–0.8886).
Additionally, for the inter-annotator consistency among annotators at different levels, emphasis was placed on the first annotations. The Dice coefficient was used to measure the concordance between the physician’s mask and the annotations from the two annotators. The mean Dice coefficient for this scenario was determined to be 0.8720 (0.8520–0.8904).
Furthermore, to examine the intra-annotator consistency of the same annotator at different time points, the Dice coefficient was calculated between the first and second annotations. The resulting mean Dice coefficients were 0.8953 for the physician, 0.8811 for annotator 1, and 0.8689 for annotator 2.
Upon thorough analysis of both the original annotations and the illustrative set, it was concluded that the annotations demonstrated stability and consistency not only within a single annotator over different instances but also across various annotators. These findings collectively provide a strong basis for accurate annotation and reproducible PSFHS, as detailed in Table 1.
Usage Notes
The whole dataset used for the PSFHS challenge of MICCAI2023 (https://ps-fh-aop-2023.grand-challenge.org/) includes two parts12,14,15: one is this PSFHS dataset (https://doi.org/10.5281/zenodo.10969427)12 and another is from the JNU-IFM dataset (https://doi.org/10.6084/m9.figshare.14371652)16. These images in the PSFHS dataset can also be used for the Intrapartum Ultrasound Grand Challenge (IUGC) 2024 of MICCAI 2024 (https://codalab.lisn.upsaclay.fr/competitions/18413).
Code availability
No novel code was used in the construction of the PSFHS dataset.
References
Sandall, J. et al. Short-term and long-term effects of caesarean section on the health of women and children. Lancet (London, England) 392, 1349–1357, https://doi.org/10.1016/s0140-6736(18)31930-5 (2018).
Seval, M. M. et al. Comparison of effects of digital vaginal examination with transperineal ultrasound during labor on pain and anxiety levels: a randomized controlled trial. Ultrasound in Obstetrics & Gynecology 48, 695–700, https://doi.org/10.1002/uog.15994 (2016).
Ghi, T. et al. ISUOG Practice Guidelines: intrapartum ultrasound. Ultrasound in Obstetrics & Gynecology 52, 128–139, https://doi.org/10.1002/uog.19072 (2018).
Ramirez Zegarra, R. & Ghi, T. Use of artificial intelligence and deep learning in fetal ultrasound imaging. Ultrasound in Obstetrics & Gynecology 62, 185–194, https://doi.org/10.1002/uog.26130 (2023).
Fiorentino, M. C., Villani, F. P., Di Cosmo, M., Frontoni, E. & Moccia, S. A review on deep-learning algorithms for fetal ultrasound-image analysis. Medical image analysis 83, 102629, https://doi.org/10.1016/j.media.2022.102629 (2023).
Chen, Z., Ou, Z., Lu, Y. & Bai, J. Direction-guided and multi-scale feature screening for fetal head–pubic symphysis segmentation and angle of progression calculation. Expert Systems with Applications 245, 123096, https://doi.org/10.1016/j.eswa.2023.123096 (2024).
Bai, J. et al. A framework for computing angle of progression from transperineal ultrasound images for evaluating fetal head descent using a novel double branch network. Frontiers in physiology 13, 940150, https://doi.org/10.3389/fphys.2022.940150 (2022).
Zhou, M. et al. The segmentation effect of style transfer on fetal head ultrasound image: a study of multi-source data. Medical & biological engineering & computing 61, 1017–1031, https://doi.org/10.1007/s11517-022-02747-1 (2023).
Lu, Y. et al. Multitask Deep Neural Network for the Fully Automatic Measurement of the Angle of Progression. Computational and mathematical methods in medicine 2022, 5192338, https://doi.org/10.1155/2022/5192338 (2022).
Lu, Y. et al. The JNU-IFM dataset for segmenting pubic symphysis-fetal head. Data in brief 41, 107904, https://doi.org/10.1016/j.dib.2022.107904 (2022).
Brunelli, E. et al. The role of the angle of progression in the prediction of the outcome of occiput posterior position in the second stage of labor. American journal of obstetrics and gynecology 225, 81.e81–81.e89, https://doi.org/10.1016/j.ajog.2021.01.017 (2021).
Bai, J. PSFHS. Zenodo. https://doi.org/10.5281/zenodo.10969427 (2024).
Jin, K. et al. FIVES: A Fundus Image Dataset for Artificial Intelligence based Vessel Segmentation. Scientific data 9, 475, https://doi.org/10.1038/s41597-022-01564-3 (2022).
Qiu, R., Zhou, M., Bai, J., Lu, Y. & Wang, H. HPSSP-Net: An Efficient Lightweight Network for Identifying Head-Pubic Symphysis Standard Plane from Intrapartum Ultrasound Images. Medical & Biological Engineering & Computing (In Press).
Bai, J., Ou, Z., Lu, Y., Ni, D. & Chen, G. Pubic Symphysis-Fetal Head Segmentation from Transperineal Ultrasound Images. International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI) 2023 (MICCAI 2023). Zenodo https://doi.org/10.5281/zenodo.7861699 (2023).
Lu, Y., Bai, J., Zhou, M. & Zhou, M. The JNU-IFM dataset for segmenting symphysis pubis-fetal head. Figshare https://doi.org/10.6084/m9.figshare.14371652 (2022).
Acknowledgements
The work is funded by the Natural Science Foundation of Guangdong Province (2024A1515011886 and 2023A1515012833), Guangzhou Municipal Science and Technology Bureau Guangzhou Key Research and Development Program (2024B03J1283 and 2024B03J1289), the Guangdong Health Technology Promotion Project (2022 NO.132), the Science and Technology Program of Guangzhou (202201010544) and the China Scholarship Council. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. We would thank these doctors and students for their contributions in data annotation (Minghong Zhou, Chao Yuan, Mengqiang Zhou, Xiaosong Jiang, Dengjiang Zhi, Ruiyu Qiu, Di Qiu, Zhanhang Song, Shen Yu, Hao Yi, Hao Liu, Jingbo Rong, Xiaoyan Xie and Jianguo Qi).
Author information
Authors and Affiliations
Contributions
Gaowen Chen: Project administration, Conducting intrapartum ultrasound, Manual segmentation, Writing. Jieyun Bai: Conceptualization, Project administration, Manual segmentation, Writing, Reviewing, Visualization. Zhanhong Ou: Conceptualization, Manual segmentation, Methodology, Reviewing. Yaosheng Lu: Conceptualization, Methodology, Project administration, Methodology, Reviewing. Huijin Wang: Conceptualization, Funding acquisition, Project administration, Reviewing.
Corresponding authors
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Chen, G., Bai, J., Ou, Z. et al. PSFHS: Intrapartum ultrasound image dataset for AI-based segmentation of pubic symphysis and fetal head. Sci Data 11, 436 (2024). https://doi.org/10.1038/s41597-024-03266-4
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41597-024-03266-4