Neural network enables objective assessment of breast symmetry

Key breast features [suprasternal notch (A), boundaries of right breast (B), left nipple-areola complex (C)]. Modified from Dilmen N. Human female breasts (published 2011; available at: https://commons.wikimedia.org/wiki/File:Breasts.svg), Wikimedia Commons, used under CC BY-SA 4.0 license. Credit: Plastic & Reconstructive Surgery (2023). DOI: 10.1097/PRS.0000000000010603

A newly developed neural network is highly accurate in identifying key landmarks important in breast surgery—opening the potential for objective assessment of breast symmetry, suggests a study in the February issue of Plastic and Reconstructive Surgery.

“Neural networks and machine learning have the potential to improve evaluation of breast symmetry in reconstructive and cosmetic breast surgery, enabling rapid, automated detection of features used by plastic surgeons,” comments lead author Nitzan Kenig, MD, of Albacete University Hospital, Spain.

Developing neural networks for objective breast assessment

Breast symmetry is a key concern in breast surgery, and is generally assessed by simple subjective evaluations by both patients and surgeons. Computer programs can provide more objective assessments, but with limitations including the need to manually enter data and lengthy calculation times.

Neural networks—an artificial intelligence technique that seeks to emulate the way the human brain processes data—are being explored for their potential to improve care in several areas of medical practice. Dr. Kenig and colleagues developed an “ad hoc convolutional neural network” to detect key breast features used in assessing breast symmetry.

Using an open-source algorithm called YOLOV3 (“You Only Look Once,” version 3), the researchers trained their neural network to identify three anatomic features used in assessing the female beast: the breast boundaries, the nipple-areola complex (nipple and surrounding tissue), and the suprasternal notch (the depression at the base of the neck, at the top of the breastbone).

The neural network was trained using 200 frontal photographs of patients who underwent breast surgery. Its performance in identifying key breast features was then tested using an additional set of 47 photographs of patients who underwent breast reconstruction after breast cancer surgery.

Potential for ‘quick, automated, objective’ evaluation of breast symmetry

After training, the neural network was highly accurate in localizing the three features, with a total detection rate of 97.7%. For the right and left breast boundaries and nipple-areola complex, accuracy was 100%. For the suprasternal notch, detection rate dipped to 87%. Processing was quick, with an average detection time of 0.52 second.

The neural network was able to detect and localize the key features even in visibly asymmetrical breast reconstructions. The high success rate confirmed that the training data set was sufficient, with no need for data augmentation techniques.

“Neural networks and machine learning have a potential of improving the evaluation of breast symmetry in the field of Plastic Surgery, by automated and quick detection of features used by surgeons in practice,” Dr. Kenig and co-authors conclude. They believe that with further advances in image detection capabilities and their applications to breast surgery, neural networks could play a role in evaluation of breast symmetry and planning of both aesthetic and reconstructive plastic surgery.

More information:
Nitzan Kenig et al, Identification of Key Breast Features Using a Neural Network: Applications of Machine Learning in the Clinical Setting of Plastic Surgery, Plastic & Reconstructive Surgery (2023). DOI: 10.1097/PRS.0000000000010603

Provided by
Wolters Kluwer Health

Citation:
Neural network enables objective assessment of breast symmetry (2024, January 26)
retrieved 26 January 2024
from https://medicalxpress.com/news/2024-01-neural-network-enables-breast-symmetry.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.