Abstract: The present invention relates to a system and method for automated detection of bone defects in wrist X-ray images. Despite the widespread prevalence of wrist bone anomalies and the availability of high-quality digital radiography images, manual detection by physicians or radiologists remains the current practice. This manual process is time-consuming and subject to human error. The proposed system leverages Convolutional Neural Networks (CNNs), specifically the YOLO v7 model, to automate bone defect detection in wrist X-ray images. By preprocessing the images and analyzing the output of the CNN model, regions of interest corresponding to potential bone defects are identified, providing valuable support to physicians in diagnosing these anomalies, particularly in emergency services. This system aims to improve the efficiency and accuracy of bone defect detection, thereby facilitating timely treatment planning and patient management.
Description:[0033]. Annotation refers to the process of labeling X-ray images to identify and delineate regions of interest, such as fractures, metal objects, and bone anomalies. This annotation task is crucial for training the YOLOv7 model to accurately detect and classify these features in new X-ray images. The annotation process involves using a suitable labeling tool, such as the YOLOv7 labeling tool or other open-source solutions, to draw bounding boxes around each desired object in the images. Each bounding box is associated with a class label, indicating whether the object is a fracture, metal object, or anomaly. By meticulously annotating the dataset, the model can learn to recognize and localize fractures and other relevant features, contributing to its ability to assist pediatric surgeons in interpreting X-ray images effectively.
[0034]. The script utilizes pd.read_csv(...) to load the dataset.csv file into a Pandas DataFrame. This file likely contains information about the X-ray images, such as filenames, patient IDs, and other relevant data. The first application of GroupShuffleSplit splits the entire dataset (70/30) into two parts: a larger training set and a combined validation/testing set. This ensures patient-aware splitting from the very beginning. The combined validation/testing set undergoes a further split using another instance of GroupShuffleSplit. This time, it's divided into the final validation and testing sets, maintaining the patient-aware approach.
[0035]. The script creates three CSV files: train_data.csv, valid_data.csv, and test_data.csv. These files share a similar structure as the original dataset but link to the newly split image sets for model training. These files likely contain information like the image filenames and any relevant labels needed by the YOLOv7 model.
[0036]. The script ensures that the necessary directories for training, validation, and testing sets exist with the structure required by the YOLOv7 model. This includes creating directories for both image and annotation files.
[0037]. Our project builds upon the foundation of real-time object detection established by previous iterations of the YOLO model family. Unlike traditional methods involving separate object proposal and classification stages, our model, treat detection as a single-stage regression problem. This approach efficiently predicts bounding boxes and their associated class probabilities simultaneously. This makes it well-suited for real-time applications where speed and efficiency are crucial.
[0038]. It embraces a unified architecture seamlessly integrating three key components: the backbone, neck, and head. This streamlined structure facilitates efficient end-to-end training and inference, optimizing the overall object detection process.
[0039]. Stage Partial Connections (CSP): These modules play a critical role in enhancing the backbone's efficiency and feature learning capabilities. They strategically split and recombine feature maps across different network layers, promoting information flow and reducing computational burden. This balance between efficiency and information richness is vital for accurate object detection.
[0040]. Our project utilizes the Task-Aligned Assigner (TAA) mechanism from the "Task-aligned One-stage Object Detection (TOOD)" approach. This mechanism plays a crucial role in selecting informative samples (objects to be detected) during training. It achieves this by assigning a weighted score to each potential object, combining two crucial factors: Predicted Classification Score (s): This score indicates the likelihood of a specific class being present in the object (e.g., "fracture" or "metal implant").Intersection over Union (IoU) (u): This metric measures the overlap between the predicted bounding box and the actual location of the object in the X-ray.
[0041]. Our wrist bone anomaly detection project utilizes a YOLOv7 model and presents a user-friendly web application interface for interacting with it. This Streamlit web application aims to provide an accessible and intuitive experience for users without requiring technical expertise. The interface features a single, dedicated upload button, allowing users to easily select and upload X-ray images in commonly supported formats like JPEG and PNG. Upon successful upload, the model processes the image and displays the results directly within the web app.
[0042]. The web application displays the results of the YOLOv7 model through visual aids called "predicted bounding boxes," highlighting the locations and sizes of potential anomalies on the X-ray image. Additionally, class labels associated with each bounding box provide specific classifications like "fracture" or "metal implant," while confidence scores represent the model's level of confidence in its predictions. This transparency allows users to assess the potential accuracy of the model's findings by understanding where anomalies are located, what type they might be, and the model's own confidence in each detection.
[0043]. This paper investigates the potential of the YOLOv7 deep learning model for detecting bone fractures in wrist X-ray images. The project utilizes the GRAZPEDWRI-DX dataset, focusing on wrist X-ray images containing various types of fractures. We evaluated the model's performance using mean Average Precision (mAP), precision, and recall. The reported mAP values range from 0.87 to 0.94 for different YOLOv7 configurations, indicating promising accuracy for detecting fractures within this dataset. Precision and recall figures also range from 0.80 to 0.90 for both metrics, suggesting a good balance between correctly identifying fractures and minimizing false positives. It is important to note that these results are based on a specific dataset and configuration. The reported mAP values indicate that it has the potential to be an accurate tool for bone fracture detection in wrist X-ray images.
[0044]. This paper presents a deep learning based approach to improve bone fracture detection in wrist X-ray images. Our study is driven by the urgent clinical demand for accurate and efficient fracture diagnosis, crucial for optimizing patient care. We acknowledge several crucial areas for further investigation and development.To enhance the model's generalizability and robustness, incorporating a greater diversity of fracture types and imaging conditions into the dataset is essential. This will ensure the model can effectively handle the variability inherent in real-world clinical cases. Real-world applications, particularly in time-sensitive clinical settings, demand rapid processing. Optimizing YOLOv7's inference speed will be critical for ensuring its suitability for such scenarios. Before considering any form of clinical deployment, collaboration with healthcare professionals is paramount. We are committed to conducting comprehensive clinical validation studies involving experts who will evaluate the model's performance on diverse clinical datasets, rigorously comparing it to established diagnostic practices. Investigating the adaptability of YOLOv7 to fractures beyond the wrist would significantly expand the model's potential application scope and contribute to its versatility in medical imaging.
This paper serves as a foundation for future development in the field of deep learning-powered fracture detection. By addressing these key areas, upholding the highest standards of clinical validation, and fostering interdisciplinary collaborations, we can work towards translating the promising potential of our model into tangible benefits for patients and healthcare professionals alike. , Claims:1.A method for automated detection of bone defects in wrist X-ray images, comprising:
a) Receiving a wrist X-ray image;
b) Pre-processing said wrist X-ray image to enhance image quality and facilitate feature extraction;
c) Applying a Convolutional Neural Network (CNN) model, specifically YOLO v7, to the pre-processed wrist X-ray image for bone defect detection;
d) Analysing the output of the CNN model to identify regions of interest corresponding to potential bone defects; and
e) Generating a report indicating the presence and location of detected bone defects within the wrist X-ray image.
2.The method as claimed in claim 1, wherein the pre-processing step includes noise reduction, contrast enhancement, and normalization techniques to optimize the input for the CNN model.
3.The method as claimed in claim 1, wherein the Convolutional Neural Network (CNN) model is trained on a dataset comprising wrist X-ray images annotated with ground truth labels of bone defects.
4.The method as claimed in claim 1, wherein the YOLO v7 CNN model utilizes single-stage object detection architecture for efficient and accurate detection of bone defects in wrist X-ray images.
5.A system for automated detection of bone defects in wrist X-ray images, comprising:
a) Means for receiving a wrist X-ray image;
b) Pre-processing means for enhancing the quality of the wrist X-ray image and facilitating feature extraction;
c) A Convolutional Neural Network (CNN) model, specifically YOLO v7, for bone defect detection applied to the preprocessed wrist X-ray image;
d) Analysis means for interpreting the output of the CNN model to identify regions of interest corresponding to potential bone defects; and
e) Reporting means for generating a report indicating the presence and location of detected bone defects within the wrist X-ray image.
6.The system as claimed in claim 5, wherein the pre-processing means includes noise reduction, contrast enhancement, and normalization modules to optimize the input for the CNN model.
7.The system as claimed in claim 5, wherein the Convolutional Neural Network (CNN) model is trained on a dataset comprising wrist X-ray images annotated with ground truth labels of bone defects.
8.The system as claimed in claim 5, wherein the YOLO v7 CNN model utilizes single-stage object detection architecture for efficient and accurate detection of bone defects in wrist X-ray images.
| # | Name | Date |
|---|---|---|
| 1 | 202441023755-REQUEST FOR EARLY PUBLICATION(FORM-9) [26-03-2024(online)].pdf | 2024-03-26 |
| 2 | 202441023755-FORM-9 [26-03-2024(online)].pdf | 2024-03-26 |
| 3 | 202441023755-FORM-9 [26-03-2024(online)]-1.pdf | 2024-03-26 |
| 4 | 202441023755-FORM 1 [26-03-2024(online)].pdf | 2024-03-26 |
| 5 | 202441023755-DRAWINGS [26-03-2024(online)].pdf | 2024-03-26 |
| 6 | 202441023755-COMPLETE SPECIFICATION [26-03-2024(online)].pdf | 2024-03-26 |
| 7 | 202441023755-FORM 3 [27-03-2024(online)].pdf | 2024-03-27 |
| 8 | 202441023755-ENDORSEMENT BY INVENTORS [27-03-2024(online)].pdf | 2024-03-27 |