AI Robots Improving Surgical Assistance Efficiency
77 patents in this list
Updated:
Current robotic surgical systems require extensive manual control and monitoring, with surgeons spending up to 30% of procedure time managing tool transitions and camera positioning. These inefficiencies compound across thousands of procedures annually, impacting both surgical team workflow and operating room utilization.
The core challenge lies in automating routine surgical tasks while maintaining absolute precision and safety in a dynamic operating environment where split-second decisions have critical consequences.
This page brings together solutions from recent research—including AI-powered tissue recognition systems, automated surgical state detection, machine learning for instrument tracking, and intelligent workflow optimization. These and other approaches focus on enhancing surgical efficiency while maintaining surgeon control over critical decision points during procedures.
1. Robotic Surgical Tool Disengagement System with Recurrent Neural Network-Based Controller Motion Detection
Verb Surgical Inc., 2024
Automatic disengagement of surgical tools from robotic arms when the user returns the hand controllers to their docks, without needing a foot pedal. A recurrent neural network classifier detects if the controller motions are docking or teleoperation. When docking is detected, the robot disengages the tools to prevent unintended movements. The classifier is trained on labeled time-series data of controller motions.
2. Machine Learning System for Surgical Video Analysis with Event Categorization and Predictive Analytics
Theator Inc., 2024
Using machine learning to analyze surgical videos and derive insights like identifying deviations from surgical planes, analyzing surgical competency, and generating statistical reports with links to video evidence. The system receives video frames from multiple surgeries, categorizes them based on events, aggregates stats within categories, and presents frames from categories. It also predicts future events in ongoing surgeries, suggests video reviews, and assigns surgeons to surgeries based on skill level and time estimates.
3. Machine Learning-Based System for Video-Based Surgical State and Instrument Detection with Shared Feature Utilization
DIGITAL SURGERY LIMITED, 2024
Automatically detecting surgical states, instrument locations, and motion profiles from video during surgery using machine learning. The method involves training machine learning models to predict surgical states and detect instruments in video frames based on input data. It uses shared features between state and instrument detection to improve both tasks. The models can also leverage temporal information from state prediction to boost instrument detection. This allows real-time augmentation of surgical video with instrument overlays and motion profiles.
4. AI System for Intraoperative Image Analysis with Dynamic Visual Guidance in Surgical Procedures
Orthogrid Systems Holdings, LLC, 2024
Artificial intelligence (AI) system for real-time surgical guidance during procedures like joint replacements, fracture reductions, and spine surgeries. The system analyzes intraoperative images to calculate surgical decision risks and provides automated guidance to optimize implant placement, fracture reduction, and alignment. It uses AI models trained on surgical data to identify anatomical landmarks, register images, and predict risks. The system displays visual guidance to surgeons, dynamically updating as the procedure progresses.
5. Robotic Surgery System with AI-Driven Tissue Identification and Autonomous Action Execution
Intuitive Surgical Operations, Inc., 2024
Using AI to enhance robotic surgery by leveraging image recognition to identify tissue types during surgery and allowing the robot to autonomously perform actions based on the identified tissue. The system involves obtaining images of the surgical site, identifying the tissue types using AI, and if the identified tissue matches a targeted type, having the robot remove it. This allows the robot to perform specific actions without manual guidance once the AI tissue recognition is confident. The AI can also help with tasks like incision placement, guide wire manipulation, and imaging output projection to improve accuracy and repeatability.
6. Robotic Surgical Instrument Tracking System with Ultrasound-Based Pose Estimation and Redundant Optical and Kinematic Switching
GLOBUS MEDICAL, INC., 2024
Tracking the position of a robotic surgical instrument using ultrasound (US) instead of optical markers. A US transducer is attached to the instrument and generates US images of the anatomy. The instrument's pose is determined by matching anatomical features in the US images to a template of the instrument. This allows accurate tracking even if the US transducer loses contact with the patient. The system switches to optical tracking if the US transducer lifts off. It also uses kinematics if the US transducer stops outputting. This provides redundancy and continuous tracking during instrument motion.
7. Robotic Surgical System with Visual Feedback-Based Needle Trajectory Estimation and Machine Learning Control Adjustment
Covidien LP, 2024
Robotic surgical system with enhanced tissue suturing guidance using visual feedback and machine learning. The system uses an imaging device to capture the position and orientation of the surgical needle inside the body during suturing. It estimates the needle tip location and trajectory from the images. This data is used to refine the robot's control signals to improve needle placement and avoid collisions. The system can also augment the video feed to highlight the needle and show its path. It can also check for tissue contact to guide the robot's motion.
8. Hierarchical Multi-Modal Position Tracking System with Inter-Subsystem Virtual Space Mapping
Dignity Health, 2024
Multi-modal position tracking system for accurately tracking objects in extended time periods for applications like surgery. The system uses hierarchical tracking sub-systems with parent, child, grandchild, etc. systems that record object positions relative to their virtual spaces. A mapping between the virtual spaces allows translating positions across the hierarchy. This provides multi-modal position estimation and verification by observing objects from multiple sub-systems and correcting them if needed. It enables accurate tracking over extended time for surgical applications by compensating for drift between sub-systems.
9. Spatio-Temporal Machine Learning System for Real-Time Detection of Anatomical Structures and Surgical Instruments in Live Video
DIGITAL SURGERY LIMITED, 2024
Using machine learning and computer vision to automatically detect critical anatomical structures and surgical instruments during live video of a surgical procedure. The technique involves training machine learning models to predict the presence and location of instruments and structures in the video frame using spatio-temporal information. The models can also predict the overall surgical state. The instrument and structure detection models share extracted features with the state prediction models to improve confidence. This allows real-time augmented visualization of the surgical view. The technique improves surgical safety and workflow by providing action guidance and alerts.
10. Computer-Assisted Surgery System with Video-Based Intraoperative Deviation Recommendation Mechanism
ORTHOSOFT ULC, 2024
Computer-assisted surgery system that can recommend deviations from standard surgical flows based on intraoperative video analysis. The system uses a processing unit and memory to monitor a surgical procedure via video feed and detect conditions requiring deviations outside the standard flow. It trains a machine learning model using video and tracking data to perform tracking without the separate tracking device. This allows the system to provide intraoperative recommendations for deviations based on video analysis alone.
11. Augmented Surgical Guidance System with AI-Based Predictive Visualization and Real-Time Tissue Tracking
Activ Surgical, Inc., 2024
Real-time augmented surgical guidance system using AI to provide predictive visualizations of critical structure locations and tissue viability during surgery. The system uses machine learning models trained on medical datasets to anticipate critical structures and tissue viability based on imaging and physiological data. It can also provide real-time guidance and decision support during surgical procedures. The AI analyzes imaging data from multiple modalities and extracts features to classify structures and assess viability. The system generates enhanced views with guidance and metrics to assist surgeons. It can also autonomously track deformable tissues during maneuvers. The AI is trained on whole surgery datasets with anatomical and physiological data.
12. Computer Vision-Based Control System for Surgical Tool Operation via Real-Time Object Recognition
DIGITAL SURGERY LIMITED, 2024
Using computer vision to improve safety and reliability of surgical procedures by controlling the operation of surgical tools based on recognized objects in the video feed. The system trains a computer vision model to recognize surgical tools from live video. During surgeries, the model is used to detect surgical tools in the video feed. When a tool is detected, it enables control of the tool to perform its function. This ensures tools are only used when visible, preventing accidental activation outside the field of view.
13. Surgical Procedure Optimization Using Simulation and AI-Driven Genetic Algorithms
HUTOM CO., LTD., 2024
Optimizing surgery by using simulation and AI to find the best tools, procedures, and entry points for minimally invasive surgeries. The method involves generating genes representing surgical procedures, simulating them in virtual bodies, evaluating optimality, and applying genetic algorithms to derive improved procedures and instrument configurations. This provides optimized surgical cue sheets and robot designs that improve efficiency and convenience in actual surgery.
14. Graph Neural Network-Based System for Surgical Video Data Interpretation with Real-Time Conceptual Analysis
THE GENERAL HOSPITAL CORPORATION, MASSACHUSETTS INSTITUTE OF TECHNOLOGY, 2024
Using graph neural networks to interpret surgical video data and provide real-time feedback to surgeons during operations. The method involves extracting numerical features from sensor data like video to represent surgical concepts like anatomy, tools, and safety. These features are passed through graph neural networks with connections representing relationships between concepts. The networks learn to reason about the concepts during surgery and provide statistical parameters representing the state of the procedure. This information can be displayed to surgeons to assist decision making and mitigate risks.
15. AI-Driven System for Automated Surgical Plan Generation Using Image-Based Patient and Environment Modeling
Shanghai United Imaging Intelligence Co., Ltd., 2024
Automatically generating surgical plans in a medical environment using AI to relieve the burden on medical professionals and enhance safety, efficiency, and effectiveness of surgeries. The system uses image analysis to create patient and environment models from captured images. It then devises a surgical plan based on these models, including movement paths for medical devices, to execute procedures. The AI identifies people and objects in images, determines patient anatomy and environment layout, and generates optimized surgical plans.
16. Haptic Feedback Generation System Using Machine Learning-Based Video Analysis for Robotic Surgery
Verb Surgical Inc., 2024
Generating haptic feedback for robotic surgery to provide tactile sensations to the surgeon during procedures. The feedback is based on analyzing video of the surgery using machine learning models trained to correlate visual appearances of tool-tissue interactions with force levels. The models predict the force applied when the video is played back, and this is converted into haptic feedback sent to the surgeon's user interface. This aims to replicate the tactile feedback a surgeon would feel during open surgery when using robotic instruments.
17. Computer-Assisted Surgical System with Co-Registered Augmented Reality and Traditional Navigation Using Dual-Tracked Markers
Smith & Nephew, Inc., 2024
Enhancing surgical workflows using a computer-assisted surgical system that provides augmented reality guidance during surgeries. The system combines traditional surgical navigation with augmented reality to improve accuracy and reduce distractions. It uses specialized markers that can be tracked by both the surgical navigation system and augmented reality. This allows co-registration of the tracking frames for seamless integration of augmented reality data into the surgeon's view. The augmented reality display shows surgical plan information directly over the patient anatomy, eliminating the need for the surgeon to look away from the surgical site. It also allows using the same markers for both tracking systems, simplifying setup and avoiding conflicts. The augmented reality system can project a virtual screen in the surgeon's line of sight to display the GUI without requiring turning the head.
18. Computer-Assisted Surgical System with Robotic Arm, Electromagnetic Tracking, and Multi-Sensor Navigation for Joint Replacement Procedures
Smith & Nephew, Inc., 2024
Computer-assisted surgical system that enhances surgical workflows by providing advanced tools and techniques to improve total joint replacement procedures. The system uses a combination of imaging, robotics, and data analysis to enable more accurate, efficient, and customized joint replacements. Features include: 1. Robotic arm with electromagnetic tracking for guided bone preparation and implantation. 2. Point probe device for high-resolution imaging of critical areas during hip surgeries. 3. Multi-sensor navigation system for accurate tracking of bone fixation devices. 4. Registration of pre-operative data to patient anatomy using the point probe. 5. 3D modeling of patient anatomy from bi-planar images. 6. Automated optimization of surgical parameters using historical data and patient goals.
19. Patterned Light-Based Object Tracking System Utilizing AI-Driven Image Processing and Machine Learning for Non-Invasive Pose Determination
Future Health Works Ltd., 2024
Minimally invasive, high precision tracking and guiding of an object like a body part during surgery without large markers. It uses a patterned light beam projected onto the object contour, image processing with AI, and machine learning to register, track, and guide the object without invasive marker placement. The light beam is segmented, quantized, reconstructed, transformed, concatenated, normalized, and analyzed by ML to determine object pose.
20. Method for Automated Virtual Segmentation and Pathway Optimization of Surgical Implants for Robotic Installation
IX Innovation LLC, 2024
Automated customization of surgical implants to enable less invasive installation using robotic surgery. The method involves using imaging data to identify less invasive installation paths and dimensions for customized implant components. The implant is virtually segmented into components that can be passed through the identified routes. The components are simulated moving and assembled at the implant site. Modifications are made if components can't pass or assemble. This allows minimally invasive robotic installation of customized implants through less damaging paths.
Request the full report with complete details of these
+57 patents for offline reading.