Estimating 2D Keypoints of Surgical Tools Using Vision-Language Models with Low-Rank Adaptation
By: Krit Duangprom, Tryphon Lambrou, Binod Bhattarai
Potential Business Impact:
Helps robots see and grab tiny surgical tools.
This paper presents a novel pipeline for 2D keypoint estima- tion of surgical tools by leveraging Vision Language Models (VLMs) fine- tuned using a low rank adjusting (LoRA) technique. Unlike traditional Convolutional Neural Network (CNN) or Transformer-based approaches, which often suffer from overfitting in small-scale medical datasets, our method harnesses the generalization capabilities of pre-trained VLMs. We carefully design prompts to create an instruction-tuning dataset and use them to align visual features with semantic keypoint descriptions. Experimental results show that with only two epochs of fine tuning, the adapted VLM outperforms the baseline models, demonstrating the ef- fectiveness of LoRA in low-resource scenarios. This approach not only improves keypoint detection performance, but also paves the way for future work in 3D surgical hands and tools pose estimation.
Similar Papers
Scaling Down to Scale Up: Towards Operationally-Efficient and Deployable Clinical Models via Cross-Modal Low-Rank Adaptation for Medical Vision-Language Models
CV and Pattern Recognition
Helps doctors find diseases in CT scans faster.
Integrating Frequency-Domain Representations with Low-Rank Adaptation in Vision-Language Models
CV and Pattern Recognition
Lets computers understand pictures even in bad light.
Challenging Vision-Language Models with Surgical Data: A New Dataset and Broad Benchmarking Study
CV and Pattern Recognition
Helps doctors see better during surgery.