高级检索
当前位置: 首页 > 详情页

CLAD-Net: cross-layer aggregation attention network for real-time endoscopic instrument detection

文献详情

资源类型:
WOS体系:
Pubmed体系:

收录情况: ◇ SCIE

机构: [1]School of Automation, Guangdong University of Technology, Guangzhou, 510006 China. [2]Department of Neurosurgery, Sun Yat-Sen University Cancer Center, Guangzhou, 510006 China. [3]Department of Gastroenterology, The First Affiliated Hospital of Zhejiang Chinese Medical University (Zhejiang Provincial Hospital of Chinese Medicine), Hangzhou, 310006 China.
出处:
ISSN:

关键词: Cross-layer feature aggregation Composite attention mechanism Refnement module Surgical instrument detection

摘要:
As medical treatments continue to advance rapidly, minimally invasive surgery (MIS) has found extensive applications across various clinical procedures. Accurate identification of medical instruments plays a vital role in comprehending surgical situations and facilitating endoscopic image-guided surgical procedures. However, the endoscopic instrument detection poses a great challenge owing to the narrow operating space, with various interfering factors (e.g. smoke, blood, body fluids) and inevitable issues (e.g. mirror reflection, visual obstruction, illumination variation) in the surgery. To promote surgical efficiency and safety in MIS, this paper proposes a cross-layer aggregated attention detection network (CLAD-Net) for accurate and real-time detection of endoscopic instruments in complex surgical scenarios. We propose a cross-layer aggregation attention module to enhance the fusion of features and raise the effectiveness of lateral propagation of feature information. We propose a composite attention mechanism (CAM) to extract contextual information at different scales and model the importance of each channel in the feature map, mitigate the information loss due to feature fusion, and effectively solve the problem of inconsistent target size and low contrast in complex contexts. Moreover, the proposed feature refinement module (RM) enhances the network's ability to extract target edge and detail information by adaptively adjusting the feature weights to fuse different layers of features. The performance of CLAD-Net was evaluated using a public laparoscopic dataset Cholec80 and another set of neuroendoscopic dataset from Sun Yat-sen University Cancer Center. From both datasets and comparisons, CLAD-Net achieves the AP0.5 of 98.9% and 98.6%, respectively, that is better than advanced detection networks. A video for the real-time detection is presented in the following link: https://github.com/A0268/video-demo.© The Author(s), under exclusive licence to Springer Nature Switzerland AG 2023. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

基金:
语种:
WOS:
PubmedID:
中科院(CAS)分区:
出版当年[2022]版:
大类 | 3 区 医学
小类 | 3 区 医学:信息
最新[2025]版:
大类 | 2 区 医学
小类 | 3 区 医学:信息
JCR分区:
出版当年[2021]版:
Q2 MEDICAL INFORMATICS
最新[2023]版:
Q1 MEDICAL INFORMATICS

影响因子: 最新[2023版] 最新五年平均 出版当年[2021版] 出版当年五年平均 出版前一年[2020版] 出版后一年[2022版]

第一作者:
第一作者机构: [1]School of Automation, Guangdong University of Technology, Guangzhou, 510006 China.
通讯作者:
推荐引用方式(GB/T 7714):
APA:
MLA:

资源点击量:2022 今日访问量:0 总访问量:648 更新日期:2024-07-01 建议使用谷歌、火狐浏览器 常见问题

版权所有©2020 广东省中医院 技术支持:重庆聚合科技有限公司 地址:广州市越秀区大德路111号