机构:[1]Jarvis Res Ctr, Tencent YouTu Lab, Shenzhen, Peoples R China[2]Pazhou Lab, Guangzhou, Peoples R China[3]Guangxi Med Univ, Life Sci Inst, Nanning, Peoples R China[4]Capital Med Univ, Beijing Tongren Hosp, Beijing, Peoples R China首都医科大学附属北京同仁医院首都医科大学附属同仁医院
Despite the great success of deep learning approaches, retinal disease classification is still challenging as the early-stage pathological regions of retinal diseases may be extremely tiny and subtle, which are difficult for networks to detect. The feature representations learnt by deep learning models focusing more on the local view may lead to indiscriminative semantic-level representation. On the contrary, if they focus more on the global semantic-level, they may ignore the discerning subtle local pathological regions. To address this issue, in this paper, we propose a hybrid framework, combining the strong global semantic representation learning capability of the vision Transformer (ViT) and the excellent capacity of local representation extraction from the conventional multiple instance learning (MIL). Particularly, a multiple instance vision Transformer (MIL-ViT) is implemented, where the vanilla ViT branch and the MIL branch generate semantic probability distributions separately, and a bag consistency loss is proposed to minimize the difference between them. Moreover, a calibrated attention mechanism is developed to embed the instance representation into the bag representation in our MIL-ViT. To further improve the feature representation capability for fundus images, we pre-train the vanilla ViT on a large-scale fundus image database. The experimental results validate that the generalization capability of the model using our pre-trained weights for fundus disease diagnosis is better than the one using ImageNet pre-trained weights. Extensive experiments on four publicly available benchmarks demonstrate that our proposed MIL-ViT outperforms latest fundus image classification methods, including various deep learning models and deep MIL methods. All our source code and pre-trained models are publicly available at https://github.com/greentreeys/MIL-VT.
基金:
This work was founded by the Key-Area Research and Development
Program of Guangdong Province, China (No. 2018B010111001),
National Key R&D Program of China (2018YFC2000702) and the Scientific
and Technical Innovation 2030—‘‘New Generation Artificial
Intelligence’’ Project (No. 2020AAA0104100).
语种:
外文
被引次数:
WOS:
中科院(CAS)分区:
出版当年[2022]版:
大类|3 区计算机科学
小类|3 区计算机:软件工程3 区计算机:信息系统
最新[2023]版:
大类|4 区计算机科学
小类|4 区计算机:信息系统4 区计算机:软件工程
JCR分区:
出版当年[2021]版:
Q2COMPUTER SCIENCE, SOFTWARE ENGINEERINGQ3COMPUTER SCIENCE, INFORMATION SYSTEMS
最新[2023]版:
Q2COMPUTER SCIENCE, INFORMATION SYSTEMSQ2COMPUTER SCIENCE, SOFTWARE ENGINEERING
第一作者机构:[1]Jarvis Res Ctr, Tencent YouTu Lab, Shenzhen, Peoples R China
通讯作者:
推荐引用方式(GB/T 7714):
Bi Qi,Sun Xu,Yu Shuang,et al.MIL-ViT: A multiple instance vision transformer for fundus image classification[J].JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION.2023,97:doi:10.1016/j.jvcir.2023.103956.
APA:
Bi, Qi,Sun, Xu,Yu, Shuang,Ma, Kai,Bian, Cheng...&Zheng, Yefeng.(2023).MIL-ViT: A multiple instance vision transformer for fundus image classification.JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION,97,
MLA:
Bi, Qi,et al."MIL-ViT: A multiple instance vision transformer for fundus image classification".JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION 97.(2023)