Abstract
Interactive medical image segmentation based on human-in-the-loop machine learning is a novel paradigm that draws on human expert knowledge to assist medical image segmentation. However, existing methods often fall into what we call interactive misunderstanding, the essence of which is the dilemma in trading off short- and long-term interaction information. To better use the interaction information at various timescales, we propose an interactive segmentation framework, called interactive MEdical image segmentation with self-adaptive Confidence CAlibration (MECCA), which combines action-based confidence learning and multi-agent reinforcement learning. A novel confidence network is learned by predicting the alignment level of the action with short-term interaction information. A confidence-based reward-shaping mechanism is then proposed to explicitly incorporate confidence in the policy gradient calculation, thus directly correcting the model’s interactive misunderstanding. MECCA also enables user-friendly interactions by reducing the interaction intensity and difficulty via label generation and interaction guidance, respectively. Numerical experiments on different segmentation tasks show that MECCA can significantly improve short- and long-term interaction information utilization efficiency with remarkably fewer labeled samples. The demo video is available at https://bit.ly/mecca-demo-video .
| Translated title of the contribution | 基于自适应置信度校准的交互式医疗图像分割框架 |
|---|---|
| Original language | English |
| Pages (from-to) | 1332-1348 |
| Number of pages | 17 |
| Journal | Frontiers of Information Technology and Electronic Engineering |
| Volume | 24 |
| Issue number | 9 |
| DOIs | |
| State | Published - Sep 2023 |
Keywords
- Confidence learning
- Interactive segmentation
- Medical image segmentation
- Multi-agent reinforcement learning
- Semi-supervised learning
- TP391.4