Super-Resolution Network with Information Distillation and Multi-Scale Attention for Medical CT Image
release_x64ockympndy3duyh3bzkzavcy
by
Tianliu Zhao, Lei Hu, Yongmei Zhang, Jianying Fang
Abstract
The CT image is an important reference for clinical diagnosis. However, due to the external influence and equipment limitation in the imaging, the CT image often has problems such as blurring, a lack of detail and unclear edges, which affect the subsequent diagnosis. In order to obtain high-quality medical CT images, we propose an information distillation and multi-scale attention network (IDMAN) for medical CT image super-resolution reconstruction. In a deep residual network, instead of only adding the convolution layer repeatedly, we introduce information distillation to make full use of the feature information. In addition, in order to better capture information and focus on more important features, we use a multi-scale attention block with multiple branches, which can automatically generate weights to adjust the network. Through these improvements, our model effectively solves the problems of insufficient feature utilization and single attention source, improves the learning ability and expression ability, and thus can reconstruct the higher quality medical CT image. We conduct a series of experiments; the results show that our method outperforms the previous algorithms and has a better performance of medical CT image reconstruction in the objective evaluation and visual effect.
In application/xml+jats
format
Archived Files and Locations
application/pdf 4.2 MB
file_lwsewhl4pza6dfsluuxqpjryvu
|
mdpi-res.com (publisher) web.archive.org (webarchive) |
application/pdf 6.9 MB
file_qzfsrhjafvdelbgu52tjbgltjm
|
mdpi-res.com (web) web.archive.org (webarchive) |
Web Captures
https://www.mdpi.com/1424-8220/21/20/6870/htm
2022-06-20 12:36:03 | 57 resources webcapture_zx2gfp6h2bgnhjf6pam43c6o3e
|
web.archive.org (webarchive) |
Open Access Publication
In DOAJ
In ISSN ROAD
In Keepers Registry
ISSN-L:
1424-8220
access all versions, variants, and formats of this works (eg, pre-prints)
Crossref Metadata (via API)
Worldcat
SHERPA/RoMEO (journal policies)
wikidata.org
CORE.ac.uk
Semantic Scholar
Google Scholar