# MAN **Repository Path**: buptybx/MAN ## Basic Information - **Project Name**: MAN - **Description**: https://github.com/icandle/MAN - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2026-01-31 - **Last Updated**: 2026-01-31 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README ##
Nankai University
**Overview:** To unleash the potential of ConvNet in super-resolution, we propose a multi-scale attention network (MAN), by coupling a classical multi-scale mechanism with emerging large kernel attention. In particular, we proposed multi-scale large kernel attention (MLKA) and gated spatial attention unit (GSAU). Experimental results illustrate that our MAN can perform on par with SwinIR and achieve varied trade-offs between state-of-the-art performance and computations. This repository contains [PyTorch](https://pytorch.org/) implementation for ***MAN*** (CVPRW 2024).1. [Requirements](#%EF%B8%8F-requirements) 2. [Datasets](#-datasets) 3. [Implementary Details](#-implementary-details) 4. [Train and Test](#%EF%B8%8F-train-and-test) 5. [Results and Models](#-results-and-models) 6. [Acknowledgments](#-acknowledgments) 7. [Citation](#-citation)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|**Params/FLOPs**| 150K/8G|1518K/114G|840K/47G|43090K/2895G|8712K/495G|
Results of our MAN-tiny/light/base models. Set5 validation set is used below to show the general performance. The visual results of five testsets are provided in the last column.
| Methods | Params | FLOPs |PSNR/SSIM (x2)|PSNR/SSIM (x3)|PSNR/SSIM (x4)|Results|
|:---------|:---------:|:--------:|:------:|:------:|:------:|:--------:|
| MAN-tiny | 150K | 8.4G | 37.91/0.9603 | 34.23/0.9258 | 32.07/0.8930 | [x2](https://pan.baidu.com/s/1mYkGvAlz0bSZuCVubkpsmg?pwd=mans)/[x3](https://pan.baidu.com/s/1RP5gGu-QPXTkH1NPH7axag?pwd=mans)/[x4](https://pan.baidu.com/s/1u22su2bT4Pq_idVxAnqWdw?pwd=mans) |
| MAN-light| 840K | 47.1G | 38.18/0.9612 | 34.65/0.9292 | 32.50/0.8988 | [x2](https://pan.baidu.com/s/1AVuPa7bsbb3qMQqMSM-IJQ?pwd=mans)/[x3](https://pan.baidu.com/s/1TRL7-Y23JddVOpEhH0ObEQ?pwd=mans)/[x4](https://pan.baidu.com/s/1T2bPZcjFRxAgMxGWtPv-Lw?pwd=mans) |
| MAN+ | 8712K | 495G | 38.44/0.9623 | 34.97/0.9315 | 32.87/0.9030 | [x2](https://pan.baidu.com/s/1pTb3Fob_7MOxMKIdopI0hQ?pwd=mans)/[x3](https://pan.baidu.com/s/1L3HEtcraU8Y9VY-HpCZdfg?pwd=mans)/[x4](https://pan.baidu.com/s/1FCNqht9zi9HecG3ExRdeWQ?pwd=mans) |
💖 Acknowledgments
---
We would thank [VAN](https://github.com/Visual-Attention-Network/VAN-Classification) and [BasicSR](https://github.com/XPixelGroup/BasicSR) for their enlightening work!
🎓 Citation
---
```
@inproceedings{wang2024multi,
title={Multi-scale Attention Network for Single Image Super-Resolution},
author={Wang, Yan and Li, Yusen and Wang, Gang and Liu, Xiaoguang},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
year={2024}
}
```
or
```
@article{wang2022multi,
title={Multi-scale Attention Network for Single Image Super-Resolution},
author={Wang, Yan and Li, Yusen and Wang, Gang and Liu, Xiaoguang},
journal={arXiv preprint arXiv:2209.14145},
year={2022}
}
```