# PSPNet
> [Pyramid Scene Parsing Network](https://arxiv.org/abs/1612.01105)
## Introduction
<!-- [ALGORITHM] -->
<a href="https://github.com/hszhao/PSPNet">Official Repo</a>
<a href="https://github.com/open-mmlab/mmsegmentation/blob/v0.17.0/mmseg/models/decode_heads/psp_head.py#L63">Code Snippet</a>
## Abstract
<!-- [ABSTRACT] -->
Scene parsing is challenging for unrestricted open vocabulary and diverse scenes. In this paper, we exploit the capability of global context information by different-region-based context aggregation through our pyramid pooling module together with the proposed pyramid scene parsing network (PSPNet). Our global prior representation is effective to produce good quality results on the scene parsing task, while PSPNet provides a superior framework for pixel-level prediction tasks. The proposed approach achieves state-of-the-art performance on various datasets. It came first in ImageNet scene parsing challenge 2016, PASCAL VOC 2012 benchmark and Cityscapes benchmark. A single PSPNet yields new record of mIoU accuracy 85.4% on PASCAL VOC 2012 and accuracy 80.2% on Cityscapes.
<!-- [IMAGE] -->
<div align=center>
<img src="https://user-images.githubusercontent.com/24582831/142902444-9f93b99e-9261-443b-a0a4-17e78eefb525.png" width="70%"/>
</div>
<div align=center >
<img alt="PSPNet-R50-D8" src="https://user-images.githubusercontent.com/47882088/209554973-66804b14-de5a-4f83-b54e-26683a91818a.jpg"/>
PSPNet-R50 D8 model structure
</div>
## Results and models
### Cityscapes
| Method | Backbone | Crop Size | Lr schd | Mem (GB) | Inf time (fps) | Device | mIoU | mIoU(ms+flip) | config | download |
| ------------- | ------------- | --------- | ------: | -------- | -------------- | ------ | ----: | ------------: | ---------------------------------------------------------------------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ |
| PSPNet | R-50-D8 | 512x1024 | 40000 | 6.1 | 4.07 | V100 | 77.85 | 79.18 | [config](https://github.com/open-mmlab/mmsegmentation/blob/main/configs/pspnet/pspnet_r50-d8_4xb2-40k_cityscapes-512x1024.py) | [model](https://download.openmmlab.com/mmsegmentation/v0.5/pspnet/pspnet_r50-d8_512x1024_40k_cityscapes/pspnet_r50-d8_512x1024_40k_cityscapes_20200605_003338-2966598c.pth) \| [log](https://download.openmmlab.com/mmsegmentation/v0.5/pspnet/pspnet_r50-d8_512x1024_40k_cityscapes/pspnet_r50-d8_512x1024_40k_cityscapes_20200605_003338.log.json) |
| PSPNet | R-101-D8 | 512x1024 | 40000 | 9.6 | 2.68 | V100 | 78.34 | 79.74 | [config](https://github.com/open-mmlab/mmsegmentation/blob/main/configs/pspnet/pspnet_r101-d8_4xb2-40k_cityscapes-512x1024.py) | [model](https://download.openmmlab.com/mmsegmentation/v0.5/pspnet/pspnet_r101-d8_512x1024_40k_cityscapes/pspnet_r101-d8_512x1024_40k_cityscapes_20200604_232751-467e7cf4.pth) \| [log](https://download.openmmlab.com/mmsegmentation/v0.5/pspnet/pspnet_r101-d8_512x1024_40k_cityscapes/pspnet_r101-d8_512x1024_40k_cityscapes_20200604_232751.log.json) |
| PSPNet | R-50-D8 | 769x769 | 40000 | 6.9 | 1.76 | V100 | 78.26 | 79.88 | [config](https://github.com/open-mmlab/mmsegmentation/blob/main/configs/pspnet/pspnet_r50-d8_4xb2-40k_cityscapes-769x769.py) | [model](https://download.openmmlab.com/mmsegmentation/v0.5/pspnet/pspnet_r50-d8_769x769_40k_cityscapes/pspnet_r50-d8_769x769_40k_cityscapes_20200606_112725-86638686.pth) \| [log](https://download.openmmlab.com/mmsegmentation/v0.5/pspnet/pspnet_r50-d8_769x769_40k_cityscapes/pspnet_r50-d8_769x769_40k_cityscapes_20200606_112725.log.json) |
| PSPNet | R-101-D8 | 769x769 | 40000 | 10.9 | 1.15 | V100 | 79.08 | 80.28 | [config](https://github.com/open-mmlab/mmsegmentation/blob/main/configs/pspnet/pspnet_r101-d8_4xb2-40k_cityscapes-769x769.py) | [model](https://download.openmmlab.com/mmsegmentation/v0.5/pspnet/pspnet_r101-d8_769x769_40k_cityscapes/pspnet_r101-d8_769x769_40k_cityscapes_20200606_112753-61c6f5be.pth) \| [log](https://download.openmmlab.com/mmsegmentation/v0.5/pspnet/pspnet_r101-d8_769x769_40k_cityscapes/pspnet_r101-d8_769x769_40k_cityscapes_20200606_112753.log.json) |
| PSPNet | R-18-D8 | 512x1024 | 80000 | 1.7 | 15.71 | V100 | 74.87 | 76.04 | [config](https://github.com/open-mmlab/mmsegmentation/blob/main/configs/pspnet/pspnet_r18-d8_4xb2-80k_cityscapes-512x1024.py) | [model](https://download.openmmlab.com/mmsegmentation/v0.5/pspnet/pspnet_r18-d8_512x1024_80k_cityscapes/pspnet_r18-d8_512x1024_80k_cityscapes_20201225_021458-09ffa746.pth) \| [log](https://download.openmmlab.com/mmsegmentation/v0.5/pspnet/pspnet_r18-d8_512x1024_80k_cityscapes/pspnet_r18-d8_512x1024_80k_cityscapes-20201225_021458.log.json) |
| PSPNet | R-50-D8 | 512x1024 | 80000 | - | - | V100 | 78.55 | 79.79 | [config](https://github.com/open-mmlab/mmsegmentation/blob/main/configs/pspnet/pspnet_r50-d8_4xb2-80k_cityscapes-512x1024.py) | [model](https://download.openmmlab.com/mmsegmentation/v0.5/pspnet/pspnet_r50-d8_512x1024_80k_cityscapes/pspnet_r50-d8_512x1024_80k_cityscapes_20200606_112131-2376f12b.pth) \| [log](https://download.openmmlab.com/mmsegmentation/v0.5/pspnet/pspnet_r50-d8_512x1024_80k_cityscapes/pspnet_r50-d8_512x1024_80k_cityscapes_20200606_112131.log.json) |
| PSPNet | R-50b-D8 rsb | 512x1024 | 80000 | 6.2 | 3.82 | V100 | 78.47 | 79.45 | [config](https://github.com/open-mmlab/mmsegmentation/blob/main/configs/pspnet/pspnet_r50-d8-rsb_4xb2-adamw-80k_cityscapes-512x1024.py) | [model](https://download.openmmlab.com/mmsegmentation/v0.5/pspnet/pspnet_r50-d8_512x1024_80k_cityscapes/pspnet_r50-d8_rsb-pretrain_512x1024_adamw_80k_cityscapes_20220315_123238-588c30be.pth) \| [log](https://download.openmmlab.com/mmsegmentation/v0.5/pspnet/pspnet_r50-d8_512x1024_80k_cityscapes/pspnet_r50-d8_rsb-pretrain_512x1024_adamw_80k_cityscapes_20220315_123238.log.json) |
| PSPNet | R-101-D8 | 512x1024 | 80000 | - | - | V100 | 79.76 | 81.01 | [config](https://github.com/open-mmlab/mmsegmentation/blob/main/configs/pspnet/pspnet_r101-d8_4xb2-80k_cityscapes-512x1024.py) | [model](https://download.openmmlab.com/mmsegmentation/v0.5/pspnet/pspnet_r101-d8_512x1024_80k_cityscapes/p
没有合适的资源?快使用搜索试试~ 我知道了~
温馨提示
mmsegmentation实现对数据集进行图片分割,数据增强,MMSegmentation是一个基于PyTorch的开源图像分割工具包,它为研究者提供了高效、灵活的框架,用以进行图像分割任务的研究与实践。在使用MMSegmentation实现对数据集的图片分割时,数据增强是一个关键步骤,它可以提升模型的泛化能力,从而得到更稳健的分割结果。 首先,数据增强是通过一系列随机变换来扩充训练数据的过程。在图像分割任务中,数据增强尤为重要,因为模型通常需要学习从复杂的图像中识别出目标区域。通过应用诸如随机裁剪、旋转、缩放、翻转等变换,数据增强可以增加模型的鲁棒性,使其能够应对各种实际场景中的图像变化。 在使用MMSegmentation进行数据增强时,用户可以通过修改配置文件来指定增强策略。这些策略可以包括基本的几何变换,如随机缩放和裁剪,以及更高级的增强方法,如色彩抖动和亮度调整。MMSegmentation还支持使用预定义的增强策略组合,或者用户可以自定义自己的增强方法。 一旦增强策略确定,MMSegmentation会在训练过程中自动应用这些策略来扩充数据集。这意味着在每个训练批次
资源推荐
资源详情
资源评论
收起资源包目录
mmsegmentation实现对数据集进行图片分割,数据增强 (1784个子文件)
make.bat 760B
make.bat 760B
CITATION.cff 291B
setup.cfg 589B
readthedocs.css 145B
readthedocs.css 145B
Dockerfile 1KB
Dockerfile 1KB
Dockerfile 528B
seg_demo.gif 930KB
.gitignore 1KB
bpe_simple_vocab_16e6.txt.gz 1.29MB
bpe_simple_vocab_16e6.txt.gz 1.29MB
biomedical.nii.gz 49KB
biomedical_ann.nii.gz 803B
MANIFEST.in 205B
MMSegmentation_Tutorial.ipynb 16KB
sam_image_demo.ipynb 3KB
inference_demo.ipynb 3KB
zhihu_qrcode.jpg 388KB
miaomiao_qrcode.jpg 220KB
3dogs.jpg 181KB
classroom__rgb_00283.jpg 49KB
gray.jpg 38KB
color.jpg 35KB
case0005_slice001.jpg 33KB
case0005_slice000.jpg 33KB
86_185913.jpg 13KB
684_2150041.jpg 11KB
bookstore_0001d_00001.jpg 11KB
00001_img.jpg 2KB
__CRyFzoDOXn6unQ6a3DnQ.jpg 2KB
00004_img.jpg 2KB
00003_img.jpg 2KB
00002_img.jpg 2KB
00000_img.jpg 2KB
7d15b18b-1e0d6e3f.jpg 387B
0004a4c0-d4dff0ad.jpg 371B
00067cfb-e535423e.jpg 367B
00054602-3bf57337.jpg 364B
7d06fefd-f7be05a6.jpg 356B
7d128593-0ccfea4c.jpg 347B
dataset.json 574B
train_samples.json 563B
LICENSE 11KB
Makefile 634B
Makefile 634B
README.md 55KB
README.md 43KB
changelog_v0.x.md 40KB
README.md 36KB
changelog.md 34KB
README.md 32KB
README.md 31KB
2_dataset_prepare.md 29KB
2_dataset_prepare.md 29KB
1_config.md 27KB
1_config.md 27KB
README.md 26KB
README_zh-CN.md 26KB
README.md 26KB
README.md 20KB
contribute_dataset.md 20KB
engine.md 18KB
engine.md 17KB
interface.md 15KB
README.md 15KB
README.md 15KB
interface.md 15KB
README.md 15KB
README.md 14KB
README.md 14KB
README.md 14KB
README.md 14KB
README.md 14KB
README.md 14KB
deploy_jetson.md 14KB
useful_tools.md 14KB
README.md 14KB
5_deployment.md 14KB
datasets.md 14KB
README.md 14KB
datasets.md 14KB
5_deployment.md 13KB
README.md 13KB
4_train_test.md 13KB
README.md 12KB
4_train_test.md 12KB
README.md 12KB
README.md 11KB
README.md 11KB
README.md 11KB
3_inference.md 11KB
README.md 11KB
README.md 11KB
3_inference.md 11KB
faq.md 10KB
README.md 10KB
README.md 10KB
README.md 10KB
共 1784 条
- 1
- 2
- 3
- 4
- 5
- 6
- 18
资源评论
yc1111yc
- 粉丝: 22
- 资源: 164
上传资源 快速赚钱
- 我的内容管理 展开
- 我的资源 快来上传第一个资源
- 我的收益 登录查看自己的收益
- 我的积分 登录查看自己的积分
- 我的C币 登录后查看C币余额
- 我的收藏
- 我的下载
- 下载帮助
安全验证
文档复制为VIP权益,开通VIP直接复制
信息提交成功