去噪算法:评价指标、计算参数量以及CBAM回顾

更新时间:2023-06-26 04:42:06 阅读: 评论:0

去噪算法:评价指标、计算参数量以及CBAM 回顾
Denoising review:
四六级口语The trials made can be found on Github.
Asssment indexes:
PSNR: peak signal to noi ratio
注意MSE是均⽅误差(已平均化),MAX表⽰图像颜⾊的最⼤值,⼀般8位图表⽰255
PSNR单位是dB.
SSIM: structural similarity index
where , and generally .
.
Note some basic np operations:
西安法语
np.split(a,4,0)
np.vstack
np.hstack, etc.
parameters & memory calculation
Note:
Total GPU memory = memory for model  & layer outputs
In training time: forward , backward X2
< VGG16 , memory of params: 528MB, memory of layers: 58.12MB/image
when training:
SGD+momentum, batchsize = 128
memory for model:
PSNR =10∗log ()=10MSE (2−1)n 2
20∗log ()
10MSE MAX L (X ,Y )=,C (X ,Y )=u +u +C X 2Y 212u u +C X Y 1,S (X ,Y )=σ+σ+C X 2Y 222σσ+C X Y 2σσ+C X Y 3
σ+C XY 3C 1=(K 1∗L ),C 2=2(K 2∗L ),C 3=2C 2/2K 1=0.01,K 2=0.03,l =255SSIM =L ∗C ∗S
1. 528MB*3 = 1.54 GB爱你在心口难开英文版
1 for params, 1 for SGD, 1 for momemtum
If u Adam, need to x4
2. Memory for outputs:
128*58.12MB *2 = 14.53GB
3. Total
14.53+1.54 = 16.07GB
FLOPS:
bevelConv = para*H*W
Linear = para
TFlops/s,可以简单写为T/s, 是数据流量的,意思是”1万亿次浮点指令每秒”,它是衡量⼀个电脑的标准。1TFlops=1024GFlops,即1T=1024G。
from torchstat import stat
model
1. Residual Den Network for Image Super-Resolution
Here, LR reprents Low Resolution Images.
upscale:
sub-pixel convolution
the best position of CBAM is after the up-sampling layer. up 0.0x dB, if patch-size up, perf down.
GRDN:
… yeah, concatenate , concatenate , concatenate .
CBAM:
The channel attention is computed as :
Where  reprents activation function, e.g. ReLU.
The spatial attention is computed as:
where, M (F )=C σ(MLP (AvgPool (F ))+MLP (MaxPool (F )))
σM (F )=s σ(f ([AvgPool (F );MaxPool (F )]))7∗7=σ(f ([F ;F ]))
7∗7avg s max s F ,F ∈avg s max s R 1×H ×W
class ChannelAttention(nn.Module):
def__init__(lf, in_planes, ratio=16):
super(ChannelAttention, lf).__init__()
lf.avg_pool = nn.AdaptiveAvgPool2d(1)
lf.max_pool = nn.AdaptiveMaxPool2d(1)
lf.fc1  = nn.Conv2d(in_planes, in_planes //16,1, bias=Fal)
lf.fc2  = nn.Conv2d(in_planes //16, in_planes,1, bias=Fal)
lf.sigmoid = nn.Sigmoid()
def forward(lf, x):
avg_out = lf.lu1(lf.fc1(lf.avg_pool(x))))itt
max_out = lf.lu1(lf.fc1(lf.max_pool(x))))
out = avg_out + max_out
return lf.sigmoid(out)
channel attention: $C\times W $ -> pool
class SpatialAttention(nn.Module):
horribledef__init__(lf, kernel_size=7):
super(SpatialAttention, lf).__init__()
asrt kernel_size in(3,7),'kernel size must be 3 or 7'
padding =3if kernel_size ==7el1
basket ca
lf.sigmoid = nn.Sigmoid()
def forward(lf, x):
百分比英文avg_out = an(x, dim=1, keepdim=True)
max_out, _ = torch.max(x, dim=1, keepdim=True)
x = torch.cat([avg_out, max_out], dim=1)
x = lf.conv1(x)
return lf.sigmoid(x)英语四级及格分数
语文表达方式
For channels, an or torch.max to implement the avg or pool for channel info.

本文发布于:2023-06-26 04:42:06,感谢您对本站的认可!

本文链接:https://www.wtabcd.cn/fanwen/fan/90/157903.html

版权声明:本站内容均来自互联网,仅供演示用,请勿用于商业和其他非法用途。如果侵犯了您的权益请与我们联系,我们将在24小时内删除。

标签:电脑   指标   数量   计算   评价   流量   误差
相关文章
留言与评论(共有 0 条评论)
   
验证码:
Copyright ©2019-2022 Comsenz Inc.Powered by © 专利检索| 网站地图