Spaces:
Runtime error
Runtime error
layer { | |
name: "data" | |
type: "Input" | |
top: "data" | |
input_param { | |
shape { | |
dim: 1 | |
dim: 1 | |
dim: 384 | |
dim: 384 | |
} | |
} | |
} | |
layer { | |
name: "data/bn" | |
type: "BatchNorm" | |
bottom: "data" | |
top: "data" | |
param { | |
lr_mult: 0.0 | |
decay_mult: 0.0 | |
} | |
param { | |
lr_mult: 0.0 | |
decay_mult: 0.0 | |
} | |
param { | |
lr_mult: 0.0 | |
decay_mult: 0.0 | |
} | |
} | |
layer { | |
name: "data/bn/scale" | |
type: "Scale" | |
bottom: "data" | |
top: "data" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 0.0 | |
} | |
param { | |
lr_mult: 1.0 | |
decay_mult: 0.0 | |
} | |
scale_param { | |
filler { | |
type: "constant" | |
value: 1.0 | |
} | |
bias_term: true | |
bias_filler { | |
type: "constant" | |
value: 0.0 | |
} | |
} | |
} | |
layer { | |
name: "stage1" | |
type: "Convolution" | |
bottom: "data" | |
top: "stage1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
param { | |
lr_mult: 1.0 | |
decay_mult: 0.0 | |
} | |
convolution_param { | |
num_output: 24 | |
bias_term: true | |
pad: 1 | |
kernel_size: 3 | |
group: 1 | |
stride: 2 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage1/bn" | |
type: "BatchNorm" | |
bottom: "stage1" | |
top: "stage1" | |
param { | |
lr_mult: 0.0 | |
decay_mult: 0.0 | |
} | |
param { | |
lr_mult: 0.0 | |
decay_mult: 0.0 | |
} | |
param { | |
lr_mult: 0.0 | |
decay_mult: 0.0 | |
} | |
} | |
layer { | |
name: "stage1/bn/scale" | |
type: "Scale" | |
bottom: "stage1" | |
top: "stage1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 0.0 | |
} | |
param { | |
lr_mult: 1.0 | |
decay_mult: 0.0 | |
} | |
scale_param { | |
filler { | |
type: "constant" | |
value: 1.0 | |
} | |
bias_term: true | |
bias_filler { | |
type: "constant" | |
value: 0.0 | |
} | |
} | |
} | |
layer { | |
name: "stage1/relu" | |
type: "ReLU" | |
bottom: "stage1" | |
top: "stage1" | |
} | |
layer { | |
name: "stage2" | |
type: "Pooling" | |
bottom: "stage1" | |
top: "stage2" | |
pooling_param { | |
pool: MAX | |
kernel_size: 3 | |
stride: 2 | |
pad: 0 | |
} | |
} | |
layer { | |
name: "stage3_1/conv1" | |
type: "Convolution" | |
bottom: "stage2" | |
top: "stage3_1/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 16 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage3_1/conv1/relu" | |
type: "ReLU" | |
bottom: "stage3_1/conv1" | |
top: "stage3_1/conv1" | |
} | |
layer { | |
name: "stage3_1/conv2" | |
type: "Convolution" | |
bottom: "stage3_1/conv1" | |
top: "stage3_1/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 16 | |
pad: 1 | |
kernel_size: 3 | |
group: 16 | |
stride: 2 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage3_1/conv3" | |
type: "Convolution" | |
bottom: "stage3_1/conv2" | |
top: "stage3_1/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 64 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage3_1/relu" | |
type: "ReLU" | |
bottom: "stage3_1/conv3" | |
top: "stage3_1/conv3" | |
} | |
layer { | |
name: "stage3_2/conv1" | |
type: "Convolution" | |
bottom: "stage3_1/conv3" | |
top: "stage3_2/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 16 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage3_2/conv1/relu" | |
type: "ReLU" | |
bottom: "stage3_2/conv1" | |
top: "stage3_2/conv1" | |
} | |
layer { | |
name: "stage3_2/conv2" | |
type: "Convolution" | |
bottom: "stage3_2/conv1" | |
top: "stage3_2/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 16 | |
pad: 1 | |
kernel_size: 3 | |
group: 16 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage3_2/conv3" | |
type: "Convolution" | |
bottom: "stage3_2/conv2" | |
top: "stage3_2/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 64 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage3_2/sum" | |
type: "Eltwise" | |
bottom: "stage3_1/conv3" | |
bottom: "stage3_2/conv3" | |
top: "stage3_2/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage3_2/relu" | |
type: "ReLU" | |
bottom: "stage3_2/sum" | |
top: "stage3_2/sum" | |
} | |
layer { | |
name: "stage3_3/conv1" | |
type: "Convolution" | |
bottom: "stage3_2/sum" | |
top: "stage3_3/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 16 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage3_3/conv1/relu" | |
type: "ReLU" | |
bottom: "stage3_3/conv1" | |
top: "stage3_3/conv1" | |
} | |
layer { | |
name: "stage3_3/conv2" | |
type: "Convolution" | |
bottom: "stage3_3/conv1" | |
top: "stage3_3/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 16 | |
pad: 1 | |
kernel_size: 3 | |
group: 16 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage3_3/conv3" | |
type: "Convolution" | |
bottom: "stage3_3/conv2" | |
top: "stage3_3/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 64 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage3_3/sum" | |
type: "Eltwise" | |
bottom: "stage3_2/sum" | |
bottom: "stage3_3/conv3" | |
top: "stage3_3/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage3_3/relu" | |
type: "ReLU" | |
bottom: "stage3_3/sum" | |
top: "stage3_3/sum" | |
} | |
layer { | |
name: "stage3_4/conv1" | |
type: "Convolution" | |
bottom: "stage3_3/sum" | |
top: "stage3_4/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 16 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage3_4/conv1/relu" | |
type: "ReLU" | |
bottom: "stage3_4/conv1" | |
top: "stage3_4/conv1" | |
} | |
layer { | |
name: "stage3_4/conv2" | |
type: "Convolution" | |
bottom: "stage3_4/conv1" | |
top: "stage3_4/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 16 | |
pad: 1 | |
kernel_size: 3 | |
group: 16 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage3_4/conv3" | |
type: "Convolution" | |
bottom: "stage3_4/conv2" | |
top: "stage3_4/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 64 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage3_4/sum" | |
type: "Eltwise" | |
bottom: "stage3_3/sum" | |
bottom: "stage3_4/conv3" | |
top: "stage3_4/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage3_4/relu" | |
type: "ReLU" | |
bottom: "stage3_4/sum" | |
top: "stage3_4/sum" | |
} | |
layer { | |
name: "stage4_1/conv1" | |
type: "Convolution" | |
bottom: "stage3_4/sum" | |
top: "stage4_1/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_1/conv1/relu" | |
type: "ReLU" | |
bottom: "stage4_1/conv1" | |
top: "stage4_1/conv1" | |
} | |
layer { | |
name: "stage4_1/conv2" | |
type: "Convolution" | |
bottom: "stage4_1/conv1" | |
top: "stage4_1/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 1 | |
kernel_size: 3 | |
group: 32 | |
stride: 2 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_1/conv3" | |
type: "Convolution" | |
bottom: "stage4_1/conv2" | |
top: "stage4_1/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_1/relu" | |
type: "ReLU" | |
bottom: "stage4_1/conv3" | |
top: "stage4_1/conv3" | |
} | |
layer { | |
name: "stage4_2/conv1" | |
type: "Convolution" | |
bottom: "stage4_1/conv3" | |
top: "stage4_2/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_2/conv1/relu" | |
type: "ReLU" | |
bottom: "stage4_2/conv1" | |
top: "stage4_2/conv1" | |
} | |
layer { | |
name: "stage4_2/conv2" | |
type: "Convolution" | |
bottom: "stage4_2/conv1" | |
top: "stage4_2/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 1 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_2/conv3" | |
type: "Convolution" | |
bottom: "stage4_2/conv2" | |
top: "stage4_2/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_2/sum" | |
type: "Eltwise" | |
bottom: "stage4_1/conv3" | |
bottom: "stage4_2/conv3" | |
top: "stage4_2/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage4_2/relu" | |
type: "ReLU" | |
bottom: "stage4_2/sum" | |
top: "stage4_2/sum" | |
} | |
layer { | |
name: "stage4_3/conv1" | |
type: "Convolution" | |
bottom: "stage4_2/sum" | |
top: "stage4_3/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_3/conv1/relu" | |
type: "ReLU" | |
bottom: "stage4_3/conv1" | |
top: "stage4_3/conv1" | |
} | |
layer { | |
name: "stage4_3/conv2" | |
type: "Convolution" | |
bottom: "stage4_3/conv1" | |
top: "stage4_3/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 1 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_3/conv3" | |
type: "Convolution" | |
bottom: "stage4_3/conv2" | |
top: "stage4_3/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_3/sum" | |
type: "Eltwise" | |
bottom: "stage4_2/sum" | |
bottom: "stage4_3/conv3" | |
top: "stage4_3/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage4_3/relu" | |
type: "ReLU" | |
bottom: "stage4_3/sum" | |
top: "stage4_3/sum" | |
} | |
layer { | |
name: "stage4_4/conv1" | |
type: "Convolution" | |
bottom: "stage4_3/sum" | |
top: "stage4_4/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_4/conv1/relu" | |
type: "ReLU" | |
bottom: "stage4_4/conv1" | |
top: "stage4_4/conv1" | |
} | |
layer { | |
name: "stage4_4/conv2" | |
type: "Convolution" | |
bottom: "stage4_4/conv1" | |
top: "stage4_4/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 1 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_4/conv3" | |
type: "Convolution" | |
bottom: "stage4_4/conv2" | |
top: "stage4_4/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_4/sum" | |
type: "Eltwise" | |
bottom: "stage4_3/sum" | |
bottom: "stage4_4/conv3" | |
top: "stage4_4/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage4_4/relu" | |
type: "ReLU" | |
bottom: "stage4_4/sum" | |
top: "stage4_4/sum" | |
} | |
layer { | |
name: "stage4_5/conv1" | |
type: "Convolution" | |
bottom: "stage4_4/sum" | |
top: "stage4_5/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_5/conv1/relu" | |
type: "ReLU" | |
bottom: "stage4_5/conv1" | |
top: "stage4_5/conv1" | |
} | |
layer { | |
name: "stage4_5/conv2" | |
type: "Convolution" | |
bottom: "stage4_5/conv1" | |
top: "stage4_5/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 1 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_5/conv3" | |
type: "Convolution" | |
bottom: "stage4_5/conv2" | |
top: "stage4_5/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_5/sum" | |
type: "Eltwise" | |
bottom: "stage4_4/sum" | |
bottom: "stage4_5/conv3" | |
top: "stage4_5/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage4_5/relu" | |
type: "ReLU" | |
bottom: "stage4_5/sum" | |
top: "stage4_5/sum" | |
} | |
layer { | |
name: "stage4_6/conv1" | |
type: "Convolution" | |
bottom: "stage4_5/sum" | |
top: "stage4_6/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_6/conv1/relu" | |
type: "ReLU" | |
bottom: "stage4_6/conv1" | |
top: "stage4_6/conv1" | |
} | |
layer { | |
name: "stage4_6/conv2" | |
type: "Convolution" | |
bottom: "stage4_6/conv1" | |
top: "stage4_6/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 1 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_6/conv3" | |
type: "Convolution" | |
bottom: "stage4_6/conv2" | |
top: "stage4_6/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_6/sum" | |
type: "Eltwise" | |
bottom: "stage4_5/sum" | |
bottom: "stage4_6/conv3" | |
top: "stage4_6/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage4_6/relu" | |
type: "ReLU" | |
bottom: "stage4_6/sum" | |
top: "stage4_6/sum" | |
} | |
layer { | |
name: "stage4_7/conv1" | |
type: "Convolution" | |
bottom: "stage4_6/sum" | |
top: "stage4_7/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_7/conv1/relu" | |
type: "ReLU" | |
bottom: "stage4_7/conv1" | |
top: "stage4_7/conv1" | |
} | |
layer { | |
name: "stage4_7/conv2" | |
type: "Convolution" | |
bottom: "stage4_7/conv1" | |
top: "stage4_7/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 1 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_7/conv3" | |
type: "Convolution" | |
bottom: "stage4_7/conv2" | |
top: "stage4_7/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_7/sum" | |
type: "Eltwise" | |
bottom: "stage4_6/sum" | |
bottom: "stage4_7/conv3" | |
top: "stage4_7/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage4_7/relu" | |
type: "ReLU" | |
bottom: "stage4_7/sum" | |
top: "stage4_7/sum" | |
} | |
layer { | |
name: "stage4_8/conv1" | |
type: "Convolution" | |
bottom: "stage4_7/sum" | |
top: "stage4_8/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_8/conv1/relu" | |
type: "ReLU" | |
bottom: "stage4_8/conv1" | |
top: "stage4_8/conv1" | |
} | |
layer { | |
name: "stage4_8/conv2" | |
type: "Convolution" | |
bottom: "stage4_8/conv1" | |
top: "stage4_8/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 1 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_8/conv3" | |
type: "Convolution" | |
bottom: "stage4_8/conv2" | |
top: "stage4_8/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage4_8/sum" | |
type: "Eltwise" | |
bottom: "stage4_7/sum" | |
bottom: "stage4_8/conv3" | |
top: "stage4_8/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage4_8/relu" | |
type: "ReLU" | |
bottom: "stage4_8/sum" | |
top: "stage4_8/sum" | |
} | |
layer { | |
name: "stage5_1/conv1" | |
type: "Convolution" | |
bottom: "stage4_8/sum" | |
top: "stage5_1/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage5_1/conv1/relu" | |
type: "ReLU" | |
bottom: "stage5_1/conv1" | |
top: "stage5_1/conv1" | |
} | |
layer { | |
name: "stage5_1/conv2" | |
type: "Convolution" | |
bottom: "stage5_1/conv1" | |
top: "stage5_1/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 2 | |
kernel_size: 3 | |
group: 32 | |
stride: 2 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 2 | |
} | |
} | |
layer { | |
name: "stage5_1/conv3" | |
type: "Convolution" | |
bottom: "stage5_1/conv2" | |
top: "stage5_1/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage5_1/relu" | |
type: "ReLU" | |
bottom: "stage5_1/conv3" | |
top: "stage5_1/conv3" | |
} | |
layer { | |
name: "stage5_2/conv1" | |
type: "Convolution" | |
bottom: "stage5_1/conv3" | |
top: "stage5_2/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage5_2/conv1/relu" | |
type: "ReLU" | |
bottom: "stage5_2/conv1" | |
top: "stage5_2/conv1" | |
} | |
layer { | |
name: "stage5_2/conv2" | |
type: "Convolution" | |
bottom: "stage5_2/conv1" | |
top: "stage5_2/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 2 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 2 | |
} | |
} | |
layer { | |
name: "stage5_2/conv3" | |
type: "Convolution" | |
bottom: "stage5_2/conv2" | |
top: "stage5_2/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage5_2/sum" | |
type: "Eltwise" | |
bottom: "stage5_1/conv3" | |
bottom: "stage5_2/conv3" | |
top: "stage5_2/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage5_2/relu" | |
type: "ReLU" | |
bottom: "stage5_2/sum" | |
top: "stage5_2/sum" | |
} | |
layer { | |
name: "stage5_3/conv1" | |
type: "Convolution" | |
bottom: "stage5_2/sum" | |
top: "stage5_3/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage5_3/conv1/relu" | |
type: "ReLU" | |
bottom: "stage5_3/conv1" | |
top: "stage5_3/conv1" | |
} | |
layer { | |
name: "stage5_3/conv2" | |
type: "Convolution" | |
bottom: "stage5_3/conv1" | |
top: "stage5_3/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 2 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 2 | |
} | |
} | |
layer { | |
name: "stage5_3/conv3" | |
type: "Convolution" | |
bottom: "stage5_3/conv2" | |
top: "stage5_3/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage5_3/sum" | |
type: "Eltwise" | |
bottom: "stage5_2/sum" | |
bottom: "stage5_3/conv3" | |
top: "stage5_3/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage5_3/relu" | |
type: "ReLU" | |
bottom: "stage5_3/sum" | |
top: "stage5_3/sum" | |
} | |
layer { | |
name: "stage5_4/conv1" | |
type: "Convolution" | |
bottom: "stage5_3/sum" | |
top: "stage5_4/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage5_4/conv1/relu" | |
type: "ReLU" | |
bottom: "stage5_4/conv1" | |
top: "stage5_4/conv1" | |
} | |
layer { | |
name: "stage5_4/conv2" | |
type: "Convolution" | |
bottom: "stage5_4/conv1" | |
top: "stage5_4/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 2 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 2 | |
} | |
} | |
layer { | |
name: "stage5_4/conv3" | |
type: "Convolution" | |
bottom: "stage5_4/conv2" | |
top: "stage5_4/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage5_4/sum" | |
type: "Eltwise" | |
bottom: "stage5_3/sum" | |
bottom: "stage5_4/conv3" | |
top: "stage5_4/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage5_4/relu" | |
type: "ReLU" | |
bottom: "stage5_4/sum" | |
top: "stage5_4/sum" | |
} | |
layer { | |
name: "stage6_1/conv4" | |
type: "Convolution" | |
bottom: "stage5_4/sum" | |
top: "stage6_1/conv4" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage6_1/conv1" | |
type: "Convolution" | |
bottom: "stage5_4/sum" | |
top: "stage6_1/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage6_1/conv1/relu" | |
type: "ReLU" | |
bottom: "stage6_1/conv1" | |
top: "stage6_1/conv1" | |
} | |
layer { | |
name: "stage6_1/conv2" | |
type: "Convolution" | |
bottom: "stage6_1/conv1" | |
top: "stage6_1/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 2 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 2 | |
} | |
} | |
layer { | |
name: "stage6_1/conv3" | |
type: "Convolution" | |
bottom: "stage6_1/conv2" | |
top: "stage6_1/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage6_1/sum" | |
type: "Eltwise" | |
bottom: "stage6_1/conv4" | |
bottom: "stage6_1/conv3" | |
top: "stage6_1/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage6_1/relu" | |
type: "ReLU" | |
bottom: "stage6_1/sum" | |
top: "stage6_1/sum" | |
} | |
layer { | |
name: "stage6_2/conv1" | |
type: "Convolution" | |
bottom: "stage6_1/sum" | |
top: "stage6_2/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage6_2/conv1/relu" | |
type: "ReLU" | |
bottom: "stage6_2/conv1" | |
top: "stage6_2/conv1" | |
} | |
layer { | |
name: "stage6_2/conv2" | |
type: "Convolution" | |
bottom: "stage6_2/conv1" | |
top: "stage6_2/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 2 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 2 | |
} | |
} | |
layer { | |
name: "stage6_2/conv3" | |
type: "Convolution" | |
bottom: "stage6_2/conv2" | |
top: "stage6_2/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage6_2/sum" | |
type: "Eltwise" | |
bottom: "stage6_1/sum" | |
bottom: "stage6_2/conv3" | |
top: "stage6_2/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage6_2/relu" | |
type: "ReLU" | |
bottom: "stage6_2/sum" | |
top: "stage6_2/sum" | |
} | |
layer { | |
name: "stage7_1/conv4" | |
type: "Convolution" | |
bottom: "stage6_2/sum" | |
top: "stage7_1/conv4" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage7_1/conv1" | |
type: "Convolution" | |
bottom: "stage6_2/sum" | |
top: "stage7_1/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage7_1/conv1/relu" | |
type: "ReLU" | |
bottom: "stage7_1/conv1" | |
top: "stage7_1/conv1" | |
} | |
layer { | |
name: "stage7_1/conv2" | |
type: "Convolution" | |
bottom: "stage7_1/conv1" | |
top: "stage7_1/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 2 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 2 | |
} | |
} | |
layer { | |
name: "stage7_1/conv3" | |
type: "Convolution" | |
bottom: "stage7_1/conv2" | |
top: "stage7_1/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage7_1/sum" | |
type: "Eltwise" | |
bottom: "stage7_1/conv4" | |
bottom: "stage7_1/conv3" | |
top: "stage7_1/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage7_1/relu" | |
type: "ReLU" | |
bottom: "stage7_1/sum" | |
top: "stage7_1/sum" | |
} | |
layer { | |
name: "stage7_2/conv1" | |
type: "Convolution" | |
bottom: "stage7_1/sum" | |
top: "stage7_2/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage7_2/conv1/relu" | |
type: "ReLU" | |
bottom: "stage7_2/conv1" | |
top: "stage7_2/conv1" | |
} | |
layer { | |
name: "stage7_2/conv2" | |
type: "Convolution" | |
bottom: "stage7_2/conv1" | |
top: "stage7_2/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 2 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 2 | |
} | |
} | |
layer { | |
name: "stage7_2/conv3" | |
type: "Convolution" | |
bottom: "stage7_2/conv2" | |
top: "stage7_2/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage7_2/sum" | |
type: "Eltwise" | |
bottom: "stage7_1/sum" | |
bottom: "stage7_2/conv3" | |
top: "stage7_2/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage7_2/relu" | |
type: "ReLU" | |
bottom: "stage7_2/sum" | |
top: "stage7_2/sum" | |
} | |
layer { | |
name: "stage8_1/conv4" | |
type: "Convolution" | |
bottom: "stage7_2/sum" | |
top: "stage8_1/conv4" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage8_1/conv1" | |
type: "Convolution" | |
bottom: "stage7_2/sum" | |
top: "stage8_1/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage8_1/conv1/relu" | |
type: "ReLU" | |
bottom: "stage8_1/conv1" | |
top: "stage8_1/conv1" | |
} | |
layer { | |
name: "stage8_1/conv2" | |
type: "Convolution" | |
bottom: "stage8_1/conv1" | |
top: "stage8_1/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 2 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 2 | |
} | |
} | |
layer { | |
name: "stage8_1/conv3" | |
type: "Convolution" | |
bottom: "stage8_1/conv2" | |
top: "stage8_1/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage8_1/sum" | |
type: "Eltwise" | |
bottom: "stage8_1/conv4" | |
bottom: "stage8_1/conv3" | |
top: "stage8_1/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage8_1/relu" | |
type: "ReLU" | |
bottom: "stage8_1/sum" | |
top: "stage8_1/sum" | |
} | |
layer { | |
name: "stage8_2/conv1" | |
type: "Convolution" | |
bottom: "stage8_1/sum" | |
top: "stage8_2/conv1" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage8_2/conv1/relu" | |
type: "ReLU" | |
bottom: "stage8_2/conv1" | |
top: "stage8_2/conv1" | |
} | |
layer { | |
name: "stage8_2/conv2" | |
type: "Convolution" | |
bottom: "stage8_2/conv1" | |
top: "stage8_2/conv2" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 32 | |
pad: 2 | |
kernel_size: 3 | |
group: 32 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 2 | |
} | |
} | |
layer { | |
name: "stage8_2/conv3" | |
type: "Convolution" | |
bottom: "stage8_2/conv2" | |
top: "stage8_2/conv3" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
convolution_param { | |
num_output: 128 | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "stage8_2/sum" | |
type: "Eltwise" | |
bottom: "stage8_1/sum" | |
bottom: "stage8_2/conv3" | |
top: "stage8_2/sum" | |
eltwise_param { | |
operation: SUM | |
} | |
} | |
layer { | |
name: "stage8_2/relu" | |
type: "ReLU" | |
bottom: "stage8_2/sum" | |
top: "stage8_2/sum" | |
} | |
layer { | |
name: "cls1/conv" | |
type: "Convolution" | |
bottom: "stage4_8/sum" | |
top: "cls1/conv" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
param { | |
lr_mult: 1.0 | |
decay_mult: 0.0 | |
} | |
convolution_param { | |
num_output: 12 | |
bias_term: true | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "cls1/permute" | |
type: "Permute" | |
bottom: "cls1/conv" | |
top: "cls1/permute" | |
permute_param { | |
order: 0 | |
order: 2 | |
order: 3 | |
order: 1 | |
} | |
} | |
layer { | |
name: "cls1/flatten" | |
type: "Flatten" | |
bottom: "cls1/permute" | |
top: "cls1/flatten" | |
flatten_param { | |
axis: 1 | |
} | |
} | |
layer { | |
name: "loc1/conv" | |
type: "Convolution" | |
bottom: "stage4_8/sum" | |
top: "loc1/conv" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
param { | |
lr_mult: 1.0 | |
decay_mult: 0.0 | |
} | |
convolution_param { | |
num_output: 24 | |
bias_term: true | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "loc1/permute" | |
type: "Permute" | |
bottom: "loc1/conv" | |
top: "loc1/permute" | |
permute_param { | |
order: 0 | |
order: 2 | |
order: 3 | |
order: 1 | |
} | |
} | |
layer { | |
name: "loc1/flatten" | |
type: "Flatten" | |
bottom: "loc1/permute" | |
top: "loc1/flatten" | |
flatten_param { | |
axis: 1 | |
} | |
} | |
layer { | |
name: "stage4_8/sum/prior_box" | |
type: "PriorBox" | |
bottom: "stage4_8/sum" | |
bottom: "data" | |
top: "stage4_8/sum/prior_box" | |
prior_box_param { | |
min_size: 50.0 | |
max_size: 100.0 | |
aspect_ratio: 2.0 | |
aspect_ratio: 0.5 | |
aspect_ratio: 3.0 | |
aspect_ratio: 0.3333333432674408 | |
flip: false | |
clip: false | |
variance: 0.10000000149011612 | |
variance: 0.10000000149011612 | |
variance: 0.20000000298023224 | |
variance: 0.20000000298023224 | |
step: 16.0 | |
} | |
} | |
layer { | |
name: "cls2/conv" | |
type: "Convolution" | |
bottom: "stage5_4/sum" | |
top: "cls2/conv" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
param { | |
lr_mult: 1.0 | |
decay_mult: 0.0 | |
} | |
convolution_param { | |
num_output: 12 | |
bias_term: true | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "cls2/permute" | |
type: "Permute" | |
bottom: "cls2/conv" | |
top: "cls2/permute" | |
permute_param { | |
order: 0 | |
order: 2 | |
order: 3 | |
order: 1 | |
} | |
} | |
layer { | |
name: "cls2/flatten" | |
type: "Flatten" | |
bottom: "cls2/permute" | |
top: "cls2/flatten" | |
flatten_param { | |
axis: 1 | |
} | |
} | |
layer { | |
name: "loc2/conv" | |
type: "Convolution" | |
bottom: "stage5_4/sum" | |
top: "loc2/conv" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
param { | |
lr_mult: 1.0 | |
decay_mult: 0.0 | |
} | |
convolution_param { | |
num_output: 24 | |
bias_term: true | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "loc2/permute" | |
type: "Permute" | |
bottom: "loc2/conv" | |
top: "loc2/permute" | |
permute_param { | |
order: 0 | |
order: 2 | |
order: 3 | |
order: 1 | |
} | |
} | |
layer { | |
name: "loc2/flatten" | |
type: "Flatten" | |
bottom: "loc2/permute" | |
top: "loc2/flatten" | |
flatten_param { | |
axis: 1 | |
} | |
} | |
layer { | |
name: "stage5_4/sum/prior_box" | |
type: "PriorBox" | |
bottom: "stage5_4/sum" | |
bottom: "data" | |
top: "stage5_4/sum/prior_box" | |
prior_box_param { | |
min_size: 100.0 | |
max_size: 150.0 | |
aspect_ratio: 2.0 | |
aspect_ratio: 0.5 | |
aspect_ratio: 3.0 | |
aspect_ratio: 0.3333333432674408 | |
flip: false | |
clip: false | |
variance: 0.10000000149011612 | |
variance: 0.10000000149011612 | |
variance: 0.20000000298023224 | |
variance: 0.20000000298023224 | |
step: 32.0 | |
} | |
} | |
layer { | |
name: "cls3/conv" | |
type: "Convolution" | |
bottom: "stage6_2/sum" | |
top: "cls3/conv" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
param { | |
lr_mult: 1.0 | |
decay_mult: 0.0 | |
} | |
convolution_param { | |
num_output: 12 | |
bias_term: true | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "cls3/permute" | |
type: "Permute" | |
bottom: "cls3/conv" | |
top: "cls3/permute" | |
permute_param { | |
order: 0 | |
order: 2 | |
order: 3 | |
order: 1 | |
} | |
} | |
layer { | |
name: "cls3/flatten" | |
type: "Flatten" | |
bottom: "cls3/permute" | |
top: "cls3/flatten" | |
flatten_param { | |
axis: 1 | |
} | |
} | |
layer { | |
name: "loc3/conv" | |
type: "Convolution" | |
bottom: "stage6_2/sum" | |
top: "loc3/conv" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
param { | |
lr_mult: 1.0 | |
decay_mult: 0.0 | |
} | |
convolution_param { | |
num_output: 24 | |
bias_term: true | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "loc3/permute" | |
type: "Permute" | |
bottom: "loc3/conv" | |
top: "loc3/permute" | |
permute_param { | |
order: 0 | |
order: 2 | |
order: 3 | |
order: 1 | |
} | |
} | |
layer { | |
name: "loc3/flatten" | |
type: "Flatten" | |
bottom: "loc3/permute" | |
top: "loc3/flatten" | |
flatten_param { | |
axis: 1 | |
} | |
} | |
layer { | |
name: "stage6_2/sum/prior_box" | |
type: "PriorBox" | |
bottom: "stage6_2/sum" | |
bottom: "data" | |
top: "stage6_2/sum/prior_box" | |
prior_box_param { | |
min_size: 150.0 | |
max_size: 200.0 | |
aspect_ratio: 2.0 | |
aspect_ratio: 0.5 | |
aspect_ratio: 3.0 | |
aspect_ratio: 0.3333333432674408 | |
flip: false | |
clip: false | |
variance: 0.10000000149011612 | |
variance: 0.10000000149011612 | |
variance: 0.20000000298023224 | |
variance: 0.20000000298023224 | |
step: 32.0 | |
} | |
} | |
layer { | |
name: "cls4/conv" | |
type: "Convolution" | |
bottom: "stage7_2/sum" | |
top: "cls4/conv" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
param { | |
lr_mult: 1.0 | |
decay_mult: 0.0 | |
} | |
convolution_param { | |
num_output: 12 | |
bias_term: true | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "cls4/permute" | |
type: "Permute" | |
bottom: "cls4/conv" | |
top: "cls4/permute" | |
permute_param { | |
order: 0 | |
order: 2 | |
order: 3 | |
order: 1 | |
} | |
} | |
layer { | |
name: "cls4/flatten" | |
type: "Flatten" | |
bottom: "cls4/permute" | |
top: "cls4/flatten" | |
flatten_param { | |
axis: 1 | |
} | |
} | |
layer { | |
name: "loc4/conv" | |
type: "Convolution" | |
bottom: "stage7_2/sum" | |
top: "loc4/conv" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
param { | |
lr_mult: 1.0 | |
decay_mult: 0.0 | |
} | |
convolution_param { | |
num_output: 24 | |
bias_term: true | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "loc4/permute" | |
type: "Permute" | |
bottom: "loc4/conv" | |
top: "loc4/permute" | |
permute_param { | |
order: 0 | |
order: 2 | |
order: 3 | |
order: 1 | |
} | |
} | |
layer { | |
name: "loc4/flatten" | |
type: "Flatten" | |
bottom: "loc4/permute" | |
top: "loc4/flatten" | |
flatten_param { | |
axis: 1 | |
} | |
} | |
layer { | |
name: "stage7_2/sum/prior_box" | |
type: "PriorBox" | |
bottom: "stage7_2/sum" | |
bottom: "data" | |
top: "stage7_2/sum/prior_box" | |
prior_box_param { | |
min_size: 200.0 | |
max_size: 300.0 | |
aspect_ratio: 2.0 | |
aspect_ratio: 0.5 | |
aspect_ratio: 3.0 | |
aspect_ratio: 0.3333333432674408 | |
flip: false | |
clip: false | |
variance: 0.10000000149011612 | |
variance: 0.10000000149011612 | |
variance: 0.20000000298023224 | |
variance: 0.20000000298023224 | |
step: 32.0 | |
} | |
} | |
layer { | |
name: "cls5/conv" | |
type: "Convolution" | |
bottom: "stage8_2/sum" | |
top: "cls5/conv" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
param { | |
lr_mult: 1.0 | |
decay_mult: 0.0 | |
} | |
convolution_param { | |
num_output: 12 | |
bias_term: true | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "cls5/permute" | |
type: "Permute" | |
bottom: "cls5/conv" | |
top: "cls5/permute" | |
permute_param { | |
order: 0 | |
order: 2 | |
order: 3 | |
order: 1 | |
} | |
} | |
layer { | |
name: "cls5/flatten" | |
type: "Flatten" | |
bottom: "cls5/permute" | |
top: "cls5/flatten" | |
flatten_param { | |
axis: 1 | |
} | |
} | |
layer { | |
name: "loc5/conv" | |
type: "Convolution" | |
bottom: "stage8_2/sum" | |
top: "loc5/conv" | |
param { | |
lr_mult: 1.0 | |
decay_mult: 1.0 | |
} | |
param { | |
lr_mult: 1.0 | |
decay_mult: 0.0 | |
} | |
convolution_param { | |
num_output: 24 | |
bias_term: true | |
pad: 0 | |
kernel_size: 1 | |
group: 1 | |
stride: 1 | |
weight_filler { | |
type: "msra" | |
} | |
dilation: 1 | |
} | |
} | |
layer { | |
name: "loc5/permute" | |
type: "Permute" | |
bottom: "loc5/conv" | |
top: "loc5/permute" | |
permute_param { | |
order: 0 | |
order: 2 | |
order: 3 | |
order: 1 | |
} | |
} | |
layer { | |
name: "loc5/flatten" | |
type: "Flatten" | |
bottom: "loc5/permute" | |
top: "loc5/flatten" | |
flatten_param { | |
axis: 1 | |
} | |
} | |
layer { | |
name: "stage8_2/sum/prior_box" | |
type: "PriorBox" | |
bottom: "stage8_2/sum" | |
bottom: "data" | |
top: "stage8_2/sum/prior_box" | |
prior_box_param { | |
min_size: 300.0 | |
max_size: 400.0 | |
aspect_ratio: 2.0 | |
aspect_ratio: 0.5 | |
aspect_ratio: 3.0 | |
aspect_ratio: 0.3333333432674408 | |
flip: false | |
clip: false | |
variance: 0.10000000149011612 | |
variance: 0.10000000149011612 | |
variance: 0.20000000298023224 | |
variance: 0.20000000298023224 | |
step: 32.0 | |
} | |
} | |
layer { | |
name: "mbox_conf" | |
type: "Concat" | |
bottom: "cls1/flatten" | |
bottom: "cls2/flatten" | |
bottom: "cls3/flatten" | |
bottom: "cls4/flatten" | |
bottom: "cls5/flatten" | |
top: "mbox_conf" | |
concat_param { | |
axis: 1 | |
} | |
} | |
layer { | |
name: "mbox_loc" | |
type: "Concat" | |
bottom: "loc1/flatten" | |
bottom: "loc2/flatten" | |
bottom: "loc3/flatten" | |
bottom: "loc4/flatten" | |
bottom: "loc5/flatten" | |
top: "mbox_loc" | |
concat_param { | |
axis: 1 | |
} | |
} | |
layer { | |
name: "mbox_priorbox" | |
type: "Concat" | |
bottom: "stage4_8/sum/prior_box" | |
bottom: "stage5_4/sum/prior_box" | |
bottom: "stage6_2/sum/prior_box" | |
bottom: "stage7_2/sum/prior_box" | |
bottom: "stage8_2/sum/prior_box" | |
top: "mbox_priorbox" | |
concat_param { | |
axis: 2 | |
} | |
} | |
layer { | |
name: "mbox_conf_reshape" | |
type: "Reshape" | |
bottom: "mbox_conf" | |
top: "mbox_conf_reshape" | |
reshape_param { | |
shape { | |
dim: 0 | |
dim: -1 | |
dim: 2 | |
} | |
} | |
} | |
layer { | |
name: "mbox_conf_softmax" | |
type: "Softmax" | |
bottom: "mbox_conf_reshape" | |
top: "mbox_conf_softmax" | |
softmax_param { | |
axis: 2 | |
} | |
} | |
layer { | |
name: "mbox_conf_flatten" | |
type: "Flatten" | |
bottom: "mbox_conf_softmax" | |
top: "mbox_conf_flatten" | |
flatten_param { | |
axis: 1 | |
} | |
} | |
layer { | |
name: "detection_output" | |
type: "DetectionOutput" | |
bottom: "mbox_loc" | |
bottom: "mbox_conf_flatten" | |
bottom: "mbox_priorbox" | |
top: "detection_output" | |
detection_output_param { | |
num_classes: 2 | |
share_location: true | |
background_label_id: 0 | |
nms_param { | |
nms_threshold: 0.44999998807907104 | |
top_k: 100 | |
} | |
code_type: CENTER_SIZE | |
keep_top_k: 100 | |
confidence_threshold: 0.20000000298023224 | |
} | |
} | |