| layer { | |
| name: "data" | |
| type: "Input" | |
| top: "data" | |
| input_param { | |
| shape { | |
| dim: 1 | |
| dim: 1 | |
| dim: 384 | |
| dim: 384 | |
| } | |
| } | |
| } | |
| layer { | |
| name: "data/bn" | |
| type: "BatchNorm" | |
| bottom: "data" | |
| top: "data" | |
| param { | |
| lr_mult: 0.0 | |
| decay_mult: 0.0 | |
| } | |
| param { | |
| lr_mult: 0.0 | |
| decay_mult: 0.0 | |
| } | |
| param { | |
| lr_mult: 0.0 | |
| decay_mult: 0.0 | |
| } | |
| } | |
| layer { | |
| name: "data/bn/scale" | |
| type: "Scale" | |
| bottom: "data" | |
| top: "data" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 0.0 | |
| } | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 0.0 | |
| } | |
| scale_param { | |
| filler { | |
| type: "constant" | |
| value: 1.0 | |
| } | |
| bias_term: true | |
| bias_filler { | |
| type: "constant" | |
| value: 0.0 | |
| } | |
| } | |
| } | |
| layer { | |
| name: "stage1" | |
| type: "Convolution" | |
| bottom: "data" | |
| top: "stage1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 0.0 | |
| } | |
| convolution_param { | |
| num_output: 24 | |
| bias_term: true | |
| pad: 1 | |
| kernel_size: 3 | |
| group: 1 | |
| stride: 2 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage1/bn" | |
| type: "BatchNorm" | |
| bottom: "stage1" | |
| top: "stage1" | |
| param { | |
| lr_mult: 0.0 | |
| decay_mult: 0.0 | |
| } | |
| param { | |
| lr_mult: 0.0 | |
| decay_mult: 0.0 | |
| } | |
| param { | |
| lr_mult: 0.0 | |
| decay_mult: 0.0 | |
| } | |
| } | |
| layer { | |
| name: "stage1/bn/scale" | |
| type: "Scale" | |
| bottom: "stage1" | |
| top: "stage1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 0.0 | |
| } | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 0.0 | |
| } | |
| scale_param { | |
| filler { | |
| type: "constant" | |
| value: 1.0 | |
| } | |
| bias_term: true | |
| bias_filler { | |
| type: "constant" | |
| value: 0.0 | |
| } | |
| } | |
| } | |
| layer { | |
| name: "stage1/relu" | |
| type: "ReLU" | |
| bottom: "stage1" | |
| top: "stage1" | |
| } | |
| layer { | |
| name: "stage2" | |
| type: "Pooling" | |
| bottom: "stage1" | |
| top: "stage2" | |
| pooling_param { | |
| pool: MAX | |
| kernel_size: 3 | |
| stride: 2 | |
| pad: 0 | |
| } | |
| } | |
| layer { | |
| name: "stage3_1/conv1" | |
| type: "Convolution" | |
| bottom: "stage2" | |
| top: "stage3_1/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 16 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage3_1/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage3_1/conv1" | |
| top: "stage3_1/conv1" | |
| } | |
| layer { | |
| name: "stage3_1/conv2" | |
| type: "Convolution" | |
| bottom: "stage3_1/conv1" | |
| top: "stage3_1/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 16 | |
| pad: 1 | |
| kernel_size: 3 | |
| group: 16 | |
| stride: 2 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage3_1/conv3" | |
| type: "Convolution" | |
| bottom: "stage3_1/conv2" | |
| top: "stage3_1/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 64 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage3_1/relu" | |
| type: "ReLU" | |
| bottom: "stage3_1/conv3" | |
| top: "stage3_1/conv3" | |
| } | |
| layer { | |
| name: "stage3_2/conv1" | |
| type: "Convolution" | |
| bottom: "stage3_1/conv3" | |
| top: "stage3_2/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 16 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage3_2/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage3_2/conv1" | |
| top: "stage3_2/conv1" | |
| } | |
| layer { | |
| name: "stage3_2/conv2" | |
| type: "Convolution" | |
| bottom: "stage3_2/conv1" | |
| top: "stage3_2/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 16 | |
| pad: 1 | |
| kernel_size: 3 | |
| group: 16 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage3_2/conv3" | |
| type: "Convolution" | |
| bottom: "stage3_2/conv2" | |
| top: "stage3_2/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 64 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage3_2/sum" | |
| type: "Eltwise" | |
| bottom: "stage3_1/conv3" | |
| bottom: "stage3_2/conv3" | |
| top: "stage3_2/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage3_2/relu" | |
| type: "ReLU" | |
| bottom: "stage3_2/sum" | |
| top: "stage3_2/sum" | |
| } | |
| layer { | |
| name: "stage3_3/conv1" | |
| type: "Convolution" | |
| bottom: "stage3_2/sum" | |
| top: "stage3_3/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 16 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage3_3/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage3_3/conv1" | |
| top: "stage3_3/conv1" | |
| } | |
| layer { | |
| name: "stage3_3/conv2" | |
| type: "Convolution" | |
| bottom: "stage3_3/conv1" | |
| top: "stage3_3/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 16 | |
| pad: 1 | |
| kernel_size: 3 | |
| group: 16 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage3_3/conv3" | |
| type: "Convolution" | |
| bottom: "stage3_3/conv2" | |
| top: "stage3_3/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 64 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage3_3/sum" | |
| type: "Eltwise" | |
| bottom: "stage3_2/sum" | |
| bottom: "stage3_3/conv3" | |
| top: "stage3_3/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage3_3/relu" | |
| type: "ReLU" | |
| bottom: "stage3_3/sum" | |
| top: "stage3_3/sum" | |
| } | |
| layer { | |
| name: "stage3_4/conv1" | |
| type: "Convolution" | |
| bottom: "stage3_3/sum" | |
| top: "stage3_4/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 16 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage3_4/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage3_4/conv1" | |
| top: "stage3_4/conv1" | |
| } | |
| layer { | |
| name: "stage3_4/conv2" | |
| type: "Convolution" | |
| bottom: "stage3_4/conv1" | |
| top: "stage3_4/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 16 | |
| pad: 1 | |
| kernel_size: 3 | |
| group: 16 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage3_4/conv3" | |
| type: "Convolution" | |
| bottom: "stage3_4/conv2" | |
| top: "stage3_4/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 64 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage3_4/sum" | |
| type: "Eltwise" | |
| bottom: "stage3_3/sum" | |
| bottom: "stage3_4/conv3" | |
| top: "stage3_4/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage3_4/relu" | |
| type: "ReLU" | |
| bottom: "stage3_4/sum" | |
| top: "stage3_4/sum" | |
| } | |
| layer { | |
| name: "stage4_1/conv1" | |
| type: "Convolution" | |
| bottom: "stage3_4/sum" | |
| top: "stage4_1/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_1/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage4_1/conv1" | |
| top: "stage4_1/conv1" | |
| } | |
| layer { | |
| name: "stage4_1/conv2" | |
| type: "Convolution" | |
| bottom: "stage4_1/conv1" | |
| top: "stage4_1/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 1 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 2 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_1/conv3" | |
| type: "Convolution" | |
| bottom: "stage4_1/conv2" | |
| top: "stage4_1/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_1/relu" | |
| type: "ReLU" | |
| bottom: "stage4_1/conv3" | |
| top: "stage4_1/conv3" | |
| } | |
| layer { | |
| name: "stage4_2/conv1" | |
| type: "Convolution" | |
| bottom: "stage4_1/conv3" | |
| top: "stage4_2/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_2/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage4_2/conv1" | |
| top: "stage4_2/conv1" | |
| } | |
| layer { | |
| name: "stage4_2/conv2" | |
| type: "Convolution" | |
| bottom: "stage4_2/conv1" | |
| top: "stage4_2/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 1 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_2/conv3" | |
| type: "Convolution" | |
| bottom: "stage4_2/conv2" | |
| top: "stage4_2/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_2/sum" | |
| type: "Eltwise" | |
| bottom: "stage4_1/conv3" | |
| bottom: "stage4_2/conv3" | |
| top: "stage4_2/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage4_2/relu" | |
| type: "ReLU" | |
| bottom: "stage4_2/sum" | |
| top: "stage4_2/sum" | |
| } | |
| layer { | |
| name: "stage4_3/conv1" | |
| type: "Convolution" | |
| bottom: "stage4_2/sum" | |
| top: "stage4_3/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_3/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage4_3/conv1" | |
| top: "stage4_3/conv1" | |
| } | |
| layer { | |
| name: "stage4_3/conv2" | |
| type: "Convolution" | |
| bottom: "stage4_3/conv1" | |
| top: "stage4_3/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 1 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_3/conv3" | |
| type: "Convolution" | |
| bottom: "stage4_3/conv2" | |
| top: "stage4_3/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_3/sum" | |
| type: "Eltwise" | |
| bottom: "stage4_2/sum" | |
| bottom: "stage4_3/conv3" | |
| top: "stage4_3/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage4_3/relu" | |
| type: "ReLU" | |
| bottom: "stage4_3/sum" | |
| top: "stage4_3/sum" | |
| } | |
| layer { | |
| name: "stage4_4/conv1" | |
| type: "Convolution" | |
| bottom: "stage4_3/sum" | |
| top: "stage4_4/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_4/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage4_4/conv1" | |
| top: "stage4_4/conv1" | |
| } | |
| layer { | |
| name: "stage4_4/conv2" | |
| type: "Convolution" | |
| bottom: "stage4_4/conv1" | |
| top: "stage4_4/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 1 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_4/conv3" | |
| type: "Convolution" | |
| bottom: "stage4_4/conv2" | |
| top: "stage4_4/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_4/sum" | |
| type: "Eltwise" | |
| bottom: "stage4_3/sum" | |
| bottom: "stage4_4/conv3" | |
| top: "stage4_4/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage4_4/relu" | |
| type: "ReLU" | |
| bottom: "stage4_4/sum" | |
| top: "stage4_4/sum" | |
| } | |
| layer { | |
| name: "stage4_5/conv1" | |
| type: "Convolution" | |
| bottom: "stage4_4/sum" | |
| top: "stage4_5/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_5/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage4_5/conv1" | |
| top: "stage4_5/conv1" | |
| } | |
| layer { | |
| name: "stage4_5/conv2" | |
| type: "Convolution" | |
| bottom: "stage4_5/conv1" | |
| top: "stage4_5/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 1 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_5/conv3" | |
| type: "Convolution" | |
| bottom: "stage4_5/conv2" | |
| top: "stage4_5/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_5/sum" | |
| type: "Eltwise" | |
| bottom: "stage4_4/sum" | |
| bottom: "stage4_5/conv3" | |
| top: "stage4_5/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage4_5/relu" | |
| type: "ReLU" | |
| bottom: "stage4_5/sum" | |
| top: "stage4_5/sum" | |
| } | |
| layer { | |
| name: "stage4_6/conv1" | |
| type: "Convolution" | |
| bottom: "stage4_5/sum" | |
| top: "stage4_6/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_6/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage4_6/conv1" | |
| top: "stage4_6/conv1" | |
| } | |
| layer { | |
| name: "stage4_6/conv2" | |
| type: "Convolution" | |
| bottom: "stage4_6/conv1" | |
| top: "stage4_6/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 1 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_6/conv3" | |
| type: "Convolution" | |
| bottom: "stage4_6/conv2" | |
| top: "stage4_6/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_6/sum" | |
| type: "Eltwise" | |
| bottom: "stage4_5/sum" | |
| bottom: "stage4_6/conv3" | |
| top: "stage4_6/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage4_6/relu" | |
| type: "ReLU" | |
| bottom: "stage4_6/sum" | |
| top: "stage4_6/sum" | |
| } | |
| layer { | |
| name: "stage4_7/conv1" | |
| type: "Convolution" | |
| bottom: "stage4_6/sum" | |
| top: "stage4_7/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_7/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage4_7/conv1" | |
| top: "stage4_7/conv1" | |
| } | |
| layer { | |
| name: "stage4_7/conv2" | |
| type: "Convolution" | |
| bottom: "stage4_7/conv1" | |
| top: "stage4_7/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 1 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_7/conv3" | |
| type: "Convolution" | |
| bottom: "stage4_7/conv2" | |
| top: "stage4_7/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_7/sum" | |
| type: "Eltwise" | |
| bottom: "stage4_6/sum" | |
| bottom: "stage4_7/conv3" | |
| top: "stage4_7/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage4_7/relu" | |
| type: "ReLU" | |
| bottom: "stage4_7/sum" | |
| top: "stage4_7/sum" | |
| } | |
| layer { | |
| name: "stage4_8/conv1" | |
| type: "Convolution" | |
| bottom: "stage4_7/sum" | |
| top: "stage4_8/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_8/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage4_8/conv1" | |
| top: "stage4_8/conv1" | |
| } | |
| layer { | |
| name: "stage4_8/conv2" | |
| type: "Convolution" | |
| bottom: "stage4_8/conv1" | |
| top: "stage4_8/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 1 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_8/conv3" | |
| type: "Convolution" | |
| bottom: "stage4_8/conv2" | |
| top: "stage4_8/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_8/sum" | |
| type: "Eltwise" | |
| bottom: "stage4_7/sum" | |
| bottom: "stage4_8/conv3" | |
| top: "stage4_8/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage4_8/relu" | |
| type: "ReLU" | |
| bottom: "stage4_8/sum" | |
| top: "stage4_8/sum" | |
| } | |
| layer { | |
| name: "stage5_1/conv1" | |
| type: "Convolution" | |
| bottom: "stage4_8/sum" | |
| top: "stage5_1/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage5_1/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage5_1/conv1" | |
| top: "stage5_1/conv1" | |
| } | |
| layer { | |
| name: "stage5_1/conv2" | |
| type: "Convolution" | |
| bottom: "stage5_1/conv1" | |
| top: "stage5_1/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 2 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 2 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 2 | |
| } | |
| } | |
| layer { | |
| name: "stage5_1/conv3" | |
| type: "Convolution" | |
| bottom: "stage5_1/conv2" | |
| top: "stage5_1/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage5_1/relu" | |
| type: "ReLU" | |
| bottom: "stage5_1/conv3" | |
| top: "stage5_1/conv3" | |
| } | |
| layer { | |
| name: "stage5_2/conv1" | |
| type: "Convolution" | |
| bottom: "stage5_1/conv3" | |
| top: "stage5_2/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage5_2/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage5_2/conv1" | |
| top: "stage5_2/conv1" | |
| } | |
| layer { | |
| name: "stage5_2/conv2" | |
| type: "Convolution" | |
| bottom: "stage5_2/conv1" | |
| top: "stage5_2/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 2 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 2 | |
| } | |
| } | |
| layer { | |
| name: "stage5_2/conv3" | |
| type: "Convolution" | |
| bottom: "stage5_2/conv2" | |
| top: "stage5_2/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage5_2/sum" | |
| type: "Eltwise" | |
| bottom: "stage5_1/conv3" | |
| bottom: "stage5_2/conv3" | |
| top: "stage5_2/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage5_2/relu" | |
| type: "ReLU" | |
| bottom: "stage5_2/sum" | |
| top: "stage5_2/sum" | |
| } | |
| layer { | |
| name: "stage5_3/conv1" | |
| type: "Convolution" | |
| bottom: "stage5_2/sum" | |
| top: "stage5_3/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage5_3/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage5_3/conv1" | |
| top: "stage5_3/conv1" | |
| } | |
| layer { | |
| name: "stage5_3/conv2" | |
| type: "Convolution" | |
| bottom: "stage5_3/conv1" | |
| top: "stage5_3/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 2 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 2 | |
| } | |
| } | |
| layer { | |
| name: "stage5_3/conv3" | |
| type: "Convolution" | |
| bottom: "stage5_3/conv2" | |
| top: "stage5_3/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage5_3/sum" | |
| type: "Eltwise" | |
| bottom: "stage5_2/sum" | |
| bottom: "stage5_3/conv3" | |
| top: "stage5_3/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage5_3/relu" | |
| type: "ReLU" | |
| bottom: "stage5_3/sum" | |
| top: "stage5_3/sum" | |
| } | |
| layer { | |
| name: "stage5_4/conv1" | |
| type: "Convolution" | |
| bottom: "stage5_3/sum" | |
| top: "stage5_4/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage5_4/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage5_4/conv1" | |
| top: "stage5_4/conv1" | |
| } | |
| layer { | |
| name: "stage5_4/conv2" | |
| type: "Convolution" | |
| bottom: "stage5_4/conv1" | |
| top: "stage5_4/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 2 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 2 | |
| } | |
| } | |
| layer { | |
| name: "stage5_4/conv3" | |
| type: "Convolution" | |
| bottom: "stage5_4/conv2" | |
| top: "stage5_4/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage5_4/sum" | |
| type: "Eltwise" | |
| bottom: "stage5_3/sum" | |
| bottom: "stage5_4/conv3" | |
| top: "stage5_4/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage5_4/relu" | |
| type: "ReLU" | |
| bottom: "stage5_4/sum" | |
| top: "stage5_4/sum" | |
| } | |
| layer { | |
| name: "stage6_1/conv4" | |
| type: "Convolution" | |
| bottom: "stage5_4/sum" | |
| top: "stage6_1/conv4" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage6_1/conv1" | |
| type: "Convolution" | |
| bottom: "stage5_4/sum" | |
| top: "stage6_1/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage6_1/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage6_1/conv1" | |
| top: "stage6_1/conv1" | |
| } | |
| layer { | |
| name: "stage6_1/conv2" | |
| type: "Convolution" | |
| bottom: "stage6_1/conv1" | |
| top: "stage6_1/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 2 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 2 | |
| } | |
| } | |
| layer { | |
| name: "stage6_1/conv3" | |
| type: "Convolution" | |
| bottom: "stage6_1/conv2" | |
| top: "stage6_1/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage6_1/sum" | |
| type: "Eltwise" | |
| bottom: "stage6_1/conv4" | |
| bottom: "stage6_1/conv3" | |
| top: "stage6_1/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage6_1/relu" | |
| type: "ReLU" | |
| bottom: "stage6_1/sum" | |
| top: "stage6_1/sum" | |
| } | |
| layer { | |
| name: "stage6_2/conv1" | |
| type: "Convolution" | |
| bottom: "stage6_1/sum" | |
| top: "stage6_2/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage6_2/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage6_2/conv1" | |
| top: "stage6_2/conv1" | |
| } | |
| layer { | |
| name: "stage6_2/conv2" | |
| type: "Convolution" | |
| bottom: "stage6_2/conv1" | |
| top: "stage6_2/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 2 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 2 | |
| } | |
| } | |
| layer { | |
| name: "stage6_2/conv3" | |
| type: "Convolution" | |
| bottom: "stage6_2/conv2" | |
| top: "stage6_2/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage6_2/sum" | |
| type: "Eltwise" | |
| bottom: "stage6_1/sum" | |
| bottom: "stage6_2/conv3" | |
| top: "stage6_2/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage6_2/relu" | |
| type: "ReLU" | |
| bottom: "stage6_2/sum" | |
| top: "stage6_2/sum" | |
| } | |
| layer { | |
| name: "stage7_1/conv4" | |
| type: "Convolution" | |
| bottom: "stage6_2/sum" | |
| top: "stage7_1/conv4" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage7_1/conv1" | |
| type: "Convolution" | |
| bottom: "stage6_2/sum" | |
| top: "stage7_1/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage7_1/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage7_1/conv1" | |
| top: "stage7_1/conv1" | |
| } | |
| layer { | |
| name: "stage7_1/conv2" | |
| type: "Convolution" | |
| bottom: "stage7_1/conv1" | |
| top: "stage7_1/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 2 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 2 | |
| } | |
| } | |
| layer { | |
| name: "stage7_1/conv3" | |
| type: "Convolution" | |
| bottom: "stage7_1/conv2" | |
| top: "stage7_1/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage7_1/sum" | |
| type: "Eltwise" | |
| bottom: "stage7_1/conv4" | |
| bottom: "stage7_1/conv3" | |
| top: "stage7_1/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage7_1/relu" | |
| type: "ReLU" | |
| bottom: "stage7_1/sum" | |
| top: "stage7_1/sum" | |
| } | |
| layer { | |
| name: "stage7_2/conv1" | |
| type: "Convolution" | |
| bottom: "stage7_1/sum" | |
| top: "stage7_2/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage7_2/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage7_2/conv1" | |
| top: "stage7_2/conv1" | |
| } | |
| layer { | |
| name: "stage7_2/conv2" | |
| type: "Convolution" | |
| bottom: "stage7_2/conv1" | |
| top: "stage7_2/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 2 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 2 | |
| } | |
| } | |
| layer { | |
| name: "stage7_2/conv3" | |
| type: "Convolution" | |
| bottom: "stage7_2/conv2" | |
| top: "stage7_2/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage7_2/sum" | |
| type: "Eltwise" | |
| bottom: "stage7_1/sum" | |
| bottom: "stage7_2/conv3" | |
| top: "stage7_2/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage7_2/relu" | |
| type: "ReLU" | |
| bottom: "stage7_2/sum" | |
| top: "stage7_2/sum" | |
| } | |
| layer { | |
| name: "stage8_1/conv4" | |
| type: "Convolution" | |
| bottom: "stage7_2/sum" | |
| top: "stage8_1/conv4" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage8_1/conv1" | |
| type: "Convolution" | |
| bottom: "stage7_2/sum" | |
| top: "stage8_1/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage8_1/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage8_1/conv1" | |
| top: "stage8_1/conv1" | |
| } | |
| layer { | |
| name: "stage8_1/conv2" | |
| type: "Convolution" | |
| bottom: "stage8_1/conv1" | |
| top: "stage8_1/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 2 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 2 | |
| } | |
| } | |
| layer { | |
| name: "stage8_1/conv3" | |
| type: "Convolution" | |
| bottom: "stage8_1/conv2" | |
| top: "stage8_1/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage8_1/sum" | |
| type: "Eltwise" | |
| bottom: "stage8_1/conv4" | |
| bottom: "stage8_1/conv3" | |
| top: "stage8_1/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage8_1/relu" | |
| type: "ReLU" | |
| bottom: "stage8_1/sum" | |
| top: "stage8_1/sum" | |
| } | |
| layer { | |
| name: "stage8_2/conv1" | |
| type: "Convolution" | |
| bottom: "stage8_1/sum" | |
| top: "stage8_2/conv1" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage8_2/conv1/relu" | |
| type: "ReLU" | |
| bottom: "stage8_2/conv1" | |
| top: "stage8_2/conv1" | |
| } | |
| layer { | |
| name: "stage8_2/conv2" | |
| type: "Convolution" | |
| bottom: "stage8_2/conv1" | |
| top: "stage8_2/conv2" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 32 | |
| pad: 2 | |
| kernel_size: 3 | |
| group: 32 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 2 | |
| } | |
| } | |
| layer { | |
| name: "stage8_2/conv3" | |
| type: "Convolution" | |
| bottom: "stage8_2/conv2" | |
| top: "stage8_2/conv3" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| convolution_param { | |
| num_output: 128 | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage8_2/sum" | |
| type: "Eltwise" | |
| bottom: "stage8_1/sum" | |
| bottom: "stage8_2/conv3" | |
| top: "stage8_2/sum" | |
| eltwise_param { | |
| operation: SUM | |
| } | |
| } | |
| layer { | |
| name: "stage8_2/relu" | |
| type: "ReLU" | |
| bottom: "stage8_2/sum" | |
| top: "stage8_2/sum" | |
| } | |
| layer { | |
| name: "cls1/conv" | |
| type: "Convolution" | |
| bottom: "stage4_8/sum" | |
| top: "cls1/conv" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 0.0 | |
| } | |
| convolution_param { | |
| num_output: 12 | |
| bias_term: true | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "cls1/permute" | |
| type: "Permute" | |
| bottom: "cls1/conv" | |
| top: "cls1/permute" | |
| permute_param { | |
| order: 0 | |
| order: 2 | |
| order: 3 | |
| order: 1 | |
| } | |
| } | |
| layer { | |
| name: "cls1/flatten" | |
| type: "Flatten" | |
| bottom: "cls1/permute" | |
| top: "cls1/flatten" | |
| flatten_param { | |
| axis: 1 | |
| } | |
| } | |
| layer { | |
| name: "loc1/conv" | |
| type: "Convolution" | |
| bottom: "stage4_8/sum" | |
| top: "loc1/conv" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 0.0 | |
| } | |
| convolution_param { | |
| num_output: 24 | |
| bias_term: true | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "loc1/permute" | |
| type: "Permute" | |
| bottom: "loc1/conv" | |
| top: "loc1/permute" | |
| permute_param { | |
| order: 0 | |
| order: 2 | |
| order: 3 | |
| order: 1 | |
| } | |
| } | |
| layer { | |
| name: "loc1/flatten" | |
| type: "Flatten" | |
| bottom: "loc1/permute" | |
| top: "loc1/flatten" | |
| flatten_param { | |
| axis: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage4_8/sum/prior_box" | |
| type: "PriorBox" | |
| bottom: "stage4_8/sum" | |
| bottom: "data" | |
| top: "stage4_8/sum/prior_box" | |
| prior_box_param { | |
| min_size: 50.0 | |
| max_size: 100.0 | |
| aspect_ratio: 2.0 | |
| aspect_ratio: 0.5 | |
| aspect_ratio: 3.0 | |
| aspect_ratio: 0.3333333432674408 | |
| flip: false | |
| clip: false | |
| variance: 0.10000000149011612 | |
| variance: 0.10000000149011612 | |
| variance: 0.20000000298023224 | |
| variance: 0.20000000298023224 | |
| step: 16.0 | |
| } | |
| } | |
| layer { | |
| name: "cls2/conv" | |
| type: "Convolution" | |
| bottom: "stage5_4/sum" | |
| top: "cls2/conv" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 0.0 | |
| } | |
| convolution_param { | |
| num_output: 12 | |
| bias_term: true | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "cls2/permute" | |
| type: "Permute" | |
| bottom: "cls2/conv" | |
| top: "cls2/permute" | |
| permute_param { | |
| order: 0 | |
| order: 2 | |
| order: 3 | |
| order: 1 | |
| } | |
| } | |
| layer { | |
| name: "cls2/flatten" | |
| type: "Flatten" | |
| bottom: "cls2/permute" | |
| top: "cls2/flatten" | |
| flatten_param { | |
| axis: 1 | |
| } | |
| } | |
| layer { | |
| name: "loc2/conv" | |
| type: "Convolution" | |
| bottom: "stage5_4/sum" | |
| top: "loc2/conv" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 0.0 | |
| } | |
| convolution_param { | |
| num_output: 24 | |
| bias_term: true | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "loc2/permute" | |
| type: "Permute" | |
| bottom: "loc2/conv" | |
| top: "loc2/permute" | |
| permute_param { | |
| order: 0 | |
| order: 2 | |
| order: 3 | |
| order: 1 | |
| } | |
| } | |
| layer { | |
| name: "loc2/flatten" | |
| type: "Flatten" | |
| bottom: "loc2/permute" | |
| top: "loc2/flatten" | |
| flatten_param { | |
| axis: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage5_4/sum/prior_box" | |
| type: "PriorBox" | |
| bottom: "stage5_4/sum" | |
| bottom: "data" | |
| top: "stage5_4/sum/prior_box" | |
| prior_box_param { | |
| min_size: 100.0 | |
| max_size: 150.0 | |
| aspect_ratio: 2.0 | |
| aspect_ratio: 0.5 | |
| aspect_ratio: 3.0 | |
| aspect_ratio: 0.3333333432674408 | |
| flip: false | |
| clip: false | |
| variance: 0.10000000149011612 | |
| variance: 0.10000000149011612 | |
| variance: 0.20000000298023224 | |
| variance: 0.20000000298023224 | |
| step: 32.0 | |
| } | |
| } | |
| layer { | |
| name: "cls3/conv" | |
| type: "Convolution" | |
| bottom: "stage6_2/sum" | |
| top: "cls3/conv" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 0.0 | |
| } | |
| convolution_param { | |
| num_output: 12 | |
| bias_term: true | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "cls3/permute" | |
| type: "Permute" | |
| bottom: "cls3/conv" | |
| top: "cls3/permute" | |
| permute_param { | |
| order: 0 | |
| order: 2 | |
| order: 3 | |
| order: 1 | |
| } | |
| } | |
| layer { | |
| name: "cls3/flatten" | |
| type: "Flatten" | |
| bottom: "cls3/permute" | |
| top: "cls3/flatten" | |
| flatten_param { | |
| axis: 1 | |
| } | |
| } | |
| layer { | |
| name: "loc3/conv" | |
| type: "Convolution" | |
| bottom: "stage6_2/sum" | |
| top: "loc3/conv" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 0.0 | |
| } | |
| convolution_param { | |
| num_output: 24 | |
| bias_term: true | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "loc3/permute" | |
| type: "Permute" | |
| bottom: "loc3/conv" | |
| top: "loc3/permute" | |
| permute_param { | |
| order: 0 | |
| order: 2 | |
| order: 3 | |
| order: 1 | |
| } | |
| } | |
| layer { | |
| name: "loc3/flatten" | |
| type: "Flatten" | |
| bottom: "loc3/permute" | |
| top: "loc3/flatten" | |
| flatten_param { | |
| axis: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage6_2/sum/prior_box" | |
| type: "PriorBox" | |
| bottom: "stage6_2/sum" | |
| bottom: "data" | |
| top: "stage6_2/sum/prior_box" | |
| prior_box_param { | |
| min_size: 150.0 | |
| max_size: 200.0 | |
| aspect_ratio: 2.0 | |
| aspect_ratio: 0.5 | |
| aspect_ratio: 3.0 | |
| aspect_ratio: 0.3333333432674408 | |
| flip: false | |
| clip: false | |
| variance: 0.10000000149011612 | |
| variance: 0.10000000149011612 | |
| variance: 0.20000000298023224 | |
| variance: 0.20000000298023224 | |
| step: 32.0 | |
| } | |
| } | |
| layer { | |
| name: "cls4/conv" | |
| type: "Convolution" | |
| bottom: "stage7_2/sum" | |
| top: "cls4/conv" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 0.0 | |
| } | |
| convolution_param { | |
| num_output: 12 | |
| bias_term: true | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "cls4/permute" | |
| type: "Permute" | |
| bottom: "cls4/conv" | |
| top: "cls4/permute" | |
| permute_param { | |
| order: 0 | |
| order: 2 | |
| order: 3 | |
| order: 1 | |
| } | |
| } | |
| layer { | |
| name: "cls4/flatten" | |
| type: "Flatten" | |
| bottom: "cls4/permute" | |
| top: "cls4/flatten" | |
| flatten_param { | |
| axis: 1 | |
| } | |
| } | |
| layer { | |
| name: "loc4/conv" | |
| type: "Convolution" | |
| bottom: "stage7_2/sum" | |
| top: "loc4/conv" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 0.0 | |
| } | |
| convolution_param { | |
| num_output: 24 | |
| bias_term: true | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "loc4/permute" | |
| type: "Permute" | |
| bottom: "loc4/conv" | |
| top: "loc4/permute" | |
| permute_param { | |
| order: 0 | |
| order: 2 | |
| order: 3 | |
| order: 1 | |
| } | |
| } | |
| layer { | |
| name: "loc4/flatten" | |
| type: "Flatten" | |
| bottom: "loc4/permute" | |
| top: "loc4/flatten" | |
| flatten_param { | |
| axis: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage7_2/sum/prior_box" | |
| type: "PriorBox" | |
| bottom: "stage7_2/sum" | |
| bottom: "data" | |
| top: "stage7_2/sum/prior_box" | |
| prior_box_param { | |
| min_size: 200.0 | |
| max_size: 300.0 | |
| aspect_ratio: 2.0 | |
| aspect_ratio: 0.5 | |
| aspect_ratio: 3.0 | |
| aspect_ratio: 0.3333333432674408 | |
| flip: false | |
| clip: false | |
| variance: 0.10000000149011612 | |
| variance: 0.10000000149011612 | |
| variance: 0.20000000298023224 | |
| variance: 0.20000000298023224 | |
| step: 32.0 | |
| } | |
| } | |
| layer { | |
| name: "cls5/conv" | |
| type: "Convolution" | |
| bottom: "stage8_2/sum" | |
| top: "cls5/conv" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 0.0 | |
| } | |
| convolution_param { | |
| num_output: 12 | |
| bias_term: true | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "cls5/permute" | |
| type: "Permute" | |
| bottom: "cls5/conv" | |
| top: "cls5/permute" | |
| permute_param { | |
| order: 0 | |
| order: 2 | |
| order: 3 | |
| order: 1 | |
| } | |
| } | |
| layer { | |
| name: "cls5/flatten" | |
| type: "Flatten" | |
| bottom: "cls5/permute" | |
| top: "cls5/flatten" | |
| flatten_param { | |
| axis: 1 | |
| } | |
| } | |
| layer { | |
| name: "loc5/conv" | |
| type: "Convolution" | |
| bottom: "stage8_2/sum" | |
| top: "loc5/conv" | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 1.0 | |
| } | |
| param { | |
| lr_mult: 1.0 | |
| decay_mult: 0.0 | |
| } | |
| convolution_param { | |
| num_output: 24 | |
| bias_term: true | |
| pad: 0 | |
| kernel_size: 1 | |
| group: 1 | |
| stride: 1 | |
| weight_filler { | |
| type: "msra" | |
| } | |
| dilation: 1 | |
| } | |
| } | |
| layer { | |
| name: "loc5/permute" | |
| type: "Permute" | |
| bottom: "loc5/conv" | |
| top: "loc5/permute" | |
| permute_param { | |
| order: 0 | |
| order: 2 | |
| order: 3 | |
| order: 1 | |
| } | |
| } | |
| layer { | |
| name: "loc5/flatten" | |
| type: "Flatten" | |
| bottom: "loc5/permute" | |
| top: "loc5/flatten" | |
| flatten_param { | |
| axis: 1 | |
| } | |
| } | |
| layer { | |
| name: "stage8_2/sum/prior_box" | |
| type: "PriorBox" | |
| bottom: "stage8_2/sum" | |
| bottom: "data" | |
| top: "stage8_2/sum/prior_box" | |
| prior_box_param { | |
| min_size: 300.0 | |
| max_size: 400.0 | |
| aspect_ratio: 2.0 | |
| aspect_ratio: 0.5 | |
| aspect_ratio: 3.0 | |
| aspect_ratio: 0.3333333432674408 | |
| flip: false | |
| clip: false | |
| variance: 0.10000000149011612 | |
| variance: 0.10000000149011612 | |
| variance: 0.20000000298023224 | |
| variance: 0.20000000298023224 | |
| step: 32.0 | |
| } | |
| } | |
| layer { | |
| name: "mbox_conf" | |
| type: "Concat" | |
| bottom: "cls1/flatten" | |
| bottom: "cls2/flatten" | |
| bottom: "cls3/flatten" | |
| bottom: "cls4/flatten" | |
| bottom: "cls5/flatten" | |
| top: "mbox_conf" | |
| concat_param { | |
| axis: 1 | |
| } | |
| } | |
| layer { | |
| name: "mbox_loc" | |
| type: "Concat" | |
| bottom: "loc1/flatten" | |
| bottom: "loc2/flatten" | |
| bottom: "loc3/flatten" | |
| bottom: "loc4/flatten" | |
| bottom: "loc5/flatten" | |
| top: "mbox_loc" | |
| concat_param { | |
| axis: 1 | |
| } | |
| } | |
| layer { | |
| name: "mbox_priorbox" | |
| type: "Concat" | |
| bottom: "stage4_8/sum/prior_box" | |
| bottom: "stage5_4/sum/prior_box" | |
| bottom: "stage6_2/sum/prior_box" | |
| bottom: "stage7_2/sum/prior_box" | |
| bottom: "stage8_2/sum/prior_box" | |
| top: "mbox_priorbox" | |
| concat_param { | |
| axis: 2 | |
| } | |
| } | |
| layer { | |
| name: "mbox_conf_reshape" | |
| type: "Reshape" | |
| bottom: "mbox_conf" | |
| top: "mbox_conf_reshape" | |
| reshape_param { | |
| shape { | |
| dim: 0 | |
| dim: -1 | |
| dim: 2 | |
| } | |
| } | |
| } | |
| layer { | |
| name: "mbox_conf_softmax" | |
| type: "Softmax" | |
| bottom: "mbox_conf_reshape" | |
| top: "mbox_conf_softmax" | |
| softmax_param { | |
| axis: 2 | |
| } | |
| } | |
| layer { | |
| name: "mbox_conf_flatten" | |
| type: "Flatten" | |
| bottom: "mbox_conf_softmax" | |
| top: "mbox_conf_flatten" | |
| flatten_param { | |
| axis: 1 | |
| } | |
| } | |
| layer { | |
| name: "detection_output" | |
| type: "DetectionOutput" | |
| bottom: "mbox_loc" | |
| bottom: "mbox_conf_flatten" | |
| bottom: "mbox_priorbox" | |
| top: "detection_output" | |
| detection_output_param { | |
| num_classes: 2 | |
| share_location: true | |
| background_label_id: 0 | |
| nms_param { | |
| nms_threshold: 0.44999998807907104 | |
| top_k: 100 | |
| } | |
| code_type: CENTER_SIZE | |
| keep_top_k: 100 | |
| confidence_threshold: 0.20000000298023224 | |
| } | |
| } | |