From 52e3ab66e1ade3a0e45f82e7d2e52a40f1493ca9 Mon Sep 17 00:00:00 2001 From: Bert Peters Date: Mon, 12 Mar 2018 13:40:13 +0100 Subject: [PATCH] Add a deduplicated version of the caffenet model. --- data/models/caffenet/model-dedup.prototxt | 220 ++++++++++++++++++++++ 1 file changed, 220 insertions(+) create mode 100644 data/models/caffenet/model-dedup.prototxt diff --git a/data/models/caffenet/model-dedup.prototxt b/data/models/caffenet/model-dedup.prototxt new file mode 100644 index 0000000..93aa4dd --- /dev/null +++ b/data/models/caffenet/model-dedup.prototxt @@ -0,0 +1,220 @@ +name: "CaffeNet" +layer { + name: "data" + type: "Input" + top: "data" + input_param { + shape { + dim: 10 + dim: 3 + dim: 227 + dim: 227 + } + } +} +layer { + name: "conv1" + type: "Convolution" + bottom: "data" + top: "conv1" + convolution_param { + num_output: 96 + kernel_size: 11 + stride: 4 + } +} +layer { + name: "relu1" + type: "ReLU" + bottom: "conv1" + top: "relu1" +} +layer { + name: "pool1" + type: "Pooling" + bottom: "relu1" + top: "pool1" + pooling_param { + pool: MAX + kernel_size: 3 + stride: 2 + } +} +layer { + name: "norm1" + type: "LRN" + bottom: "pool1" + top: "norm1" + lrn_param { + local_size: 5 + alpha: 0.0001 + beta: 0.75 + } +} +layer { + name: "conv2" + type: "Convolution" + bottom: "norm1" + top: "conv2" + convolution_param { + num_output: 256 + pad: 2 + kernel_size: 5 + group: 2 + } +} +layer { + name: "relu2" + type: "ReLU" + bottom: "conv2" + top: "relu2" +} +layer { + name: "pool2" + type: "Pooling" + bottom: "relu2" + top: "pool2" + pooling_param { + pool: MAX + kernel_size: 3 + stride: 2 + } +} +layer { + name: "norm2" + type: "LRN" + bottom: "pool2" + top: "norm2" + lrn_param { + local_size: 5 + alpha: 0.0001 + beta: 0.75 + } +} +layer { + name: "conv3" + type: "Convolution" + bottom: "norm2" + top: "conv3" + convolution_param { + num_output: 384 + pad: 1 + kernel_size: 3 + } +} +layer { + name: "relu3" + type: "ReLU" + bottom: "conv3" + top: "relu3" +} +layer { + name: "conv4" + type: "Convolution" + bottom: "relu3" + top: "conv4" + convolution_param { + num_output: 384 + pad: 1 + kernel_size: 3 + group: 2 + } +} +layer { + name: "relu4" + type: "ReLU" + bottom: "conv4" + top: "relu4" +} +layer { + name: "conv5" + type: "Convolution" + bottom: "relu4" + top: "conv5" + convolution_param { + num_output: 256 + pad: 1 + kernel_size: 3 + group: 2 + } +} +layer { + name: "relu5" + type: "ReLU" + bottom: "conv5" + top: "relu5" +} +layer { + name: "pool5" + type: "Pooling" + bottom: "relu5" + top: "pool5" + pooling_param { + pool: MAX + kernel_size: 3 + stride: 2 + } +} +layer { + name: "fc6" + type: "InnerProduct" + bottom: "pool5" + top: "fc6" + inner_product_param { + num_output: 4096 + } +} +layer { + name: "relu6" + type: "ReLU" + bottom: "fc6" + top: "relu6" +} +layer { + name: "drop6" + type: "Dropout" + bottom: "relu6" + top: "drop6" + dropout_param { + dropout_ratio: 0.5 + } +} +layer { + name: "fc7" + type: "InnerProduct" + bottom: "drop6" + top: "fc7" + inner_product_param { + num_output: 4096 + } +} +layer { + name: "relu7" + type: "ReLU" + bottom: "fc7" + top: "relu7" +} +layer { + name: "drop7" + type: "Dropout" + bottom: "relu7" + top: "drop7" + dropout_param { + dropout_ratio: 0.5 + } +} +layer { + name: "fc8" + type: "InnerProduct" + bottom: "drop7" + top: "fc8" + inner_product_param { + num_output: 1000 + } +} +layer { + name: "prob" + type: "Softmax" + bottom: "fc8" + top: "prob" +}