chibi@1604:~/caffe$ cd python; python classify.py --raw_scale 255 ../101_ObjectCategories/dolphin/image_0055.jpg ../result.npy; cd .. I0311 19:06:37.452085 6289 gpu_memory.cpp:105] GPUMemory::Manager initialized I0311 19:06:37.453488 6289 gpu_memory.cpp:107] Total memory: 25364004864, Free: 24854396928, dev_info[0]: total=25364004864 free=24854396928 CPU mode W0311 19:06:37.453549 6289 _caffe.cpp:172] DEPRECATION WARNING - deprecated use of Python interface W0311 19:06:37.453603 6289 _caffe.cpp:173] Use this instead (with the named "weights" parameter): W0311 19:06:37.453610 6289 _caffe.cpp:175] Net('../models/bvlc_reference_caffenet/deploy.prototxt', 1, weights='../models/bvlc_reference_caffenet/bvlc_reference_caffenet.caffemodel') I0311 19:06:37.455462 6289 net.cpp:83] Initializing net from parameters: name: "CaffeNet" state { phase: TEST level: 0 } layer { name: "data" type: "Input" top: "data" input_param { shape { dim: 10 dim: 3 dim: 227 dim: 227 } } } layer { name: "conv1" type: "Convolution" bottom: "data" top: "conv1" convolution_param { num_output: 96 kernel_size: 11 stride: 4 } } layer { name: "relu1" type: "ReLU" bottom: "conv1" top: "conv1" } layer { name: "pool1" type: "Pooling" bottom: "conv1" top: "pool1" pooling_param { pool: MAX kernel_size: 3 stride: 2 } } layer { name: "norm1" type: "LRN" bottom: "pool1" top: "norm1" lrn_param { local_size: 5 alpha: 0.0001 beta: 0.75 } } layer { name: "conv2" type: "Convolution" bottom: "norm1" top: "conv2" convolution_param { num_output: 256 pad: 2 kernel_size: 5 group: 2 } } layer { name: "relu2" type: "ReLU" bottom: "conv2" top: "conv2" } layer { name: "pool2" type: "Pooling" bottom: "conv2" top: "pool2" pooling_param { pool: MAX kernel_size: 3 stride: 2 } } layer { name: "norm2" type: "LRN" bottom: "pool2" top: "norm2" lrn_param { local_size: 5 alpha: 0.0001 beta: 0.75 } } layer { name: "conv3" type: "Convolution" bottom: "norm2" top: "conv3" convolution_param { num_output: 384 pad: 1 kernel_size: 3 } } layer { name: "relu3" type: "ReLU" bottom: "conv3" top: "conv3" } layer { name: "conv4" type: "Convolution" bottom: "conv3" top: "conv4" convolution_param { num_output: 384 pad: 1 kernel_size: 3 group: 2 } } layer { name: "relu4" type: "ReLU" bottom: "conv4" top: "conv4" } layer { name: "conv5" type: "Convolution" bottom: "conv4" top: "conv5" convolution_param { num_output: 256 pad: 1 kernel_size: 3 group: 2 } } layer { name: "relu5" type: "ReLU" bottom: "conv5" top: "conv5" } layer { name: "pool5" type: "Pooling" bottom: "conv5" top: "pool5" pooling_param { pool: MAX kernel_size: 3 stride: 2 } } layer { name: "fc6" type: "InnerProduct" bottom: "pool5" top: "fc6" inner_product_param { num_output: 4096 } } layer { name: "relu6" type: "ReLU" bottom: "fc6" top: "fc6" } layer { name: "drop6" type: "Dropout" bottom: "fc6" top: "fc6" dropout_param { dropout_ratio: 0.5 } } layer { name: "fc7" type: "InnerProduct" bottom: "fc6" top: "fc7" inner_product_param { num_output: 4096 } } layer { name: "relu7" type: "ReLU" bottom: "fc7" top: "fc7" } layer { name: "drop7" type: "Dropout" bottom: "fc7" top: "fc7" dropout_param { dropout_ratio: 0.5 } } layer { name: "fc8" type: "InnerProduct" bottom: "fc7" top: "fc8" inner_product_param { num_output: 1000 } } layer { name: "prob" type: "Softmax" bottom: "fc8" top: "prob" } I0311 19:06:37.455623 6289 net.cpp:113] Using FLOAT as default forward math type I0311 19:06:37.455631 6289 net.cpp:119] Using FLOAT as default backward math type I0311 19:06:37.455637 6289 layer_factory.hpp:172] Creating layer 'data' of type 'Input' I0311 19:06:37.455643 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.455655 6289 net.cpp:202] Created Layer data (0) I0311 19:06:37.455663 6289 net.cpp:544] data -> data I0311 19:06:37.455683 6289 net.cpp:262] Setting up data I0311 19:06:37.455691 6289 net.cpp:269] TEST Top shape for layer 0 'data' 10 3 227 227 (1545870) I0311 19:06:37.455710 6289 layer_factory.hpp:172] Creating layer 'conv1' of type 'Convolution' I0311 19:06:37.455716 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.455731 6289 net.cpp:202] Created Layer conv1 (1) I0311 19:06:37.455737 6289 net.cpp:574] conv1 <- data I0311 19:06:37.455744 6289 net.cpp:544] conv1 -> conv1 I0311 19:06:37.456061 6289 net.cpp:262] Setting up conv1 I0311 19:06:37.456070 6289 net.cpp:269] TEST Top shape for layer 1 'conv1' 10 96 55 55 (2904000) I0311 19:06:37.456081 6289 layer_factory.hpp:172] Creating layer 'relu1' of type 'ReLU' I0311 19:06:37.456086 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.456094 6289 net.cpp:202] Created Layer relu1 (2) I0311 19:06:37.456099 6289 net.cpp:574] relu1 <- conv1 I0311 19:06:37.456104 6289 net.cpp:529] relu1 -> conv1 (in-place) I0311 19:06:37.456110 6289 net.cpp:262] Setting up relu1 I0311 19:06:37.456116 6289 net.cpp:269] TEST Top shape for layer 2 'relu1' 10 96 55 55 (2904000) I0311 19:06:37.456121 6289 layer_factory.hpp:172] Creating layer 'pool1' of type 'Pooling' I0311 19:06:37.456125 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.456133 6289 net.cpp:202] Created Layer pool1 (3) I0311 19:06:37.456137 6289 net.cpp:574] pool1 <- conv1 I0311 19:06:37.456142 6289 net.cpp:544] pool1 -> pool1 I0311 19:06:37.456153 6289 net.cpp:262] Setting up pool1 I0311 19:06:37.456159 6289 net.cpp:269] TEST Top shape for layer 3 'pool1' 10 96 27 27 (699840) I0311 19:06:37.456164 6289 layer_factory.hpp:172] Creating layer 'norm1' of type 'LRN' I0311 19:06:37.456169 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.456179 6289 net.cpp:202] Created Layer norm1 (4) I0311 19:06:37.456183 6289 net.cpp:574] norm1 <- pool1 I0311 19:06:37.456188 6289 net.cpp:544] norm1 -> norm1 I0311 19:06:37.456198 6289 net.cpp:262] Setting up norm1 I0311 19:06:37.456204 6289 net.cpp:269] TEST Top shape for layer 4 'norm1' 10 96 27 27 (699840) I0311 19:06:37.456209 6289 layer_factory.hpp:172] Creating layer 'conv2' of type 'Convolution' I0311 19:06:37.456213 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.456223 6289 net.cpp:202] Created Layer conv2 (5) I0311 19:06:37.456228 6289 net.cpp:574] conv2 <- norm1 I0311 19:06:37.456233 6289 net.cpp:544] conv2 -> conv2 I0311 19:06:37.458699 6289 net.cpp:262] Setting up conv2 I0311 19:06:37.458709 6289 net.cpp:269] TEST Top shape for layer 5 'conv2' 10 256 27 27 (1866240) I0311 19:06:37.458719 6289 layer_factory.hpp:172] Creating layer 'relu2' of type 'ReLU' I0311 19:06:37.458724 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.458732 6289 net.cpp:202] Created Layer relu2 (6) I0311 19:06:37.458737 6289 net.cpp:574] relu2 <- conv2 I0311 19:06:37.458742 6289 net.cpp:529] relu2 -> conv2 (in-place) I0311 19:06:37.458748 6289 net.cpp:262] Setting up relu2 I0311 19:06:37.458753 6289 net.cpp:269] TEST Top shape for layer 6 'relu2' 10 256 27 27 (1866240) I0311 19:06:37.458758 6289 layer_factory.hpp:172] Creating layer 'pool2' of type 'Pooling' I0311 19:06:37.458763 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.458770 6289 net.cpp:202] Created Layer pool2 (7) I0311 19:06:37.458775 6289 net.cpp:574] pool2 <- conv2 I0311 19:06:37.458781 6289 net.cpp:544] pool2 -> pool2 I0311 19:06:37.458787 6289 net.cpp:262] Setting up pool2 I0311 19:06:37.458792 6289 net.cpp:269] TEST Top shape for layer 7 'pool2' 10 256 13 13 (432640) I0311 19:06:37.458797 6289 layer_factory.hpp:172] Creating layer 'norm2' of type 'LRN' I0311 19:06:37.458801 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.458819 6289 net.cpp:202] Created Layer norm2 (8) I0311 19:06:37.458824 6289 net.cpp:574] norm2 <- pool2 I0311 19:06:37.458830 6289 net.cpp:544] norm2 -> norm2 I0311 19:06:37.458837 6289 net.cpp:262] Setting up norm2 I0311 19:06:37.458843 6289 net.cpp:269] TEST Top shape for layer 8 'norm2' 10 256 13 13 (432640) I0311 19:06:37.458848 6289 layer_factory.hpp:172] Creating layer 'conv3' of type 'Convolution' I0311 19:06:37.458854 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.458863 6289 net.cpp:202] Created Layer conv3 (9) I0311 19:06:37.458868 6289 net.cpp:574] conv3 <- norm2 I0311 19:06:37.458873 6289 net.cpp:544] conv3 -> conv3 I0311 19:06:37.465615 6289 net.cpp:262] Setting up conv3 I0311 19:06:37.465625 6289 net.cpp:269] TEST Top shape for layer 9 'conv3' 10 384 13 13 (648960) I0311 19:06:37.465644 6289 layer_factory.hpp:172] Creating layer 'relu3' of type 'ReLU' I0311 19:06:37.465649 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.465656 6289 net.cpp:202] Created Layer relu3 (10) I0311 19:06:37.465661 6289 net.cpp:574] relu3 <- conv3 I0311 19:06:37.465665 6289 net.cpp:529] relu3 -> conv3 (in-place) I0311 19:06:37.465670 6289 net.cpp:262] Setting up relu3 I0311 19:06:37.465677 6289 net.cpp:269] TEST Top shape for layer 10 'relu3' 10 384 13 13 (648960) I0311 19:06:37.465682 6289 layer_factory.hpp:172] Creating layer 'conv4' of type 'Convolution' I0311 19:06:37.465687 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.465696 6289 net.cpp:202] Created Layer conv4 (11) I0311 19:06:37.465701 6289 net.cpp:574] conv4 <- conv3 I0311 19:06:37.465706 6289 net.cpp:544] conv4 -> conv4 I0311 19:06:37.470995 6289 net.cpp:262] Setting up conv4 I0311 19:06:37.471005 6289 net.cpp:269] TEST Top shape for layer 11 'conv4' 10 384 13 13 (648960) I0311 19:06:37.471014 6289 layer_factory.hpp:172] Creating layer 'relu4' of type 'ReLU' I0311 19:06:37.471019 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.471024 6289 net.cpp:202] Created Layer relu4 (12) I0311 19:06:37.471029 6289 net.cpp:574] relu4 <- conv4 I0311 19:06:37.471035 6289 net.cpp:529] relu4 -> conv4 (in-place) I0311 19:06:37.471040 6289 net.cpp:262] Setting up relu4 I0311 19:06:37.471047 6289 net.cpp:269] TEST Top shape for layer 12 'relu4' 10 384 13 13 (648960) I0311 19:06:37.471052 6289 layer_factory.hpp:172] Creating layer 'conv5' of type 'Convolution' I0311 19:06:37.471058 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.471069 6289 net.cpp:202] Created Layer conv5 (13) I0311 19:06:37.471074 6289 net.cpp:574] conv5 <- conv4 I0311 19:06:37.471079 6289 net.cpp:544] conv5 -> conv5 I0311 19:06:37.474608 6289 net.cpp:262] Setting up conv5 I0311 19:06:37.474617 6289 net.cpp:269] TEST Top shape for layer 13 'conv5' 10 256 13 13 (432640) I0311 19:06:37.474627 6289 layer_factory.hpp:172] Creating layer 'relu5' of type 'ReLU' I0311 19:06:37.474632 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.474637 6289 net.cpp:202] Created Layer relu5 (14) I0311 19:06:37.474642 6289 net.cpp:574] relu5 <- conv5 I0311 19:06:37.474647 6289 net.cpp:529] relu5 -> conv5 (in-place) I0311 19:06:37.474653 6289 net.cpp:262] Setting up relu5 I0311 19:06:37.474658 6289 net.cpp:269] TEST Top shape for layer 14 'relu5' 10 256 13 13 (432640) I0311 19:06:37.474664 6289 layer_factory.hpp:172] Creating layer 'pool5' of type 'Pooling' I0311 19:06:37.474668 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.474678 6289 net.cpp:202] Created Layer pool5 (15) I0311 19:06:37.474683 6289 net.cpp:574] pool5 <- conv5 I0311 19:06:37.474686 6289 net.cpp:544] pool5 -> pool5 I0311 19:06:37.474696 6289 net.cpp:262] Setting up pool5 I0311 19:06:37.474701 6289 net.cpp:269] TEST Top shape for layer 15 'pool5' 10 256 6 6 (92160) I0311 19:06:37.474716 6289 layer_factory.hpp:172] Creating layer 'fc6' of type 'InnerProduct' I0311 19:06:37.474722 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.474731 6289 net.cpp:202] Created Layer fc6 (16) I0311 19:06:37.474737 6289 net.cpp:574] fc6 <- pool5 I0311 19:06:37.474742 6289 net.cpp:544] fc6 -> fc6 I0311 19:06:37.775234 6289 net.cpp:262] Setting up fc6 I0311 19:06:37.775267 6289 net.cpp:269] TEST Top shape for layer 16 'fc6' 10 4096 (40960) I0311 19:06:37.775280 6289 layer_factory.hpp:172] Creating layer 'relu6' of type 'ReLU' I0311 19:06:37.775287 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.775300 6289 net.cpp:202] Created Layer relu6 (17) I0311 19:06:37.775305 6289 net.cpp:574] relu6 <- fc6 I0311 19:06:37.775312 6289 net.cpp:529] relu6 -> fc6 (in-place) I0311 19:06:37.775319 6289 net.cpp:262] Setting up relu6 I0311 19:06:37.775324 6289 net.cpp:269] TEST Top shape for layer 17 'relu6' 10 4096 (40960) I0311 19:06:37.775329 6289 layer_factory.hpp:172] Creating layer 'drop6' of type 'Dropout' I0311 19:06:37.775334 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.775349 6289 net.cpp:202] Created Layer drop6 (18) I0311 19:06:37.775354 6289 net.cpp:574] drop6 <- fc6 I0311 19:06:37.775358 6289 net.cpp:529] drop6 -> fc6 (in-place) I0311 19:06:37.775365 6289 net.cpp:262] Setting up drop6 I0311 19:06:37.775372 6289 net.cpp:269] TEST Top shape for layer 18 'drop6' 10 4096 (40960) I0311 19:06:37.775377 6289 layer_factory.hpp:172] Creating layer 'fc7' of type 'InnerProduct' I0311 19:06:37.775382 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.775388 6289 net.cpp:202] Created Layer fc7 (19) I0311 19:06:37.775393 6289 net.cpp:574] fc7 <- fc6 I0311 19:06:37.775400 6289 net.cpp:544] fc7 -> fc7 I0311 19:06:37.908700 6289 net.cpp:262] Setting up fc7 I0311 19:06:37.908728 6289 net.cpp:269] TEST Top shape for layer 19 'fc7' 10 4096 (40960) I0311 19:06:37.908740 6289 layer_factory.hpp:172] Creating layer 'relu7' of type 'ReLU' I0311 19:06:37.908751 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.908763 6289 net.cpp:202] Created Layer relu7 (20) I0311 19:06:37.908771 6289 net.cpp:574] relu7 <- fc7 I0311 19:06:37.908778 6289 net.cpp:529] relu7 -> fc7 (in-place) I0311 19:06:37.908785 6289 net.cpp:262] Setting up relu7 I0311 19:06:37.908790 6289 net.cpp:269] TEST Top shape for layer 20 'relu7' 10 4096 (40960) I0311 19:06:37.908797 6289 layer_factory.hpp:172] Creating layer 'drop7' of type 'Dropout' I0311 19:06:37.908802 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.908809 6289 net.cpp:202] Created Layer drop7 (21) I0311 19:06:37.908813 6289 net.cpp:574] drop7 <- fc7 I0311 19:06:37.908818 6289 net.cpp:529] drop7 -> fc7 (in-place) I0311 19:06:37.908824 6289 net.cpp:262] Setting up drop7 I0311 19:06:37.908829 6289 net.cpp:269] TEST Top shape for layer 21 'drop7' 10 4096 (40960) I0311 19:06:37.908834 6289 layer_factory.hpp:172] Creating layer 'fc8' of type 'InnerProduct' I0311 19:06:37.908839 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.908848 6289 net.cpp:202] Created Layer fc8 (22) I0311 19:06:37.908852 6289 net.cpp:574] fc8 <- fc7 I0311 19:06:37.908857 6289 net.cpp:544] fc8 -> fc8 I0311 19:06:37.941224 6289 net.cpp:262] Setting up fc8 I0311 19:06:37.941238 6289 net.cpp:269] TEST Top shape for layer 22 'fc8' 10 1000 (10000) I0311 19:06:37.941248 6289 layer_factory.hpp:172] Creating layer 'prob' of type 'Softmax' I0311 19:06:37.941253 6289 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0311 19:06:37.941267 6289 net.cpp:202] Created Layer prob (23) I0311 19:06:37.941293 6289 net.cpp:574] prob <- fc8 I0311 19:06:37.941300 6289 net.cpp:544] prob -> prob I0311 19:06:37.941314 6289 net.cpp:262] Setting up prob I0311 19:06:37.941320 6289 net.cpp:269] TEST Top shape for layer 23 'prob' 10 1000 (10000) I0311 19:06:37.941325 6289 net.cpp:340] prob does not need backward computation. I0311 19:06:37.941330 6289 net.cpp:340] fc8 does not need backward computation. I0311 19:06:37.941335 6289 net.cpp:340] drop7 does not need backward computation. I0311 19:06:37.941340 6289 net.cpp:340] relu7 does not need backward computation. I0311 19:06:37.941344 6289 net.cpp:340] fc7 does not need backward computation. I0311 19:06:37.941349 6289 net.cpp:340] drop6 does not need backward computation. I0311 19:06:37.941354 6289 net.cpp:340] relu6 does not need backward computation. I0311 19:06:37.941359 6289 net.cpp:340] fc6 does not need backward computation. I0311 19:06:37.941363 6289 net.cpp:340] pool5 does not need backward computation. I0311 19:06:37.941367 6289 net.cpp:340] relu5 does not need backward computation. I0311 19:06:37.941372 6289 net.cpp:340] conv5 does not need backward computation. I0311 19:06:37.941377 6289 net.cpp:340] relu4 does not need backward computation. I0311 19:06:37.941381 6289 net.cpp:340] conv4 does not need backward computation. I0311 19:06:37.941386 6289 net.cpp:340] relu3 does not need backward computation. I0311 19:06:37.941391 6289 net.cpp:340] conv3 does not need backward computation. I0311 19:06:37.941396 6289 net.cpp:340] norm2 does not need backward computation. I0311 19:06:37.941402 6289 net.cpp:340] pool2 does not need backward computation. I0311 19:06:37.941406 6289 net.cpp:340] relu2 does not need backward computation. I0311 19:06:37.941411 6289 net.cpp:340] conv2 does not need backward computation. I0311 19:06:37.941416 6289 net.cpp:340] norm1 does not need backward computation. I0311 19:06:37.941421 6289 net.cpp:340] pool1 does not need backward computation. I0311 19:06:37.941426 6289 net.cpp:340] relu1 does not need backward computation. I0311 19:06:37.941431 6289 net.cpp:340] conv1 does not need backward computation. I0311 19:06:37.941435 6289 net.cpp:340] data does not need backward computation. I0311 19:06:37.941439 6289 net.cpp:382] This network produces output prob I0311 19:06:37.941459 6289 net.cpp:405] Top memory (TEST) required for data: 68681400 diff: 68681400 I0311 19:06:37.941464 6289 net.cpp:408] Bottom memory (TEST) required for data: 68641400 diff: 68641400 I0311 19:06:37.941468 6289 net.cpp:411] Shared (in-place) memory (TEST) by data: 26658560 diff: 26658560 I0311 19:06:37.941473 6289 net.cpp:414] Parameters memory (TEST) required for data: 243860896 diff: 243860896 I0311 19:06:37.941478 6289 net.cpp:417] Parameters shared memory (TEST) by data: 0 diff: 0 I0311 19:06:37.941491 6289 net.cpp:423] Network initialization done. I0311 19:06:38.073019 6289 upgrade_proto.cpp:43] Attempting to upgrade input file specified using deprecated transformation parameters: ../models/bvlc_reference_caffenet/bvlc_reference_caffenet.caffemodel I0311 19:06:38.073046 6289 upgrade_proto.cpp:46] Successfully upgraded file specified using deprecated data transformation parameters. W0311 19:06:38.073051 6289 upgrade_proto.cpp:48] Note that future Caffe releases will only support transform_param messages for transformation fields. I0311 19:06:38.073067 6289 upgrade_proto.cpp:52] Attempting to upgrade input file specified using deprecated V1LayerParameter: ../models/bvlc_reference_caffenet/bvlc_reference_caffenet.caffemodel I0311 19:06:38.293721 6289 upgrade_proto.cpp:60] Successfully upgraded file specified using deprecated V1LayerParameter I0311 19:06:38.299028 6289 net.cpp:1141] Copying source layer data Type:Data #blobs=0 I0311 19:06:38.299038 6289 net.cpp:1141] Copying source layer conv1 Type:Convolution #blobs=2 I0311 19:06:38.299386 6289 net.cpp:1141] Copying source layer relu1 Type:ReLU #blobs=0 I0311 19:06:38.299391 6289 net.cpp:1141] Copying source layer pool1 Type:Pooling #blobs=0 I0311 19:06:38.299415 6289 net.cpp:1141] Copying source layer norm1 Type:LRN #blobs=0 I0311 19:06:38.299420 6289 net.cpp:1141] Copying source layer conv2 Type:Convolution #blobs=2 I0311 19:06:38.302163 6289 net.cpp:1141] Copying source layer relu2 Type:ReLU #blobs=0 I0311 19:06:38.302170 6289 net.cpp:1141] Copying source layer pool2 Type:Pooling #blobs=0 I0311 19:06:38.302173 6289 net.cpp:1141] Copying source layer norm2 Type:LRN #blobs=0 I0311 19:06:38.302188 6289 net.cpp:1141] Copying source layer conv3 Type:Convolution #blobs=2 I0311 19:06:38.309876 6289 net.cpp:1141] Copying source layer relu3 Type:ReLU #blobs=0 I0311 19:06:38.309883 6289 net.cpp:1141] Copying source layer conv4 Type:Convolution #blobs=2 I0311 19:06:38.315716 6289 net.cpp:1141] Copying source layer relu4 Type:ReLU #blobs=0 I0311 19:06:38.315722 6289 net.cpp:1141] Copying source layer conv5 Type:Convolution #blobs=2 I0311 19:06:38.319622 6289 net.cpp:1141] Copying source layer relu5 Type:ReLU #blobs=0 I0311 19:06:38.319628 6289 net.cpp:1141] Copying source layer pool5 Type:Pooling #blobs=0 I0311 19:06:38.319630 6289 net.cpp:1141] Copying source layer fc6 Type:InnerProduct #blobs=2 I0311 19:06:38.645920 6289 net.cpp:1141] Copying source layer relu6 Type:ReLU #blobs=0 I0311 19:06:38.645944 6289 net.cpp:1141] Copying source layer drop6 Type:Dropout #blobs=0 I0311 19:06:38.645947 6289 net.cpp:1141] Copying source layer fc7 Type:InnerProduct #blobs=2 I0311 19:06:38.791030 6289 net.cpp:1141] Copying source layer relu7 Type:ReLU #blobs=0 I0311 19:06:38.791052 6289 net.cpp:1141] Copying source layer drop7 Type:Dropout #blobs=0 I0311 19:06:38.791055 6289 net.cpp:1141] Copying source layer fc8 Type:InnerProduct #blobs=2 I0311 19:06:38.826484 6289 net.cpp:1133] Ignoring source layer loss (10, 3, 227, 227) Loading file: ../101_ObjectCategories/dolphin/image_0055.jpg Classifying 1 inputs. Done in 0.49 s. Saving results into ../result.npy chibi@1604:~/caffe$ python show_result.py data/ilsvrc12/synset_words.txt result.npy #1 | n02071294 killer whale, killer, orca, grampus, sea wolf, Orcinus orca | 93.3% #2 | n02066245 grey whale, gray whale, devilfish, Eschrichtius gibbosus, Eschrichtius robustus | 5.5% #3 | n02074367 dugong, Dugong dugon | 0.3% chibi@1604:~/caffe$