chibi@2204:~/caffe$ cd python; python3 classify.py --raw_scale 255 ../101_ObjectCategories/cougar_face/image_0026.jpg ../result.npy; cd .. /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4 warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}" WARNING: Logging before InitGoogleLogging() is written to STDERR I0221 04:56:06.749579 2506 gpu_memory.cpp:82] GPUMemory::Manager initialized CPU mode W0221 04:56:06.889393 2506 _caffe.cpp:172] DEPRECATION WARNING - deprecated use of Python interface W0221 04:56:06.889556 2506 _caffe.cpp:173] Use this instead (with the named "weights" parameter): W0221 04:56:06.889565 2506 _caffe.cpp:175] Net('/home/chibi/caffe/python/../models/bvlc_reference_caffenet/deploy.prototxt', 1, weights='/home/chibi/caffe/python/../models/bvlc_reference_caffenet/bvlc_reference_caffenet.caffemodel') I0221 04:56:06.889802 2506 net.cpp:86] Initializing net from parameters: name: "CaffeNet" state { phase: TEST level: 0 } layer { name: "data" type: "Input" top: "data" input_param { shape { dim: 10 dim: 3 dim: 227 dim: 227 } } } layer { name: "conv1" type: "Convolution" bottom: "data" top: "conv1" convolution_param { num_output: 96 kernel_size: 11 stride: 4 } } layer { name: "relu1" type: "ReLU" bottom: "conv1" top: "conv1" } layer { name: "pool1" type: "Pooling" bottom: "conv1" top: "pool1" pooling_param { pool: MAX kernel_size: 3 stride: 2 } } layer { name: "norm1" type: "LRN" bottom: "pool1" top: "norm1" lrn_param { local_size: 5 alpha: 0.0001 beta: 0.75 } } layer { name: "conv2" type: "Convolution" bottom: "norm1" top: "conv2" convolution_param { num_output: 256 pad: 2 kernel_size: 5 group: 2 } } layer { name: "relu2" type: "ReLU" bottom: "conv2" top: "conv2" } layer { name: "pool2" type: "Pooling" bottom: "conv2" top: "pool2" pooling_param { pool: MAX kernel_size: 3 stride: 2 } } layer { name: "norm2" type: "LRN" bottom: "pool2" top: "norm2" lrn_param { local_size: 5 alpha: 0.0001 beta: 0.75 } } layer { name: "conv3" type: "Convolution" bottom: "norm2" top: "conv3" convolution_param { num_output: 384 pad: 1 kernel_size: 3 } } layer { name: "relu3" type: "ReLU" bottom: "conv3" top: "conv3" } layer { name: "conv4" type: "Convolution" bottom: "conv3" top: "conv4" convolution_param { num_output: 384 pad: 1 kernel_size: 3 group: 2 } } layer { name: "relu4" type: "ReLU" bottom: "conv4" top: "conv4" } layer { name: "conv5" type: "Convolution" bottom: "conv4" top: "conv5" convolution_param { num_output: 256 pad: 1 kernel_size: 3 group: 2 } } layer { name: "relu5" type: "ReLU" bottom: "conv5" top: "conv5" } layer { name: "pool5" type: "Pooling" bottom: "conv5" top: "pool5" pooling_param { pool: MAX kernel_size: 3 stride: 2 } } layer { name: "fc6" type: "InnerProduct" bottom: "pool5" top: "fc6" inner_product_param { num_output: 4096 } } layer { name: "relu6" type: "ReLU" bottom: "fc6" top: "fc6" } layer { name: "drop6" type: "Dropout" bottom: "fc6" top: "fc6" dropout_param { dropout_ratio: 0.5 } } layer { name: "fc7" type: "InnerProduct" bottom: "fc6" top: "fc7" inner_product_param { num_output: 4096 } } layer { name: "relu7" type: "ReLU" bottom: "fc7" top: "fc7" } layer { name: "drop7" type: "Dropout" bottom: "fc7" top: "fc7" dropout_param { dropout_ratio: 0.5 } } layer { name: "fc8" type: "InnerProduct" bottom: "fc7" top: "fc8" inner_product_param { num_output: 1000 } } layer { name: "prob" type: "Softmax" bottom: "fc8" top: "prob" } I0221 04:56:06.890007 2506 net.cpp:116] Using FLOAT as default forward math type I0221 04:56:06.890017 2506 net.cpp:122] Using FLOAT as default backward math type I0221 04:56:06.890021 2506 layer_factory.hpp:172] Creating layer 'data' of type 'Input' I0221 04:56:06.890026 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.890040 2506 net.cpp:205] Created Layer data (0) I0221 04:56:06.890050 2506 net.cpp:547] data -> data I0221 04:56:06.890074 2506 net.cpp:265] Setting up data I0221 04:56:06.890080 2506 net.cpp:272] TEST Top shape for layer 0 'data' 10 3 227 227 (1545870) I0221 04:56:06.890095 2506 layer_factory.hpp:172] Creating layer 'conv1' of type 'Convolution' I0221 04:56:06.890105 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.890146 2506 net.cpp:205] Created Layer conv1 (1) I0221 04:56:06.890156 2506 net.cpp:577] conv1 <- data I0221 04:56:06.890167 2506 net.cpp:547] conv1 -> conv1 I0221 04:56:06.891309 2506 net.cpp:265] Setting up conv1 I0221 04:56:06.891316 2506 net.cpp:272] TEST Top shape for layer 1 'conv1' 10 96 55 55 (2904000) I0221 04:56:06.891333 2506 layer_factory.hpp:172] Creating layer 'relu1' of type 'ReLU' I0221 04:56:06.891337 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.891345 2506 net.cpp:205] Created Layer relu1 (2) I0221 04:56:06.891351 2506 net.cpp:577] relu1 <- conv1 I0221 04:56:06.891360 2506 net.cpp:532] relu1 -> conv1 (in-place) I0221 04:56:06.891369 2506 net.cpp:265] Setting up relu1 I0221 04:56:06.891378 2506 net.cpp:272] TEST Top shape for layer 2 'relu1' 10 96 55 55 (2904000) I0221 04:56:06.891389 2506 layer_factory.hpp:172] Creating layer 'pool1' of type 'Pooling' I0221 04:56:06.891398 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.891412 2506 net.cpp:205] Created Layer pool1 (3) I0221 04:56:06.891422 2506 net.cpp:577] pool1 <- conv1 I0221 04:56:06.891433 2506 net.cpp:547] pool1 -> pool1 I0221 04:56:06.891453 2506 net.cpp:265] Setting up pool1 I0221 04:56:06.891462 2506 net.cpp:272] TEST Top shape for layer 3 'pool1' 10 96 27 27 (699840) I0221 04:56:06.891474 2506 layer_factory.hpp:172] Creating layer 'norm1' of type 'LRN' I0221 04:56:06.891484 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.891502 2506 net.cpp:205] Created Layer norm1 (4) I0221 04:56:06.891511 2506 net.cpp:577] norm1 <- pool1 I0221 04:56:06.891521 2506 net.cpp:547] norm1 -> norm1 I0221 04:56:06.891541 2506 net.cpp:265] Setting up norm1 I0221 04:56:06.891552 2506 net.cpp:272] TEST Top shape for layer 4 'norm1' 10 96 27 27 (699840) I0221 04:56:06.891564 2506 layer_factory.hpp:172] Creating layer 'conv2' of type 'Convolution' I0221 04:56:06.891574 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.891592 2506 net.cpp:205] Created Layer conv2 (5) I0221 04:56:06.891601 2506 net.cpp:577] conv2 <- norm1 I0221 04:56:06.891613 2506 net.cpp:547] conv2 -> conv2 I0221 04:56:06.893671 2506 net.cpp:265] Setting up conv2 I0221 04:56:06.893679 2506 net.cpp:272] TEST Top shape for layer 5 'conv2' 10 256 27 27 (1866240) I0221 04:56:06.893687 2506 layer_factory.hpp:172] Creating layer 'relu2' of type 'ReLU' I0221 04:56:06.893692 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.893697 2506 net.cpp:205] Created Layer relu2 (6) I0221 04:56:06.893702 2506 net.cpp:577] relu2 <- conv2 I0221 04:56:06.893707 2506 net.cpp:532] relu2 -> conv2 (in-place) I0221 04:56:06.893713 2506 net.cpp:265] Setting up relu2 I0221 04:56:06.893723 2506 net.cpp:272] TEST Top shape for layer 6 'relu2' 10 256 27 27 (1866240) I0221 04:56:06.893731 2506 layer_factory.hpp:172] Creating layer 'pool2' of type 'Pooling' I0221 04:56:06.893738 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.893752 2506 net.cpp:205] Created Layer pool2 (7) I0221 04:56:06.893761 2506 net.cpp:577] pool2 <- conv2 I0221 04:56:06.893771 2506 net.cpp:547] pool2 -> pool2 I0221 04:56:06.893783 2506 net.cpp:265] Setting up pool2 I0221 04:56:06.893790 2506 net.cpp:272] TEST Top shape for layer 7 'pool2' 10 256 13 13 (432640) I0221 04:56:06.893802 2506 layer_factory.hpp:172] Creating layer 'norm2' of type 'LRN' I0221 04:56:06.893810 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.893826 2506 net.cpp:205] Created Layer norm2 (8) I0221 04:56:06.893834 2506 net.cpp:577] norm2 <- pool2 I0221 04:56:06.893842 2506 net.cpp:547] norm2 -> norm2 I0221 04:56:06.893867 2506 net.cpp:265] Setting up norm2 I0221 04:56:06.893877 2506 net.cpp:272] TEST Top shape for layer 8 'norm2' 10 256 13 13 (432640) I0221 04:56:06.893889 2506 layer_factory.hpp:172] Creating layer 'conv3' of type 'Convolution' I0221 04:56:06.893898 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.893919 2506 net.cpp:205] Created Layer conv3 (9) I0221 04:56:06.893929 2506 net.cpp:577] conv3 <- norm2 I0221 04:56:06.893939 2506 net.cpp:547] conv3 -> conv3 I0221 04:56:06.902036 2506 net.cpp:265] Setting up conv3 I0221 04:56:06.902045 2506 net.cpp:272] TEST Top shape for layer 9 'conv3' 10 384 13 13 (648960) I0221 04:56:06.902053 2506 layer_factory.hpp:172] Creating layer 'relu3' of type 'ReLU' I0221 04:56:06.902058 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.902063 2506 net.cpp:205] Created Layer relu3 (10) I0221 04:56:06.902068 2506 net.cpp:577] relu3 <- conv3 I0221 04:56:06.902072 2506 net.cpp:532] relu3 -> conv3 (in-place) I0221 04:56:06.902081 2506 net.cpp:265] Setting up relu3 I0221 04:56:06.902088 2506 net.cpp:272] TEST Top shape for layer 10 'relu3' 10 384 13 13 (648960) I0221 04:56:06.902098 2506 layer_factory.hpp:172] Creating layer 'conv4' of type 'Convolution' I0221 04:56:06.902105 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.902113 2506 net.cpp:205] Created Layer conv4 (11) I0221 04:56:06.902117 2506 net.cpp:577] conv4 <- conv3 I0221 04:56:06.902124 2506 net.cpp:547] conv4 -> conv4 I0221 04:56:06.906562 2506 net.cpp:265] Setting up conv4 I0221 04:56:06.906571 2506 net.cpp:272] TEST Top shape for layer 11 'conv4' 10 384 13 13 (648960) I0221 04:56:06.906579 2506 layer_factory.hpp:172] Creating layer 'relu4' of type 'ReLU' I0221 04:56:06.906584 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.906589 2506 net.cpp:205] Created Layer relu4 (12) I0221 04:56:06.906594 2506 net.cpp:577] relu4 <- conv4 I0221 04:56:06.906597 2506 net.cpp:532] relu4 -> conv4 (in-place) I0221 04:56:06.906605 2506 net.cpp:265] Setting up relu4 I0221 04:56:06.906613 2506 net.cpp:272] TEST Top shape for layer 12 'relu4' 10 384 13 13 (648960) I0221 04:56:06.906623 2506 layer_factory.hpp:172] Creating layer 'conv5' of type 'Convolution' I0221 04:56:06.906630 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.906639 2506 net.cpp:205] Created Layer conv5 (13) I0221 04:56:06.906644 2506 net.cpp:577] conv5 <- conv4 I0221 04:56:06.906651 2506 net.cpp:547] conv5 -> conv5 I0221 04:56:06.909503 2506 net.cpp:265] Setting up conv5 I0221 04:56:06.909511 2506 net.cpp:272] TEST Top shape for layer 13 'conv5' 10 256 13 13 (432640) I0221 04:56:06.909519 2506 layer_factory.hpp:172] Creating layer 'relu5' of type 'ReLU' I0221 04:56:06.909523 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.909528 2506 net.cpp:205] Created Layer relu5 (14) I0221 04:56:06.909533 2506 net.cpp:577] relu5 <- conv5 I0221 04:56:06.909538 2506 net.cpp:532] relu5 -> conv5 (in-place) I0221 04:56:06.909545 2506 net.cpp:265] Setting up relu5 I0221 04:56:06.909554 2506 net.cpp:272] TEST Top shape for layer 14 'relu5' 10 256 13 13 (432640) I0221 04:56:06.909562 2506 layer_factory.hpp:172] Creating layer 'pool5' of type 'Pooling' I0221 04:56:06.909570 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.909576 2506 net.cpp:205] Created Layer pool5 (15) I0221 04:56:06.909583 2506 net.cpp:577] pool5 <- conv5 I0221 04:56:06.909592 2506 net.cpp:547] pool5 -> pool5 I0221 04:56:06.909610 2506 net.cpp:265] Setting up pool5 I0221 04:56:06.909619 2506 net.cpp:272] TEST Top shape for layer 15 'pool5' 10 256 6 6 (92160) I0221 04:56:06.909632 2506 layer_factory.hpp:172] Creating layer 'fc6' of type 'InnerProduct' I0221 04:56:06.909641 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:06.909683 2506 net.cpp:205] Created Layer fc6 (16) I0221 04:56:06.909693 2506 net.cpp:577] fc6 <- pool5 I0221 04:56:06.909701 2506 net.cpp:547] fc6 -> fc6 I0221 04:56:07.150792 2506 net.cpp:265] Setting up fc6 I0221 04:56:07.150812 2506 net.cpp:272] TEST Top shape for layer 16 'fc6' 10 4096 (40960) I0221 04:56:07.150827 2506 layer_factory.hpp:172] Creating layer 'relu6' of type 'ReLU' I0221 04:56:07.150832 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:07.150841 2506 net.cpp:205] Created Layer relu6 (17) I0221 04:56:07.150849 2506 net.cpp:577] relu6 <- fc6 I0221 04:56:07.150861 2506 net.cpp:532] relu6 -> fc6 (in-place) I0221 04:56:07.150869 2506 net.cpp:265] Setting up relu6 I0221 04:56:07.150875 2506 net.cpp:272] TEST Top shape for layer 17 'relu6' 10 4096 (40960) I0221 04:56:07.150884 2506 layer_factory.hpp:172] Creating layer 'drop6' of type 'Dropout' I0221 04:56:07.150894 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:07.150909 2506 net.cpp:205] Created Layer drop6 (18) I0221 04:56:07.150919 2506 net.cpp:577] drop6 <- fc6 I0221 04:56:07.150928 2506 net.cpp:532] drop6 -> fc6 (in-place) I0221 04:56:07.150938 2506 net.cpp:265] Setting up drop6 I0221 04:56:07.150945 2506 net.cpp:272] TEST Top shape for layer 18 'drop6' 10 4096 (40960) I0221 04:56:07.150954 2506 layer_factory.hpp:172] Creating layer 'fc7' of type 'InnerProduct' I0221 04:56:07.150961 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:07.150975 2506 net.cpp:205] Created Layer fc7 (19) I0221 04:56:07.150985 2506 net.cpp:577] fc7 <- fc6 I0221 04:56:07.150995 2506 net.cpp:547] fc7 -> fc7 I0221 04:56:07.258231 2506 net.cpp:265] Setting up fc7 I0221 04:56:07.258250 2506 net.cpp:272] TEST Top shape for layer 19 'fc7' 10 4096 (40960) I0221 04:56:07.258263 2506 layer_factory.hpp:172] Creating layer 'relu7' of type 'ReLU' I0221 04:56:07.258268 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:07.258280 2506 net.cpp:205] Created Layer relu7 (20) I0221 04:56:07.258288 2506 net.cpp:577] relu7 <- fc7 I0221 04:56:07.258299 2506 net.cpp:532] relu7 -> fc7 (in-place) I0221 04:56:07.258308 2506 net.cpp:265] Setting up relu7 I0221 04:56:07.258314 2506 net.cpp:272] TEST Top shape for layer 20 'relu7' 10 4096 (40960) I0221 04:56:07.258322 2506 layer_factory.hpp:172] Creating layer 'drop7' of type 'Dropout' I0221 04:56:07.258332 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:07.258345 2506 net.cpp:205] Created Layer drop7 (21) I0221 04:56:07.258353 2506 net.cpp:577] drop7 <- fc7 I0221 04:56:07.258366 2506 net.cpp:532] drop7 -> fc7 (in-place) I0221 04:56:07.258378 2506 net.cpp:265] Setting up drop7 I0221 04:56:07.258384 2506 net.cpp:272] TEST Top shape for layer 21 'drop7' 10 4096 (40960) I0221 04:56:07.258396 2506 layer_factory.hpp:172] Creating layer 'fc8' of type 'InnerProduct' I0221 04:56:07.258404 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:07.258419 2506 net.cpp:205] Created Layer fc8 (22) I0221 04:56:07.258426 2506 net.cpp:577] fc8 <- fc7 I0221 04:56:07.258435 2506 net.cpp:547] fc8 -> fc8 I0221 04:56:07.284632 2506 net.cpp:265] Setting up fc8 I0221 04:56:07.284650 2506 net.cpp:272] TEST Top shape for layer 22 'fc8' 10 1000 (10000) I0221 04:56:07.284662 2506 layer_factory.hpp:172] Creating layer 'prob' of type 'Softmax' I0221 04:56:07.284668 2506 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:07.284680 2506 net.cpp:205] Created Layer prob (23) I0221 04:56:07.284689 2506 net.cpp:577] prob <- fc8 I0221 04:56:07.284700 2506 net.cpp:547] prob -> prob I0221 04:56:07.284732 2506 net.cpp:265] Setting up prob I0221 04:56:07.284741 2506 net.cpp:272] TEST Top shape for layer 23 'prob' 10 1000 (10000) I0221 04:56:07.284778 2506 net.cpp:343] prob does not need backward computation. I0221 04:56:07.284788 2506 net.cpp:343] fc8 does not need backward computation. I0221 04:56:07.284796 2506 net.cpp:343] drop7 does not need backward computation. I0221 04:56:07.284806 2506 net.cpp:343] relu7 does not need backward computation. I0221 04:56:07.284816 2506 net.cpp:343] fc7 does not need backward computation. I0221 04:56:07.284824 2506 net.cpp:343] drop6 does not need backward computation. I0221 04:56:07.284833 2506 net.cpp:343] relu6 does not need backward computation. I0221 04:56:07.284843 2506 net.cpp:343] fc6 does not need backward computation. I0221 04:56:07.284852 2506 net.cpp:343] pool5 does not need backward computation. I0221 04:56:07.284862 2506 net.cpp:343] relu5 does not need backward computation. I0221 04:56:07.284870 2506 net.cpp:343] conv5 does not need backward computation. I0221 04:56:07.284879 2506 net.cpp:343] relu4 does not need backward computation. I0221 04:56:07.284888 2506 net.cpp:343] conv4 does not need backward computation. I0221 04:56:07.284897 2506 net.cpp:343] relu3 does not need backward computation. I0221 04:56:07.284906 2506 net.cpp:343] conv3 does not need backward computation. I0221 04:56:07.284916 2506 net.cpp:343] norm2 does not need backward computation. I0221 04:56:07.284926 2506 net.cpp:343] pool2 does not need backward computation. I0221 04:56:07.284935 2506 net.cpp:343] relu2 does not need backward computation. I0221 04:56:07.284945 2506 net.cpp:343] conv2 does not need backward computation. I0221 04:56:07.284955 2506 net.cpp:343] norm1 does not need backward computation. I0221 04:56:07.284965 2506 net.cpp:343] pool1 does not need backward computation. I0221 04:56:07.284973 2506 net.cpp:343] relu1 does not need backward computation. I0221 04:56:07.284981 2506 net.cpp:343] conv1 does not need backward computation. I0221 04:56:07.284991 2506 net.cpp:343] data does not need backward computation. I0221 04:56:07.285000 2506 net.cpp:385] This network produces output prob I0221 04:56:07.285029 2506 net.cpp:408] Top memory (TEST) required for data: 68681400 diff: 68681400 I0221 04:56:07.285038 2506 net.cpp:411] Bottom memory (TEST) required for data: 68641400 diff: 68641400 I0221 04:56:07.285045 2506 net.cpp:414] Shared (in-place) memory (TEST) by data: 26658560 diff: 26658560 I0221 04:56:07.285053 2506 net.cpp:417] Parameters memory (TEST) required for data: 243860896 diff: 42272 I0221 04:56:07.285061 2506 net.cpp:420] Parameters shared memory (TEST) by data: 0 diff: 0 I0221 04:56:07.285070 2506 net.cpp:426] Network initialization done. I0221 04:56:07.561777 2506 upgrade_proto.cpp:43] Attempting to upgrade input file specified using deprecated transformation parameters: /home/chibi/caffe/python/../models/bvlc_reference_caffenet/bvlc_reference_caffenet.caffemodel I0221 04:56:07.561800 2506 upgrade_proto.cpp:46] Successfully upgraded file specified using deprecated data transformation parameters. W0221 04:56:07.561805 2506 upgrade_proto.cpp:48] Note that future Caffe releases will only support transform_param messages for transformation fields. I0221 04:56:07.561828 2506 upgrade_proto.cpp:52] Attempting to upgrade input file specified using deprecated V1LayerParameter: /home/chibi/caffe/python/../models/bvlc_reference_caffenet/bvlc_reference_caffenet.caffemodel I0221 04:56:07.812788 2506 upgrade_proto.cpp:60] Successfully upgraded file specified using deprecated V1LayerParameter I0221 04:56:07.821203 2506 net.cpp:1143] Copying source layer data Type:Data #blobs=0 I0221 04:56:07.821214 2506 net.cpp:1143] Copying source layer conv1 Type:Convolution #blobs=2 I0221 04:56:07.821424 2506 net.cpp:1143] Copying source layer relu1 Type:ReLU #blobs=0 I0221 04:56:07.821431 2506 net.cpp:1143] Copying source layer pool1 Type:Pooling #blobs=0 I0221 04:56:07.821435 2506 net.cpp:1143] Copying source layer norm1 Type:LRN #blobs=0 I0221 04:56:07.821439 2506 net.cpp:1143] Copying source layer conv2 Type:Convolution #blobs=2 I0221 04:56:07.823042 2506 net.cpp:1143] Copying source layer relu2 Type:ReLU #blobs=0 I0221 04:56:07.823050 2506 net.cpp:1143] Copying source layer pool2 Type:Pooling #blobs=0 I0221 04:56:07.823055 2506 net.cpp:1143] Copying source layer norm2 Type:LRN #blobs=0 I0221 04:56:07.823060 2506 net.cpp:1143] Copying source layer conv3 Type:Convolution #blobs=2 I0221 04:56:07.827580 2506 net.cpp:1143] Copying source layer relu3 Type:ReLU #blobs=0 I0221 04:56:07.827589 2506 net.cpp:1143] Copying source layer conv4 Type:Convolution #blobs=2 I0221 04:56:07.830938 2506 net.cpp:1143] Copying source layer relu4 Type:ReLU #blobs=0 I0221 04:56:07.830946 2506 net.cpp:1143] Copying source layer conv5 Type:Convolution #blobs=2 I0221 04:56:07.833213 2506 net.cpp:1143] Copying source layer relu5 Type:ReLU #blobs=0 I0221 04:56:07.833221 2506 net.cpp:1143] Copying source layer pool5 Type:Pooling #blobs=0 I0221 04:56:07.833225 2506 net.cpp:1143] Copying source layer fc6 Type:InnerProduct #blobs=2 I0221 04:56:08.019860 2506 net.cpp:1143] Copying source layer relu6 Type:ReLU #blobs=0 I0221 04:56:08.019868 2506 net.cpp:1143] Copying source layer drop6 Type:Dropout #blobs=0 I0221 04:56:08.019873 2506 net.cpp:1143] Copying source layer fc7 Type:InnerProduct #blobs=2 I0221 04:56:08.102897 2506 net.cpp:1143] Copying source layer relu7 Type:ReLU #blobs=0 I0221 04:56:08.102907 2506 net.cpp:1143] Copying source layer drop7 Type:Dropout #blobs=0 I0221 04:56:08.102911 2506 net.cpp:1143] Copying source layer fc8 Type:InnerProduct #blobs=2 I0221 04:56:08.124135 2506 net.cpp:1135] Ignoring source layer loss (10, 3, 227, 227) Loading file: ../101_ObjectCategories/cougar_face/image_0026.jpg Classifying 1 inputs. Done in 0.28 s. Saving results into ../result.npy chibi@2204:~/caffe$ python3 show_result_py3.py data/ilsvrc12/synset_words.txt result.npy #1 | n02129165 lion, king of beasts, Panthera leo | 47.6% #2 | n02125311 cougar, puma, catamount, mountain lion, painter, panther, Felis concolor | 24.4% #3 | n02492660 howler monkey, howler | 7.2% chibi@2204:~/caffe$ cd python; python3 classify.py --raw_scale 255 ../101_ObjectCategories/cougar_face/image_0006.jpg ../result.npy; cd .. /usr/lib/python3/dist-packages/scipy/__init__.py:146: UserWarning: A NumPy version >=1.17.3 and <1.25.0 is required for this version of SciPy (detected version 1.26.4 warnings.warn(f"A NumPy version >={np_minversion} and <{np_maxversion}" WARNING: Logging before InitGoogleLogging() is written to STDERR I0221 04:56:47.647929 2536 gpu_memory.cpp:82] GPUMemory::Manager initialized CPU mode W0221 04:56:47.787454 2536 _caffe.cpp:172] DEPRECATION WARNING - deprecated use of Python interface W0221 04:56:47.787621 2536 _caffe.cpp:173] Use this instead (with the named "weights" parameter): W0221 04:56:47.787629 2536 _caffe.cpp:175] Net('/home/chibi/caffe/python/../models/bvlc_reference_caffenet/deploy.prototxt', 1, weights='/home/chibi/caffe/python/../models/bvlc_reference_caffenet/bvlc_reference_caffenet.caffemodel') I0221 04:56:47.787848 2536 net.cpp:86] Initializing net from parameters: name: "CaffeNet" state { phase: TEST level: 0 } layer { name: "data" type: "Input" top: "data" input_param { shape { dim: 10 dim: 3 dim: 227 dim: 227 } } } layer { name: "conv1" type: "Convolution" bottom: "data" top: "conv1" convolution_param { num_output: 96 kernel_size: 11 stride: 4 } } layer { name: "relu1" type: "ReLU" bottom: "conv1" top: "conv1" } layer { name: "pool1" type: "Pooling" bottom: "conv1" top: "pool1" pooling_param { pool: MAX kernel_size: 3 stride: 2 } } layer { name: "norm1" type: "LRN" bottom: "pool1" top: "norm1" lrn_param { local_size: 5 alpha: 0.0001 beta: 0.75 } } layer { name: "conv2" type: "Convolution" bottom: "norm1" top: "conv2" convolution_param { num_output: 256 pad: 2 kernel_size: 5 group: 2 } } layer { name: "relu2" type: "ReLU" bottom: "conv2" top: "conv2" } layer { name: "pool2" type: "Pooling" bottom: "conv2" top: "pool2" pooling_param { pool: MAX kernel_size: 3 stride: 2 } } layer { name: "norm2" type: "LRN" bottom: "pool2" top: "norm2" lrn_param { local_size: 5 alpha: 0.0001 beta: 0.75 } } layer { name: "conv3" type: "Convolution" bottom: "norm2" top: "conv3" convolution_param { num_output: 384 pad: 1 kernel_size: 3 } } layer { name: "relu3" type: "ReLU" bottom: "conv3" top: "conv3" } layer { name: "conv4" type: "Convolution" bottom: "conv3" top: "conv4" convolution_param { num_output: 384 pad: 1 kernel_size: 3 group: 2 } } layer { name: "relu4" type: "ReLU" bottom: "conv4" top: "conv4" } layer { name: "conv5" type: "Convolution" bottom: "conv4" top: "conv5" convolution_param { num_output: 256 pad: 1 kernel_size: 3 group: 2 } } layer { name: "relu5" type: "ReLU" bottom: "conv5" top: "conv5" } layer { name: "pool5" type: "Pooling" bottom: "conv5" top: "pool5" pooling_param { pool: MAX kernel_size: 3 stride: 2 } } layer { name: "fc6" type: "InnerProduct" bottom: "pool5" top: "fc6" inner_product_param { num_output: 4096 } } layer { name: "relu6" type: "ReLU" bottom: "fc6" top: "fc6" } layer { name: "drop6" type: "Dropout" bottom: "fc6" top: "fc6" dropout_param { dropout_ratio: 0.5 } } layer { name: "fc7" type: "InnerProduct" bottom: "fc6" top: "fc7" inner_product_param { num_output: 4096 } } layer { name: "relu7" type: "ReLU" bottom: "fc7" top: "fc7" } layer { name: "drop7" type: "Dropout" bottom: "fc7" top: "fc7" dropout_param { dropout_ratio: 0.5 } } layer { name: "fc8" type: "InnerProduct" bottom: "fc7" top: "fc8" inner_product_param { num_output: 1000 } } layer { name: "prob" type: "Softmax" bottom: "fc8" top: "prob" } I0221 04:56:47.788025 2536 net.cpp:116] Using FLOAT as default forward math type I0221 04:56:47.788033 2536 net.cpp:122] Using FLOAT as default backward math type I0221 04:56:47.788038 2536 layer_factory.hpp:172] Creating layer 'data' of type 'Input' I0221 04:56:47.788041 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.788056 2536 net.cpp:205] Created Layer data (0) I0221 04:56:47.788064 2536 net.cpp:547] data -> data I0221 04:56:47.788080 2536 net.cpp:265] Setting up data I0221 04:56:47.788086 2536 net.cpp:272] TEST Top shape for layer 0 'data' 10 3 227 227 (1545870) I0221 04:56:47.788094 2536 layer_factory.hpp:172] Creating layer 'conv1' of type 'Convolution' I0221 04:56:47.788100 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.788122 2536 net.cpp:205] Created Layer conv1 (1) I0221 04:56:47.788128 2536 net.cpp:577] conv1 <- data I0221 04:56:47.788133 2536 net.cpp:547] conv1 -> conv1 I0221 04:56:47.788409 2536 net.cpp:265] Setting up conv1 I0221 04:56:47.788415 2536 net.cpp:272] TEST Top shape for layer 1 'conv1' 10 96 55 55 (2904000) I0221 04:56:47.788432 2536 layer_factory.hpp:172] Creating layer 'relu1' of type 'ReLU' I0221 04:56:47.788437 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.788444 2536 net.cpp:205] Created Layer relu1 (2) I0221 04:56:47.788448 2536 net.cpp:577] relu1 <- conv1 I0221 04:56:47.788453 2536 net.cpp:532] relu1 -> conv1 (in-place) I0221 04:56:47.788460 2536 net.cpp:265] Setting up relu1 I0221 04:56:47.788463 2536 net.cpp:272] TEST Top shape for layer 2 'relu1' 10 96 55 55 (2904000) I0221 04:56:47.788470 2536 layer_factory.hpp:172] Creating layer 'pool1' of type 'Pooling' I0221 04:56:47.788473 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.788479 2536 net.cpp:205] Created Layer pool1 (3) I0221 04:56:47.788484 2536 net.cpp:577] pool1 <- conv1 I0221 04:56:47.788489 2536 net.cpp:547] pool1 -> pool1 I0221 04:56:47.788501 2536 net.cpp:265] Setting up pool1 I0221 04:56:47.788506 2536 net.cpp:272] TEST Top shape for layer 3 'pool1' 10 96 27 27 (699840) I0221 04:56:47.788512 2536 layer_factory.hpp:172] Creating layer 'norm1' of type 'LRN' I0221 04:56:47.788516 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.788527 2536 net.cpp:205] Created Layer norm1 (4) I0221 04:56:47.788530 2536 net.cpp:577] norm1 <- pool1 I0221 04:56:47.788535 2536 net.cpp:547] norm1 -> norm1 I0221 04:56:47.788545 2536 net.cpp:265] Setting up norm1 I0221 04:56:47.788550 2536 net.cpp:272] TEST Top shape for layer 4 'norm1' 10 96 27 27 (699840) I0221 04:56:47.788555 2536 layer_factory.hpp:172] Creating layer 'conv2' of type 'Convolution' I0221 04:56:47.788560 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.788568 2536 net.cpp:205] Created Layer conv2 (5) I0221 04:56:47.788573 2536 net.cpp:577] conv2 <- norm1 I0221 04:56:47.788578 2536 net.cpp:547] conv2 -> conv2 I0221 04:56:47.790549 2536 net.cpp:265] Setting up conv2 I0221 04:56:47.790555 2536 net.cpp:272] TEST Top shape for layer 5 'conv2' 10 256 27 27 (1866240) I0221 04:56:47.790562 2536 layer_factory.hpp:172] Creating layer 'relu2' of type 'ReLU' I0221 04:56:47.790567 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.790572 2536 net.cpp:205] Created Layer relu2 (6) I0221 04:56:47.790576 2536 net.cpp:577] relu2 <- conv2 I0221 04:56:47.790580 2536 net.cpp:532] relu2 -> conv2 (in-place) I0221 04:56:47.790585 2536 net.cpp:265] Setting up relu2 I0221 04:56:47.790589 2536 net.cpp:272] TEST Top shape for layer 6 'relu2' 10 256 27 27 (1866240) I0221 04:56:47.790596 2536 layer_factory.hpp:172] Creating layer 'pool2' of type 'Pooling' I0221 04:56:47.790599 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.790606 2536 net.cpp:205] Created Layer pool2 (7) I0221 04:56:47.790611 2536 net.cpp:577] pool2 <- conv2 I0221 04:56:47.790614 2536 net.cpp:547] pool2 -> pool2 I0221 04:56:47.790620 2536 net.cpp:265] Setting up pool2 I0221 04:56:47.790624 2536 net.cpp:272] TEST Top shape for layer 7 'pool2' 10 256 13 13 (432640) I0221 04:56:47.790629 2536 layer_factory.hpp:172] Creating layer 'norm2' of type 'LRN' I0221 04:56:47.790634 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.790643 2536 net.cpp:205] Created Layer norm2 (8) I0221 04:56:47.790647 2536 net.cpp:577] norm2 <- pool2 I0221 04:56:47.790652 2536 net.cpp:547] norm2 -> norm2 I0221 04:56:47.790669 2536 net.cpp:265] Setting up norm2 I0221 04:56:47.790674 2536 net.cpp:272] TEST Top shape for layer 8 'norm2' 10 256 13 13 (432640) I0221 04:56:47.790680 2536 layer_factory.hpp:172] Creating layer 'conv3' of type 'Convolution' I0221 04:56:47.790684 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.790693 2536 net.cpp:205] Created Layer conv3 (9) I0221 04:56:47.790699 2536 net.cpp:577] conv3 <- norm2 I0221 04:56:47.790704 2536 net.cpp:547] conv3 -> conv3 I0221 04:56:47.796380 2536 net.cpp:265] Setting up conv3 I0221 04:56:47.796389 2536 net.cpp:272] TEST Top shape for layer 9 'conv3' 10 384 13 13 (648960) I0221 04:56:47.796397 2536 layer_factory.hpp:172] Creating layer 'relu3' of type 'ReLU' I0221 04:56:47.796402 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.796408 2536 net.cpp:205] Created Layer relu3 (10) I0221 04:56:47.796412 2536 net.cpp:577] relu3 <- conv3 I0221 04:56:47.796416 2536 net.cpp:532] relu3 -> conv3 (in-place) I0221 04:56:47.796422 2536 net.cpp:265] Setting up relu3 I0221 04:56:47.796427 2536 net.cpp:272] TEST Top shape for layer 10 'relu3' 10 384 13 13 (648960) I0221 04:56:47.796432 2536 layer_factory.hpp:172] Creating layer 'conv4' of type 'Convolution' I0221 04:56:47.796437 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.796443 2536 net.cpp:205] Created Layer conv4 (11) I0221 04:56:47.796447 2536 net.cpp:577] conv4 <- conv3 I0221 04:56:47.796452 2536 net.cpp:547] conv4 -> conv4 I0221 04:56:47.800704 2536 net.cpp:265] Setting up conv4 I0221 04:56:47.800711 2536 net.cpp:272] TEST Top shape for layer 11 'conv4' 10 384 13 13 (648960) I0221 04:56:47.800719 2536 layer_factory.hpp:172] Creating layer 'relu4' of type 'ReLU' I0221 04:56:47.800724 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.800729 2536 net.cpp:205] Created Layer relu4 (12) I0221 04:56:47.800732 2536 net.cpp:577] relu4 <- conv4 I0221 04:56:47.800737 2536 net.cpp:532] relu4 -> conv4 (in-place) I0221 04:56:47.800741 2536 net.cpp:265] Setting up relu4 I0221 04:56:47.800747 2536 net.cpp:272] TEST Top shape for layer 12 'relu4' 10 384 13 13 (648960) I0221 04:56:47.800752 2536 layer_factory.hpp:172] Creating layer 'conv5' of type 'Convolution' I0221 04:56:47.800756 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.800767 2536 net.cpp:205] Created Layer conv5 (13) I0221 04:56:47.800772 2536 net.cpp:577] conv5 <- conv4 I0221 04:56:47.800776 2536 net.cpp:547] conv5 -> conv5 I0221 04:56:47.803615 2536 net.cpp:265] Setting up conv5 I0221 04:56:47.803623 2536 net.cpp:272] TEST Top shape for layer 13 'conv5' 10 256 13 13 (432640) I0221 04:56:47.803632 2536 layer_factory.hpp:172] Creating layer 'relu5' of type 'ReLU' I0221 04:56:47.803635 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.803640 2536 net.cpp:205] Created Layer relu5 (14) I0221 04:56:47.803644 2536 net.cpp:577] relu5 <- conv5 I0221 04:56:47.803649 2536 net.cpp:532] relu5 -> conv5 (in-place) I0221 04:56:47.803655 2536 net.cpp:265] Setting up relu5 I0221 04:56:47.803660 2536 net.cpp:272] TEST Top shape for layer 14 'relu5' 10 256 13 13 (432640) I0221 04:56:47.803665 2536 layer_factory.hpp:172] Creating layer 'pool5' of type 'Pooling' I0221 04:56:47.803669 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.803675 2536 net.cpp:205] Created Layer pool5 (15) I0221 04:56:47.803680 2536 net.cpp:577] pool5 <- conv5 I0221 04:56:47.803685 2536 net.cpp:547] pool5 -> pool5 I0221 04:56:47.803692 2536 net.cpp:265] Setting up pool5 I0221 04:56:47.803696 2536 net.cpp:272] TEST Top shape for layer 15 'pool5' 10 256 6 6 (92160) I0221 04:56:47.803701 2536 layer_factory.hpp:172] Creating layer 'fc6' of type 'InnerProduct' I0221 04:56:47.803706 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:47.803738 2536 net.cpp:205] Created Layer fc6 (16) I0221 04:56:47.803743 2536 net.cpp:577] fc6 <- pool5 I0221 04:56:47.803748 2536 net.cpp:547] fc6 -> fc6 I0221 04:56:48.045320 2536 net.cpp:265] Setting up fc6 I0221 04:56:48.045338 2536 net.cpp:272] TEST Top shape for layer 16 'fc6' 10 4096 (40960) I0221 04:56:48.045351 2536 layer_factory.hpp:172] Creating layer 'relu6' of type 'ReLU' I0221 04:56:48.045356 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:48.045368 2536 net.cpp:205] Created Layer relu6 (17) I0221 04:56:48.045374 2536 net.cpp:577] relu6 <- fc6 I0221 04:56:48.045379 2536 net.cpp:532] relu6 -> fc6 (in-place) I0221 04:56:48.045387 2536 net.cpp:265] Setting up relu6 I0221 04:56:48.045392 2536 net.cpp:272] TEST Top shape for layer 17 'relu6' 10 4096 (40960) I0221 04:56:48.045395 2536 layer_factory.hpp:172] Creating layer 'drop6' of type 'Dropout' I0221 04:56:48.045400 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:48.045406 2536 net.cpp:205] Created Layer drop6 (18) I0221 04:56:48.045410 2536 net.cpp:577] drop6 <- fc6 I0221 04:56:48.045415 2536 net.cpp:532] drop6 -> fc6 (in-place) I0221 04:56:48.045421 2536 net.cpp:265] Setting up drop6 I0221 04:56:48.045426 2536 net.cpp:272] TEST Top shape for layer 18 'drop6' 10 4096 (40960) I0221 04:56:48.045431 2536 layer_factory.hpp:172] Creating layer 'fc7' of type 'InnerProduct' I0221 04:56:48.045435 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:48.045442 2536 net.cpp:205] Created Layer fc7 (19) I0221 04:56:48.045446 2536 net.cpp:577] fc7 <- fc6 I0221 04:56:48.045451 2536 net.cpp:547] fc7 -> fc7 I0221 04:56:48.153349 2536 net.cpp:265] Setting up fc7 I0221 04:56:48.153367 2536 net.cpp:272] TEST Top shape for layer 19 'fc7' 10 4096 (40960) I0221 04:56:48.153379 2536 layer_factory.hpp:172] Creating layer 'relu7' of type 'ReLU' I0221 04:56:48.153385 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:48.153396 2536 net.cpp:205] Created Layer relu7 (20) I0221 04:56:48.153402 2536 net.cpp:577] relu7 <- fc7 I0221 04:56:48.153409 2536 net.cpp:532] relu7 -> fc7 (in-place) I0221 04:56:48.153414 2536 net.cpp:265] Setting up relu7 I0221 04:56:48.153419 2536 net.cpp:272] TEST Top shape for layer 20 'relu7' 10 4096 (40960) I0221 04:56:48.153424 2536 layer_factory.hpp:172] Creating layer 'drop7' of type 'Dropout' I0221 04:56:48.153429 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:48.153434 2536 net.cpp:205] Created Layer drop7 (21) I0221 04:56:48.153439 2536 net.cpp:577] drop7 <- fc7 I0221 04:56:48.153445 2536 net.cpp:532] drop7 -> fc7 (in-place) I0221 04:56:48.153450 2536 net.cpp:265] Setting up drop7 I0221 04:56:48.153453 2536 net.cpp:272] TEST Top shape for layer 21 'drop7' 10 4096 (40960) I0221 04:56:48.153458 2536 layer_factory.hpp:172] Creating layer 'fc8' of type 'InnerProduct' I0221 04:56:48.153463 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:48.153469 2536 net.cpp:205] Created Layer fc8 (22) I0221 04:56:48.153474 2536 net.cpp:577] fc8 <- fc7 I0221 04:56:48.153478 2536 net.cpp:547] fc8 -> fc8 I0221 04:56:48.179706 2536 net.cpp:265] Setting up fc8 I0221 04:56:48.179725 2536 net.cpp:272] TEST Top shape for layer 22 'fc8' 10 1000 (10000) I0221 04:56:48.179737 2536 layer_factory.hpp:172] Creating layer 'prob' of type 'Softmax' I0221 04:56:48.179742 2536 layer_factory.hpp:184] Layer's types are Ftype:FLOAT Btype:FLOAT Fmath:FLOAT Bmath:FLOAT I0221 04:56:48.179756 2536 net.cpp:205] Created Layer prob (23) I0221 04:56:48.179762 2536 net.cpp:577] prob <- fc8 I0221 04:56:48.179768 2536 net.cpp:547] prob -> prob I0221 04:56:48.179790 2536 net.cpp:265] Setting up prob I0221 04:56:48.179795 2536 net.cpp:272] TEST Top shape for layer 23 'prob' 10 1000 (10000) I0221 04:56:48.179826 2536 net.cpp:343] prob does not need backward computation. I0221 04:56:48.179832 2536 net.cpp:343] fc8 does not need backward computation. I0221 04:56:48.179837 2536 net.cpp:343] drop7 does not need backward computation. I0221 04:56:48.179842 2536 net.cpp:343] relu7 does not need backward computation. I0221 04:56:48.179847 2536 net.cpp:343] fc7 does not need backward computation. I0221 04:56:48.179852 2536 net.cpp:343] drop6 does not need backward computation. I0221 04:56:48.179857 2536 net.cpp:343] relu6 does not need backward computation. I0221 04:56:48.179862 2536 net.cpp:343] fc6 does not need backward computation. I0221 04:56:48.179867 2536 net.cpp:343] pool5 does not need backward computation. I0221 04:56:48.179872 2536 net.cpp:343] relu5 does not need backward computation. I0221 04:56:48.179876 2536 net.cpp:343] conv5 does not need backward computation. I0221 04:56:48.179880 2536 net.cpp:343] relu4 does not need backward computation. I0221 04:56:48.179885 2536 net.cpp:343] conv4 does not need backward computation. I0221 04:56:48.179889 2536 net.cpp:343] relu3 does not need backward computation. I0221 04:56:48.179894 2536 net.cpp:343] conv3 does not need backward computation. I0221 04:56:48.179899 2536 net.cpp:343] norm2 does not need backward computation. I0221 04:56:48.179903 2536 net.cpp:343] pool2 does not need backward computation. I0221 04:56:48.179910 2536 net.cpp:343] relu2 does not need backward computation. I0221 04:56:48.179915 2536 net.cpp:343] conv2 does not need backward computation. I0221 04:56:48.179920 2536 net.cpp:343] norm1 does not need backward computation. I0221 04:56:48.179925 2536 net.cpp:343] pool1 does not need backward computation. I0221 04:56:48.179930 2536 net.cpp:343] relu1 does not need backward computation. I0221 04:56:48.179934 2536 net.cpp:343] conv1 does not need backward computation. I0221 04:56:48.179939 2536 net.cpp:343] data does not need backward computation. I0221 04:56:48.179944 2536 net.cpp:385] This network produces output prob I0221 04:56:48.179961 2536 net.cpp:408] Top memory (TEST) required for data: 68681400 diff: 68681400 I0221 04:56:48.179966 2536 net.cpp:411] Bottom memory (TEST) required for data: 68641400 diff: 68641400 I0221 04:56:48.179970 2536 net.cpp:414] Shared (in-place) memory (TEST) by data: 26658560 diff: 26658560 I0221 04:56:48.179975 2536 net.cpp:417] Parameters memory (TEST) required for data: 243860896 diff: 42272 I0221 04:56:48.179979 2536 net.cpp:420] Parameters shared memory (TEST) by data: 0 diff: 0 I0221 04:56:48.179984 2536 net.cpp:426] Network initialization done. I0221 04:56:48.457369 2536 upgrade_proto.cpp:43] Attempting to upgrade input file specified using deprecated transformation parameters: /home/chibi/caffe/python/../models/bvlc_reference_caffenet/bvlc_reference_caffenet.caffemodel I0221 04:56:48.457391 2536 upgrade_proto.cpp:46] Successfully upgraded file specified using deprecated data transformation parameters. W0221 04:56:48.457396 2536 upgrade_proto.cpp:48] Note that future Caffe releases will only support transform_param messages for transformation fields. I0221 04:56:48.457417 2536 upgrade_proto.cpp:52] Attempting to upgrade input file specified using deprecated V1LayerParameter: /home/chibi/caffe/python/../models/bvlc_reference_caffenet/bvlc_reference_caffenet.caffemodel I0221 04:56:48.708874 2536 upgrade_proto.cpp:60] Successfully upgraded file specified using deprecated V1LayerParameter I0221 04:56:48.717226 2536 net.cpp:1143] Copying source layer data Type:Data #blobs=0 I0221 04:56:48.717237 2536 net.cpp:1143] Copying source layer conv1 Type:Convolution #blobs=2 I0221 04:56:48.717425 2536 net.cpp:1143] Copying source layer relu1 Type:ReLU #blobs=0 I0221 04:56:48.717430 2536 net.cpp:1143] Copying source layer pool1 Type:Pooling #blobs=0 I0221 04:56:48.717435 2536 net.cpp:1143] Copying source layer norm1 Type:LRN #blobs=0 I0221 04:56:48.717439 2536 net.cpp:1143] Copying source layer conv2 Type:Convolution #blobs=2 I0221 04:56:48.719022 2536 net.cpp:1143] Copying source layer relu2 Type:ReLU #blobs=0 I0221 04:56:48.719028 2536 net.cpp:1143] Copying source layer pool2 Type:Pooling #blobs=0 I0221 04:56:48.719033 2536 net.cpp:1143] Copying source layer norm2 Type:LRN #blobs=0 I0221 04:56:48.719036 2536 net.cpp:1143] Copying source layer conv3 Type:Convolution #blobs=2 I0221 04:56:48.723485 2536 net.cpp:1143] Copying source layer relu3 Type:ReLU #blobs=0 I0221 04:56:48.723492 2536 net.cpp:1143] Copying source layer conv4 Type:Convolution #blobs=2 I0221 04:56:48.726866 2536 net.cpp:1143] Copying source layer relu4 Type:ReLU #blobs=0 I0221 04:56:48.726871 2536 net.cpp:1143] Copying source layer conv5 Type:Convolution #blobs=2 I0221 04:56:48.729117 2536 net.cpp:1143] Copying source layer relu5 Type:ReLU #blobs=0 I0221 04:56:48.729122 2536 net.cpp:1143] Copying source layer pool5 Type:Pooling #blobs=0 I0221 04:56:48.729126 2536 net.cpp:1143] Copying source layer fc6 Type:InnerProduct #blobs=2 I0221 04:56:48.915156 2536 net.cpp:1143] Copying source layer relu6 Type:ReLU #blobs=0 I0221 04:56:48.915163 2536 net.cpp:1143] Copying source layer drop6 Type:Dropout #blobs=0 I0221 04:56:48.915167 2536 net.cpp:1143] Copying source layer fc7 Type:InnerProduct #blobs=2 I0221 04:56:48.998550 2536 net.cpp:1143] Copying source layer relu7 Type:ReLU #blobs=0 I0221 04:56:48.998557 2536 net.cpp:1143] Copying source layer drop7 Type:Dropout #blobs=0 I0221 04:56:48.998561 2536 net.cpp:1143] Copying source layer fc8 Type:InnerProduct #blobs=2 I0221 04:56:49.018901 2536 net.cpp:1135] Ignoring source layer loss (10, 3, 227, 227) Loading file: ../101_ObjectCategories/cougar_face/image_0006.jpg Classifying 1 inputs. Done in 0.27 s. Saving results into ../result.npy chibi@2204:~/caffe$ python3 show_result_py3.py data/ilsvrc12/synset_words.txt result.npy #1 | n02125311 cougar, puma, catamount, mountain lion, painter, panther, Felis concolor | 90.5% #2 | n02127052 lynx, catamount | 3.4% #3 | n02114712 red wolf, maned wolf, Canis rufus, Canis niger | 1.7% chibi@2204:~/caffe$