-
Notifications
You must be signed in to change notification settings - Fork 196
/
test_caffe_manual_output.log
121 lines (93 loc) · 1.85 KB
/
test_caffe_manual_output.log
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
read image mean succeeded
196608
1 3 256 256
110.177 110.459
Input blob size:
10 3 227 227
#Layers: 23
[1] layer name: conv1, type: Convolution
#Blobs: 2
96 3 11 11
-0.00121359 0.00323653
96 1 1 1
-0.504212 -0.188764
[2] layer name: relu1, type: ReLU
#Blobs: 0
[3] layer name: pool1, type: Pooling
#Blobs: 0
[4] layer name: norm1, type: LRN
#Blobs: 0
[5] layer name: conv2, type: Convolution
#Blobs: 2
256 48 5 5
-0.0111258 0.0218567
256 1 1 1
0.982103 0.994344
[6] layer name: relu2, type: ReLU
#Blobs: 0
[7] layer name: pool2, type: Pooling
#Blobs: 0
[8] layer name: norm2, type: LRN
#Blobs: 0
[9] layer name: conv3, type: Convolution
#Blobs: 2
384 256 3 3
-0.000527019 0.00534925
384 1 1 1
-0.000547366 -0.00652369
[10] layer name: relu3, type: ReLU
#Blobs: 0
[11] layer name: conv4, type: Convolution
#Blobs: 2
384 192 3 3
0.00378311 0.000618855
384 1 1 1
0.884884 1.02955
[12] layer name: relu4, type: ReLU
#Blobs: 0
[13] layer name: conv5, type: Convolution
#Blobs: 2
256 192 3 3
-0.0134502 -0.0174268
256 1 1 1
0.982957 1.04293
[14] layer name: relu5, type: ReLU
#Blobs: 0
[15] layer name: pool5, type: Pooling
#Blobs: 0
[16] layer name: fc6, type: InnerProduct
#Blobs: 2
4096 9216 1 1
0.00639847 0.00915686
4096 1 1 1
0.983698 1.00962
[17] layer name: relu6, type: ReLU
#Blobs: 0
[18] layer name: drop6, type: Dropout
#Blobs: 0
[19] layer name: fc7, type: InnerProduct
#Blobs: 2
4096 4096 1 1
0.0125212 -0.0134137
4096 1 1 1
1.09399 0.999266
[20] layer name: relu7, type: ReLU
#Blobs: 0
[21] layer name: drop7, type: Dropout
#Blobs: 0
[22] layer name: fc8, type: InnerProduct
#Blobs: 2
1000 4096 1 1
0.000160601 -0.010393
1000 1 1 1
-0.196648 -0.100678
[23] layer name: prob, type: Softmax
#Blobs: 0
weights and bias from layer: conv1
-0.00121359 0.00323653
-0.504212 -0.188764
new weights and bias from layer: conv1
1.1111 2.2222
#Features: 2904000
0 0
END