forked from ChenglongChen/batch_normalization
-
Notifications
You must be signed in to change notification settings - Fork 0
/
lenet_BN_sgd_solver.prototxt
42 lines (35 loc) · 1003 Bytes
/
lenet_BN_sgd_solver.prototxt
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
# random seed to ensure reproducible results (seems don't effect GPU...)
random_seed: 1234
# model file
net: "lenet_BN_train_valid.prototxt"
# test_iter specifies how many forward passes the test should carry out.
# In the case of MNIST, we have test batch size 100 and 100 test iterations,
# covering the full 10,000 testing images.
test_iter: 100
# Carry out testing every 500 training iterations.
test_interval: 500
# Display every 100 iterations
display: 100
# The maximum number of iterations
max_iter: 10000
# snapshot intermediate results
snapshot: 5000
snapshot_prefix: "lenet"
# solver mode: CPU or GPU
solver_mode: GPU
# solver type (NOTE: only SGD and NESTEROV support accumulate gradients)
solver_type: SGD
#solver_type: NESTEROV
#solver_type: ADAGRAD
#solver_type: ADADELTA
# for ADADELTA to avoid dividing by zero
delta: 1e-8
# learning rate policy
base_lr: 0.01
momentum: 0.9
# The learning rate policy
lr_policy: "inv"
gamma: 0.0001
power: 0.75
# weight decay
weight_decay: 0.0005