Skip to content

Commit 5e36304

Browse files
authored
adjust batch size (#812)
1 parent 34d4471 commit 5e36304

File tree

2 files changed

+9
-16
lines changed

2 files changed

+9
-16
lines changed

demo/distillation/distill.py

+8-15
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@
2222
parser = argparse.ArgumentParser(description=__doc__)
2323
add_arg = functools.partial(add_arguments, argparser=parser)
2424
# yapf: disable
25-
add_arg('batch_size', int, 64, "Minibatch size.")
25+
add_arg('batch_size', int, 256, "Minibatch size.")
2626
add_arg('use_gpu', bool, True, "Whether to use GPU or not.")
2727
add_arg('save_inference', bool, False, "Whether to save inference model.")
2828
add_arg('total_images', int, 1281167, "Training image number.")
@@ -45,12 +45,7 @@
4545

4646

4747
def piecewise_decay(args):
48-
if args.use_gpu:
49-
devices_num = paddle.fluid.core.get_cuda_device_count()
50-
else:
51-
devices_num = int(os.environ.get('CPU_NUM', 1))
52-
step = int(
53-
math.ceil(float(args.total_images) / args.batch_size) / devices_num)
48+
step = int(math.ceil(float(args.total_images) / args.batch_size))
5449
bd = [step * e for e in args.step_epochs]
5550
lr = [args.lr * (0.1**i) for i in range(len(bd) + 1)]
5651
learning_rate = paddle.optimizer.lr.PiecewiseDecay(
@@ -63,12 +58,7 @@ def piecewise_decay(args):
6358

6459

6560
def cosine_decay(args):
66-
if args.use_gpu:
67-
devices_num = paddle.fluid.core.get_cuda_device_count()
68-
else:
69-
devices_num = int(os.environ.get('CPU_NUM', 1))
70-
step = int(
71-
math.ceil(float(args.total_images) / args.batch_size) / devices_num)
61+
step = int(math.ceil(float(args.total_images) / args.batch_size))
7262
learning_rate = paddle.optimizer.lr.CosineAnnealingDecay(
7363
learning_rate=args.lr, T_max=step * args.num_epochs, verbose=False)
7464
optimizer = paddle.optimizer.Momentum(
@@ -108,7 +98,10 @@ def compress(args):
10898
places = paddle.static.cuda_places(
10999
) if args.use_gpu else paddle.static.cpu_places()
110100
place = places[0]
111-
101+
if args.use_gpu:
102+
devices_num = paddle.fluid.core.get_cuda_device_count()
103+
else:
104+
devices_num = int(os.environ.get('CPU_NUM', 1))
112105
with paddle.static.program_guard(student_program, s_startup):
113106
with paddle.fluid.unique_name.guard():
114107
image = paddle.static.data(
@@ -120,7 +113,7 @@ def compress(args):
120113
places=places,
121114
feed_list=[image, label],
122115
drop_last=True,
123-
batch_size=args.batch_size,
116+
batch_size=int(args.batch_size / devices_num),
124117
return_list=False,
125118
shuffle=True,
126119
use_shared_memory=True,

demo/dygraph/quant/train.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,7 @@
4444
parser = argparse.ArgumentParser(description=__doc__)
4545
add_arg = functools.partial(add_arguments, argparser=parser)
4646
# yapf: disable
47-
add_arg('batch_size', int, 256, "Single Card Minibatch size.")
47+
add_arg('batch_size', int, 64, "Single Card Minibatch size.")
4848
add_arg('use_gpu', bool, True, "Whether to use GPU or not.")
4949
add_arg('model', str, "mobilenet_v3", "The target model.")
5050
add_arg('pretrained_model', str, "MobileNetV3_large_x1_0_ssld_pretrained", "Whether to use pretrained model.")

0 commit comments

Comments
 (0)