Skip to content

Commit 2435f4d

Browse files
committed
Fix
1 parent 34c3538 commit 2435f4d

27 files changed

+75
-214
lines changed

test/collective/fleet/test_dgc_optimizer.py

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,6 @@ def check_dgc_momentum_optimizer(
4545
mul_x = block.create_parameter(
4646
dtype="float32",
4747
shape=[dims[0], dims[1]],
48-
lod_level=0,
4948
name="mul.x",
5049
optimize_attr={'learning_rate': 1.1},
5150
regularizer=(
@@ -55,12 +54,11 @@ def check_dgc_momentum_optimizer(
5554
),
5655
)
5756
mul_y = block.create_var(
58-
dtype="float32", shape=[dims[1], dims[2]], lod_level=0, name="mul.y"
57+
dtype="float32", shape=[dims[1], dims[2]], name="mul.y"
5958
)
6059
mul_out = block.create_var(
6160
dtype="float32",
6261
shape=[dims[0], dims[2]],
63-
lod_level=0,
6462
name="mul.out",
6563
)
6664
block.append_op(
@@ -94,9 +92,7 @@ def check_dgc_momentum_optimizer(
9492
dgc_momentum_optimizer._optimizer.get_velocity_str
9593
)
9694

97-
mean_out = block.create_var(
98-
dtype="float32", shape=[1], lod_level=0, name="mean.out"
99-
)
95+
mean_out = block.create_var(dtype="float32", shape=[1], name="mean.out")
10096
block.append_op(
10197
type="mean", inputs={"X": mul_out}, outputs={"Out": mean_out}
10298
)

test/deprecated/ir/test_ir_fc_fuse_pass_deprecated.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ class FCFusePassTest(PassTest):
3131
def setUp(self):
3232
with base.program_guard(self.main_program, self.startup_program):
3333
data = paddle.static.data(
34-
name="data", shape=[32, 128], dtype="float32", lod_level=0
34+
name="data", shape=[32, 128], dtype="float32"
3535
)
3636
tmp_0 = paddle.static.nn.fc(
3737
x=data, size=128, num_flatten_dims=1, activation="relu"

test/deprecated/ir/test_ir_preln_residual_bias_fuse_pass_deprecated.py

Lines changed: 4 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -27,13 +27,9 @@ def setUp(self):
2727
with paddle.static.program_guard(
2828
self.main_program, self.startup_program
2929
):
30-
x = paddle.static.data(
31-
name="x", shape=[128, 768], dtype="float32", lod_level=0
32-
)
30+
x = paddle.static.data(name="x", shape=[128, 768], dtype="float32")
3331
bias = paddle.static.create_parameter(shape=[768], dtype='float32')
34-
y = paddle.static.data(
35-
name="y", shape=[128, 768], dtype="float32", lod_level=0
36-
)
32+
y = paddle.static.data(name="y", shape=[128, 768], dtype="float32")
3733
x = x + bias
3834
elementwise_out = x + y
3935
out = paddle.static.nn.layer_norm(input=elementwise_out)
@@ -63,12 +59,8 @@ def setUp(self):
6359
with paddle.static.program_guard(
6460
self.main_program, self.startup_program
6561
):
66-
x = paddle.static.data(
67-
name="x", shape=[128, 768], dtype="float32", lod_level=0
68-
)
69-
y = paddle.static.data(
70-
name="y", shape=[128, 768], dtype="float32", lod_level=0
71-
)
62+
x = paddle.static.data(name="x", shape=[128, 768], dtype="float32")
63+
y = paddle.static.data(name="y", shape=[128, 768], dtype="float32")
7264
elementwise_out = x + y
7365
out = paddle.static.nn.layer_norm(input=elementwise_out)
7466

test/deprecated/ir/test_ir_skip_layernorm_pass_deprecated.py

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -27,12 +27,8 @@ class SkipLayerNormFusePassTest(PassTest):
2727
def setUp(self):
2828
paddle.enable_static()
2929
with base.program_guard(self.main_program, self.startup_program):
30-
x = paddle.static.data(
31-
name="x", shape=[128, 768], dtype="float32", lod_level=0
32-
)
33-
y = paddle.static.data(
34-
name="y", shape=[128, 768], dtype="float32", lod_level=0
35-
)
30+
x = paddle.static.data(name="x", shape=[128, 768], dtype="float32")
31+
y = paddle.static.data(name="y", shape=[128, 768], dtype="float32")
3632
elementwise_out = paddle.add(x=x, y=y)
3733
out = paddle.static.nn.layer_norm(input=elementwise_out)
3834

test/deprecated/legacy_test/dist_fleet_ctr.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -79,7 +79,6 @@ def net(self, args, is_train=True, batch_size=4, lr=0.01):
7979
name="click",
8080
shape=[-1, 1],
8181
dtype="int64",
82-
lod_level=0,
8382
)
8483

8584
datas = [dnn_data, lr_data, label]

test/deprecated/legacy_test/test_dataset.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -444,16 +444,16 @@ def test_in_memory_dataset_masterpatch1(self):
444444
startup_program = base.Program()
445445
with base.program_guard(train_program, startup_program):
446446
var1 = paddle.static.data(
447-
name="slot1", shape=[-1, 1], dtype="int64", lod_level=0
447+
name="slot1", shape=[-1, 1], dtype="int64"
448448
)
449449
var2 = paddle.static.data(
450-
name="slot2", shape=[-1, 1], dtype="int64", lod_level=0
450+
name="slot2", shape=[-1, 1], dtype="int64"
451451
)
452452
var3 = paddle.static.data(
453-
name="slot3", shape=[-1, 1], dtype="float32", lod_level=0
453+
name="slot3", shape=[-1, 1], dtype="float32"
454454
)
455455
var4 = paddle.static.data(
456-
name="slot4", shape=[-1, 1], dtype="float32", lod_level=0
456+
name="slot4", shape=[-1, 1], dtype="float32"
457457
)
458458
slots_vars = [var1, var2, var3, var4]
459459

test/deprecated/legacy_test/test_infer_no_need_buffer_slots_deprecated.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -26,12 +26,12 @@ def net(self):
2626
x1 = (
2727
base.default_main_program()
2828
.global_block()
29-
.create_var(dtype="float32", shape=[1], lod_level=0, name="x1")
29+
.create_var(dtype="float32", shape=[1], name="x1")
3030
)
3131
x2 = (
3232
base.default_main_program()
3333
.global_block()
34-
.create_var(dtype="float32", shape=[1], lod_level=0, name="x2")
34+
.create_var(dtype="float32", shape=[1], name="x2")
3535
)
3636
x = paddle.add(x1, x2)
3737
return x

test/deprecated/legacy_test/test_layers_deprecated.py

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1400,9 +1400,7 @@ def test_simple_conv2d(self):
14001400
def test_shuffle_batch(self):
14011401
# TODO(minqiyang): dygraph do not support lod now
14021402
with self.static_graph():
1403-
x = paddle.static.data(
1404-
name='X', shape=[-1, 4, 50], dtype='float32', lod_level=0
1405-
)
1403+
x = paddle.static.data(name='X', shape=[-1, 4, 50], dtype='float32')
14061404
out1 = shuffle_batch(x)
14071405
paddle.seed(1000)
14081406
out2 = shuffle_batch(x)

test/deprecated/legacy_test/test_optimizer_deprecated.py

Lines changed: 30 additions & 68 deletions
Original file line numberDiff line numberDiff line change
@@ -37,18 +37,17 @@ def check_sgd_optimizer(optimizer_attr):
3737
mul_x = block.create_parameter(
3838
dtype="float32",
3939
shape=[5, 10],
40-
lod_level=0,
4140
name="mul.x",
4241
optimize_attr=optimizer_attr,
4342
)
4443
mul_y = block.create_var(
45-
dtype="float32", shape=[10, 8], lod_level=0, name="mul.y"
44+
dtype="float32", shape=[10, 8], name="mul.y"
4645
)
4746
mul_out = block.create_var(
48-
dtype="float32", shape=[5, 8], lod_level=0, name="mul.out"
47+
dtype="float32", shape=[5, 8], name="mul.out"
4948
)
5049
mean_out = block.create_var(
51-
dtype="float32", shape=[1], lod_level=0, name="mean.out"
50+
dtype="float32", shape=[1], name="mean.out"
5251
)
5352
block.append_op(
5453
type="mul",
@@ -81,18 +80,17 @@ def check_sgd_optimizer(optimizer_attr):
8180
mul_x = block.create_parameter(
8281
dtype="float32",
8382
shape=[5, 10],
84-
lod_level=0,
8583
name="mul.x",
8684
optimize_attr=optimizer_attr,
8785
)
8886
mul_y = block.create_var(
89-
dtype="float32", shape=[10, 8], lod_level=0, name="mul.y"
87+
dtype="float32", shape=[10, 8], name="mul.y"
9088
)
9189
mul_out = block.create_var(
92-
dtype="float32", shape=[5, 8], lod_level=0, name="mul.out"
90+
dtype="float32", shape=[5, 8], name="mul.out"
9391
)
9492
mean_out = block.create_var(
95-
dtype="float32", shape=[1], lod_level=0, name="mean.out"
93+
dtype="float32", shape=[1], name="mean.out"
9694
)
9795
block.append_op(
9896
type="mul",
@@ -133,15 +131,12 @@ def test_vanilla_momentum_optimizer(self):
133131
mul_x = block.create_parameter(
134132
dtype="float32",
135133
shape=[5, 10],
136-
lod_level=0,
137134
name="mul.x",
138135
optimize_attr={'learning_rate': 1.1},
139136
)
140-
mul_y = block.create_var(
141-
dtype="float32", shape=[10, 8], lod_level=0, name="mul.y"
142-
)
137+
mul_y = block.create_var(dtype="float32", shape=[10, 8], name="mul.y")
143138
mul_out = block.create_var(
144-
dtype="float32", shape=[5, 8], lod_level=0, name="mul.out"
139+
dtype="float32", shape=[5, 8], name="mul.out"
145140
)
146141
block.append_op(
147142
type="mul",
@@ -153,9 +148,7 @@ def test_vanilla_momentum_optimizer(self):
153148
momentum_optimizer = self.MockMomentum(
154149
learning_rate=learning_rate, momentum=0.2
155150
)
156-
mean_out = block.create_var(
157-
dtype="float32", shape=[1], lod_level=0, name="mean.out"
158-
)
151+
mean_out = block.create_var(dtype="float32", shape=[1], name="mean.out")
159152
block.append_op(
160153
type="mean", inputs={"X": mul_out}, outputs={"Out": mean_out}
161154
)
@@ -192,25 +185,20 @@ def test_nesterov_momentum_optimizer(self):
192185
mul_x = block.create_parameter(
193186
dtype="float32",
194187
shape=[5, 10],
195-
lod_level=0,
196188
name="mul.x",
197189
optimize_attr={'learning_rate': 1.1},
198190
)
199-
mul_y = block.create_var(
200-
dtype="float32", shape=[10, 8], lod_level=0, name="mul.y"
201-
)
191+
mul_y = block.create_var(dtype="float32", shape=[10, 8], name="mul.y")
202192
mul_out = block.create_var(
203-
dtype="float32", shape=[5, 8], lod_level=0, name="mul.out"
193+
dtype="float32", shape=[5, 8], name="mul.out"
204194
)
205195
block.append_op(
206196
type="mul",
207197
inputs={"X": mul_x, "Y": mul_y},
208198
outputs={"Out": mul_out},
209199
attrs={"x_num_col_dims": 1},
210200
)
211-
mean_out = block.create_var(
212-
dtype="float32", shape=[1], lod_level=0, name="mean.out"
213-
)
201+
mean_out = block.create_var(dtype="float32", shape=[1], name="mean.out")
214202
block.append_op(
215203
type="mean", inputs={"X": mul_out}, outputs={"Out": mean_out}
216204
)
@@ -263,25 +251,20 @@ def test_adam_optimizer(self):
263251
mul_x = block.create_parameter(
264252
dtype="float32",
265253
shape=[5, 10],
266-
lod_level=0,
267254
name="mul.x",
268255
optimize_attr={'learning_rate': 1.1},
269256
)
270-
mul_y = block.create_var(
271-
dtype="float32", shape=[10, 8], lod_level=0, name="mul.y"
272-
)
257+
mul_y = block.create_var(dtype="float32", shape=[10, 8], name="mul.y")
273258
mul_out = block.create_var(
274-
dtype="float32", shape=[5, 8], lod_level=0, name="mul.out"
259+
dtype="float32", shape=[5, 8], name="mul.out"
275260
)
276261
block.append_op(
277262
type="mul",
278263
inputs={"X": mul_x, "Y": mul_y},
279264
outputs={"Out": mul_out},
280265
attrs={"x_num_col_dims": 1},
281266
)
282-
mean_out = block.create_var(
283-
dtype="float32", shape=[1], lod_level=0, name="mean.out"
284-
)
267+
mean_out = block.create_var(dtype="float32", shape=[1], name="mean.out")
285268
block.append_op(
286269
type="mean", inputs={"X": mul_out}, outputs={"Out": mean_out}
287270
)
@@ -321,45 +304,32 @@ def net(self, return_input=False, with_dropout=False, with_seed=False):
321304
program = framework.Program()
322305
block = program.global_block()
323306
mul_x = block.create_parameter(
324-
dtype="float32", shape=[5, 10], lod_level=0, name="mul.x"
325-
)
326-
mul_y = block.create_var(
327-
dtype="float32", shape=[10, 8], lod_level=0, name="mul.y"
307+
dtype="float32", shape=[5, 10], name="mul.x"
328308
)
309+
mul_y = block.create_var(dtype="float32", shape=[10, 8], name="mul.y")
329310
mul_out = block.create_var(
330-
dtype="float32", shape=[5, 8], lod_level=0, name="mul.out"
311+
dtype="float32", shape=[5, 8], name="mul.out"
331312
)
332313

333314
if with_dropout is True:
334315
mul_out_drop = block.create_var(
335316
dtype="float32",
336317
shape=[5, 8],
337-
lod_level=0,
338318
name="mul.out.dropout",
339319
)
340320
mul_out_mask = block.create_var(
341-
dtype="uint8", shape=[5, 8], lod_level=0, name="mul.out.mask"
321+
dtype="uint8", shape=[5, 8], name="mul.out.mask"
342322
)
343323
if with_seed is True:
344324
seed_out = block.create_var(
345325
dtype="int32", shape=[1], name="seed.out"
346326
)
347327

348-
b1 = block.create_parameter(
349-
dtype="float32", shape=[5, 8], lod_level=0, name="b1"
350-
)
351-
b1_out = block.create_var(
352-
dtype="float32", shape=[5, 8], lod_level=0, name="b1_out"
353-
)
354-
b2 = block.create_parameter(
355-
dtype="float32", shape=[5, 8], lod_level=0, name="b2"
356-
)
357-
b2_out = block.create_var(
358-
dtype="float32", shape=[5, 8], lod_level=0, name="b2_out"
359-
)
360-
mean_out = block.create_var(
361-
dtype="float32", shape=[1], lod_level=0, name="mean.out"
362-
)
328+
b1 = block.create_parameter(dtype="float32", shape=[5, 8], name="b1")
329+
b1_out = block.create_var(dtype="float32", shape=[5, 8], name="b1_out")
330+
b2 = block.create_parameter(dtype="float32", shape=[5, 8], name="b2")
331+
b2_out = block.create_var(dtype="float32", shape=[5, 8], name="b2_out")
332+
mean_out = block.create_var(dtype="float32", shape=[1], name="mean.out")
363333
block.append_op(
364334
type="mul",
365335
inputs={"X": mul_x, "Y": mul_y},
@@ -927,23 +897,15 @@ def net(self):
927897
program = framework.Program()
928898
block = program.global_block()
929899
mul_x = block.create_parameter(
930-
dtype="float32", shape=[5, 10], lod_level=0, name="mul.x"
931-
)
932-
mul_y = block.create_var(
933-
dtype="float32", shape=[10, 8], lod_level=0, name="mul.y"
900+
dtype="float32", shape=[5, 10], name="mul.x"
934901
)
902+
mul_y = block.create_var(dtype="float32", shape=[10, 8], name="mul.y")
935903
mul_out = block.create_var(
936-
dtype="float32", shape=[5, 8], lod_level=0, name="mul.out"
937-
)
938-
b1 = block.create_parameter(
939-
dtype="float32", shape=[5, 8], lod_level=0, name="b1"
940-
)
941-
b1_out = block.create_var(
942-
dtype="float32", shape=[5, 8], lod_level=0, name="b1_out"
943-
)
944-
mean_out = block.create_var(
945-
dtype="float32", shape=[1], lod_level=0, name="mean.out"
904+
dtype="float32", shape=[5, 8], name="mul.out"
946905
)
906+
b1 = block.create_parameter(dtype="float32", shape=[5, 8], name="b1")
907+
b1_out = block.create_var(dtype="float32", shape=[5, 8], name="b1_out")
908+
mean_out = block.create_var(dtype="float32", shape=[1], name="mean.out")
947909
block.append_op(
948910
type="mul",
949911
inputs={"X": mul_x, "Y": mul_y},

0 commit comments

Comments
 (0)