Skip to content

Commit

Permalink
[Cleanup][C-8] Replace Program.random_seed (#61518)
Browse files Browse the repository at this point in the history
  • Loading branch information
co63oc authored Feb 4, 2024
1 parent b5f6235 commit 22cf91f
Show file tree
Hide file tree
Showing 5 changed files with 8 additions and 13 deletions.
3 changes: 1 addition & 2 deletions test/legacy_test/test_fused_ec_moe_op.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,11 +19,10 @@

import paddle
import paddle.nn.functional as F
from paddle.base.framework import default_main_program
from paddle.incubate.nn.functional import fused_ec_moe
from paddle.nn.layer.common import Linear

default_main_program().random_seed = 42
paddle.seed(42)


class TestFusedEcMoEOp(OpTest):
Expand Down
7 changes: 3 additions & 4 deletions test/legacy_test/test_fused_feedforward_op.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,6 @@
import paddle
import paddle.incubate.nn.functional as incubate_f
import paddle.nn.functional as F
from paddle.base.framework import default_main_program
from paddle.nn.layer import transformer
from paddle.nn.layer.common import Dropout, Linear
from paddle.nn.layer.norm import LayerNorm
Expand Down Expand Up @@ -159,7 +158,7 @@ def FusedFFN(self):
return out, x.grad

def test_out_and_grad(self):
default_main_program().random_seed = 42
paddle.seed(42)
base_out, base_grad = self.Base()
fused_out, fused_grad = self.FusedFFN()
np.testing.assert_allclose(
Expand Down Expand Up @@ -233,7 +232,7 @@ def run_fused_feedforward(
):
main = paddle.static.Program()
startup = paddle.static.Program()
main.random_seed = 42
paddle.seed(42)
with paddle.static.program_guard(main, startup):
x = paddle.static.data(
name='x',
Expand Down Expand Up @@ -314,7 +313,7 @@ def run_base_ffn(
):
main = paddle.static.Program()
startup = paddle.static.Program()
main.random_seed = 42
paddle.seed(42)
with paddle.static.program_guard(main, startup):
x = paddle.static.data(
name='x',
Expand Down
3 changes: 1 addition & 2 deletions test/legacy_test/test_fused_feedforward_pass.py
Original file line number Diff line number Diff line change
Expand Up @@ -103,9 +103,8 @@ def get_value(self, use_pass=False):
)

main_prog = paddle.static.Program()
main_prog.random_seed = 1234
startup_prog = paddle.static.Program()
startup_prog.random_seed = 1234
paddle.seed(1234)

with paddle.static.program_guard(main_prog, startup_prog):
data = paddle.static.data(
Expand Down
3 changes: 1 addition & 2 deletions test/legacy_test/test_fused_multi_transformer_int8_op.py
Original file line number Diff line number Diff line change
Expand Up @@ -21,12 +21,11 @@
import paddle.nn.functional as F
from paddle import _legacy_C_ops, tensor
from paddle.base import core
from paddle.base.framework import default_main_program
from paddle.nn.layer.common import Dropout
from paddle.nn.layer.norm import LayerNorm
from paddle.nn.layer.transformer import _convert_attention_mask

default_main_program().random_seed = 42
paddle.seed(42)
np.random.seed(0)


Expand Down
5 changes: 2 additions & 3 deletions test/xpu/test_fused_feedforward_op_xpu.py
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,6 @@
import paddle
import paddle.incubate.nn.functional as incubate_f
import paddle.nn.functional as F
from paddle.base.framework import default_main_program
from paddle.nn.layer import transformer
from paddle.nn.layer.common import Dropout, Linear
from paddle.nn.layer.norm import LayerNorm
Expand Down Expand Up @@ -165,7 +164,7 @@ def FusedFFN(self):
return out, x.grad

def test_out_and_grad(self):
default_main_program().random_seed = 42
paddle.seed(42)
base_out, base_grad = self.Base()
fused_out, fused_grad = self.FusedFFN()
np.testing.assert_allclose(
Expand Down Expand Up @@ -199,7 +198,7 @@ def getShape(self):
class APITestStaticFusedFFN(unittest.TestCase):
def test_static(self):
paddle.enable_static()
default_main_program().random_seed = 42
paddle.seed(42)
dtype = "float32"
layer_norm_dtype = "float32"
batch_size = 1
Expand Down

0 comments on commit 22cf91f

Please sign in to comment.