Skip to content
Open
Changes from 1 commit
Commits
Show all changes
43 commits
Select commit Hold shift + click to select a range
cb03cb2
disable enable_use_compute_stream in data_parallel
CPFLAME Jun 21, 2022
296c31c
set loss_dict to rank0 in graph_base
CPFLAME Jun 21, 2022
917370f
add scale_mask_softmax_fusion
CPFLAME Jun 21, 2022
5863532
change libai set_stage(stage_id, placement)
CPFLAME Jun 21, 2022
c563f52
fix all2all in linear and bert
CPFLAME Jun 21, 2022
972bbb4
tuning pipeline stage for layers
CPFLAME Jun 21, 2022
ff16bc4
set ONEFLOW_FUSE_OPTIMIZER_UPDATE_CAST
CPFLAME Jun 21, 2022
33f76bd
add rdma in LiBai
CPFLAME Jun 21, 2022
f648488
set input placement to cpu stage
CPFLAME Jun 21, 2022
05dbef9
reformat code
CPFLAME Jun 21, 2022
e0725f3
fix transformer_layer && flake code
CPFLAME Jun 21, 2022
fce3f7c
fix input_placement_device in get_batch
CPFLAME Jun 22, 2022
f3cb41e
fix input_placement_device in eager trainer get_batch
CPFLAME Jun 22, 2022
1c90aa3
fix t5 fused_scale_mask_softmax_dropout bug
CPFLAME Jun 22, 2022
16ca685
add multihead_attn_fusion in libai_bench
CPFLAME Jun 22, 2022
ba77596
set roberta config (scale_mask_softmax_fusion=True)
xiezipeng-ML Jun 23, 2022
5b38a60
update roberta_pretrain.py
xiezipeng-ML Jun 23, 2022
324302e
update robberta config
xiezipeng-ML Jun 23, 2022
e848dbc
finish model_test
CPFLAME Jun 23, 2022
76bae8b
Merge branch 'libai_bench_merge_main' of github.com:Oneflow-Inc/libai…
CPFLAME Jun 23, 2022
c640cb4
refine roberta
xiezipeng-ML Jun 23, 2022
b8793b0
update roberta.py
xiezipeng-ML Jun 23, 2022
3ae0c67
reformat code
CPFLAME Jun 23, 2022
a2e0dc2
Merge branch 'libai_bench_merge_main' of github.com:Oneflow-Inc/libai…
CPFLAME Jun 23, 2022
6b7d878
Merge branch 'main' of github.com:Oneflow-Inc/libai into libai_bench_…
CPFLAME Jun 24, 2022
f8a41f1
Merge branch 'libai_bench_merge_main' of github.com:Oneflow-Inc/libai…
CPFLAME Jun 24, 2022
71ba2db
disable_straighten_algorithm
CPFLAME Jun 24, 2022
93e5836
Merge branch 'libai_bench_merge_main' of github.com:Oneflow-Inc/libai…
CPFLAME Jun 24, 2022
24105ad
delete disable_straighten_algorithm
CPFLAME Jun 29, 2022
f23e60e
delete disable_straighten_algorithm
CPFLAME Jun 29, 2022
9a4eb0b
Merge branch 'libai_bench_merge_main' of github.com:Oneflow-Inc/libai…
CPFLAME Jun 29, 2022
ed3fcd6
concate local tensor for grad acc in graph
CPFLAME Jul 1, 2022
80b0b31
Merge branch 'main' of github.com:Oneflow-Inc/libai into libai_bench_…
CPFLAME Jul 1, 2022
27a8f58
Merge branch 'libai_bench_merge_main' of github.com:Oneflow-Inc/libai…
CPFLAME Jul 1, 2022
e032a6b
flake code
CPFLAME Jul 1, 2022
9fc504c
Merge branch 'libai_bench_merge_main' of github.com:Oneflow-Inc/libai…
CPFLAME Jul 1, 2022
68e93a6
merge main
CPFLAME Jul 5, 2022
7640b1a
set multihead_attn_fusion to model_config
CPFLAME Jul 5, 2022
cd24ea0
Merge branch 'main' of github.com:Oneflow-Inc/libai into libai_bench
CPFLAME Jul 6, 2022
d10d1d9
Merge branches 'libai_bench' and 'main' of github.com:Oneflow-Inc/lib…
CPFLAME Jul 8, 2022
77a85c2
Merge branch 'main' of github.com:Oneflow-Inc/libai into libai_bench
CPFLAME Jul 18, 2022
b1c7d32
Merge branch 'main' of github.com:Oneflow-Inc/libai into libai_bench
CPFLAME Aug 2, 2022
231ddd8
Merge branch 'main' of github.com:Oneflow-Inc/libai into libai_bench
CPFLAME Mar 31, 2023
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Next Next commit
disable enable_use_compute_stream in data_parallel
  • Loading branch information
CPFLAME committed Jun 21, 2022
commit cb03cb2f5485f3deab26da778a689e6f59f216ad
9 changes: 4 additions & 5 deletions libai/models/utils/graph_base.py
Original file line number Diff line number Diff line change
Expand Up @@ -89,12 +89,11 @@ def __init__(
"python3 -m pip install --pre oneflow -f https://staging.oneflow.info/branch/release-auto_parallel-v0.1/[PLATFORM]" # noqa
)

# Enable compute_stream for computation and communication with the same cuda stream.
# Enable cuda stream for computation and communication as the same stream.
# This will reduce memory when using model parallelism.
# if dist_util.is_tensor_model_parallel() or dist_util.is_pipeline_model_parallel():

# Enable compute_stream by default.
flow.boxing.nccl.enable_use_compute_stream(True)
dist_util = dist.get_dist_util()
if dist_util.is_tensor_model_parallel() or dist_util.is_pipeline_model_parallel():
flow.boxing.nccl.enable_use_compute_stream(True)

def build(self, **kwargs):
if self.is_train:
Expand Down