Skip to content

Commit

Permalink
fix incorrect sharding without zero (#5545)
Browse files Browse the repository at this point in the history
Co-authored-by: Edenzzzz <[email protected]>
  • Loading branch information
Edenzzzz and Edenzzzz committed Apr 2, 2024
1 parent e614aa3 commit 7e0ec5a
Showing 1 changed file with 4 additions and 2 deletions.
6 changes: 4 additions & 2 deletions colossalai/shardformer/shard/shard_config.py
Original file line number Diff line number Diff line change
Expand Up @@ -74,8 +74,10 @@ def _turn_on_all_optimization(self):
self.enable_fused_normalization = True
self.enable_flash_attention = True
self.enable_jit_fused = True
self.enable_sequence_parallelism = True
self.enable_sequence_overlap = True
# This can cause non-in-place param sharding when used without ZeRO.
# It may also slow down training when seq len is small. Plz enable manually.
# self.enable_sequence_parallelism = True
# self.enable_sequence_overlap = True

def _infer(self):
"""
Expand Down

0 comments on commit 7e0ec5a

Please sign in to comment.