Skip to content

Commit

Permalink
fix typos
Browse files Browse the repository at this point in the history
  • Loading branch information
Jackmin801 authored and samsja committed Jul 11, 2024
1 parent 5e23acd commit debf131
Show file tree
Hide file tree
Showing 3 changed files with 3 additions and 3 deletions.
2 changes: 1 addition & 1 deletion CONTRIBUTING.md
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
# Development workflow

This is the develpment workflow of prime intellect to build upon hivemind
This is the development workflow of prime intellect to build upon hivemind

## Install dependencies

Expand Down
2 changes: 1 addition & 1 deletion open_diloco/train_fsdp.py
Original file line number Diff line number Diff line change
Expand Up @@ -116,7 +116,7 @@ def cast_str_to_list(cls, values: dict[str, Any]) -> dict[str, Any]:
class Config(BaseConfig):
path_model: str = "PrimeIntellect/llama-150m-fresh"
torch_compile: bool = True
attn_implementation: str = "flash_attention_2"
attn_implementation: str = "sdpa"
# Data
dataset_name_or_path: str = "allenai/c4"
seq_length: int = 1024
Expand Down
2 changes: 1 addition & 1 deletion open_diloco/utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -117,7 +117,7 @@ def get_compression_kwargs(hivemind_compression: str) -> dict:

def found_inf_grad(optimizer: torch.optim.Optimizer, scaler: torch.cuda.amp.GradScaler) -> bool:
"""
this function check if the scaler has found inf grad for the optimizer. It does by looking up the optimzier state
this function check if the scaler has found inf grad for the optimizer. It does by looking up the optimizer state
regsited inside the scaler. Code is mostly copied/inspired by the torch GradScaler codebase.
"""
if not scaler._enabled:
Expand Down

0 comments on commit debf131

Please sign in to comment.