Skip to content

Commit

Permalink
removed hard coded BLOCK_SIZE
Browse files Browse the repository at this point in the history
  • Loading branch information
Virginia Adams committed Feb 14, 2024
1 parent 4ed3217 commit 3de9e60
Show file tree
Hide file tree
Showing 2 changed files with 1 addition and 2 deletions.
2 changes: 1 addition & 1 deletion src/levanter/models/flash_attention.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@
from levanter.models.attention import AttentionMask, materialize_mask


# TODO: tune
# TODO: Tune
BLOCK_SIZE = 128


Expand Down
1 change: 0 additions & 1 deletion src/levanter/models/gpt2.py
Original file line number Diff line number Diff line change
Expand Up @@ -195,7 +195,6 @@ def __call__(self, x: NamedArray, mask: Optional[AttentionMask | NamedArray], la
prng=k_drop,
attention_dtype=jnp.float32 if self.config.upcast_attn else None,
)
print(f"\n\nATTENTION OUTPUT: {attn_output}\n\n")
attn_output = self.c_proj(attn_output, key=k_out)

if self.config.upcast_attn:
Expand Down

0 comments on commit 3de9e60

Please sign in to comment.