Skip to content

optimize the generation of attention mask#331

Open
imh966 wants to merge 1 commit into
deepspeedai:mainfrom
imh966:fix_attention_mask
Open

optimize the generation of attention mask#331
imh966 wants to merge 1 commit into
deepspeedai:mainfrom
imh966:fix_attention_mask

Conversation

@imh966
Copy link
Copy Markdown

@imh966 imh966 commented Jan 13, 2024

Hi, I found that the attention mask tensor is created on cpu, leading to inefficient operations on attention mask and an extra H2D operation.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant