Skip to content

upload for Multi-graphics inference#10

Open
qqtang-code wants to merge 1 commit intoByteDance-Seed:mainfrom
qqtang-code:main
Open

upload for Multi-graphics inference#10
qqtang-code wants to merge 1 commit intoByteDance-Seed:mainfrom
qqtang-code:main

Conversation

@qqtang-code
Copy link
Copy Markdown

On the 3090 or 4090 graphics card, the maximum block_size = 64 is supported. At the same time, you need to add with torch.cuda.device(x.device): and change the mask position to support multi-card reasoning

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants