Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Check failed: operand_dim < ins->operand(0)->shape().rank() (2 vs. 2)Does not support this kind of Gather. #958

Open
caixiiaoyang opened this issue Sep 28, 2023 · 2 comments

Comments

@caixiiaoyang
Copy link

Please describe the bug
Aborted (core dumped)
Please describe the expected behavior
I have two A100 GPUs, when I use alpa.PipeshardParallel(), the model runs fine, when I use alpa.ShardParallel(), I get a core dumped.
This error occurs during the auto_sharding process. Check failed: operand_dim < ins->operand(0)->shape().rank() (2 vs. 2)Does not support this kind of Gather.I would like to know under what circumstances this error occurs. Can you provide some troubleshooting ideas and specific errors, as shown in the screenshot below?
Screenshots
image

Code snippet to reproduce the problem

Additional information
Add any other context about the problem here or include any logs that would be helpful to diagnose the problem.

@zigzagcai
Copy link

zigzagcai commented Nov 22, 2023

I also met this issue when trying to use alpa.ShardParallel() or alpa.PipeshardParallel() to auto parallelize my llama model.

image

@caixiiaoyang
Copy link
Author

I also met this issue when trying to use alpa.ShardParallel() or alpa.PipeshardParallel() to auto parallelize my llama model.

image

I also encountered this problem in the process of parallelizing llama. Is your problem solved?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants