Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

run on windows #26

Open
ziyanxzy opened this issue Jan 6, 2025 · 3 comments
Open

run on windows #26

ziyanxzy opened this issue Jan 6, 2025 · 3 comments

Comments

@ziyanxzy
Copy link

ziyanxzy commented Jan 6, 2025

compiled_fn = self.call_user_compiler(gm)
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^

File "C:\Users\SAS\miniforge3\envs\inf\Lib\site-packages\torch_dynamo\output_graph.py", line 1416, in call_user_compiler
return self._call_user_compiler(gm)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\SAS\miniforge3\envs\inf\Lib\site-packages\torch_dynamo\output_graph.py", line 1465, in _call_user_compiler
raise BackendCompilerFailed(self.compiler_fn, e) from e
torch._dynamo.exc.BackendCompilerFailed: backend='inductor' raised:
ImportError: DLL load failed while importing cuda_utils: The specified module could not be found.

You can suppress this exception and fall back to eager by setting:
import torch._dynamo
torch._dynamo.config.suppress_errors = True

@JeyesHan
Copy link
Collaborator

JeyesHan commented Jan 6, 2025

@ziyanxzy Thanks for your attention to Infinity. It seems that the error comes from Flex_Attn. Maybe you could run the code with --use_flex_attn=False
Besides, note that this codebase runs on Linux machines. We don't know whether it could be run on Windows. We are glad that you could share your experiences here, regardless of whether they were successful or unsuccessful.

@ziyanxzy
Copy link
Author

ziyanxzy commented Jan 7, 2025

Thanks han. i have use_flex_attn=0,but still have triton issue. can i do not use triton kernel in infinity?

@ziyanxzy
Copy link
Author

ziyanxzy commented Jan 7, 2025

and if i donot use flash_attn, it still has flash_attn_varlen_kvpacked_func function in crossattention, can i replace it?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants