ImportError: /home/operationgpt/anaconda3/envs/lyj_py10_torch230/lib/python3.10/site-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZN3c104cuda9SetDeviceEi 解决方式:重装flash attention 卸载已有的flash-attn, 输入pip uninstall flash-attn, 然后输入y 查看自己对应...
从异常上看,提示flash_attn_2_cuda.cpython-38-x86_64-linux-gnu.so这个库异常,这种未定义符号的异常,一般都是编译so时和当前环境不一致导致的 具体到flash_attn这个库,如果不是从源码编译,其对cuda版本和torch版本都是有要求的,所以在官方github的release上可以看到官方会提供很多不同cuda和torch版本的whl文件,...
ImportError: /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c1010TensorImpl27throw_data_ptr_access_errorEv Sorry, something went wrong. FrancescoSaverioZuppichinichanged the titleTests don't workAug 15, 2023 ...
File "/usr/local/lib/python3.10/dist-packages/flash_attn/flash_attn_interface.py", line 10, in import flash_attn_2_cuda as flash_attn_cuda ImportError: /usr/local/lib/python3.10/dist-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZN3c104cuda14ExchangeD...
安装:pip install flash_attn-2.3.5+cu116torch1.13cxx11abiFalse-cp310-cp310-linux_x86_64.whl -i https://mirrors.aliyun.com/pypi/simple/,加个镜像提速没毛病 注意:abiTrue的不行,False的可以,就很奇怪,True的会报错:...-linux-gnu.so: undefined symbol: _ZN3c104cuda9SetDeviceEi... 问题处理:...
RuntimeError: Failed to import transformers.models.llama.modeling_llama because of the following error (look up to see its traceback):/opt/miniconda3/envs/llama_xyj/lib/python3.10/site-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZN3c104cuda9SetDeviceEi ...
RuntimeError( RuntimeError: Failed to import transformers.models.qwen2.modeling_qwen2 because of the following error (look up to see its traceback): /usr/local/app/.local/lib/python3.10/site-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZN3c104cuda9Set...
installing all from scratch. torch: 2.3.0 flash-attn: 2.5.7 exllama: 0.0.19 Still getting the error: flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZN3c104cuda9SetDeviceEi
I build flash_attn from source code with pytorch.2.3.0 code >>> from flash_attn import flash_attn_2_cuda A100-80G cuda 12.1 my env: accelerate==0.29.3 aiofiles==23.2.1 aiohttp==3.9.5 aiosignal==1.3.1 altair==5.3.0 annotated-types==0.6.0 ...
get. (not sure if it's the same problem or not) import flash_attn_2_cuda as flash_attn_cuda ImportError: /home/chris/ai/data/text-generation-webui/installer_files/env/lib/python3.10/site-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: _ZNK3c106SymIntl...