Attempting to compiler the following IR causes LLVM to crash (compiler explorer): target triple = "mips64el-unknown-linux-gnuabi64" define fp128 @foo() { call { i8, i128 } @bar() ret fp128 0xL00000000000000000000000000000000 } define { i...
Function signature fuzzer(ongoing) Static call, closure call, method call, interface call,reflect.{ValueOf(target),MakeFunc(x, target),Method(x)}.{Call,Interface}, called fromdefer, called fromgo, called as a finalizer {Big,Small} {argument,result} in {memory,registers} and {not addressed,...
[rank0]: RuntimeError: Output 0 of ViewBackward0 is a view and its base or another view of its base has been modified inplace. This view is the output of a function that returns multiple views. Such functions do not allow the output views to be modified inplace. You should replace t...
ERROR:asyncio:Exception in callback functools.partial(<function _raise_exception_on_finish at 0x7f351ca30ee0>, error_callback=<bound method AsyncLLMEngine._error_callback of <vllm.engine.russianblue_async_llm_engine.RussianBlueAsyncLLMEngine object at 0x7f3510194100>>) handle: <Handle functools...
🐛 Describe the bug When calling into cuSparseLt function from different threads, we can get this error. Notably this happens in the BW pass. I have a small repro code - if the BW pass runs on the main thread, the error does not show up. ...
<function at remote 0x7ffff72d5990>, 'test_gradient_checkpoint_scope_twice': <functio...(truncated)) at ../Python/pythonrun.c:1291 #16 0x00005555557ae3fb in run_mod (mod=<optimized out>, filename=<optimized out>, globals={'__name__': '__main__', '__doc__': '\nTest :mod:...
My question is: will cuSPARSELt offer a better acceleration than CUTLASS? The error: RuntimeError: cuSPARSELT not supported on your machine. occurred when I ran the function to_sparse_semi_structured() The code inside this function is: if self._FORCE_CUTLASS: ... else: compressed_tensor =...
run_endpoint_function( File "/home/yangjie/.conda/qwen4jay/lib/python3.10/site-packages/fastapi/routing.py", line 191, in run_endpoint_function return await dependant.call(**values) File "/data1/yangjie/vllm/vllm/entrypoints/openai/api_server.py", line 218, in create_embedding generator =...
48 bits virtual Byte Order: Little Endian CPU(s): 128 On-line CPU(s) list: 0-127 Vendor ID: GenuineIntel Model name: Intel(R) Xeon(R) Platinum 8369B CPU @ 2.90GHz CPU family: 6 Model: 106 Thread(s) per core: 2 Core(s) per socket: 32 Socket(s): 2 Stepping: 6 BogoMIPS ...
inkey_averagesreturnself._function_events.key_averages(File"python3.10/site-packages/torch/autograd/profiler_util.py",line332,inkey_averagesstats[get_key(evt,group_by_input_shapes,group_by_stack_n)].add(evt)File"lib/python3.10/site-packages/torch/autograd/profiler_util.py",line699,inaddself....