"torch.logsumexp": (_always_executable, logsumexp), "torch.log_softmax": (_always_executable, log_softmax), "torch.softmax": (_always_executable, softmax), "torch.nn.functional.scaled_dot_product_attention": ( _scaled_dot_product_attention_check,107 changes: 102 additions & 5 deletions...