I wonder if they are actually used on CUDA. And also provide a code (model/input) that can reproduce this behavior. titaiwangms added the onnx-needs-info label Jul 11, 2022 titaiwangms changed the title ONNX export | tensors spread across devices cuda and cpu [ONNX ] export | ...
Hello, I'm experiencing an issue similar to the one described in this comment. When exporting to ONNX, I encounter an error indicating that tensors are on different devices (CPU and CUDA). The error log is as follows: ERROR:root:Exportin...
The functionality to collect and display information about CUDA graphs was added in one version of Nsys. Then a release or two later, we saw that people were wanting to construct data mining on those fields and we added them to the export. Basically you have a version where we supported th...
print('Loading weights from checkpoint (' + args.trained_model + ')') if args.cuda: net.load_state_dict(copyStateDict(torch.load(args.trained_model))) else: net.load_state_dict(copyStateDict(torch.load(args.trained_model, map_location='cpu'))) if args.cuda: net = net.cuda() input...
os.environ["CUDA_VISIBLE_DEVICES"]="" importtensorflowastf fromtacotron.modelsimportcreate_model fromtacotron_hparamsimporthparams importshutil #with tf.device('/cpu:0'): inputs=tf.placeholder(tf.int32,[1,None],'inputs') input_lengths=tf.placeholder(tf.int32,[1],'input_lengths') ...
# devices: 1Device 0Name: NVIDIA GeForce RTX 2060Preferred: TRUEPower Envelope: DISCRETEAttachment: UNKNOWN# attached displays: 0GPU accessible RAM: 6,442 MBVRAM: 6,442 MBDedicated System RAM: 0 MBShared System RAM: 0 MBAPI version: 3.0 (OpenCL 3.0 CUDA)...
If you cannot interact, then the Save Dialog is probably open but not visible. If this is the case, it made have been moved to another location like the bottom of the screen or to a different monitor. You can test this by click the Enter or Esc key after clickin...
🐛 Describe the bug The exported model failed to do inference on cuda. import torch ep = torch.export.load('retina.pt2') gm = ep.module() gm(torch.rand(1, 3, 800, 1216)) # success gm = ep.module().cuda() gm(torch.rand(1, 3, 800, 1216).cud...
import os #os.environ["CUDA_VISIBLE_DEVICES"] = "-1" import tensorflow as tf from tensorflow import keras from tensorflow.keras import layers from tensorflow.keras.applications import DenseNet121, MobileNetV3Small img_inputs = keras.Input(shape=(128, 128, 3)) backend = MobileNetV3Small(include...
when I run docker build -f docker/ubuntu-18.04.Dockerfile --build-arg CUDA_VERSION=10.1 --tag=tensorrt .,It occur error like: Sending build context to Docker daemon 170.8MB Step 1/10 : ARG CUDA_VERSION=10.1 Step 2/10 : FROM nvcr.io/nvidia/cuda:${CUDA_VERSION}-cudnn7-devel...