Category "nvidia"

tensorflow: Fail to find dnn implementation

I'm trying to run my code Keras CuDNNGRU on tensorflow using gpu but it always get error "Fail to find dnn implementation" even though I already installed CUDA

pytools.prefork.ExecError: error invoking 'nvcc --version': [Errno 2] No such file or directory

I have install pycuda and I am trying to test it with code below. import pycuda.driver as cuda import pycuda.autoinit from pycuda.compiler import SourceModule

GPU memory is empty, but CUDA out of memory error occurs

During training this code with ray tune(1 gpu for 1 trial), after few hours of training (about 20 trials) CUDA out of memory error occurred from GPU:0,1. And ev

Pytorch CUDA error: no kernel image is available for execution on the device on RTX 3090 with cuda 11.1

If I run the following: import torch import sys print('A', sys.version) print('B', torch.__version__) print('C', torch.cuda.is_available()) print('D', torch.bac

GPU-accelerated video processing with ffmpeg

I want to use ffmpeg to accelerate video encode and decode with an NVIDIA GPU. From NVIDIA's website: NVIDIA GPUs contain one or more hardware-based decod

How to generate trace on armv8 Linux - CoreSight ETM - NVIDIA DRIVE AGX

I'd like to analyze worst case execution time of the programs. For this, there is a tool and expecting trace output of the program(specifically CoreSight - Embe

Primusrun/Optirun allegedly cannot locate/open config directory: “/etc/bumblebee/xorg.conf.d”

I have recently switched from open source drivers to nvidia, to bumblebee as instructed by ubuntuforums.org users to better use my two gpu's capabilities. It al

Nvidia NVML Driver/library version mismatch [closed]

When I run nvidia-smi, I get the following message: Failed to initialize NVML: Driver/library version mismatch An hour ago I received the sa

tensorflow/stream_executor/cuda/cuda_driver.cc:328] failed call to cuInit: CUDA_ERROR_UNKNOWN: unknown error

I am trying to use GPU with Tensorflow. My Tensorflow version is 2.4.1 and I am using Cuda version 11.2. Here is the output of nvidia-smi. +--------------------

Using String parameter for nvidia triton

I'm trying to deploy a simple model on the Triton Inference Server. It is loaded well but I'm having trouble formatting the input to do a proper inference reque

Could not load dynamic library 'libcublas.so.10'; dlerror: libcublas.so.10: cannot open shared object file: No such file or directory;

When I try to run a python script , which uses tensorflow, it shows following error ... 2020-10-04 16:01:44.994797: I tensorflow/stream_executor/platform/defaul