I am using numba cuda to calculate a function. The code is simply to add up all the values into one result, but numba cuda gives me a different result from nu
Recently I'm searching for info about if PCIe devices are involved in the uefi secure boot, and if so, how it is done. From the uefi specification, the main boo
I have been having this error, and I don't know why, especially since I am following someone's code exactly and the person had no error when running this img_sh
I am training my neural network built with PyTorch under Google Colab Pro+ (Tesla P100-PCIE GPU) but encounters the following strange phenomenon: The amount of
I'm developing a web component using WebGL2 (and three.js) with OES_element_index_uint enabled. I'm drawing a geometry using indexed vertices and I'm seeing th
I want to allocate 4.5GB to my openCL program in android phone wiht 8GB memory. But I found the memory size from CL_DEVICE_GLOBAL_MEM_SIZE is much lower than th
I want to allocate 4.5GB to my openCL program in android phone wiht 8GB memory. But I found the memory size from CL_DEVICE_GLOBAL_MEM_SIZE is much lower than th
The result is below,I run the project stylegan2, but it fails. So I need help. The link is https://github.com/NVlabs/stylegan2 File "/home/ubuntu/worksp
Adding a GPU to a TF takes a long time (about 5 minutes). 2020-10-13 20:40:44.526254: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successful
docker stack deploy with GPU , but can't find nvidia devices description: When I use docker-compose up start the program, the code works well! But when I use
I am running T5-base-grammar-correction for grammer correction on my dataframe with text column from happytransformer import HappyTextToText from happytransform
I am using ROCM's hip for programming. When I use hipLaunchKernelGGL, everything works fine. But when I use hipLaunchKernel(), I always get bizarre errors. Erro
Purpose: I want to render an image in the screen and save it in my disk. Description: I have a render target view. I have a input shader resource view with its
I'm having a weird error executing an opencl kernel, When I'm trying to build the opencl kernel using the clBuildProgram() execution err = clBuildProgram(progra
There is an object detection pre-trained model i.e. Yolov3/v4-tiny, when the algorithm is implemented in python, everything looked good, there is no lag while p
My system has a GPU. When I run Tensorflow on it, TF automatically detects GPU and starts running the thread on the GPU. How can I change this? I.e. how can I r
model = lgbm.LGBMClassifier(n_estimators=1250, num_leaves=128,learning_rate=0.009,verbose=1)`enter code here` using the LGBM classifier is there way to use t
When I run nvidia-smi, I get the following message: Failed to initialize NVML: Driver/library version mismatch An hour ago I received the sa
I am trying to use GPU with Tensorflow. My Tensorflow version is 2.4.1 and I am using Cuda version 11.2. Here is the output of nvidia-smi. +--------------------
I have done some GPU programming in the past and realized that unlike in typical sequential algorithms, where for example you may have an O(nl