cuda

On windows11, nvcc cannot show the correct version of CUDA

On windows11, nvcc cannot show the correct version of CUDA Question: The version I originally installed is 11.6. To match the tensorflow2.11.0, I had to install the v11.2 of CUDA, during which I first uinstall the newer version of CUDA(every thing about it) and then install the earlier version that is 11.2. However, after the …

Total answers: 1

Why is pip not letting me install torch==1.9.1+cu111 in a new conda env when I have another conda env that has exactly that version?

Why is pip not letting me install torch==1.9.1+cu111 in a new conda env when I have another conda env that has exactly that version? Question: When I run the pip install in the new conda env: (base) brando9~ $ pip install torch==1.9.1+cu111 torchvision==0.10.1+cu111 torchaudio==0.9.1 -f https://download.pytorch.org/whl/torch_stable.html Looking in links: https://download.pytorch.org/whl/torch_stable.html ERROR: Could not find a …

Total answers: 2

Can we really do parallel computing using numba python library in python?

Can we really do parallel computing using numba python library in python? Question: I am new to CUDA and was going through Running Python script on GPU. Performance with GPU is better than that without GPU (without GPU: 3.525673059999974, with GPU: 0.07701390800002628) for the following code executed in a colab notebook: from numba import jit, …

Total answers: 1

importing dask_cuda results in parse_memory_limit error

importing dask_cuda results in parse_memory_limit error Question: I’m trying to import dask_cuda as the examples: from dask_cuda import LocalCUDACluster from dask.distributed import Client But I receive the following error: ImportError Traceback (most recent call last) Input In [3], in <cell line: 1>() —-> 1 from dask_cuda import LocalCUDACluster File ~/miniconda3/lib/python3.8/site-packages/dask_cuda/__init__.py:5, in <module> 2 import dask.dataframe.shuffle …

Total answers: 1

PyTorch: CUDA is not available

PyTorch: CUDA is not available Question: I’m trying to run PyTorch on a NVIDIA Jetson Nano and my project requires me to use CUDA. I’m running on Ubuntu 18.04 and Python 3.10.6. I followed this guide to install CUDA 11.6. Then using the instructions on PyTorch.org I installed PyTorch using this command: pip install torch==1.12.0 …

Total answers: 1

How to run PyTorch on GPU with CUDA 10.2 on Windows 7?

How to run PyTorch on GPU with CUDA 10.2 on Windows 7? Question: I want to test run some finite-difference time domain (FDTD) simulations using parallelization on a GPU and compare the speed to CPU-based simulation runs. This is basically my very first attempt at using GPU acceleration for scientific computing. My system configuration: CPU: …

Total answers: 1

Errors in PyCuda indexing Numpy array of integers

Errors in PyCuda indexing Numpy array of integers Question: I am moving my first steps into PyCuda to perform some parallel computation and I came across a behavior I do not understand. I started from the very basic tutorial that can be found on PyCuda official website (a simple script to double all elements of …

Total answers: 1

I can't get output numbers with ctypes cuda

I can't get output numbers with ctypes cuda Question: cuda1.cu #include <iostream> using namespace std ; # define DELLEXPORT extern "C" __declspec(dllexport) __global__ void kernel(long* answer = 0){ *answer = threadIdx.x + (blockIdx.x * blockDim.x); } DELLEXPORT void resoult(long* h_answer){ long* d_answer = 0; cudaMalloc(&d_answer, sizeof(long)); kernel<<<10,1000>>>(d_answer); cudaMemcpy(&h_answer, d_answer, sizeof(long), cudaMemcpyDeviceToHost); cudaFree(d_answer); } main.py import …

Total answers: 1