gpu

GPU Benchmark

What software for GPU benchmarking do you recommend? Also, any suggestions on how to use it? ...

Monitor the Graphics card usage

What software can I use to monitor how much of the graphics card is used when I run a certain application? I want to see how much my application uses the GPU ...

GPU monitoring software

I'm interested in a GPU monitoring software that would log the processes and the memory (video card memory) being used by the processes. I'm opened to all kind of solutions, even a software + some scripts to help me log this kind of data. I need the software to run under Windows Vista ...

How to obtain OpenCL SDK?

I was perusing http://www.khronos.org/ web site and only found headers for OpenCL (not OpenGL which I don't care about). How can I obtain OpenCL SDK? ...

Can I exploit GPU to do AES encryption from .NET? If so, how?

Interesting paper from Trinity College of Dublin: AES Encryption Implementation and Analysis on Commodity Graphics Processing Units Their technique uses openGL to enlist the GPU to do the numeric transforms required by AES. How difficult would it be to expose this capability - performing stream encryption - via a managed .NET library?...

C#: Perform Operations on GPU, not CPU (Calculate Pi)

Hello, I've recently read a lot about software (mostly scientific/math and encryption related) that moves part of their calculation onto the GPU which causes a 100-1000 (!) fold increase in speed for supported operations. Is there a library, API or other way to run something on the GPU via C#? I'm thinking of simple Pi calculation. I h...

cudaMemcpyToSymbol do not copy data

I want to use constant memory which will be accessed by all threads across all of my kernels. The declaration is something like this extern constant float smooth [8 * 1024]; I am copying data to this variable using cudaMemcpyToSymbol("smooth", smooth_local, smooth_size, 0, cudaMemcpyHostToDevice); smooth_size = 7K bytes It was givi...

Extracting GPU information using C++ (and hopefully WMI?)

Hi everyone! I'd like to know how to extract GPU using C++. However, the information that I'm interested in is not the data that is available from the Win32_VideoController class (ahh.. the Murphy's Laws of Programming). Specifically, I'm interested in the (Memory and Shader clocks). I'd like to know how to get at that information. I'...

Getting GPU clock speeds with SetupDiEnumDeviceInfo

Hi everyone! I posted a question earlier regarding obtaining GPU clock speeds, but I guess the thread appeared as already answered since someone had replied to it. I'd been advised by one of your members to try extracting GPU clock speeds using SetupDiEnumDeviceInfo. However, I looked around at some examples, like this one: http://ww...

Any open source library for sparse linear algebra in OpenCL?

I am looking for some sparse linear algebra OpenCL kernels such as blas vector/vector operations and matrix / vector operations but with sparse data structures. Ideally that library would feature most of scipy.sparse but using OpenCL kernels instead of scalar C code wrapped in python ndarrays. After some googling I could not find anythi...

How to scale cholesky factorization on multiple GPUs

Hello folks, I have implemented Cholesky Factorization for solving large linear equation on GPU using ATI Stream SDK. Now I want to exploit computation power of more and more GPUs and I want to run this code on multiple GPUs. Currently I have One Machine and One GPU installed on it and cholesky factorization is running properly. I want...

Is GPGPU a hack ?

Hello Folks, I had started working on GPGPU some days ago and successfully implemented cholesky factorization with good performacne and I attended a conference on High Performance Computing where some people said that "GPGPU is a Hack". I am still confused what does it mean and why they were saying it hack. One said that this is hack b...

ATI Stream SDK on ubuntu 9.04

Hello All, I have used ATI Stream SDK on windows XP SP3 and implemented one algorithm on GPU. But Now I am interested in scaling this algorithm on multiple GPUs on mutiple machines I switched to UBUNTU to use MPI ( To send messages ). I googled this but I got references for installation on SLES and RHEL but I am looking for UBUNTU 9.04...

high precision math on GPU

I'm interested in implementing an algorithm on the GPU using HLSL, but one of my main concerns is that I would like a variable level of precision. Are there techniques out there to emulate 64bit precision and higher that could be implemented on the GPU. Thanks! ...

Apply a ColorMatrix making use of GPU

I have a C# application that recolors an image using ColorMatrix. I understand that ColorMatrix doesn't make use of the GPU. What would be the best avenue to explore if I wanted to use the GPU to recolor the image? Pointers to where look in any suggested libraries would be appreciated, examples even more so! ...

Nvidia GPU Computing SDK and Xcode 3.1

I'm trying to compile some code for OpenCL (and the NV Computing SDK) in Xcode. I (think) I've set the search paths correctly, but when I try to compile the source code from the example (oclDeviceQuery, which builds perfectly using sudo make) I get the following error: /Developer/SDKs/MacOSX10.6.sdk/System/Library/Frameworks/OpenCL.fram...

Death of the Cell processor

Hi, in the last times I heard lots of people claiming that the Cell processor is dead, mainly due to the following reasons: a) lack of support in the new playstation 3, as the user can not install linux b) the increasing processing power of the GPU's and its costs sinking c) the existence of a unified programming approach (openCL) for ...

Why do I get a CL_MEM_OBJECT_ALLOCATION_FAILURE?

I'm allocating a cl_mem buffer on a GPU and work on it, which works fine until a certain size is exceeded. In that case the allocation itself succeeds, but execution or copying does not. I do want to use the device's memory for faster operation so I allocate like: buf = clCreateBuffer (cxGPUContext, CL_MEM_WRITE_ONLY, buf_size, NULL, &c...

Do graphic cards have instruction sets of their own?

Do graphic cards have instruction sets of their own? I assume they do, but I have been wondering if it is proprietary or if there is some sort of open standard. Is every GPU instruction preceded by a CPU instruction or is it seamless? That is does OpenGL or DirectX call on the driver layer via the CPU which then sends a GPU instruction...

How to tell the graphic card memory usage?

My application seems to be slow, but in terms of CPU and RAM, it seems that it is OK. So, I want to know how much memory of the graphic card I am using. I've seen some questions about this on SO, but they talk about Linux or NVidia. I would like to have this information for ATI cards on Windows. Thanks. ...