gpu

How to determine if an application is using the GPU

I'm looking for a way to determine how to know whether an application is using the GPU with Objective-C. I want to be able to determine if any applications currently running on the system have work going on on the GPU (ie: a reason why the latest MacBook Pros would switch to the discrete graphics over the Intel HD graphics). I've tried ...

Nvidia Tesla vs 480 for CUDA programming

I am doing research on CUDA programming. i have the option to buy a single NVidia Tesla or buy around 4-5 NVidia 480? what do you recommend? ...

Solving problems involving more complex data structures with CUDA

So I read a bit about CUDA and GPU programming. I noticed a few things such that access to global memory is slow (therefore shared memory should be used) and that the execution path of threads in a warp should not diverge. I also looked at the (dense) matrix multiplication example, described in the programmers manual and the nbody probl...

How to synchronize cuda threads when they are in the same loop and we need to synchronize them to execute only a limited part

Hi all, I have written a code and Now I want to implement this on cuda GPU but I'm new to synchronization so please help me with this, It's little urgent to me. Below I'm presenting the code and I want to that LOOP1 to be executed by all threads (heance I want to this portion to take advantage of cuda and the remaining portion (the porti...

Double precision floating point in CUDA

Does CUDA support double precision floating point numbers ? Also need reasons for the same. ...

Periodic GPU performance problem

Hi folks! I have a WinForms application that uses XNA to animate 3D models in a control. The app have been doing just fine for months but recently I've started to experience periodic pauses in the animation. Setting out to investigate what is going on I have established these facts: It happens on my machine only, other machines works ...

Can GPU capabilities impact virtual machine performance?

While this many not seem like a programming question directly, it impacts my development activities and so it seems like it belongs here. It seems that more and more developers are turning to virtual environments for development activities on their computers, SharePoint development being a prime example. Also, as a trainer, I have virt...

DirectCompute versus OpenCL for GPU programming?

I have some (financial) tasks which should map well to GPU computing, but I'm not really sure if I should go with OpenCL or DirectCompute. I did some GPU computing, but it was a long time ago (3 years). I did it through OpenGL since there was not really any alternative back then. I've seen some OpenCL presentations and it looks really n...

Memory Size Allocation on GPU - opengl texture loading question

I'm loading pixels from an image which is 32 w by 32 height. The format I'm loading them in is ARGB via java. When I bind this to the video card, I can expect that the video card might use somewhere around 32*32*4 bytes, or 4K. Similarly, 1024 w, 1024 h would be 1024*1024*4 = 4MB. Is my understanding correct? Now I understand wh...

60K+ Sprites on the 360?

Hey everyone, Just wondering - throwing ideas in my head - about starting a new XNA project for the 360. I would like it to be retro-old school, and emulating scanlines and color palettes and such. As part of this idea, what I would ideally like to do is manually draw each and every pixel of the screen. So, worst-case scenario I would...

Performance differences between different CUDA SDK's?

If I want to re-write my application so that it leverages the power of nVidia's CUDA SDK, are there any differences at all in runtime performance between the different SDK offerings: C++, Java, Python? Is there any difference at all between these 3 SDK's, besides the obvious language being used? ...

Crunching number or scientific simulation applications that use matrix multiplication intensively

Hi, For my studies, we have code for matrix multiplication, for sizes between 1000-10000. It looks pretty fast, and uses GPU for calculations. As homework we need to find number crunching applications, with available source code, whose bottlenecks are in matrix multiplication. We will connect the program with the GPU code for matrix mul...

CUDA: injecting my own PTX function?

I would like to be able to use a feature in PTX 1.3 which is not yet implemented it the C interface. Is there a way to write my own function in PTX and inject into an existing binary? The feature I'm looking for is getting the value of %smid ...

Is it possible to achieve Huffman decoding in GPU?

We have a database encoded with Huffman coding. The aim here is to copy on the GPU it with its associated decoder; then on the GPU, decod the database and do stuff on this decoded database without copying back it on the CPU. I am far to be a Huffman specialist, but the few I know shows that it seems to be an algorithm essentially based ...

WPF/Silverlight performance while using GPU intensive applications

I might be mixing apples and oranges in this question since I'm noob in mentioned areas, so please try to understand what I mean. I read that WPF (and Silverlight, as an extension to it) uses GPU acceleration to render itself on the screen. If so, what happens when I have GPU-intensive application already running (say, game in windowed...

Next programming paradigm for CBE/GPU in the next years

Hi, in the last five years, there has been a rise in the use of GPU and CBE for parallelization of applications. Around 2005-2007 verything seemed to be programmed by hand, C, etc. Afterwards new unifying alternatives emerged like CUDA for GPU and lastly OpenCL. What do you think will be the programming paradigm for GPU/CBE in the for...

Learning GPGPU programming

My hands have been itching to learn GPGPU programming for some time. I finally have some time on my hands so I want to use it as wisely as possible. I'm really interested in your guys experience with GPGPU programming, any pointers, references to good literature, links to sites, interesting projects etc. My interests lie mainly in scie...

GPU Emulator for CUDA programming without the hardware

Question: Is there an emulator for a Geforce card that would allow me to program and test CUDA without having the actual hardware? Info: I'm looking to speed up a few simulations of mine in CUDA, but my problem is that I'm not always around my desktop for doing this development. I would like to do some work on my netbook instead, but...

Sending char ** data types to device

Hi, heres sumthin ive been battling for 4 days now. I have an array of character pointers which i want to send to device. Can somebody tell me how... Heres what ive tried so far: char **a; char **b; *a[0]="Foo1"; *a[1]=="Foo2"; cudaMalloc(void**)?,sizeof(?); cudamemcpy(b,a,sizeof(?),cudaMemcpyHostToDevice); How do i pass in the param...

WPF high performance drawing - should I group by brush color?

In Direct2D they recommend drawing similar things together, to avoid unnecessary GPU state changes. They also do some drawing operation reordering behind the scene just for that. I have to draw a lot of rectangles which can have one of two colors. I'm thinking of doing the drawing in two passes, one for the rectangles with the first col...