Hi all,
I have a GeForce 285 and 580. The 285 is my primary since I am using that with the monitor and the 580 totally dedicated to Cuda computing.
I wrote my code, i specified “compute_20,sm_20” and also
int deviceCount, device;
cudaDeviceProp prop;
// Num. of Device Count
cudaGetDeviceCount(&deviceCount);
printf("\nDevice found:\n");
for(i=0; i<deviceCount; i++) {
cudaGetDeviceProperties(&prop, i);
printf("%s\n", prop.name);
}
// Device Selection
device = 0;
cudaSetDevice(device);
// Current Device Detection
cudaGetDeviceProperties(&prop, device);
printf("Using device %d: %s \n", device, prop.name);
To select the 580 (that is 0, 285 is 1)
Well, when I run it, if I set a thread/block until 512 everything fine, but from 513 going over I get error…
I guess it is using anyway the 285 since it has the 512 threads/block limit, while on 580 this is 1024.
What could I do?