Is it possible to synchronize two CUDA streams without blocking the host? I know there's cudaStreamWaitEvent
, which is non-blocking. But what about the creation and destruction of the events using cudaEventCreate
and cudaEventDestroy
.
The documentation for cudaEventDestroy
says:
In case event has been recorded but has not yet been completed when cudaEventDestroy() is called, the function will return immediately and the resources associated with event will be released automatically once the device has completed event.
What I don't understand here is what the difference is between a recorded event and a completed event. Also this seems to imply that the call is blocking if the event has not yet been recorded.
Anyone who can shed some light on this?
You're on the right track by using
cudaStreamWaitEvent
. Creating events does carry some cost, but they can be created during your application start-up to prevent the creation time from being costly during your GPU routines.An event is recorded when you you put the event into a stream. It is completed after all activity that was put into the stream before the event has completed. Recording the event basically puts a marker into your stream, which is the thing that enables
cudaStreamWaitEvent
to stop forward progress on the stream until the event has completed.