65 void add(
const float *value)
69 void add(
const Type type,
const void *value,
size_t size)
74 values[
count] = (
void *)value;
78 template<
typename T,
typename... Args>
void add(
const T *first, Args... args)
154 LOG(FATAL) <<
"Request of GPU interop of a device which does not support it.";
static DBVT_INLINE btScalar size(const btDbvtVolume &a)
virtual int num_sort_partition_elements() const
DeviceKernelMask last_kernels_enqueued_
virtual int num_concurrent_busy_states(const size_t state_size) const =0
virtual void copy_from_device(device_memory &mem)=0
bool is_per_kernel_performance_
virtual bool supports_local_atomic_sort() const
virtual int num_concurrent_states(const size_t state_size) const =0
virtual void init_execution()=0
virtual void copy_to_device(device_memory &mem)=0
map< DeviceKernelMask, double > stats_kernel_time_
virtual unique_ptr< DeviceGraphicsInterop > graphics_interop_create()
string debug_active_kernels()
virtual bool synchronize()=0
virtual bool enqueue(DeviceKernel kernel, const int work_size, DeviceKernelArguments const &args)=0
virtual void * native_queue()
void debug_init_execution()
DeviceQueue(Device *device)
void debug_enqueue_begin(DeviceKernel kernel, const int work_size)
virtual void zero_to_device(device_memory &mem)=0
#define CCL_NAMESPACE_END
uint64_t DeviceKernelMask
ccl_gpu_kernel_postfix ccl_global const int ccl_global float const int work_size
static const int MAX_ARGS
void add(const device_ptr *value)
DeviceKernelArguments(const T *arg)
void add(const T *first, Args... args)
void add(const Type type, const void *value, size_t size)
void add(const float *value)
void add(const int32_t *value)
DeviceKernelArguments(const T *first, Args... args)
void add(const KernelFilmConvert *value)