23 MetalDeviceQueue(MetalDevice *device);
52 id<MTLComputeCommandEncoder> get_compute_encoder(
DeviceKernel kernel);
53 id<MTLBlitCommandEncoder> get_blit_encoder();
55 MetalDevice *metal_device_;
56 MetalBufferPool temp_buffer_pool_;
58 API_AVAILABLE(macos(11.0), ios(14.0))
59 MTLCommandBufferDescriptor *command_buffer_desc_ =
nullptr;
60 id<MTLDevice> mtlDevice_ = nil;
61 id<MTLCommandQueue> mtlCommandQueue_ = nil;
62 id<MTLCommandBuffer> mtlCommandBuffer_ = nil;
63 id<MTLComputeCommandEncoder> mtlComputeEncoder_ = nil;
64 id<MTLBlitCommandEncoder> mtlBlitEncoder_ = nil;
65 API_AVAILABLE(macos(10.14), ios(14.0))
66 id<MTLSharedEvent> shared_event_ = nil;
67 API_AVAILABLE(macos(10.14), ios(14.0))
68 MTLSharedEventListener *shared_event_listener_ = nil;
71 dispatch_queue_t event_queue_;
72 dispatch_semaphore_t wait_semaphore_;
79 std::vector<CopyBack> copy_back_mem_;
82 uint64_t command_buffers_submitted_ = 0;
83 uint64_t command_buffers_completed_ = 0;
86 void close_compute_encoder();
87 void close_blit_encoder();
89 bool verbose_tracing_ =
false;
90 bool label_command_encoders_ =
false;
99 std::vector<TimingData> command_encoder_labels_;
100 bool profiling_enabled_ =
false;
103 id<MTLCounterSampleBuffer> counter_sample_buffer_ = nil;
104 std::atomic<uint64_t> counter_sample_buffer_curr_idx_ = 0;
106 void flush_timing_stats();
114 double last_completion_time_ = 0.0;
118 id<MTLCaptureScope> mtlCaptureScope_ = nil;
120 int capture_dispatch_counter_ = 0;
121 bool capture_samples_ =
false;
122 int capture_reset_counter_ = 0;
123 bool is_capturing_ =
false;
124 bool is_capturing_to_disk_ =
false;
125 bool has_captured_to_disk_ =
false;
static DBVT_INLINE btScalar size(const btDbvtVolume &a)
virtual int num_sort_partition_elements() const
virtual int num_concurrent_busy_states(const size_t state_size) const =0
virtual void copy_from_device(device_memory &mem)=0
virtual bool supports_local_atomic_sort() const
virtual int num_concurrent_states(const size_t state_size) const =0
virtual void init_execution()=0
virtual void copy_to_device(device_memory &mem)=0
virtual bool synchronize()=0
virtual bool enqueue(DeviceKernel kernel, const int work_size, DeviceKernelArguments const &args)=0
virtual void * native_queue()
virtual void zero_to_device(device_memory &mem)=0
#define CCL_NAMESPACE_END
ccl_gpu_kernel_postfix ccl_global const int ccl_global float const int work_size
unsigned __int64 uint64_t