5#include "testing/testing.h"
20#define ITEMS_NUM 10000
26 int *data = (
int *)userdata;
28 *((
int *)tls->userdata_chunk) += index;
33 void *__restrict join_v,
34 void *__restrict userdata_chunk)
36 int *join = (
int *)join_v;
37 int *chunk = (
int *)userdata_chunk;
51 settings.min_iter_per_thread = 1;
53 settings.userdata_chunk = &
sum;
54 settings.userdata_chunk_size =
sizeof(
sum);
78 int *data = (
int *)item;
79 int *
count = (
int *)userdata;
81 EXPECT_TRUE(data !=
nullptr);
111 if (data[i] ==
nullptr) {
118 for (i = 0; i <
ITEMS_NUM - 5; i += 23) {
119 for (
int j = 0; j < 5; j++) {
120 if (data[i + j] !=
nullptr) {
122 data[i + j] =
nullptr;
137 if (data[i] !=
nullptr) {
157 int *data = (
int *)item;
159 EXPECT_TRUE(data !=
nullptr);
160 if (task_data->accumulate_items ==
nullptr) {
161 task_data->accumulate_items = MEM_cnew<ListBase>(__func__);
171 void *__restrict chunk_join,
172 void *__restrict chunk)
177 if (data_chunk->accumulate_items !=
nullptr) {
178 if (join_chunk->accumulate_items ==
nullptr) {
179 join_chunk->accumulate_items = MEM_cnew<ListBase>(__func__);
210 tls_data.accumulate_items =
nullptr;
212 settings.userdata_chunk = &tls_data;
213 settings.userdata_chunk_size =
sizeof(tls_data);
223 int number_accum = 0;
225 int *data = (
int *)link->data;
226 number_accum += *
data;
239 std::atomic<int> counter = 0;
241 [&]() { counter++; },
242 [&]() { counter++; },
243 [&]() { counter++; },
244 [&]() { counter++; },
245 [&]() { counter++; });
EXPECT_EQ(BLI_expr_pylike_eval(expr, nullptr, 0, &result), EXPR_PYLIKE_INVALID)
#define LISTBASE_FOREACH(type, var, list)
void void void BLI_movelisttolist(struct ListBase *dst, struct ListBase *src) ATTR_NONNULL(1
void void BLI_freelistN(struct ListBase *listbase) ATTR_NONNULL(1)
void BLI_addtail(struct ListBase *listbase, void *vlink) ATTR_NONNULL(1)
struct LinkData * BLI_genericNodeN(void *data)
int BLI_listbase_count(const struct ListBase *listbase) ATTR_WARN_UNUSED_RESULT ATTR_NONNULL(1)
void * BLI_mempool_alloc(BLI_mempool *pool) ATTR_MALLOC ATTR_WARN_UNUSED_RESULT ATTR_RETURNS_NONNULL ATTR_NONNULL(1)
void BLI_mempool_free(BLI_mempool *pool, void *addr) ATTR_NONNULL(1
BLI_mempool * BLI_mempool_create(unsigned int esize, unsigned int elem_num, unsigned int pchunk, unsigned int flag) ATTR_MALLOC ATTR_WARN_UNUSED_RESULT ATTR_RETURNS_NONNULL
void BLI_mempool_destroy(BLI_mempool *pool) ATTR_NONNULL(1)
struct MempoolIterData MempoolIterData
void BLI_task_parallel_range(int start, int stop, void *userdata, TaskParallelRangeFunc func, const TaskParallelSettings *settings)
BLI_INLINE void BLI_parallel_range_settings_defaults(TaskParallelSettings *settings)
void BLI_task_parallel_mempool(struct BLI_mempool *mempool, void *userdata, TaskParallelMempoolFunc func, const TaskParallelSettings *settings)
BLI_INLINE void BLI_parallel_mempool_settings_defaults(TaskParallelSettings *settings)
static void task_mempool_iter_func(void *userdata, MempoolIterData *item, const TaskParallelTLS *__restrict)
static void task_mempool_iter_tls_reduce(const void *__restrict, void *__restrict chunk_join, void *__restrict chunk)
static void task_range_iter_func(void *userdata, int index, const TaskParallelTLS *__restrict tls)
static void task_mempool_iter_tls_free(const void *, void *__restrict userdata_chunk)
static void task_range_iter_reduce_func(const void *__restrict, void *__restrict join_v, void *__restrict userdata_chunk)
static void task_mempool_iter_tls_func(void *, MempoolIterData *item, const TaskParallelTLS *__restrict tls)
struct TaskMemPool_Chunk { ListBase *accumulate_items;} TaskMemPool_Chunk
void BLI_threadapi_init(void)
void BLI_threadapi_exit(void)
Read Guarded memory(de)allocation.
Provides wrapper around system-specific atomic primitives, and some extensions (faked-atomic operatio...
ATOMIC_INLINE uint32_t atomic_sub_and_fetch_uint32(uint32_t *p, uint32_t x)
static T sum(const btAlignedObjectArray< T > &items)
void MEM_freeN(void *vmemh)
void parallel_invoke(Functions &&...functions)