/nptl/ |
A D | tst-thread-affinity-sched.c | 25 getaffinity (size_t size, cpu_set_t *set) in getaffinity() argument 27 return sched_getaffinity (0, size, set); in getaffinity() 31 setaffinity (size_t size, const cpu_set_t *set) in setaffinity() argument 33 return sched_setaffinity (0, size, set); in setaffinity()
|
A D | tst-thread-affinity-pthread2.c | 32 size_t size; member 43 (task->thread, task->size, task->set); in affinity_access_thread() 46 (task->thread, task->size, task->set); in affinity_access_thread() 51 run_affinity_access_thread (cpu_set_t *set, size_t size, bool get) in run_affinity_access_thread() argument 57 .size = size, in run_affinity_access_thread() 84 setaffinity (size_t size, const cpu_set_t *set) in setaffinity() argument 86 return run_affinity_access_thread ((cpu_set_t *) set, size, false); in setaffinity() 90 getaffinity (size_t size, cpu_set_t *set) in getaffinity() argument 92 return run_affinity_access_thread (set, size, true); in getaffinity()
|
A D | tst-thread-affinity-pthread.c | 26 setaffinity (size_t size, const cpu_set_t *set) in setaffinity() argument 28 int ret = pthread_setaffinity_np (pthread_self (), size, set); in setaffinity() 38 getaffinity (size_t size, cpu_set_t *set) in getaffinity() argument 40 int ret = pthread_getaffinity_np (pthread_self (), size, set); in getaffinity()
|
A D | pthread_attr_getstacksize.c | 28 size_t size = iattr->stacksize; in __pthread_attr_getstacksize() local 32 if (size == 0) in __pthread_attr_getstacksize() 35 size = __default_pthread_attr.internal.stacksize; in __pthread_attr_getstacksize() 38 *stacksize = size; in __pthread_attr_getstacksize()
|
A D | allocatestack.c | 59 size_t size = *sizep; in get_cached_stack() local 193 assert (freesize < size); in advise_stack_range() 204 assert (freesize < size); in advise_stack_range() 219 size_t size; in allocate_stack() local 230 size = attr->stacksize; in allocate_stack() 265 assert (size > adj); in allocate_stack() 337 assert (size != 0); in allocate_stack() 354 size += guardsize; in allocate_stack() 363 reqsize = size; in allocate_stack() 399 __munmap (mem, size); in allocate_stack() [all …]
|
A D | tst-setgroups.c | 44 int size; in do_test() local 51 size = getgroups (sizeof (list) / sizeof (list[0]), list); in do_test() 52 if (size < 0) in do_test() 57 if (setgroups (size, list) < 0) in do_test()
|
A D | alloca_cutoff.c | 26 __libc_alloca_cutoff (size_t size) in __libc_alloca_cutoff() argument 28 return size <= (MIN (__MAX_ALLOCA_CUTOFF, in __libc_alloca_cutoff()
|
A D | tst-tls5.h | 8 size_t size; member 21 tls_registry[i].size = sizeof (x); \
|
A D | tst-tls5.c | 52 tls_registry[cnt].size = sizeof (struct pthread); in do_test() 63 tls_registry[i].size, tls_registry[i].align); in do_test() 70 && (tls_registry[i - 1].addr + tls_registry[i - 1].size in do_test() 80 max_addr = MAX (tls_registry[i].addr + tls_registry[i].size, in do_test()
|
A D | pthread_getattr_np.c | 173 size_t size = 16; in __pthread_getattr_np() local 178 size <<= 1; in __pthread_getattr_np() 180 void *newp = realloc (cpuset, size); in __pthread_getattr_np() 188 ret = __pthread_getaffinity_np (thread_id, size, cpuset); in __pthread_getattr_np() 191 while (ret == EINVAL && size < 1024 * 1024); in __pthread_getattr_np() 194 ret = __pthread_attr_setaffinity_np (attr, size, cpuset); in __pthread_getattr_np()
|
A D | tpp.c | 73 size_t size = sizeof *tpp; in libc_hidden_def() local 74 size += (fifo_max_prio - fifo_min_prio + 1) in libc_hidden_def() 76 tpp = calloc (size, 1); in libc_hidden_def()
|
A D | pthread_cond_common.c | 166 __condvar_set_orig_size (pthread_cond_t *cond, unsigned int size) in __condvar_set_orig_size() argument 175 | (size << 2); in __condvar_set_orig_size() 178 atomic_store_relaxed (&cond->__data.__g1_orig_size, (size << 2) | 2); in __condvar_set_orig_size()
|