/xen/xen/xsm/flask/ss/ |
A D | sidtab.c | 56 cur = cur->next; in sidtab_insert() 59 if ( cur && sid == cur->sid ) in sidtab_insert() 110 cur = cur->next; in sidtab_search() 119 cur = cur->next; in sidtab_search() 120 if ( !cur || sid != cur->sid ) in sidtab_search() 144 cur = cur->next; in sidtab_map() 179 cur = cur->next; in sidtab_map_remove_on_error() 187 cur = cur->next; in sidtab_map_remove_on_error() 208 cur = cur->next; in sidtab_search_context() 268 cur = cur->next; in sidtab_hash_eval() [all …]
|
A D | avtab.c | 73 for ( prev = NULL, cur = h->htable[hvalue]; cur; in avtab_insert() 74 prev = cur, cur = cur->next) in avtab_insert() 113 for ( prev = NULL, cur = h->htable[hvalue]; cur; in avtab_insert_nonunique() 114 prev = cur, cur = cur->next ) in avtab_insert_nonunique() 146 for ( cur = h->htable[hvalue]; cur; cur = cur->next ) in avtab_search() 181 for ( cur = h->htable[hvalue]; cur; cur = cur->next ) in avtab_search_node() 187 return cur; in avtab_search_node() 211 for ( cur = node->next; cur; cur = cur->next ) in avtab_search_node_next() 246 cur = cur->next; in avtab_destroy() 312 if ( cur ) in avtab_hash_eval() [all …]
|
A D | hashtab.c | 48 while ( cur && h->keycmp(h, key, cur->key) > 0 ) in hashtab_insert() 50 prev = cur; in hashtab_insert() 51 cur = cur->next; in hashtab_insert() 54 if ( cur && (h->keycmp(h, key, cur->key) == 0) ) in hashtab_insert() 87 while ( cur != NULL && h->keycmp(h, key, cur->key) > 0 ) in hashtab_search() 88 cur = cur->next; in hashtab_search() 90 if ( cur == NULL || (h->keycmp(h, key, cur->key) != 0) ) in hashtab_search() 110 cur = cur->next; in hashtab_destroy() 138 ret = apply(cur->key, cur->datum, args); in hashtab_map() 141 cur = cur->next; in hashtab_map() [all …]
|
A D | conditional.c | 31 struct cond_expr *cur; in cond_evaluate_expr() local 35 for ( cur = expr; cur != NULL; cur = cur->next ) in cond_evaluate_expr() 97 struct cond_av_list* cur; in evaluate_cond_node() local 106 for ( cur = node->true_list; cur != NULL; cur = cur->next ) in evaluate_cond_node() 114 for ( cur = node->false_list; cur != NULL; cur = cur->next ) in evaluate_cond_node() 139 for ( cur = list; cur != NULL; cur = next ) in cond_av_list_destroy() 141 next = cur->next; in cond_av_list_destroy() 143 xfree(cur); in cond_av_list_destroy() 168 for ( cur = list; cur != NULL; cur = next ) in cond_list_destroy() 170 next = cur->next; in cond_list_destroy() [all …]
|
/xen/xen/arch/x86/cpu/mcheck/ |
A D | vmce.c | 179 struct vcpu *cur = current; in vmce_rdmsr() local 184 spin_lock(&cur->arch.vmce.lock); in vmce_rdmsr() 196 *val = cur->arch.vmce.mcg_cap; in vmce_rdmsr() 216 cur, *val); in vmce_rdmsr() 222 cur); in vmce_rdmsr() 227 ret = mce_bank_msr(cur, msr) ? bank_mce_rdmsr(cur, msr, val) : 0; in vmce_rdmsr() 231 spin_unlock(&cur->arch.vmce.lock); in vmce_rdmsr() 310 struct vcpu *cur = current; in vmce_wrmsr() local 313 spin_lock(&cur->arch.vmce.lock); in vmce_wrmsr() 324 cur, val); in vmce_wrmsr() [all …]
|
/xen/tools/ocaml/xenstored/ |
A D | quota.ml | 29 cur: (Xenctrl.domid, int) Hashtbl.t; (* current domains quota *) RecordField 33 if Hashtbl.mem quota.cur domid 40 let copy quota = { quota with cur = (Hashtbl.copy quota.cur) } 42 let del quota id = Hashtbl.remove quota.cur id 49 if id > 0 && Hashtbl.mem quota.cur id then 50 let entry = Hashtbl.find quota.cur id in 60 let get_entry quota id = Hashtbl.find quota.cur id 64 then Hashtbl.remove quota.cur id 66 if Hashtbl.mem quota.cur id then 67 Hashtbl.replace quota.cur id nb [all …]
|
/xen/xen/arch/x86/hvm/ |
A D | save.c | 178 if ( ctxt.cur < desc->length || in hvm_save_one() 179 off > ctxt.cur - desc->length ) in hvm_save_one() 284 ASSERT(h->cur <= h->size); in hvm_save() 355 if ( h->size - h->cur < len + sizeof (*d) ) in _hvm_init_entry() 366 h->cur += sizeof(*d); in _hvm_init_entry() 374 memcpy(&h->data[h->cur], src, src_len); in _hvm_write_entry() 375 h->cur += src_len; in _hvm_write_entry() 384 if ( sizeof(*d) > h->size - h->cur) in _hvm_check_entry() 404 h->cur += sizeof(*d); in _hvm_check_entry() 417 memcpy(dest, &h->data[h->cur], d->length); in _hvm_read_entry() [all …]
|
/xen/tools/libxc/ |
A D | xc_netbsd.c | 29 off_t cur = 0; in discard_file_cache() local 45 if ( ( cur = lseek(fd, 0, SEEK_CUR)) == (off_t)-1 ) in discard_file_cache() 46 cur = 0; in discard_file_cache() 47 cur &= ~(PAGE_SIZE - 1); in discard_file_cache() 51 if ( posix_fadvise(fd, 0, cur, POSIX_FADV_DONTNEED) < 0 ) in discard_file_cache()
|
A D | xc_freebsd.c | 25 off_t cur = 0; in discard_file_cache() local 38 if ( (cur = lseek(fd, 0, SEEK_CUR)) == (off_t)-1 ) in discard_file_cache() 39 cur = 0; in discard_file_cache() 40 cur &= ~(XC_PAGE_SIZE-1); in discard_file_cache() 44 if ( posix_fadvise(fd, 0, cur, POSIX_FADV_DONTNEED) < 0 ) in discard_file_cache()
|
A D | xc_linux.c | 25 off_t cur = 0; in discard_file_cache() local 41 if ( (cur = lseek(fd, 0, SEEK_CUR)) == (off_t)-1 ) in discard_file_cache() 42 cur = 0; in discard_file_cache() 43 cur &= ~(XC_PAGE_SIZE-1); in discard_file_cache() 47 if ( posix_fadvise64(fd, 0, cur, POSIX_FADV_DONTNEED) < 0 ) in discard_file_cache()
|
A D | xc_dom_bzimageloader.c | 410 const unsigned char *cur = dom->kernel_blob; in xc_try_lzo1x_decode() local 432 if ( left < 16 || memcmp(cur, magic, 9) ) in xc_try_lzo1x_decode() 440 version = lzo_read_16(cur + 9); in xc_try_lzo1x_decode() 441 cur += 16; in xc_try_lzo1x_decode() 447 ++cur; in xc_try_lzo1x_decode() 464 ret += 1 + cur[ret] + 4; in xc_try_lzo1x_decode() 471 cur += ret; in xc_try_lzo1x_decode() 483 dst_len = lzo_read_32(cur); in xc_try_lzo1x_decode() 502 src_len = lzo_read_32(cur + 4); in xc_try_lzo1x_decode() 526 ret = lzo1x_decompress_safe(cur, src_len, in xc_try_lzo1x_decode() [all …]
|
A D | xc_offline_page.c | 46 int cur; member 183 if (backup->max == backup->cur) in backup_ptes() 193 backup->entries[backup->cur].table_mfn = table_mfn; in backup_ptes() 194 backup->entries[backup->cur++].offset = offset; in backup_ptes() 247 for (index = 0; index < backup->cur; index ++) in __update_pte() 252 if (index != backup->cur) in __update_pte() 558 old_ptes.cur = 0; in xc_exchange_page()
|
A D | xc_cpuid_x86.c | 294 xen_cpuid_leaf_t *host = NULL, *max = NULL, *cur = NULL; in xc_cpuid_xend_policy() local 316 (cur = calloc(nr_leaves, sizeof(*cur))) == NULL ) in xc_cpuid_xend_policy() 325 rc = xc_get_domain_cpu_policy(xch, domid, &nr_cur, cur, &nr_msrs, NULL); in xc_cpuid_xend_policy() 361 xen_cpuid_leaf_t *cur_leaf = find_leaf(cur, nr_cur, xend); in xc_cpuid_xend_policy() 408 rc = xc_set_domain_cpu_policy(xch, domid, nr_cur, cur, 0, NULL, in xc_cpuid_xend_policy() 421 free(cur); in xc_cpuid_xend_policy()
|
/xen/xen/arch/x86/pv/ |
A D | callback.c | 348 struct trap_info cur; in do_set_trap_table() local 362 if ( copy_from_guest(&cur, traps, 1) ) in do_set_trap_table() 368 if ( cur.address == 0 ) in do_set_trap_table() 371 if ( !is_canonical_address(cur.address) ) in do_set_trap_table() 374 fixup_guest_code_selector(curr->domain, cur.cs); in do_set_trap_table() 376 memcpy(&dst[cur.vector], &cur, sizeof(cur)); in do_set_trap_table() 394 struct compat_trap_info cur; in compat_set_trap_table() local 407 if ( copy_from_guest(&cur, traps, 1) ) in compat_set_trap_table() 413 if ( cur.address == 0 ) in compat_set_trap_table() 416 fixup_guest_code_selector(curr->domain, cur.cs); in compat_set_trap_table() [all …]
|
/xen/xen/arch/x86/guest/hyperv/ |
A D | tlb.c | 41 unsigned long cur = (unsigned long)va; in fill_gva_list() local 43 unsigned long end = cur + (PAGE_SIZE << order); in fill_gva_list() 47 unsigned long diff = end - cur; in fill_gva_list() 49 gva_list[n] = cur & PAGE_MASK; in fill_gva_list() 58 cur += HV_TLB_FLUSH_UNIT; in fill_gva_list() 63 cur = end; in fill_gva_list() 67 } while ( cur < end ); in fill_gva_list()
|
/xen/xen/common/ |
A D | list_sort.c | 132 struct list_head *cur = list; in list_sort() local 134 cur->next = NULL; in list_sort() 137 cur = merge(priv, cmp, part[lev], cur); in list_sort() 148 part[lev] = cur; in list_sort()
|
A D | rcupdate.c | 53 long cur; /* Current batch number. */ member 62 .cur = -300, 349 rcp->completed == rcp->cur) { in rcu_start_batch() 356 rcp->cur++; in rcu_start_batch() 381 rcp->completed = rcp->cur; in cpu_quiet() 394 if (rdp->quiescbatch != rcp->cur) { in rcu_check_quiescent_state() 397 rdp->quiescbatch = rcp->cur; in rcu_check_quiescent_state() 415 if (likely(rdp->quiescbatch == rcp->cur)) in rcu_check_quiescent_state() 448 rdp->batch = rcp->cur + 1; in __rcu_process_callbacks() 504 if (rdp->quiescbatch != rcp->cur || rdp->qs_pending) in __rcu_pending() [all …]
|
A D | kexec.c | 108 const char *cur; in parse_crashkernel() local 119 cur = NULL; in parse_crashkernel() 125 ranges[idx].start = parse_size_and_unit(cur = str + !!idx, &str); in parse_crashkernel() 126 if ( cur == str ) in parse_crashkernel() 138 ranges[idx].end = parse_size_and_unit(cur = str, &str); in parse_crashkernel() 139 if ( cur == str ) in parse_crashkernel() 158 ranges[idx].size = parse_size_and_unit(cur = str + 1, &str); in parse_crashkernel() 159 if ( cur == str ) in parse_crashkernel() 168 kexec_crash_area.size = parse_size_and_unit(cur = str, &str); in parse_crashkernel() 169 if ( cur != str ) in parse_crashkernel() [all …]
|
A D | vsprintf.c | 273 unsigned int cur, rbot, rtop; in print_bitmap_list() local 276 rbot = cur = find_first_bit(bitmap, nr_bits); in print_bitmap_list() 277 while ( cur < nr_bits ) in print_bitmap_list() 279 rtop = cur; in print_bitmap_list() 280 cur = find_next_bit(bitmap, nr_bits, cur + 1); in print_bitmap_list() 282 if ( cur < nr_bits && cur <= rtop + 1 ) in print_bitmap_list() 303 rbot = cur; in print_bitmap_list()
|
/xen/tools/xl/ |
A D | bash-completion | 7 local cur opts xl 9 cur="${COMP_WORDS[COMP_CWORD]}" 13 …d' | awk '/^ [^ ]/ {print $1}' | sed 's/$/ ,/g'` && COMPREPLY=( $(compgen -W "${opts}" -- ${cur}) )
|
/xen/xen/arch/x86/acpi/cpufreq/ |
A D | cpufreq.c | 271 struct perf_pair readin, cur, *saved; in get_measured_perf() local 304 cur.aperf.whole = readin.aperf.whole - saved->aperf.whole; in get_measured_perf() 305 cur.mperf.whole = readin.mperf.whole - saved->mperf.whole; in get_measured_perf() 309 if (unlikely(((unsigned long)(-1) / 100) < cur.aperf.whole)) { in get_measured_perf() 311 cur.aperf.whole >>= shift_count; in get_measured_perf() 312 cur.mperf.whole >>= shift_count; in get_measured_perf() 315 if (cur.aperf.whole && cur.mperf.whole) in get_measured_perf() 316 perf_percent = (cur.aperf.whole * 100) / cur.mperf.whole; in get_measured_perf() 453 policy->cur = freqs.new; in acpi_cpufreq_target() 586 policy->cur = acpi_cpufreq_guess_freq(data, policy->cpu); in acpi_cpufreq_cpu_init() [all …]
|
/xen/xen/drivers/cpufreq/ |
A D | cpufreq_ondemand.c | 147 if (max_load_freq > (uint64_t) dbs_tuners_ins.up_threshold * policy->cur) { in dbs_check_cpu() 149 if (policy->cur == max) in dbs_check_cpu() 157 if (policy->cur == policy->min) in dbs_check_cpu() 166 < (uint64_t) (dbs_tuners_ins.up_threshold - 10) * policy->cur) { in dbs_check_cpu() 228 if ((!cpu_online(cpu)) || (!policy->cur)) in cpufreq_governor_dbs() 299 if (policy->max < this_dbs_info->cur_policy->cur) in cpufreq_governor_dbs() 302 else if (policy->min > this_dbs_info->cur_policy->cur) in cpufreq_governor_dbs()
|
A D | utility.c | 79 pxpt->u.cur = to; in cpufreq_statistic_update() 365 unsigned int prev_freq = policy->cur; in __cpufreq_driver_target() 370 TRACE_2D(TRC_PM_FREQ_CHANGE, prev_freq/1000, policy->cur/1000); in __cpufreq_driver_target() 391 return policy->cur; in cpufreq_driver_getavg()
|
/xen/xen/include/asm-x86/hvm/ |
A D | save.h | 27 uint32_t cur; member 64 = (struct hvm_save_descriptor *)&(_h)->data[(_h)->cur]; \ 89 const struct hvm_save_descriptor *d = (const void *)&h->data[h->cur]; in hvm_load_instance()
|
/xen/xen/drivers/acpi/ |
A D | pmstat.c | 102 cpufreq_residency_update(op->cpuid, pxpt->u.cur); in do_get_pm_info() 122 op->u.getpx.cur = pxpt->u.cur; in do_get_pm_info() 258 cpufreq_driver.get ? cpufreq_driver.get(op->cpuid) : policy->cur; in get_cpufreq_para() 261 op->u.get_para.scaling_cur_freq = policy->cur; in get_cpufreq_para() 281 op->u.get_para.u.userspace.scaling_setspeed = policy->cur; in get_cpufreq_para()
|