/linux/drivers/net/ethernet/mellanox/mlxsw/ |
A D | spectrum_span.c | 46 struct mlxsw_sp_span *span; member 90 span = kzalloc(struct_size(span, entries, entries_count), GFP_KERNEL); in mlxsw_sp_span_init() 91 if (!span) in mlxsw_sp_span_init() 99 span->mlxsw_sp = mlxsw_sp; in mlxsw_sp_span_init() 100 mlxsw_sp->span = span; in mlxsw_sp_span_init() 117 kfree(mlxsw_sp->span); in mlxsw_sp_span_init() 131 kfree(mlxsw_sp->span); in mlxsw_sp_span_fini() 823 policer_id >= span->policer_id_base + span->entries_count) in mlxsw_sp_span_policer_id_base_set() 1018 struct mlxsw_sp_span *span = mlxsw_sp->span; in mlxsw_sp_span_entry_ops() local 1476 struct mlxsw_sp_span *span = trigger_entry->span; in mlxsw_sp_span_trigger_ops_set() local [all …]
|
/linux/drivers/net/ethernet/marvell/prestera/ |
A D | prestera_span.c | 52 list_for_each_entry(entry, &span->entries, list) { in prestera_span_entry_find_by_id() 66 list_for_each_entry(entry, &span->entries, list) { in prestera_span_entry_find_by_port() 219 struct prestera_span *span; in prestera_span_init() local 221 span = kzalloc(sizeof(*span), GFP_KERNEL); in prestera_span_init() 222 if (!span) in prestera_span_init() 225 INIT_LIST_HEAD(&span->entries); in prestera_span_init() 227 sw->span = span; in prestera_span_init() 228 span->sw = sw; in prestera_span_init() 235 struct prestera_span *span = sw->span; in prestera_span_fini() local 237 WARN_ON(!list_empty(&span->entries)); in prestera_span_fini() [all …]
|
/linux/drivers/scsi/megaraid/ |
A D | megaraid_sas_fp.c | 151 return &map->raidMap.ldSpanMap[ld].spanBlock[span].span; in MR_LdSpanPtrGet() 381 for (span = 0; span < raid->spanDepth; span++, pSpanBlock++) { in MR_GetSpanBlock() 398 return span; in MR_GetSpanBlock() 442 for (span = 0; span < raid->spanDepth; span++) in mr_spanset_get_span_block() 463 return span; in mr_spanset_get_span_block() 509 for (span = 0, span_offset = 0; span < raid->spanDepth; span++) in get_row_from_strip() 562 for (span = 0; span < raid->spanDepth; span++) in get_strip_from_row() 625 for (span = 0, span_offset = 0; span < raid->spanDepth; span++) in get_arm_from_strip() 832 span = 0; in MR_GetPhyParams() 927 span = 0; in mr_get_phy_params_r56_rmw() [all …]
|
A D | mbox_defs.h | 649 adap_span_40ld_t span[SPAN_DEPTH_8_SPANS]; member 662 adap_span_8ld_t span[SPAN_DEPTH_8_SPANS]; member 675 adap_span_8ld_t span[SPAN_DEPTH_4_SPANS]; member
|
/linux/tools/testing/selftests/resctrl/ |
A D | mbm_test.c | 18 show_bw_info(unsigned long *bw_imc, unsigned long *bw_resc, int span) in show_bw_info() argument 43 ksft_print_msg("Span (MB): %d\n", span); in show_bw_info() 50 static int check_results(int span) in check_results() argument 82 ret = show_bw_info(bw_imc, bw_resc, span); in check_results() 117 int mbm_bw_change(int span, int cpu_no, char *bw_report, char **benchmark_cmd) in mbm_bw_change() argument 123 .span = span, in mbm_bw_change() 138 ret = check_results(span); in mbm_bw_change()
|
A D | resctrl_tests.c | 57 static void run_mbm_test(bool has_ben, char **benchmark_cmd, int span, in run_mbm_test() argument 71 res = mbm_bw_change(span, cpu_no, bw_report, benchmark_cmd); in run_mbm_test() 76 static void run_mba_test(bool has_ben, char **benchmark_cmd, int span, in run_mba_test() argument 89 sprintf(benchmark_cmd[1], "%d", span); in run_mba_test() 131 int c, cpu_no = 1, span = 250, argc_new = argc, i, no_of_bits = 0; in main() local 226 sprintf(benchmark_cmd[1], "%d", span); in main() 245 run_mbm_test(has_ben, benchmark_cmd, span, cpu_no, bw_report); in main() 248 run_mba_test(has_ben, benchmark_cmd, span, cpu_no, bw_report); in main()
|
A D | resctrl.h | 61 unsigned long span; member 96 int run_fill_buf(unsigned long span, int malloc_and_init_memory, int memflush, 99 int mbm_bw_change(int span, int cpu_no, char *bw_report, char **benchmark_cmd);
|
A D | cmt_test.c | 73 return show_cache_info(sum_llc_occu_resc, no_of_bits, param->span, in check_results() 124 .span = cache_size * n / count_of_bits, in cmt_resctrl_val() 130 sprintf(benchmark_cmd[1], "%lu", param.span); in cmt_resctrl_val()
|
A D | cat_test.c | 90 return show_cache_info(sum_llc_perf_miss, no_of_bits, param->span / 64, in check_results() 158 param.span = cache_size * (count_of_bits - n) / count_of_bits; in cat_perf_miss_val() 179 param.span = cache_size * n / count_of_bits; in cat_perf_miss_val()
|
A D | fill_buf.c | 193 int run_fill_buf(unsigned long span, int malloc_and_init_memory, in run_fill_buf() argument 196 unsigned long long cache_size = span; in run_fill_buf()
|
A D | resctrlfs.c | 309 unsigned long span, buffer_span; in run_benchmark() local 326 span = strtoul(benchmark_cmd[1], NULL, 10); in run_benchmark() 333 buffer_span = span * MB; in run_benchmark() 335 buffer_span = span; in run_benchmark()
|
/linux/Documentation/scheduler/ |
A D | sched-domains.rst | 10 Each scheduling domain spans a number of CPUs (stored in the ->span field). 11 A domain's span MUST be a superset of it child's span (this restriction could 12 be relaxed if the need arises), and a base domain for CPU i MUST span at least 13 i. The top domain for each CPU will generally span all CPUs in the system 16 explicitly set. A sched domain's span means "balance process load among these 22 domain's span. The group pointed to by the ->groups pointer MUST contain the CPU 58 The "base" domain will "span" the first level of the hierarchy. In the case 59 of SMT, you'll span all siblings of the physical CPU, with each group being 62 In SMP, the parent of the base domain will span all physical CPUs in the 64 of the SMP domain will span the entire machine, with each group having the
|
/linux/drivers/iio/dac/ |
A D | ad5791.c | 68 int (*get_lin_comp) (unsigned int span); 210 static int ad5791_get_lin_comp(unsigned int span) in ad5791_get_lin_comp() argument 212 if (span <= 10000) in ad5791_get_lin_comp() 214 else if (span <= 12000) in ad5791_get_lin_comp() 216 else if (span <= 16000) in ad5791_get_lin_comp() 218 else if (span <= 19000) in ad5791_get_lin_comp() 224 static int ad5780_get_lin_comp(unsigned int span) in ad5780_get_lin_comp() argument 226 if (span <= 10000) in ad5780_get_lin_comp()
|
/linux/drivers/platform/surface/aggregator/ |
A D | ssh_parser.h | 138 struct ssam_span *span) in sshp_buf_span_from() argument 140 span->ptr = buf->ptr + offset; in sshp_buf_span_from() 141 span->len = buf->len - offset; in sshp_buf_span_from()
|
/linux/arch/sh/mm/ |
A D | pmb.c | 146 unsigned long span; in pmb_mapping_exists() local 169 span = pmbe->size; in pmb_mapping_exists() 176 span += iter->size; in pmb_mapping_exists() 181 if (size <= span) { in pmb_mapping_exists() 657 unsigned long span, newsize; in pmb_merge() local 661 span = newsize = head->size; in pmb_merge() 665 span += tail->size; in pmb_merge() 667 if (pmb_size_valid(span)) { in pmb_merge() 668 newsize = span; in pmb_merge()
|
/linux/net/mac802154/ |
A D | rx.c | 37 __le16 span, sshort; in ieee802154_subif_frame() local 42 span = wpan_dev->pan_id; in ieee802154_subif_frame() 55 if (mac_cb(skb)->dest.pan_id != span && in ieee802154_subif_frame() 64 if (mac_cb(skb)->dest.pan_id != span && in ieee802154_subif_frame()
|
/linux/include/linux/ |
A D | energy_model.h | 123 struct em_data_callback *cb, cpumask_t *span, 270 struct em_data_callback *cb, cpumask_t *span, in em_dev_register_perf_domain() argument
|
/linux/Documentation/devicetree/bindings/interrupt-controller/ |
A D | msi-controller.yaml | 37 A list of <phandle intspec span> tuples, where "phandle" is the 39 interrupt specifier and "span" is the size of the
|
/linux/include/linux/sched/ |
A D | topology.h | 160 unsigned long span[]; member 165 return to_cpumask(sd->span); in sched_domain_span()
|
/linux/drivers/perf/ |
A D | arm_smmuv3_pmu.c | 253 int idx, u32 span, u32 sid) in smmu_pmu_set_event_filter() argument 258 evtyper = get_event(event) | span << SMMU_PMCG_SID_SPAN_SHIFT; in smmu_pmu_set_event_filter() 279 u32 span, sid; in smmu_pmu_apply_event_filter() local 283 span = filter_en ? get_filter_span(event) : in smmu_pmu_apply_event_filter() 294 smmu_pmu_set_event_filter(event, idx, span, sid); in smmu_pmu_apply_event_filter()
|
/linux/drivers/usb/host/ |
A D | ehci-sched.c | 1190 iso_sched->span = urb->number_of_packets * stream->uperiod; in itd_sched_init() 1255 num_itds = 1 + (sched->span + 7) / 8; in itd_urb_transaction() 1478 u32 now, base, next, start, period, span, now2; in iso_stream_schedule() local 1487 span = sched->span; in iso_stream_schedule() 1489 span <<= 3; in iso_stream_schedule() 1614 if (skip >= span) { /* Entirely in the past? */ in iso_stream_schedule() 1616 urb, start + base, span - period, now2 + base, in iso_stream_schedule() 1620 skip = span - period; in iso_stream_schedule() 1624 skip = span; /* Skip the entire URB */ in iso_stream_schedule() 1641 if (unlikely(start + span - period >= mod + wrap)) { in iso_stream_schedule() [all …]
|
/linux/Documentation/sphinx-static/ |
A D | theme_overrides.css | 106 span.menuselection { 111 code.kbd, code.kbd span {
|
/linux/kernel/sched/ |
A D | topology.c | 460 free_cpumask_var(rd->span); in free_rootdomain() 478 cpumask_clear_cpu(rq->cpu, old_rd->span); in rq_attach_root() 492 cpumask_set_cpu(rq->cpu, rd->span); in rq_attach_root() 517 if (!zalloc_cpumask_var(&rd->span, GFP_KERNEL)) in init_rootdomain() 550 free_cpumask_var(rd->span); in init_rootdomain() 995 const struct cpumask *span = sched_domain_span(sd); in build_overlap_sched_groups() local 1003 for_each_cpu_wrap(i, span, cpu) { in build_overlap_sched_groups() 1056 !cpumask_subset(sched_domain_span(sibling->child), span)) in build_overlap_sched_groups() 1207 const struct cpumask *span = sched_domain_span(sd); in build_sched_groups() local 1216 for_each_cpu_wrap(i, span, cpu) { in build_sched_groups()
|
/linux/drivers/s390/scsi/ |
A D | zfcp_qdio.c | 56 unsigned long long now, span; in zfcp_qdio_account() local 60 span = (now - qdio->req_q_time) >> 12; in zfcp_qdio_account() 62 qdio->req_q_util += used * span; in zfcp_qdio_account()
|
/linux/Documentation/litmus-tests/rcu/ |
A D | RCU+sync+read.litmus | 11 * other things) that an RCU read-side critical section cannot span a grace period.
|