/linux/arch/mips/cavium-octeon/executive/ |
A D | cvmx-cmd-queue.c | 100 int max_depth, int fpa_pool, in cvmx_cmd_queue_initialize() argument 117 if ((max_depth < 0) || (max_depth > 1 << 20)) in cvmx_cmd_queue_initialize() 119 } else if (max_depth != 0) in cvmx_cmd_queue_initialize() 129 if (max_depth != (int)qstate->max_depth) { in cvmx_cmd_queue_initialize() 133 (int)qstate->max_depth); in cvmx_cmd_queue_initialize() 170 qstate->max_depth = max_depth; in cvmx_cmd_queue_initialize()
|
/linux/arch/mips/include/asm/octeon/ |
A D | cvmx-cmd-queue.h | 133 uint32_t max_depth; member 171 int max_depth, int fpa_pool, 343 if (CVMX_CMD_QUEUE_ENABLE_MAX_DEPTH && unlikely(qptr->max_depth)) { in cvmx_cmd_queue_write() 345 (cvmx_cmd_queue_length(queue_id) > (int)qptr->max_depth)) { in cvmx_cmd_queue_write() 441 if (CVMX_CMD_QUEUE_ENABLE_MAX_DEPTH && unlikely(qptr->max_depth)) { in cvmx_cmd_queue_write2() 443 (cvmx_cmd_queue_length(queue_id) > (int)qptr->max_depth)) { in cvmx_cmd_queue_write2() 542 if (CVMX_CMD_QUEUE_ENABLE_MAX_DEPTH && unlikely(qptr->max_depth)) { in cvmx_cmd_queue_write3() 544 (cvmx_cmd_queue_length(queue_id) > (int)qptr->max_depth)) { in cvmx_cmd_queue_write3()
|
/linux/drivers/staging/media/hantro/ |
A D | rockchip_vpu_hw.c | 48 .max_depth = 2, 76 .max_depth = 2, 89 .max_depth = 2, 102 .max_depth = 2, 122 .max_depth = 2, 135 .max_depth = 2, 148 .max_depth = 2, 168 .max_depth = 2, 181 .max_depth = 2, 194 .max_depth = 2,
|
A D | sama5d4_vdec_hw.c | 29 .max_depth = 2, 42 .max_depth = 2, 55 .max_depth = 2,
|
A D | imx8m_vpu_hw.c | 96 .max_depth = 2, 109 .max_depth = 2, 122 .max_depth = 2, 142 .max_depth = 2,
|
/linux/drivers/acpi/acpica/ |
A D | utcache.c | 34 u16 max_depth, struct acpi_memory_list **return_cache) in acpi_os_create_cache() argument 56 cache->max_depth = max_depth; in acpi_os_create_cache() 165 if (cache->current_depth >= cache->max_depth) { in acpi_os_release_object()
|
A D | nsdump.c | 621 u32 max_depth, in acpi_ns_dump_objects() argument 646 (void)acpi_ns_walk_namespace(type, start_handle, max_depth, in acpi_ns_dump_objects() 745 u32 max_depth, in acpi_ns_dump_object_paths() argument 767 (void)acpi_ns_walk_namespace(type, start_handle, max_depth, in acpi_ns_dump_object_paths() 775 (void)acpi_ns_walk_namespace(type, start_handle, max_depth, in acpi_ns_dump_object_paths() 826 void acpi_ns_dump_tables(acpi_handle search_base, u32 max_depth) in acpi_ns_dump_tables() argument 850 acpi_ns_dump_objects(ACPI_TYPE_ANY, ACPI_DISPLAY_OBJECTS, max_depth, in acpi_ns_dump_tables()
|
A D | dbstats.c | 75 if (list->max_depth > 0) { in acpi_db_list_info() 79 list->max_depth, list->max_depth - list->current_depth, in acpi_db_list_info() 83 if (list->max_depth > 0) { in acpi_db_list_info()
|
A D | acnamesp.h | 79 u32 max_depth, 179 void acpi_ns_dump_tables(acpi_handle search_base, u32 max_depth); 196 u32 max_depth, 202 u32 max_depth,
|
A D | dbnames.c | 175 u32 max_depth = ACPI_UINT32_MAX; in acpi_db_dump_namespace() local 188 max_depth = strtoul(depth_arg, NULL, 0); in acpi_db_dump_namespace() 206 acpi_ns_dump_objects(ACPI_TYPE_ANY, ACPI_DISPLAY_SUMMARY, max_depth, in acpi_db_dump_namespace() 256 u32 max_depth = ACPI_UINT32_MAX; in acpi_db_dump_namespace_by_owner() local 264 max_depth = strtoul(depth_arg, NULL, 0); in acpi_db_dump_namespace_by_owner() 273 acpi_ns_dump_objects(ACPI_TYPE_ANY, ACPI_DISPLAY_SUMMARY, max_depth, in acpi_db_dump_namespace_by_owner()
|
A D | nswalk.c | 152 u32 max_depth, in acpi_ns_walk_namespace() argument 279 (level < max_depth) && (status != AE_CTRL_DEPTH)) { in acpi_ns_walk_namespace()
|
A D | nsxfeval.c | 556 u32 max_depth, in acpi_walk_namespace() argument 568 (!max_depth) || (!descending_callback && !ascending_callback)) { in acpi_walk_namespace() 606 status = acpi_ns_walk_namespace(type, start_object, max_depth, in acpi_walk_namespace()
|
/linux/block/ |
A D | blk-iolatency.c | 278 return rq_wait_inc_below(rqw, iolat->rq_depth.max_depth); in iolat_acquire_inflight() 370 unsigned long old = iolat->rq_depth.max_depth; in scale_change() 382 iolat->rq_depth.max_depth = old; in scale_change() 387 iolat->rq_depth.max_depth = max(old, 1UL); in scale_change() 445 if (iolat->rq_depth.max_depth == 1 && direction < 0) { in check_scale_change() 453 iolat->rq_depth.max_depth = UINT_MAX; in check_scale_change() 908 if (iolat->rq_depth.max_depth == UINT_MAX) in iolatency_ssd_stat() 916 iolat->rq_depth.max_depth); in iolatency_ssd_stat() 934 if (iolat->rq_depth.max_depth == UINT_MAX) in iolatency_pd_stat() 939 iolat->rq_depth.max_depth, avg_lat, cur_win); in iolatency_pd_stat() [all …]
|
A D | blk-rq-qos.c | 130 rqd->max_depth = 1; in rq_depth_calc_max_depth() 132 rqd->max_depth = 2; in rq_depth_calc_max_depth() 157 rqd->max_depth = depth; in rq_depth_calc_max_depth() 190 if (rqd->max_depth == 1) in rq_depth_scale_down()
|
A D | blk-wbt.c | 294 rwb->wb_background, rwb->wb_normal, rqd->max_depth); in rwb_trace_step() 301 } else if (rwb->rq_depth.max_depth <= 2) { in calc_wb_limits() 302 rwb->wb_normal = rwb->rq_depth.max_depth; in calc_wb_limits() 305 rwb->wb_normal = (rwb->rq_depth.max_depth + 1) / 2; in calc_wb_limits() 306 rwb->wb_background = (rwb->rq_depth.max_depth + 3) / 4; in calc_wb_limits() 477 limit = rwb->rq_depth.max_depth; in get_limit()
|
/linux/arch/um/drivers/ |
A D | vector_kern.c | 254 % qi->max_depth; in vector_advancehead() 284 % qi->max_depth; in vector_advancetail() 349 if (queue_depth < qi->max_depth) { in vector_enqueue() 424 if (send_len + qi->head > qi->max_depth) in vector_send() 425 send_len = qi->max_depth - qi->head; in vector_send() 496 for (i = 0; i < qi->max_depth; i++) { in destroy_queue() 505 for (i = 0; i < qi->max_depth; i++) { in destroy_queue() 537 result->max_depth = max_size; in create_queue() 1448 ring->rx_max_pending = vp->rx_queue->max_depth; in vector_get_ringparam() 1450 ring->rx_pending = vp->rx_queue->max_depth; in vector_get_ringparam() [all …]
|
A D | vector_kern.h | 46 int queue_depth, head, tail, max_depth, max_iov_frags; member
|
/linux/drivers/firewire/ |
A D | core-topology.c | 148 if (node->ports[i]->max_depth > depths[0]) { in update_hop_count() 150 depths[0] = node->ports[i]->max_depth; in update_hop_count() 151 } else if (node->ports[i]->max_depth > depths[1]) in update_hop_count() 152 depths[1] = node->ports[i]->max_depth; in update_hop_count() 155 node->max_depth = depths[0] + 1; in update_hop_count()
|
A D | core.h | 184 u8 max_depth:4; /* Maximum depth to any leaf node */ member
|
/linux/fs/befs/ |
A D | befs_fs_types.h | 220 fs32 max_depth; member 230 u32 max_depth; member
|
A D | btree.c | 153 sup->max_depth = fs32_to_cpu(sb, od_sup->max_depth); in befs_bt_read_super()
|
/linux/fs/f2fs/ |
A D | dir.c | 394 unsigned int max_depth; in __f2fs_find_entry() local 407 max_depth = F2FS_I(dir)->i_current_depth; in __f2fs_find_entry() 408 if (unlikely(max_depth > MAX_DIR_HASH_DEPTH)) { in __f2fs_find_entry() 410 dir->i_ino, max_depth); in __f2fs_find_entry() 411 max_depth = MAX_DIR_HASH_DEPTH; in __f2fs_find_entry() 412 f2fs_i_depth_write(dir, max_depth); in __f2fs_find_entry() 415 for (level = 0; level < max_depth; level++) { in __f2fs_find_entry()
|
/linux/tools/perf/util/ |
A D | callchain.h | 76 u64 max_depth; member 182 root->max_depth = 0; in callchain_init()
|
/linux/kernel/bpf/ |
A D | stackmap.c | 264 u32 max_depth = map->value_size / stack_map_data_size(map); in __bpf_get_stackid() local 266 u32 init_nr = sysctl_perf_event_max_stack - max_depth; in __bpf_get_stackid() 340 u32 max_depth = map->value_size / stack_map_data_size(map); in BPF_CALL_3() local 342 u32 init_nr = sysctl_perf_event_max_stack - max_depth; in BPF_CALL_3()
|
/linux/drivers/net/ethernet/netronome/nfp/bpf/ |
A D | verifier.c | 701 unsigned int max_depth = 0, depth = 0, frame = 0; in nfp_bpf_get_stack_usage() local 717 max_depth = max(max_depth, depth); in nfp_bpf_get_stack_usage() 743 return max_depth; in nfp_bpf_get_stack_usage()
|