/linux/block/ |
A D | blk-mq-debugfs.c | 208 { "requeue_list", 0400, .seq_ops = &queue_requeue_list_seq_ops }, 634 if (attr->seq_ops) { in blk_mq_debugfs_open() 635 ret = seq_open(file, attr->seq_ops); in blk_mq_debugfs_open() 670 {"dispatch", 0400, .seq_ops = &hctx_dispatch_seq_ops}, 685 {"default_rq_list", 0400, .seq_ops = &ctx_default_rq_list_seq_ops}, 686 {"read_rq_list", 0400, .seq_ops = &ctx_read_rq_list_seq_ops}, 687 {"poll_rq_list", 0400, .seq_ops = &ctx_poll_rq_list_seq_ops},
|
A D | blk-mq-debugfs.h | 15 const struct seq_operations *seq_ops; member
|
A D | mq-deadline.c | 1097 .seq_ops = &deadline_##name##_fifo_seq_ops} 1116 {"dispatch0", 0400, .seq_ops = &deadline_dispatch0_seq_ops}, 1117 {"dispatch1", 0400, .seq_ops = &deadline_dispatch1_seq_ops}, 1118 {"dispatch2", 0400, .seq_ops = &deadline_dispatch2_seq_ops},
|
A D | kyber-iosched.c | 998 {#name "_rqs", 0400, .seq_ops = &kyber_##name##_rqs_seq_ops}, \
|
/linux/fs/proc/ |
A D | proc_net.c | 57 p = __seq_open_private(file, PDE(inode)->seq_ops, state_size); in seq_open_net() 115 p->seq_ops = ops; in proc_create_net_data() 159 p->seq_ops = ops; in proc_create_net_data_write()
|
A D | generic.c | 592 return seq_open_private(file, de->seq_ops, de->state_size); in proc_seq_open() 593 return seq_open(file, de->seq_ops); in proc_seq_open() 623 p->seq_ops = ops; in proc_create_seq_private()
|
A D | internal.h | 48 const struct seq_operations *seq_ops; member
|
A D | base.c | 3008 const struct seq_operations *seq_ops) in proc_id_map_open() argument 3025 ret = seq_open(file, seq_ops); in proc_id_map_open()
|
/linux/kernel/bpf/ |
A D | task_iter.c | 532 .seq_ops = &task_seq_ops, 550 .seq_ops = &task_file_seq_ops, 570 .seq_ops = &task_vma_seq_ops,
|
A D | prog_iter.c | 85 .seq_ops = &bpf_prog_seq_ops,
|
A D | map_iter.c | 85 .seq_ops = &bpf_map_seq_ops,
|
A D | bpf_iter.c | 597 priv_data = __seq_open_private(file, seq_info->seq_ops, in prepare_seq_file()
|
A D | arraymap.c | 642 .seq_ops = &bpf_array_map_seq_ops,
|
A D | hashtab.c | 2046 .seq_ops = &bpf_hash_map_seq_ops,
|
/linux/kernel/trace/ |
A D | trace_events.c | 2128 const struct seq_operations *seq_ops) in ftrace_event_open() argument 2137 ret = seq_open(file, seq_ops); in ftrace_event_open() 2159 const struct seq_operations *seq_ops = &show_event_seq_ops; in ftrace_event_avail_open() local 2162 return ftrace_event_open(inode, file, seq_ops); in ftrace_event_avail_open() 2168 const struct seq_operations *seq_ops = &show_set_event_seq_ops; in ftrace_event_set_open() local 2180 ret = ftrace_event_open(inode, file, seq_ops); in ftrace_event_set_open() 2189 const struct seq_operations *seq_ops = &show_set_pid_seq_ops; in ftrace_event_set_pid_open() local 2201 ret = ftrace_event_open(inode, file, seq_ops); in ftrace_event_set_pid_open() 2210 const struct seq_operations *seq_ops = &show_set_no_pid_seq_ops; in ftrace_event_set_npid_open() local 2222 ret = ftrace_event_open(inode, file, seq_ops); in ftrace_event_set_npid_open()
|
A D | ftrace.c | 5943 const struct seq_operations *seq_ops; member 6116 fgd->seq_ops = &ftrace_graph_seq_ops; in ftrace_graph_open() 6144 fgd->seq_ops = &ftrace_graph_seq_ops; in ftrace_graph_notrace_open() 7532 const struct seq_operations *seq_ops; in pid_open() local 7547 seq_ops = &ftrace_pid_sops; in pid_open() 7550 seq_ops = &ftrace_no_pid_sops; in pid_open() 7558 ret = seq_open(file, seq_ops); in pid_open()
|
/linux/net/core/ |
A D | sock_map.c | 786 .seq_ops = &sock_map_seq_ops, 1378 .seq_ops = &sock_hash_seq_ops,
|
A D | bpf_sk_storage.c | 915 .seq_ops = &bpf_sk_storage_map_seq_ops,
|
/linux/include/linux/ |
A D | bpf.h | 56 const struct seq_operations *seq_ops; member
|
/linux/net/netlink/ |
A D | af_netlink.c | 2839 .seq_ops = &netlink_seq_ops,
|
/linux/net/unix/ |
A D | af_unix.c | 3362 .seq_ops = &bpf_iter_unix_seq_ops,
|
/linux/net/ipv4/ |
A D | tcp_ipv4.c | 3278 .seq_ops = &bpf_iter_tcp_seq_ops,
|
A D | udp.c | 3300 .seq_ops = &bpf_iter_udp_seq_ops,
|
/linux/net/ipv6/ |
A D | route.c | 6644 .seq_ops = &ipv6_route_seq_ops,
|