/linux/drivers/thunderbolt/ |
A D | xdomain.c | 610 if (xd) in tb_xdp_handle_request() 615 if (xd) { in tb_xdp_handle_request() 629 if (xd && device_is_registered(&xd->dev)) { in tb_xdp_handle_request() 1073 xd->remote_uuid, xd->properties_retries, in tb_xdomain_get_properties() 1136 xd->vendor, xd->device); in tb_xdomain_get_properties() 1137 if (xd->vendor_name && xd->device_name) in tb_xdomain_get_properties() 1138 dev_info(&xd->dev, "%s %s\n", xd->vendor_name, in tb_xdomain_get_properties() 1385 xd = kzalloc(sizeof(*xd), GFP_KERNEL); in tb_xdomain_alloc() 1386 if (!xd) in tb_xdomain_alloc() 1717 if (xd) in switch_find_xdomain() [all …]
|
A D | icm.c | 690 if (!xd) in add_xdomain() 824 if (xd) { in icm_fr_device_connected() 925 if (xd) { in icm_fr_xdomain_connected() 928 xd_phy_port = phy_port_from_route(xd->route, xd->depth); in icm_fr_xdomain_connected() 953 if (!xd) { in icm_fr_xdomain_connected() 961 if (xd) { in icm_fr_xdomain_connected() 1002 if (xd) { in icm_fr_xdomain_disconnected() 1259 if (xd) { in __icm_tr_device_connected() 1341 if (xd) { in icm_tr_xdomain_connected() 1354 if (xd) { in icm_tr_xdomain_connected() [all …]
|
A D | dma_test.c | 94 struct tb_xdomain *xd; member 125 tb_xdomain_release_in_hopid(dt->xd, dt->rx_hopid); in dma_test_free_rings() 139 struct tb_xdomain *xd = dt->xd; in dma_test_start_rings() local 160 ret = tb_xdomain_alloc_out_hopid(xd, -1); in dma_test_start_rings() 185 ret = tb_xdomain_alloc_in_hopid(xd, -1); in dma_test_start_rings() 194 ret = tb_xdomain_enable_paths(dt->xd, dt->tx_hopid, in dma_test_start_rings() 468 return tb_xdomain_lane_bonding_enable(dt->xd); in dma_test_set_bonding() 470 tb_xdomain_lane_bonding_disable(dt->xd); in dma_test_set_bonding() 493 dt->xd->link_width != dt->link_width) { in dma_test_check_errors() 638 struct tb_xdomain *xd = tb_service_parent(svc); in dma_test_probe() local [all …]
|
A D | tb.c | 186 struct tb_xdomain *xd; in tb_scan_xdomain() local 194 if (xd) { in tb_scan_xdomain() 195 tb_xdomain_put(xd); in tb_scan_xdomain() 201 if (xd) { in tb_scan_xdomain() 204 tb_xdomain_add(xd); in tb_scan_xdomain() 1092 sw = tb_to_switch(xd->dev.parent); in tb_approve_xdomain_paths() 1126 sw = tb_to_switch(xd->dev.parent); in __tb_disconnect_xdomain_paths() 1146 if (!xd->is_unplugged) { in tb_disconnect_xdomain_paths() 1228 xd->is_unplugged = true; in tb_handle_hotplug() 1229 tb_xdomain_remove(xd); in tb_handle_hotplug() [all …]
|
A D | domain.c | 805 int tb_domain_approve_xdomain_paths(struct tb *tb, struct tb_xdomain *xd, in tb_domain_approve_xdomain_paths() argument 812 return tb->cm_ops->approve_xdomain_paths(tb, xd, transmit_path, in tb_domain_approve_xdomain_paths() 832 int tb_domain_disconnect_xdomain_paths(struct tb *tb, struct tb_xdomain *xd, in tb_domain_disconnect_xdomain_paths() argument 839 return tb->cm_ops->disconnect_xdomain_paths(tb, xd, transmit_path, in tb_domain_disconnect_xdomain_paths() 845 struct tb_xdomain *xd; in disconnect_xdomain() local 849 xd = tb_to_xdomain(dev); in disconnect_xdomain() 850 if (xd && xd->tb == tb) in disconnect_xdomain() 851 ret = tb_xdomain_disable_all_paths(xd); in disconnect_xdomain()
|
/linux/fs/jffs2/ |
A D | xattr.c | 247 xd->hashkey = xattr_datum_hashkey(xd->xprefix, xd->xname, xd->xvalue, xd->value_len); in do_load_xattr_datum() 258 xd->xid, xd->xprefix, xd->xname); in do_load_xattr_datum() 331 xd->xid, xd->version, xd->xprefix, xd->xname); in save_xattr_datum() 561 ref->xd = xd; in create_xattr_ref() 581 xd = ref->xd; in delete_xattr_ref() 623 xd = ref->xd; in jffs2_xattr_free_inode() 852 ref->xd = xd; in jffs2_build_xattr_subsystem() 983 xd = ref->xd; in jffs2_listxattr() 1050 xd = ref->xd; in do_jffs2_getxattr() 1120 xd = ref->xd; in do_jffs2_setxattr() [all …]
|
A D | malloc.c | 279 struct jffs2_xattr_datum *xd; in jffs2_alloc_xattr_datum() local 280 xd = kmem_cache_zalloc(xattr_datum_cache, GFP_KERNEL); in jffs2_alloc_xattr_datum() 281 dbg_memalloc("%p\n", xd); in jffs2_alloc_xattr_datum() 282 if (!xd) in jffs2_alloc_xattr_datum() 285 xd->class = RAWNODE_CLASS_XATTR_DATUM; in jffs2_alloc_xattr_datum() 286 xd->node = (void *)xd; in jffs2_alloc_xattr_datum() 287 INIT_LIST_HEAD(&xd->xindex); in jffs2_alloc_xattr_datum() 288 return xd; in jffs2_alloc_xattr_datum() 291 void jffs2_free_xattr_datum(struct jffs2_xattr_datum *xd) in jffs2_free_xattr_datum() argument 293 dbg_memalloc("%p\n", xd); in jffs2_free_xattr_datum() [all …]
|
A D | scan.c | 332 struct jffs2_xattr_datum *xd; in jffs2_scan_xattr_node() local 359 if (IS_ERR(xd)) in jffs2_scan_xattr_node() 360 return PTR_ERR(xd); in jffs2_scan_xattr_node() 362 if (xd->version > version) { in jffs2_scan_xattr_node() 365 raw->next_in_ino = xd->node->next_in_ino; in jffs2_scan_xattr_node() 366 xd->node->next_in_ino = raw; in jffs2_scan_xattr_node() 368 xd->version = version; in jffs2_scan_xattr_node() 369 xd->xprefix = rx->xprefix; in jffs2_scan_xattr_node() 370 xd->name_len = rx->name_len; in jffs2_scan_xattr_node() 372 xd->data_crc = je32_to_cpu(rx->data_crc); in jffs2_scan_xattr_node() [all …]
|
A D | xattr.h | 59 struct jffs2_xattr_datum *xd; /* reference to jffs2_xattr_datum */ member 84 extern int jffs2_garbage_collect_xattr_datum(struct jffs2_sb_info *c, struct jffs2_xattr_datum *xd, 89 extern void jffs2_release_xattr_datum(struct jffs2_sb_info *c, struct jffs2_xattr_datum *xd);
|
A D | summary.c | 492 struct jffs2_xattr_datum *xd; in jffs2_sum_process_sum_data() local 501 xd = jffs2_setup_xattr_datum(c, je32_to_cpu(spx->xid), in jffs2_sum_process_sum_data() 503 if (IS_ERR(xd)) in jffs2_sum_process_sum_data() 504 return PTR_ERR(xd); in jffs2_sum_process_sum_data() 505 if (xd->version > je32_to_cpu(spx->version)) { in jffs2_sum_process_sum_data() 510 raw->next_in_ino = xd->node->next_in_ino; in jffs2_sum_process_sum_data() 511 xd->node->next_in_ino = raw; in jffs2_sum_process_sum_data() 513 xd->version = je32_to_cpu(spx->version); in jffs2_sum_process_sum_data() 515 PAD(je32_to_cpu(spx->totlen)), (void *)xd); in jffs2_sum_process_sum_data()
|
/linux/arch/powerpc/sysdev/xive/ |
A D | common.c | 385 xd->stale_p = false; in xive_do_source_eoi() 434 xd->stale_p = true; in xive_irq_eoi() 835 xd->stale_p = !xd->saved_p; in xive_irq_set_vcpu_affinity() 864 if (xd->saved_p) { in xive_irq_set_vcpu_affinity() 922 if (!xd->saved_p) in xive_irq_set_vcpu_affinity() 982 if (xd->eoi_mmio == xd->trig_mmio) in xive_cleanup_irq_data() 999 if (!xd) in xive_irq_alloc_data() 1003 kfree(xd); in xive_irq_alloc_data() 1025 if (!xd) in xive_irq_free_data() 1029 kfree(xd); in xive_irq_free_data() [all …]
|
/linux/drivers/dma/ |
A D | uniphier-xdmac.c | 139 src_addr = xd->nodes[xd->cur_node].src; in uniphier_xdmac_chan_start() 140 dst_addr = xd->nodes[xd->cur_node].dst; in uniphier_xdmac_chan_start() 141 its = xd->nodes[xd->cur_node].burst_size; in uniphier_xdmac_chan_start() 142 tnum = xd->nodes[xd->cur_node].nr_burst; in uniphier_xdmac_chan_start() 222 if (xd) in uniphier_xdmac_start() 226 xc->xd = xd; in uniphier_xdmac_start() 249 if (xc->xd->cur_node >= xc->xd->nr_node) { in uniphier_xdmac_chan_irq() 294 xd = kzalloc(struct_size(xd, nodes, nr), GFP_NOWAIT); in uniphier_xdmac_prep_dma_memcpy() 295 if (!xd) in uniphier_xdmac_prep_dma_memcpy() 351 if (!xd) in uniphier_xdmac_prep_slave_sg() [all …]
|
/linux/include/linux/ |
A D | thunderbolt.h | 281 struct tb_xdomain *xd; in tb_xdomain_find_by_uuid_locked() local 284 xd = tb_xdomain_find_by_uuid(tb, uuid); in tb_xdomain_find_by_uuid_locked() 287 return xd; in tb_xdomain_find_by_uuid_locked() 293 struct tb_xdomain *xd; in tb_xdomain_find_by_route_locked() local 296 xd = tb_xdomain_find_by_route(tb, route); in tb_xdomain_find_by_route_locked() 299 return xd; in tb_xdomain_find_by_route_locked() 304 if (xd) in tb_xdomain_get() 305 get_device(&xd->dev); in tb_xdomain_get() 306 return xd; in tb_xdomain_get() 311 if (xd) in tb_xdomain_put() [all …]
|
/linux/arch/powerpc/kvm/ |
A D | book3s_xive_template.c | 64 val =__x_readq(__x_eoi_page(xd) + offset); in GLUE() 75 if (xd->flags & XIVE_IRQ_FLAG_STORE_EOI) in GLUE() 77 else if (xd->flags & XIVE_IRQ_FLAG_LSI) { in GLUE() 99 if ((eoi_val & 1) && __x_trig_page(xd)) in GLUE() 100 __x_writeq(0, __x_trig_page(xd)); in GLUE() 382 struct xive_irq_data *xd; in GLUE() local 423 if (!(xd->flags & XIVE_IRQ_FLAG_LSI)) in GLUE() 427 GLUE(X_PFX,source_eoi)(hw_num, xd); in GLUE() 502 struct xive_irq_data *xd; in GLUE() local 569 GLUE(X_PFX,source_eoi)(hw_num, xd); in GLUE() [all …]
|
A D | book3s_hv_rm_xive.c | 40 #define __x_eoi_page(xd) ((void __iomem *)((xd)->eoi_page)) argument 41 #define __x_trig_page(xd) ((void __iomem *)((xd)->trig_page)) argument
|
A D | book3s_xive.c | 46 #define __x_eoi_page(xd) ((void __iomem *)((xd)->eoi_mmio)) argument 47 #define __x_trig_page(xd) ((void __iomem *)((xd)->trig_mmio)) argument 227 if (WARN_ON(!xd->trig_mmio)) in xive_irq_trigger() 230 out_be64(xd->trig_mmio, 0); in xive_irq_trigger() 481 struct xive_irq_data *xd; in xive_lock_and_mask() local 541 struct xive_irq_data *xd; in xive_finish_unmask() local 1192 xd->stale_p = false; in xive_cleanup_single_escalation() 1195 xd->stale_p = true; in xive_cleanup_single_escalation() 2235 xd->eoi_page); in kvmppc_xive_debug_show_queues() 2250 struct xive_irq_data *xd; in kvmppc_xive_debug_show_sources() local [all …]
|
A D | book3s_xive_native.c | 30 static u8 xive_vm_esb_load(struct xive_irq_data *xd, u32 offset) in xive_vm_esb_load() argument 40 val = in_be64(xd->eoi_mmio + offset); in xive_vm_esb_load() 236 struct xive_irq_data *xd; in xive_native_esb_fault() local 264 kvmppc_xive_select_irq(state, &hw_num, &xd); in xive_native_esb_fault() 272 page = page_offset % 2 ? xd->eoi_page : xd->trig_page; in xive_native_esb_fault() 515 struct xive_irq_data *xd; in kvmppc_xive_native_sync_source() local 533 kvmppc_xive_select_irq(state, &hw_num, &xd); in kvmppc_xive_native_sync_source() 863 struct xive_irq_data *xd; in kvmppc_xive_native_sync_sources() local 886 kvmppc_xive_select_irq(state, &hw_num, &xd); in kvmppc_xive_native_sync_sources()
|
/linux/drivers/net/ |
A D | thunderbolt.c | 178 struct tb_xdomain *xd; member 236 struct tb_xdomain *xd = net->xd; in tbnet_login_response() local 253 struct tb_xdomain *xd = net->xd; in tbnet_login_request() local 256 tbnet_fill_header(&request.hdr, xd->route, sequence, xd->local_uuid, in tbnet_login_request() 273 struct tb_xdomain *xd = net->xd; in tbnet_logout_response() local 287 struct tb_xdomain *xd = net->xd; in tbnet_logout_request() local 290 tbnet_fill_header(&request.hdr, xd->route, 0, xd->local_uuid, in tbnet_logout_request() 419 if (route != net->xd->route) in tbnet_handle_packet() 867 struct tb_xdomain *xd = net->xd; in tbnet_open() local 1204 const struct tb_xdomain *xd = net->xd; in tbnet_generate_mac() local [all …]
|
/linux/tools/testing/selftests/arm64/fp/ |
A D | sve-test.S | 91 .macro _adrz xd, xn, nrtmp 92 ldr \xd, =zref 94 madd \xd, x\nrtmp, \xn, \xd 98 .macro _adrp xd, xn, nrtmp 99 ldr \xd, =pref 103 madd \xd, x\nrtmp, \xn, \xd
|
A D | fpsimd-test.S | 66 .macro _adrv xd, xn, nrtmp 67 ldr \xd, =vref 69 madd \xd, x\nrtmp, \xn, \xd
|
/linux/arch/powerpc/platforms/powernv/ |
A D | vas.c | 54 struct xive_irq_data *xd; in init_vas_instance() local 124 xd = irq_get_handler_data(vinst->virq); in init_vas_instance() 125 if (!xd) { in init_vas_instance() 131 vinst->irq_port = xd->trig_page; in init_vas_instance()
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
A D | gddr5.c | 37 int pd, lf, xd, vh, vr, vo, l3; in nvkm_gddr5_calc() local 41 xd = !ram->next->bios.ramcfg_DLLoff; in nvkm_gddr5_calc() 81 ram->mr[1] |= (xd & 0x01) << 7; in nvkm_gddr5_calc()
|
/linux/drivers/misc/ocxl/ |
A D | afu_irq.c | 198 struct xive_irq_data *xd; in ocxl_afu_irq_get_addr() local 205 xd = irq_get_handler_data(irq->virq); in ocxl_afu_irq_get_addr() 206 addr = xd ? xd->trig_page : 0; in ocxl_afu_irq_get_addr()
|
/linux/fs/jfs/ |
A D | jfs_dtree.h | 20 pxd_t xd; member 53 pxd_t xd; /* 8: child extent descriptor */ member
|
/linux/drivers/staging/rts5208/ |
A D | Makefile | 5 rtsx_card.o general.o sd.o xd.o ms.o spi.o
|