Home
last modified time | relevance | path

Searched refs:xd (Results 1 – 25 of 85) sorted by relevance

1234

/linux-6.1.9/drivers/thunderbolt/
Dxdomain.c164 int tb_xdomain_response(struct tb_xdomain *xd, const void *response, in tb_xdomain_response() argument
167 return __tb_xdomain_response(xd->tb->ctl, response, size, type); in tb_xdomain_response()
216 int tb_xdomain_request(struct tb_xdomain *xd, const void *request, in tb_xdomain_request() argument
221 return __tb_xdomain_request(xd->tb->ctl, request, request_size, in tb_xdomain_request()
415 struct tb_xdomain *xd, u8 sequence, const struct tb_xdp_properties *req) in tb_xdp_properties_response() argument
427 if (!uuid_equal(xd->local_uuid, &req->dst_uuid)) { in tb_xdp_properties_response()
428 tb_xdp_error_response(ctl, xd->route, sequence, in tb_xdp_properties_response()
433 mutex_lock(&xd->lock); in tb_xdp_properties_response()
435 if (req->offset >= xd->local_property_block_len) { in tb_xdp_properties_response()
436 mutex_unlock(&xd->lock); in tb_xdp_properties_response()
[all …]
Dicm.c560 static int icm_fr_approve_xdomain_paths(struct tb *tb, struct tb_xdomain *xd, in icm_fr_approve_xdomain_paths() argument
570 request.link_info = xd->depth << ICM_LINK_INFO_DEPTH_SHIFT | xd->link; in icm_fr_approve_xdomain_paths()
571 memcpy(&request.remote_uuid, xd->remote_uuid, sizeof(*xd->remote_uuid)); in icm_fr_approve_xdomain_paths()
590 static int icm_fr_disconnect_xdomain_paths(struct tb *tb, struct tb_xdomain *xd, in icm_fr_disconnect_xdomain_paths() argument
597 phy_port = tb_phy_port_from_link(xd->link); in icm_fr_disconnect_xdomain_paths()
685 struct tb_xdomain *xd; in add_xdomain() local
689 xd = tb_xdomain_alloc(sw->tb, &sw->dev, route, local_uuid, remote_uuid); in add_xdomain()
690 if (!xd) in add_xdomain()
693 xd->link = link; in add_xdomain()
694 xd->depth = depth; in add_xdomain()
[all …]
Ddma_test.c94 struct tb_xdomain *xd; member
125 tb_xdomain_release_in_hopid(dt->xd, dt->rx_hopid); in dma_test_free_rings()
130 tb_xdomain_release_out_hopid(dt->xd, dt->tx_hopid); in dma_test_free_rings()
139 struct tb_xdomain *xd = dt->xd; in dma_test_start_rings() local
152 ring = tb_ring_alloc_tx(xd->tb->nhi, -1, DMA_TEST_TX_RING_SIZE, in dma_test_start_rings()
160 ret = tb_xdomain_alloc_out_hopid(xd, -1); in dma_test_start_rings()
175 ring = tb_ring_alloc_rx(xd->tb->nhi, -1, DMA_TEST_RX_RING_SIZE, in dma_test_start_rings()
185 ret = tb_xdomain_alloc_in_hopid(xd, -1); in dma_test_start_rings()
194 ret = tb_xdomain_enable_paths(dt->xd, dt->tx_hopid, in dma_test_start_rings()
220 ret = tb_xdomain_disable_paths(dt->xd, dt->tx_hopid, in dma_test_stop_rings()
[all …]
Dtb.c203 static int tb_port_configure_xdomain(struct tb_port *port, struct tb_xdomain *xd) in tb_port_configure_xdomain() argument
206 return usb4_port_configure_xdomain(port, xd); in tb_port_configure_xdomain()
224 struct tb_xdomain *xd; in tb_scan_xdomain() local
231 xd = tb_xdomain_find_by_route(tb, route); in tb_scan_xdomain()
232 if (xd) { in tb_scan_xdomain()
233 tb_xdomain_put(xd); in tb_scan_xdomain()
237 xd = tb_xdomain_alloc(tb, &sw->dev, route, tb->root_switch->uuid, in tb_scan_xdomain()
239 if (xd) { in tb_scan_xdomain()
240 tb_port_at(route, sw)->xdomain = xd; in tb_scan_xdomain()
241 tb_port_configure_xdomain(port, xd); in tb_scan_xdomain()
[all …]
Ddomain.c797 int tb_domain_approve_xdomain_paths(struct tb *tb, struct tb_xdomain *xd, in tb_domain_approve_xdomain_paths() argument
804 return tb->cm_ops->approve_xdomain_paths(tb, xd, transmit_path, in tb_domain_approve_xdomain_paths()
824 int tb_domain_disconnect_xdomain_paths(struct tb *tb, struct tb_xdomain *xd, in tb_domain_disconnect_xdomain_paths() argument
831 return tb->cm_ops->disconnect_xdomain_paths(tb, xd, transmit_path, in tb_domain_disconnect_xdomain_paths()
837 struct tb_xdomain *xd; in disconnect_xdomain() local
841 xd = tb_to_xdomain(dev); in disconnect_xdomain()
842 if (xd && xd->tb == tb) in disconnect_xdomain()
843 ret = tb_xdomain_disable_all_paths(xd); in disconnect_xdomain()
Dtb.h487 int (*approve_xdomain_paths)(struct tb *tb, struct tb_xdomain *xd,
490 int (*disconnect_xdomain_paths)(struct tb *tb, struct tb_xdomain *xd,
727 int tb_domain_approve_xdomain_paths(struct tb *tb, struct tb_xdomain *xd,
730 int tb_domain_disconnect_xdomain_paths(struct tb *tb, struct tb_xdomain *xd,
1143 void tb_xdomain_add(struct tb_xdomain *xd);
1144 void tb_xdomain_remove(struct tb_xdomain *xd);
1148 static inline struct tb_switch *tb_xdomain_parent(struct tb_xdomain *xd) in tb_xdomain_parent() argument
1150 return tb_to_switch(xd->dev.parent); in tb_xdomain_parent()
1200 int usb4_port_configure_xdomain(struct tb_port *port, struct tb_xdomain *xd);
1295 void tb_xdomain_debugfs_init(struct tb_xdomain *xd);
[all …]
Ddebugfs.c986 static void margining_xdomain_init(struct tb_xdomain *xd) in margining_xdomain_init() argument
991 parent_sw = tb_xdomain_parent(xd); in margining_xdomain_init()
992 downstream = tb_port_at(xd->route, parent_sw); in margining_xdomain_init()
997 static void margining_xdomain_remove(struct tb_xdomain *xd) in margining_xdomain_remove() argument
1002 parent_sw = tb_xdomain_parent(xd); in margining_xdomain_remove()
1003 downstream = tb_port_at(xd->route, parent_sw); in margining_xdomain_remove()
1009 static inline void margining_xdomain_init(struct tb_xdomain *xd) { } in margining_xdomain_init() argument
1010 static inline void margining_xdomain_remove(struct tb_xdomain *xd) { } in margining_xdomain_remove() argument
1531 void tb_xdomain_debugfs_init(struct tb_xdomain *xd) in tb_xdomain_debugfs_init() argument
1533 margining_xdomain_init(xd); in tb_xdomain_debugfs_init()
[all …]
/linux-6.1.9/fs/jffs2/
Dxattr.c71 static int is_xattr_datum_unchecked(struct jffs2_sb_info *c, struct jffs2_xattr_datum *xd) in is_xattr_datum_unchecked() argument
77 for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) { in is_xattr_datum_unchecked()
87 static void unload_xattr_datum(struct jffs2_sb_info *c, struct jffs2_xattr_datum *xd) in unload_xattr_datum() argument
90 D1(dbg_xattr("%s: xid=%u, version=%u\n", __func__, xd->xid, xd->version)); in unload_xattr_datum()
91 if (xd->xname) { in unload_xattr_datum()
92 c->xdatum_mem_usage -= (xd->name_len + 1 + xd->value_len); in unload_xattr_datum()
93 kfree(xd->xname); in unload_xattr_datum()
96 list_del_init(&xd->xindex); in unload_xattr_datum()
97 xd->hashkey = 0; in unload_xattr_datum()
98 xd->xname = NULL; in unload_xattr_datum()
[all …]
Dmalloc.c279 struct jffs2_xattr_datum *xd; in jffs2_alloc_xattr_datum() local
280 xd = kmem_cache_zalloc(xattr_datum_cache, GFP_KERNEL); in jffs2_alloc_xattr_datum()
281 dbg_memalloc("%p\n", xd); in jffs2_alloc_xattr_datum()
282 if (!xd) in jffs2_alloc_xattr_datum()
285 xd->class = RAWNODE_CLASS_XATTR_DATUM; in jffs2_alloc_xattr_datum()
286 xd->node = (void *)xd; in jffs2_alloc_xattr_datum()
287 INIT_LIST_HEAD(&xd->xindex); in jffs2_alloc_xattr_datum()
288 return xd; in jffs2_alloc_xattr_datum()
291 void jffs2_free_xattr_datum(struct jffs2_xattr_datum *xd) in jffs2_free_xattr_datum() argument
293 dbg_memalloc("%p\n", xd); in jffs2_free_xattr_datum()
[all …]
Dxattr.h59 struct jffs2_xattr_datum *xd; /* reference to jffs2_xattr_datum */ member
84 extern int jffs2_garbage_collect_xattr_datum(struct jffs2_sb_info *c, struct jffs2_xattr_datum *xd,
89 extern void jffs2_release_xattr_datum(struct jffs2_sb_info *c, struct jffs2_xattr_datum *xd);
Dscan.c334 struct jffs2_xattr_datum *xd; in jffs2_scan_xattr_node() local
360 xd = jffs2_setup_xattr_datum(c, xid, version); in jffs2_scan_xattr_node()
361 if (IS_ERR(xd)) in jffs2_scan_xattr_node()
362 return PTR_ERR(xd); in jffs2_scan_xattr_node()
364 if (xd->version > version) { in jffs2_scan_xattr_node()
367 raw->next_in_ino = xd->node->next_in_ino; in jffs2_scan_xattr_node()
368 xd->node->next_in_ino = raw; in jffs2_scan_xattr_node()
370 xd->version = version; in jffs2_scan_xattr_node()
371 xd->xprefix = rx->xprefix; in jffs2_scan_xattr_node()
372 xd->name_len = rx->name_len; in jffs2_scan_xattr_node()
[all …]
Dsummary.c492 struct jffs2_xattr_datum *xd; in jffs2_sum_process_sum_data() local
501 xd = jffs2_setup_xattr_datum(c, je32_to_cpu(spx->xid), in jffs2_sum_process_sum_data()
503 if (IS_ERR(xd)) in jffs2_sum_process_sum_data()
504 return PTR_ERR(xd); in jffs2_sum_process_sum_data()
505 if (xd->version > je32_to_cpu(spx->version)) { in jffs2_sum_process_sum_data()
510 raw->next_in_ino = xd->node->next_in_ino; in jffs2_sum_process_sum_data()
511 xd->node->next_in_ino = raw; in jffs2_sum_process_sum_data()
513 xd->version = je32_to_cpu(spx->version); in jffs2_sum_process_sum_data()
515 PAD(je32_to_cpu(spx->totlen)), (void *)xd); in jffs2_sum_process_sum_data()
/linux-6.1.9/arch/powerpc/sysdev/xive/
Dcommon.c92 static bool xive_is_store_eoi(struct xive_irq_data *xd) in xive_is_store_eoi() argument
94 return xd->flags & XIVE_IRQ_FLAG_STORE_EOI && xive_store_eoi; in xive_is_store_eoi()
217 static notrace u8 xive_esb_read(struct xive_irq_data *xd, u32 offset) in xive_esb_read() argument
221 if (offset == XIVE_ESB_SET_PQ_10 && xive_is_store_eoi(xd)) in xive_esb_read()
224 if ((xd->flags & XIVE_IRQ_FLAG_H_INT_ESB) && xive_ops->esb_rw) in xive_esb_read()
225 val = xive_ops->esb_rw(xd->hw_irq, offset, 0, 0); in xive_esb_read()
227 val = in_be64(xd->eoi_mmio + offset); in xive_esb_read()
232 static void xive_esb_write(struct xive_irq_data *xd, u32 offset, u64 data) in xive_esb_write() argument
234 if ((xd->flags & XIVE_IRQ_FLAG_H_INT_ESB) && xive_ops->esb_rw) in xive_esb_write()
235 xive_ops->esb_rw(xd->hw_irq, offset, data, 1); in xive_esb_write()
[all …]
/linux-6.1.9/drivers/dma/
Duniphier-xdmac.c89 struct uniphier_xdmac_desc *xd; member
132 struct uniphier_xdmac_desc *xd) in uniphier_xdmac_chan_start() argument
140 src_addr = xd->nodes[xd->cur_node].src; in uniphier_xdmac_chan_start()
141 dst_addr = xd->nodes[xd->cur_node].dst; in uniphier_xdmac_chan_start()
142 its = xd->nodes[xd->cur_node].burst_size; in uniphier_xdmac_chan_start()
143 tnum = xd->nodes[xd->cur_node].nr_burst; in uniphier_xdmac_chan_start()
149 if (xd->dir == DMA_DEV_TO_MEM) { in uniphier_xdmac_chan_start()
158 if (xd->dir == DMA_MEM_TO_DEV) { in uniphier_xdmac_chan_start()
220 struct uniphier_xdmac_desc *xd; in uniphier_xdmac_start() local
222 xd = uniphier_xdmac_next_desc(xc); in uniphier_xdmac_start()
[all …]
/linux-6.1.9/include/linux/
Dthunderbolt.h260 int tb_xdomain_lane_bonding_enable(struct tb_xdomain *xd);
261 void tb_xdomain_lane_bonding_disable(struct tb_xdomain *xd);
262 int tb_xdomain_alloc_in_hopid(struct tb_xdomain *xd, int hopid);
263 void tb_xdomain_release_in_hopid(struct tb_xdomain *xd, int hopid);
264 int tb_xdomain_alloc_out_hopid(struct tb_xdomain *xd, int hopid);
265 void tb_xdomain_release_out_hopid(struct tb_xdomain *xd, int hopid);
266 int tb_xdomain_enable_paths(struct tb_xdomain *xd, int transmit_path,
269 int tb_xdomain_disable_paths(struct tb_xdomain *xd, int transmit_path,
273 static inline int tb_xdomain_disable_all_paths(struct tb_xdomain *xd) in tb_xdomain_disable_all_paths() argument
275 return tb_xdomain_disable_paths(xd, -1, -1, -1, -1); in tb_xdomain_disable_all_paths()
[all …]
/linux-6.1.9/drivers/net/
Dthunderbolt.c180 struct tb_xdomain *xd; member
242 struct tb_xdomain *xd = net->xd; in tbnet_login_response() local
245 tbnet_fill_header(&reply.hdr, route, sequence, xd->local_uuid, in tbnet_login_response()
246 xd->remote_uuid, TBIP_LOGIN_RESPONSE, sizeof(reply), in tbnet_login_response()
251 return tb_xdomain_response(xd, &reply, sizeof(reply), in tbnet_login_response()
259 struct tb_xdomain *xd = net->xd; in tbnet_login_request() local
262 tbnet_fill_header(&request.hdr, xd->route, sequence, xd->local_uuid, in tbnet_login_request()
263 xd->remote_uuid, TBIP_LOGIN, sizeof(request), in tbnet_login_request()
269 return tb_xdomain_request(xd, &request, sizeof(request), in tbnet_login_request()
279 struct tb_xdomain *xd = net->xd; in tbnet_logout_response() local
[all …]
/linux-6.1.9/arch/powerpc/kvm/
Dbook3s_xive.c33 #define __x_eoi_page(xd) ((void __iomem *)((xd)->eoi_mmio)) argument
34 #define __x_trig_page(xd) ((void __iomem *)((xd)->trig_mmio)) argument
80 static u8 xive_vm_esb_load(struct xive_irq_data *xd, u32 offset) in xive_vm_esb_load() argument
84 if (offset == XIVE_ESB_SET_PQ_10 && xd->flags & XIVE_IRQ_FLAG_STORE_EOI) in xive_vm_esb_load()
87 val = __raw_readq(__x_eoi_page(xd) + offset); in xive_vm_esb_load()
95 static void xive_vm_source_eoi(u32 hw_irq, struct xive_irq_data *xd) in xive_vm_source_eoi() argument
98 if (xd->flags & XIVE_IRQ_FLAG_STORE_EOI) in xive_vm_source_eoi()
99 __raw_writeq(0, __x_eoi_page(xd) + XIVE_ESB_STORE_EOI); in xive_vm_source_eoi()
100 else if (xd->flags & XIVE_IRQ_FLAG_LSI) { in xive_vm_source_eoi()
106 __raw_readq(__x_eoi_page(xd) + XIVE_ESB_LOAD_EOI); in xive_vm_source_eoi()
[all …]
Dbook3s_xive_native.c30 static u8 xive_vm_esb_load(struct xive_irq_data *xd, u32 offset) in xive_vm_esb_load() argument
40 val = in_be64(xd->eoi_mmio + offset); in xive_vm_esb_load()
236 struct xive_irq_data *xd; in xive_native_esb_fault() local
264 kvmppc_xive_select_irq(state, &hw_num, &xd); in xive_native_esb_fault()
272 page = page_offset % 2 ? xd->eoi_page : xd->trig_page; in xive_native_esb_fault()
515 struct xive_irq_data *xd; in kvmppc_xive_native_sync_source() local
533 kvmppc_xive_select_irq(state, &hw_num, &xd); in kvmppc_xive_native_sync_source()
863 struct xive_irq_data *xd; in kvmppc_xive_native_sync_sources() local
886 kvmppc_xive_select_irq(state, &hw_num, &xd); in kvmppc_xive_native_sync_sources()
/linux-6.1.9/arch/powerpc/platforms/powernv/
Dvas.c54 struct xive_irq_data *xd; in init_vas_instance() local
124 xd = irq_get_handler_data(vinst->virq); in init_vas_instance()
125 if (!xd) { in init_vas_instance()
131 vinst->irq_port = xd->trig_page; in init_vas_instance()
/linux-6.1.9/tools/testing/selftests/arm64/fp/
Dsve-test.S92 .macro _adrz xd, xn, nrtmp
93 ldr \xd, =zref
95 madd \xd, x\nrtmp, \xn, \xd
99 .macro _adrp xd, xn, nrtmp
100 ldr \xd, =pref
104 madd \xd, x\nrtmp, \xn, \xd
Dfpsimd-test.S66 .macro _adrv xd, xn, nrtmp
67 ldr \xd, =vref
69 madd \xd, x\nrtmp, \xn, \xd
Dza-test.S76 .macro _adrza xd, xn, nrtmp
77 ldr \xd, =zaref
79 madd \xd, x\nrtmp, \xn, \xd
/linux-6.1.9/drivers/gpu/drm/nouveau/nvkm/subdev/fb/
Dgddr5.c37 int pd, lf, xd, vh, vr, vo, l3; in nvkm_gddr5_calc() local
41 xd = !ram->next->bios.ramcfg_DLLoff; in nvkm_gddr5_calc()
81 ram->mr[1] |= (xd & 0x01) << 7; in nvkm_gddr5_calc()
/linux-6.1.9/drivers/misc/ocxl/
Dafu_irq.c199 struct xive_irq_data *xd; in ocxl_afu_irq_get_addr() local
206 xd = irq_get_handler_data(irq->virq); in ocxl_afu_irq_get_addr()
207 addr = xd ? xd->trig_page : 0; in ocxl_afu_irq_get_addr()
/linux-6.1.9/fs/jfs/
Djfs_dtree.h20 pxd_t xd; member
53 pxd_t xd; /* 8: child extent descriptor */ member

1234