/linux-6.6.21/drivers/thunderbolt/ |
D | dma_port.c | 181 ret = dma_port_read(sw->tb->ctl, &type, tb_route(sw), ports[i], in dma_find_port() 249 ret = dma_port_read(sw->tb->ctl, &in, tb_route(sw), dma->port, in dma_port_wait_for_completion() 285 ret = dma_port_write(sw->tb->ctl, &in, tb_route(sw), dma->port, in dma_port_request() 294 ret = dma_port_read(sw->tb->ctl, &out, tb_route(sw), dma->port, in dma_port_request() 320 return dma_port_read(sw->tb->ctl, buf, tb_route(sw), dma->port, in dma_port_flash_read_block() 333 ret = dma_port_write(sw->tb->ctl, buf, tb_route(sw), dma->port, in dma_port_flash_write_block() 426 ret = dma_port_read(sw->tb->ctl, &out, tb_route(sw), dma->port, in dma_port_flash_update_auth_status()
|
D | switch.c | 275 if (!tb_route(sw)) { in nvm_authenticate() 828 return (tb_route(parent) & mask) == (tb_route(sw) & mask); in tb_switch_is_reachable() 861 next = tb_port_at(tb_route(end->sw), prev->sw); in tb_next_port_on_path() 1507 res = tb_cfg_reset(sw->tb->ctl, tb_route(sw)); in tb_switch_reset() 1643 if (tb_route(sw)) { in tb_switch_set_authorized() 2054 if (tb_route(sw) && in switch_attr_is_visible() 2063 if (tb_route(sw)) in switch_attr_is_visible() 2075 if (tb_route(sw)) in switch_attr_is_visible() 2129 if (!tb_route(sw)) { in tb_switch_uevent() 2355 dev_set_name(&sw->dev, "%u-%llx", tb->index, tb_route(sw)); in tb_switch_alloc() [all …]
|
D | tb.h | 559 static inline u64 tb_route(const struct tb_switch *sw) in tb_route() function 639 tb_route(sw), in tb_sw_read() 653 tb_route(sw), in tb_sw_write() 667 tb_route(port->sw), in tb_port_read() 681 tb_route(port->sw), in tb_port_write() 698 tb_route(__sw), ## arg); \ 709 tb_route(__port->sw), __port->port, ## arg); \ 862 if (WARN_ON(!tb_route(sw))) in tb_switch_downstream_port() 864 return tb_port_at(tb_route(sw), tb_switch_parent(sw)); in tb_switch_downstream_port() 1150 return tb_route(port->sw) in tb_downstream_route()
|
D | clx.c | 194 if (!tb_route(sw)) in tb_switch_clx_init() 220 if (!tb_route(sw)) in tb_switch_pm_secondary_resolve() 242 if (!tb_route(sw)) in tb_switch_mask_clx_objections()
|
D | path.c | 170 path->name, tb_route(src->sw), src->port); in tb_path_discover() 472 path->name, tb_route(path->hops[0].in_port->sw), in tb_path_deactivate() 474 tb_route(path->hops[path->path_length - 1].out_port->sw), in tb_path_deactivate() 501 path->name, tb_route(path->hops[0].in_port->sw), in tb_path_activate() 503 tb_route(path->hops[path->path_length - 1].out_port->sw), in tb_path_activate()
|
D | tb.c | 566 port = tb_port_at(tb_route(sw), tb->root_switch); in tb_find_first_usb3_tunnel() 584 tb_route(src_port->sw), src_port->port, tb_route(dst_port->sw), in tb_available_bandwidth() 780 if (tb_route(parent)) { in tb_tunnel_usb3() 820 if (tb_route(parent)) in tb_tunnel_usb3() 828 if (tb_route(parent)) in tb_tunnel_usb3() 842 if (tb_route(sw)) { in tb_create_usb3_tunnels() 893 tb_queue_hotplug(port->sw->tb, tb_route(port->sw), port->port, in tb_scan_port() 1117 } else if (!tb_route(sw)) { in tb_find_pcie_down() 1253 host_port = tb_route(in->sw) ? in tb_find_dp_out() 1254 tb_port_at(tb_route(in->sw), tb->root_switch) : NULL; in tb_find_dp_out() [all …]
|
D | test.c | 511 KUNIT_EXPECT_EQ(test, tb_route(p->sw), test_data[i].route); in tb_test_path_single_hop_walk() 523 KUNIT_EXPECT_EQ(test, tb_route(p->sw), test_data[i].route); in tb_test_path_single_hop_walk() 570 KUNIT_EXPECT_EQ(test, tb_route(p->sw), test_data[i].route); in tb_test_path_daisy_chain_walk() 582 KUNIT_EXPECT_EQ(test, tb_route(p->sw), test_data[i].route); in tb_test_path_daisy_chain_walk() 633 KUNIT_EXPECT_EQ(test, tb_route(p->sw), test_data[i].route); in tb_test_path_simple_tree_walk() 645 KUNIT_EXPECT_EQ(test, tb_route(p->sw), test_data[i].route); in tb_test_path_simple_tree_walk() 717 KUNIT_EXPECT_EQ(test, tb_route(p->sw), test_data[i].route); in tb_test_path_complex_tree_walk() 729 KUNIT_EXPECT_EQ(test, tb_route(p->sw), test_data[i].route); in tb_test_path_complex_tree_walk() 820 KUNIT_EXPECT_EQ(test, tb_route(p->sw), test_data[i].route); in tb_test_path_max_length_walk() 832 KUNIT_EXPECT_EQ(test, tb_route(p->sw), test_data[i].route); in tb_test_path_max_length_walk() [all …]
|
D | tmu.c | 377 if (tb_route(sw)) { in tmu_mode_init() 452 if (!tb_route(sw)) in tb_switch_tmu_post_time() 565 if (tb_route(sw)) { in tb_switch_tmu_disable() 961 if (tb_route(sw)) { in tb_switch_tmu_enable()
|
D | lc.c | 177 if (!tb_route(sw)) in tb_lc_start_lane_initialization() 375 if (!tb_route(sw)) in tb_lc_set_wake()
|
D | icm.c | 633 u64 route = tb_route(sw); in add_switch() 642 tb_port_at(tb_route(sw), parent_sw)->remote = NULL; in add_switch() 782 route = tb_route(sw); in icm_fr_device_connected() 1236 if (tb_route(sw) == route && !!sw->authorized == authorized) { in __icm_tr_device_connected() 2012 if (tb_route(sw)) in icm_unplug_children() 2121 if (tb_route(sw)) in icm_runtime_suspend_switch() 2128 if (tb_route(sw)) { in icm_runtime_resume_switch() 2266 u64 route = tb_route(sw); in icm_usb4_switch_op() 2337 tb_route(sw), auth->reply.hdr.flags, auth->reply.status); in icm_usb4_switch_nvm_authenticate_status()
|
D | usb4.c | 169 if (tb_route(sw)) { in usb4_switch_check_wakes() 249 if (!tb_route(sw)) in usb4_switch_setup() 314 if (!tb_route(sw)) in usb4_switch_configuration_valid() 414 u64 route = tb_route(sw); in usb4_switch_set_wake() 800 if (!tb_route(sw) && max_dma < 0) { in usb4_switch_credits_init() 1985 if (tb_route(port->sw)) in usb4_usb3_port_cm_request()
|
D | tunnel.c | 65 tb_route(__tunnel->src_port->sw), \ 67 tb_route(__tunnel->dst_port->sw), \ 663 if (tb_route(out->sw) && tb_switch_is_titan_ridge(out->sw)) { in tb_dp_xchg_caps() 1948 if (!tb_route(down->sw)) { in tb_tunnel_discover_usb3() 2053 if (!tb_route(down->sw)) { in tb_tunnel_alloc_usb3()
|
D | eeprom.c | 709 if (!tb_route(sw)) in tb_drom_read()
|
D | debugfs.c | 972 u64 route = tb_route(sw); in margining_switch_init() 989 u64 route = tb_route(sw); in margining_switch_remove()
|