Home
last modified time | relevance | path

Searched refs:to_dev (Results 1 – 19 of 19) sorted by relevance

/linux-6.6.21/drivers/net/ethernet/mellanox/mlxsw/
Dspectrum_span.c140 const struct net_device *to_dev, in mlxsw_sp1_span_entry_cpu_parms() argument
169 const struct net_device *to_dev, in mlxsw_sp_span_entry_phys_parms() argument
172 sparmsp->dest_port = netdev_priv(to_dev); in mlxsw_sp_span_entry_phys_parms()
412 mlxsw_sp_span_gretap4_route(const struct net_device *to_dev, in mlxsw_sp_span_gretap4_route() argument
415 struct ip_tunnel *tun = netdev_priv(to_dev); in mlxsw_sp_span_gretap4_route()
424 parms = mlxsw_sp_ipip_netdev_parms4(to_dev); in mlxsw_sp_span_gretap4_route()
426 0, 0, dev_net(to_dev), parms.link, tun->fwmark, 0, in mlxsw_sp_span_gretap4_route()
451 const struct net_device *to_dev, in mlxsw_sp_span_entry_gretap4_parms() argument
454 struct ip_tunnel_parm tparm = mlxsw_sp_ipip_netdev_parms4(to_dev); in mlxsw_sp_span_entry_gretap4_parms()
462 if (!(to_dev->flags & IFF_UP) || in mlxsw_sp_span_entry_gretap4_parms()
[all …]
Dspectrum_span.h56 const struct net_device *to_dev; member
71 const struct net_device *to_dev; member
80 bool (*can_handle)(const struct net_device *to_dev);
82 const struct net_device *to_dev,
95 const struct net_device *to_dev);
Dspectrum_matchall.c36 if (!mall_entry->mirror.to_dev) { in mlxsw_sp_mall_port_mirror_add()
41 agent_parms.to_dev = mall_entry->mirror.to_dev; in mlxsw_sp_mall_port_mirror_add()
271 mall_entry->mirror.to_dev = act->dev; in mlxsw_sp_mall_replace()
418 .to_dev = NULL, /* Mirror to CPU. */ in mlxsw_sp2_mall_sample_add()
Dspectrum_acl_flex_actions.c144 agent_parms.to_dev = out_dev; in mlxsw_sp_act_mirror_add()
Dspectrum_qdisc.c1834 .to_dev = mall_entry->mirror.to_dev, in mlxsw_sp_qevent_mirror_configure()
2031 mall_entry->mirror.to_dev = act->dev; in mlxsw_sp_qevent_mall_replace()
Dspectrum.h1194 const struct net_device *to_dev; member
/linux-6.6.21/drivers/dma/ioat/
Ddma.c80 dev_err(to_dev(ioat_chan), "Err(%d): %s\n", in ioat_print_chanerrs()
177 dev_dbg(to_dev(ioat_chan), in __ioat_issue_pending()
213 dev_err(to_dev(ioat_chan), in __ioat_start_null_desc()
218 dev_dbg(to_dev(ioat_chan), in __ioat_start_null_desc()
256 dev_dbg(to_dev(ioat_chan), in __ioat_restart_chan()
317 dev_dbg(to_dev(ioat_chan), "%s: cookie: %d\n", __func__, cookie); in ioat_tx_submit_unlock()
391 descs->virt = dma_alloc_coherent(to_dev(ioat_chan), in ioat_alloc_ring()
398 dma_free_coherent(to_dev(ioat_chan), in ioat_alloc_ring()
420 dma_free_coherent(to_dev(ioat_chan), in ioat_alloc_ring()
472 dev_dbg(to_dev(ioat_chan), "%s: num_descs: %d (%x:%x:%x)\n", in ioat_check_space_lock()
[all …]
Dprep.c279 struct device *dev = to_dev(ioat_chan); in dump_pq_desc_dbg()
306 struct device *dev = to_dev(ioat_chan); in dump_pq16_desc_dbg()
358 dev_dbg(to_dev(ioat_chan), "%s\n", __func__); in __ioat_prep_pq_lock()
480 dev_dbg(to_dev(ioat_chan), "%s\n", __func__); in __ioat_prep_pq16_lock()
507 dev_err(to_dev(ioat_chan), in __ioat_prep_pq16_lock()
Ddma.h23 #define to_dev(ioat_chan) (&(ioat_chan)->ioat_dma->pdev->dev) macro
222 struct device *dev = to_dev(ioat_chan); in __dump_desc_dbg()
Dsysfs.c88 dev_warn(to_dev(ioat_chan), in ioat_kobject_add()
Dinit.c637 dev_dbg(to_dev(ioat_chan), "freeing %d idle descriptors\n", descs); in ioat_free_chan_resources()
644 dev_err(to_dev(ioat_chan), "Freeing %d in use descriptors!\n", in ioat_free_chan_resources()
654 dma_free_coherent(to_dev(ioat_chan), IOAT_CHUNK_SIZE, in ioat_free_chan_resources()
758 dev_WARN(to_dev(ioat_chan), in ioat_alloc_chan_resources()
/linux-6.6.21/samples/bpf/
Dxdp_sample.bpf.c207 const struct net_device *to_dev, int sent, int drops, int err) in BPF_PROG() argument
214 idx_out = to_dev->ifindex; in BPF_PROG()
238 const struct net_device *to_dev, int sent, int drops, int err) in BPF_PROG() argument
246 idx_out = to_dev->ifindex; in BPF_PROG()
/linux-6.6.21/drivers/nvdimm/
Dbtt.c27 static struct device *to_dev(struct arena_info *arena) in to_dev() function
68 dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->infooff, 512), in btt_info_write()
70 dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->info2off, 512), in btt_info_write()
100 dev_err_ratelimited(to_dev(arena), in __btt_map_write()
141 dev_err_ratelimited(to_dev(arena), in btt_map_write()
159 dev_err_ratelimited(to_dev(arena), in btt_map_read()
340 dev_err(to_dev(arena), in btt_log_read()
423 dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->mapoff, 512), in btt_map_init()
429 dev_WARN_ONCE(to_dev(arena), size < 512, in btt_map_init()
467 dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->logoff, 512), in btt_log_init()
[all …]
Dpmem.c35 static struct device *to_dev(struct pmem_device *pmem) in to_dev() function
46 return to_nd_region(to_dev(pmem)->parent); in to_region()
102 long cleared = nvdimm_clear_poison(to_dev(pmem), phys, len); in __pmem_clear_poison()
/linux-6.6.21/tools/testing/selftests/net/forwarding/
Dmirror_lib.sh7 local to_dev=$1; shift
12 action mirred egress mirror dev $to_dev
/linux-6.6.21/include/trace/events/
Dxdp.h271 const struct net_device *to_dev,
274 TP_ARGS(from_dev, to_dev, sent, drops, err),
288 __entry->to_ifindex = to_dev->ifindex;
/linux-6.6.21/tools/testing/selftests/bpf/progs/
Dcore_kern.c44 *from_dev, const struct net_device *to_dev, int sent, int drops, in BPF_PROG() argument
/linux-6.6.21/drivers/gpu/drm/amd/amdkfd/
Dkfd_crat.c1144 struct kfd_topology_device *dev, *to_dev; in kfd_parse_subtype_iolink() local
1194 to_dev = kfd_topology_device_by_proximity_domain_no_lock(id_to); in kfd_parse_subtype_iolink()
1195 if (!to_dev) in kfd_parse_subtype_iolink()
1205 to_dev->node_props.io_links_count++; in kfd_parse_subtype_iolink()
1206 list_add_tail(&props2->list, &to_dev->io_link_props); in kfd_parse_subtype_iolink()
Dkfd_topology.c1219 static void kfd_set_iolink_non_coherent(struct kfd_topology_device *to_dev, in kfd_set_iolink_non_coherent() argument
1224 if (!to_dev->gpu && in kfd_set_iolink_non_coherent()
1228 if (to_dev->gpu) { in kfd_set_iolink_non_coherent()
1234 KFD_GC_VERSION(to_dev->gpu) == IP_VERSION(9, 4, 0))) { in kfd_set_iolink_non_coherent()