Home
last modified time | relevance | path

Searched refs:shmem (Results 1 – 25 of 72) sorted by relevance

123

/linux-6.6.21/drivers/gpu/drm/
Ddrm_gem_shmem_helper.c55 struct drm_gem_shmem_object *shmem; in __drm_gem_shmem_create() local
65 shmem = to_drm_gem_shmem_obj(obj); in __drm_gem_shmem_create()
67 shmem = kzalloc(sizeof(*shmem), GFP_KERNEL); in __drm_gem_shmem_create()
68 if (!shmem) in __drm_gem_shmem_create()
70 obj = &shmem->base; in __drm_gem_shmem_create()
78 shmem->map_wc = false; /* dma-buf mappings use always writecombine */ in __drm_gem_shmem_create()
91 INIT_LIST_HEAD(&shmem->madv_list); in __drm_gem_shmem_create()
105 return shmem; in __drm_gem_shmem_create()
138 void drm_gem_shmem_free(struct drm_gem_shmem_object *shmem) in drm_gem_shmem_free() argument
140 struct drm_gem_object *obj = &shmem->base; in drm_gem_shmem_free()
[all …]
/linux-6.6.21/include/drm/
Ddrm_gem_shmem_helper.h100 void drm_gem_shmem_free(struct drm_gem_shmem_object *shmem);
102 void drm_gem_shmem_put_pages(struct drm_gem_shmem_object *shmem);
103 int drm_gem_shmem_pin(struct drm_gem_shmem_object *shmem);
104 void drm_gem_shmem_unpin(struct drm_gem_shmem_object *shmem);
105 int drm_gem_shmem_vmap(struct drm_gem_shmem_object *shmem,
107 void drm_gem_shmem_vunmap(struct drm_gem_shmem_object *shmem,
109 int drm_gem_shmem_mmap(struct drm_gem_shmem_object *shmem, struct vm_area_struct *vma);
111 int drm_gem_shmem_madvise(struct drm_gem_shmem_object *shmem, int madv);
113 static inline bool drm_gem_shmem_is_purgeable(struct drm_gem_shmem_object *shmem) in drm_gem_shmem_is_purgeable() argument
115 return (shmem->madv > 0) && in drm_gem_shmem_is_purgeable()
[all …]
/linux-6.6.21/drivers/firmware/arm_scmi/
Dshmem.c35 void shmem_tx_prepare(struct scmi_shared_mem __iomem *shmem, in shmem_tx_prepare() argument
55 spin_until_cond((ioread32(&shmem->channel_status) & in shmem_tx_prepare()
58 if (!(ioread32(&shmem->channel_status) & in shmem_tx_prepare()
67 iowrite32(0x0, &shmem->channel_status); in shmem_tx_prepare()
69 &shmem->flags); in shmem_tx_prepare()
70 iowrite32(sizeof(shmem->msg_header) + xfer->tx.len, &shmem->length); in shmem_tx_prepare()
71 iowrite32(pack_scmi_header(&xfer->hdr), &shmem->msg_header); in shmem_tx_prepare()
73 memcpy_toio(shmem->msg_payload, xfer->tx.buf, xfer->tx.len); in shmem_tx_prepare()
76 u32 shmem_read_header(struct scmi_shared_mem __iomem *shmem) in shmem_read_header() argument
78 return ioread32(&shmem->msg_header); in shmem_read_header()
[all …]
Dmailbox.c32 struct scmi_shared_mem __iomem *shmem; member
41 shmem_tx_prepare(smbox->shmem, m, smbox->cinfo); in tx_prepare()
57 if (cl->knows_txdone && !shmem_channel_free(smbox->shmem)) { in rx_callback()
62 scmi_rx_callback(smbox->cinfo, shmem_read_header(smbox->shmem), NULL); in rx_callback()
165 struct device_node *shmem; in mailbox_chan_setup() local
182 shmem = of_parse_phandle(cdev->of_node, "shmem", idx); in mailbox_chan_setup()
183 if (!of_device_is_compatible(shmem, "arm,scmi-shmem")) { in mailbox_chan_setup()
184 of_node_put(shmem); in mailbox_chan_setup()
188 ret = of_address_to_resource(shmem, 0, &res); in mailbox_chan_setup()
189 of_node_put(shmem); in mailbox_chan_setup()
[all …]
Dcommon.h307 void shmem_tx_prepare(struct scmi_shared_mem __iomem *shmem,
309 u32 shmem_read_header(struct scmi_shared_mem __iomem *shmem);
310 void shmem_fetch_response(struct scmi_shared_mem __iomem *shmem,
312 void shmem_fetch_notification(struct scmi_shared_mem __iomem *shmem,
314 void shmem_clear_channel(struct scmi_shared_mem __iomem *shmem);
315 bool shmem_poll_done(struct scmi_shared_mem __iomem *shmem,
317 bool shmem_channel_free(struct scmi_shared_mem __iomem *shmem);
Dsmc.c58 struct scmi_shared_mem __iomem *shmem; member
73 shmem_read_header(scmi_info->shmem), NULL); in smc_msg_done_isr()
155 scmi_info->shmem = devm_ioremap(dev, res.start, size); in smc_chan_setup()
156 if (!scmi_info->shmem) { in smc_chan_setup()
223 shmem_tx_prepare(scmi_info->shmem, xfer, cinfo); in smc_send_message()
242 shmem_fetch_response(scmi_info->shmem, xfer); in smc_fetch_response()
Doptee.c125 struct scmi_shared_mem __iomem *shmem; member
344 shmem_clear_channel(channel->req.shmem); in scmi_optee_clear_channel()
388 channel->req.shmem = devm_ioremap(dev, res.start, size); in setup_static_shmem()
389 if (!channel->req.shmem) { in setup_static_shmem()
499 shmem_tx_prepare(channel->req.shmem, xfer, cinfo); in scmi_optee_send_message()
517 shmem_fetch_response(channel->req.shmem, xfer); in scmi_optee_fetch_response()
/linux-6.6.21/drivers/gpu/drm/panfrost/
Dpanfrost_gem_shrinker.c23 struct drm_gem_shmem_object *shmem; in panfrost_gem_shrinker_count() local
29 list_for_each_entry(shmem, &pfdev->shrinker_list, madv_list) { in panfrost_gem_shrinker_count()
30 if (drm_gem_shmem_is_purgeable(shmem)) in panfrost_gem_shrinker_count()
31 count += shmem->base.size >> PAGE_SHIFT; in panfrost_gem_shrinker_count()
41 struct drm_gem_shmem_object *shmem = to_drm_gem_shmem_obj(obj); in panfrost_gem_purge() local
51 if (!dma_resv_trylock(shmem->base.resv)) in panfrost_gem_purge()
58 dma_resv_unlock(shmem->base.resv); in panfrost_gem_purge()
70 struct drm_gem_shmem_object *shmem, *tmp; in panfrost_gem_shrinker_scan() local
76 list_for_each_entry_safe(shmem, tmp, &pfdev->shrinker_list, madv_list) { in panfrost_gem_shrinker_scan()
79 if (drm_gem_shmem_is_purgeable(shmem) && in panfrost_gem_shrinker_scan()
[all …]
Dpanfrost_gem.c240 struct drm_gem_shmem_object *shmem; in panfrost_gem_create() local
247 shmem = drm_gem_shmem_create(dev, size); in panfrost_gem_create()
248 if (IS_ERR(shmem)) in panfrost_gem_create()
249 return ERR_CAST(shmem); in panfrost_gem_create()
251 bo = to_panfrost_bo(&shmem->base); in panfrost_gem_create()
/linux-6.6.21/drivers/net/ethernet/microsoft/mana/
Dshm_channel.c145 u64 *shmem; in mana_smc_setup_hwc() local
185 shmem = (u64 *)ptr; in mana_smc_setup_hwc()
187 *shmem = frame_addr & PAGE_FRAME_L48_MASK; in mana_smc_setup_hwc()
193 shmem = (u64 *)ptr; in mana_smc_setup_hwc()
195 *shmem = frame_addr & PAGE_FRAME_L48_MASK; in mana_smc_setup_hwc()
201 shmem = (u64 *)ptr; in mana_smc_setup_hwc()
203 *shmem = frame_addr & PAGE_FRAME_L48_MASK; in mana_smc_setup_hwc()
209 shmem = (u64 *)ptr; in mana_smc_setup_hwc()
211 *shmem = frame_addr & PAGE_FRAME_L48_MASK; in mana_smc_setup_hwc()
/linux-6.6.21/Documentation/devicetree/bindings/firmware/
Darm,scmi.yaml38 with shmem address(4KB-page, offset) as parameters
85 The effective combination in numbers of mboxes and shmem descriptors let
88 1 mbox / 1 shmem => SCMI TX over 1 mailbox bidirectional channel
89 2 mbox / 2 shmem => SCMI TX and RX over 2 mailbox bidirectional channels
90 2 mbox / 1 shmem => SCMI TX over 2 mailbox unidirectional channels
91 3 mbox / 2 shmem => SCMI TX and RX over 3 mailbox unidirectional channels
92 Any other combination of mboxes and shmem is invalid.
96 shmem:
271 shmem:
299 - shmem
[all …]
Darm,scpi.yaml48 shmem:
170 - shmem
178 shmem = <&cpu_scp_hpri>; /* HP-NonSecure */
225 compatible = "arm,scp-shmem";
230 compatible = "arm,scp-shmem";
241 shmem = <&cpu_scp_lpri>, <&cpu_scp_hpri>;
Dnvidia,tegra186-bpmp.yaml63 See "../reserved-memory/nvidia,tegra264-bpmp-shmem.yaml" for bindings for
81 shmem:
130 - shmem
182 shmem = <&cpu_bpmp_tx>, <&cpu_bpmp_rx>;
/linux-6.6.21/Documentation/devicetree/bindings/reserved-memory/
Dnvidia,tegra264-bpmp-shmem.yaml4 $id: http://devicetree.org/schemas/reserved-memory/nvidia,tegra264-bpmp-shmem.yaml#
17 The sub-node is named shmem@<address>.
24 const: nvidia,tegra264-bpmp-shmem
41 dram_cpu_bpmp_mail: shmem@f1be0000 {
42 compatible = "nvidia,tegra264-bpmp-shmem";
/linux-6.6.21/Documentation/devicetree/bindings/mailbox/
Dmailbox.txt26 - shmem : List of phandle pointing to the shared memory(SHM) area between the
39 Example with shared memory(shmem):
49 cl_shmem: shmem@0 {
50 compatible = "client-shmem";
58 shmem = <&cl_shmem>;
Darm,mhu.yaml111 shmem = <&cpu_scp_hpri>; /* HP-NonSecure */
145 shmem = <&cpu_scp_lpri0>,
163 shmem = <&cpu_scp_hpri0>,
/linux-6.6.21/arch/arm64/boot/dts/arm/
Djuno-scmi.dtsi49 shmem = <&cpu_scp_lpri0 &cpu_scp_lpri1>;
63 shmem = <&cpu_scp_hpri0 &cpu_scp_hpri1>;
205 compatible = "arm,scmi-shmem";
210 compatible = "arm,scmi-shmem";
215 compatible = "arm,scmi-shmem";
220 compatible = "arm,scmi-shmem";
/linux-6.6.21/drivers/net/arcnet/
Dcom90xx.c58 static int com90xx_found(int ioaddr, int airq, u_long shmem, void __iomem *);
88 static int shmem; variable
93 module_param(shmem, int, 0);
107 if (!io && !irq && !shmem && !*device && com90xx_skip_probe) in com90xx_probe()
131 if (shmem) in com90xx_probe()
132 shmems[numshmems++] = shmem; in com90xx_probe()
460 static int __init com90xx_found(int ioaddr, int airq, u_long shmem, in com90xx_found() argument
473 release_mem_region(shmem, MIRROR_SIZE); in com90xx_found()
485 check_mirror(shmem - MIRROR_SIZE, MIRROR_SIZE) == 0 && in com90xx_found()
486 check_mirror(shmem - 2 * MIRROR_SIZE, MIRROR_SIZE) == 1) in com90xx_found()
[all …]
Darc-rimi.c126 unsigned long first_mirror, last_mirror, shmem; in arcrimi_found() local
146 shmem = dev->mem_start; in arcrimi_found()
159 check_mirror(shmem - MIRROR_SIZE, MIRROR_SIZE) == 0 && in arcrimi_found()
160 check_mirror(shmem - 2 * MIRROR_SIZE, MIRROR_SIZE) == 1) in arcrimi_found()
163 first_mirror = shmem - mirror_size; in arcrimi_found()
168 last_mirror = shmem + mirror_size; in arcrimi_found()
194 release_mem_region(shmem, MIRROR_SIZE); in arcrimi_found()
/linux-6.6.21/drivers/gpu/drm/virtio/
Dvirtgpu_object.c123 struct virtio_gpu_object_shmem *shmem; in virtio_gpu_create_object() local
126 shmem = kzalloc(sizeof(*shmem), GFP_KERNEL); in virtio_gpu_create_object()
127 if (!shmem) in virtio_gpu_create_object()
130 dshmem = &shmem->base.base; in virtio_gpu_create_object()
/linux-6.6.21/drivers/gpu/drm/lima/
Dlima_gem.c108 struct drm_gem_shmem_object *shmem; in lima_gem_create_handle() local
113 shmem = drm_gem_shmem_create(dev, size); in lima_gem_create_handle()
114 if (IS_ERR(shmem)) in lima_gem_create_handle()
115 return PTR_ERR(shmem); in lima_gem_create_handle()
117 obj = &shmem->base; in lima_gem_create_handle()
131 struct sg_table *sgt = drm_gem_shmem_get_pages_sgt(shmem); in lima_gem_create_handle()
/linux-6.6.21/Documentation/admin-guide/cgroup-v1/
Dmemcg_test.rst111 The best way to understand shmem's page state transition is to read
112 mm/shmem.c.
114 But brief explanation of the behavior of memcg around shmem will be
119 - radix-tree of shmem's inode.
126 - A new page is added to shmem's radix-tree.
158 Historically, memcg's shmem handling was poor and we saw some amount
159 of troubles here. This is because shmem is page-cache but can be
160 SwapCache. Test with shmem/tmpfs is always good test.
/linux-6.6.21/drivers/net/ethernet/8390/
Dmac8390.c738 long shmem = (start_page - WD_START_PG)<<8; in sane_block_output() local
740 memcpy_toio((void __iomem *)dev->mem_start + shmem, buf, count); in sane_block_output()
780 long shmem = (start_page - WD_START_PG)<<8; in dayna_block_output() local
782 dayna_memcpy_tocard(dev, shmem, buf, count); in dayna_block_output()
819 long shmem = (start_page - WD_START_PG)<<8; in slow_sane_block_output() local
821 word_memcpy_tocard(dev->mem_start + shmem, buf, count); in slow_sane_block_output()
/linux-6.6.21/Documentation/devicetree/bindings/sram/
Dsram.yaml87 - amlogic,meson-gxbb-scp-shmem
88 - amlogic,meson-axg-scp-shmem
89 - arm,juno-scp-shmem
90 - arm,scmi-shmem
91 - arm,scp-shmem
/linux-6.6.21/drivers/tee/amdtee/
Dcore.c345 struct shmem_desc shmem; in amdtee_map_shmem() local
357 shmem.kaddr = shm->kaddr; in amdtee_map_shmem()
358 shmem.size = shm->size; in amdtee_map_shmem()
364 rc = handle_map_shmem(count, &shmem, &buf_id); in amdtee_map_shmem()

123