/linux-6.1.9/drivers/platform/surface/aggregator/ |
D | ssh_packet_layer.c | 335 ptl_info(packet->ptl, "packet error injection: dropping ACK packet %p\n", in __ssh_ptl_should_drop_ack_packet() 347 ptl_info(packet->ptl, "packet error injection: dropping NAK packet %p\n", in __ssh_ptl_should_drop_nak_packet() 359 ptl_info(packet->ptl, in __ssh_ptl_should_drop_dsq_packet() 387 static int ssh_ptl_write_buf(struct ssh_ptl *ptl, struct ssh_packet *packet, in ssh_ptl_write_buf() argument 395 ptl_info(packet->ptl, in ssh_ptl_write_buf() 402 return serdev_device_write_buf(ptl->serdev, buf, count); in ssh_ptl_write_buf() 419 ptl_info(packet->ptl, in ssh_ptl_tx_inject_invalid_data() 431 static void ssh_ptl_rx_inject_invalid_syn(struct ssh_ptl *ptl, in ssh_ptl_rx_inject_invalid_syn() argument 448 static void ssh_ptl_rx_inject_invalid_data(struct ssh_ptl *ptl, in ssh_ptl_rx_inject_invalid_data() argument 486 static inline int ssh_ptl_write_buf(struct ssh_ptl *ptl, in ssh_ptl_write_buf() argument [all …]
|
D | ssh_packet_layer.h | 139 int ssh_ptl_init(struct ssh_ptl *ptl, struct serdev_device *serdev, 142 void ssh_ptl_destroy(struct ssh_ptl *ptl); 151 static inline struct device *ssh_ptl_get_device(struct ssh_ptl *ptl) in ssh_ptl_get_device() argument 153 return ptl->serdev ? &ptl->serdev->dev : NULL; in ssh_ptl_get_device() 156 int ssh_ptl_tx_start(struct ssh_ptl *ptl); 157 int ssh_ptl_tx_stop(struct ssh_ptl *ptl); 158 int ssh_ptl_rx_start(struct ssh_ptl *ptl); 159 int ssh_ptl_rx_stop(struct ssh_ptl *ptl); 160 void ssh_ptl_shutdown(struct ssh_ptl *ptl); 162 int ssh_ptl_submit(struct ssh_ptl *ptl, struct ssh_packet *p); [all …]
|
D | ssh_request_layer.h | 66 struct ssh_ptl ptl; member 94 #define rtl_dbg(r, fmt, ...) ptl_dbg(&(r)->ptl, fmt, ##__VA_ARGS__) 95 #define rtl_info(p, fmt, ...) ptl_info(&(p)->ptl, fmt, ##__VA_ARGS__) 96 #define rtl_warn(r, fmt, ...) ptl_warn(&(r)->ptl, fmt, ##__VA_ARGS__) 97 #define rtl_err(r, fmt, ...) ptl_err(&(r)->ptl, fmt, ##__VA_ARGS__) 112 return ssh_ptl_get_device(&rtl->ptl); in ssh_rtl_get_device() 123 struct ssh_ptl *ptl; in ssh_request_rtl() local 125 ptl = READ_ONCE(rqst->packet.ptl); in ssh_request_rtl() 126 return likely(ptl) ? to_ssh_rtl(ptl, ptl) : NULL; in ssh_request_rtl()
|
D | ssh_request_layer.c | 258 status = ssh_ptl_submit(&rtl->ptl, &rqst->packet); in ssh_rtl_tx_try_process_one() 391 if (cmpxchg(&rqst->packet.ptl, NULL, &rtl->ptl)) { in ssh_rtl_submit() 619 if (flags == fixed && !READ_ONCE(r->packet.ptl)) { in ssh_rtl_cancel_nonpending() 674 if (!READ_ONCE(r->packet.ptl)) { in ssh_rtl_cancel_pending() 911 struct ssh_rtl *rtl = to_ssh_rtl(p, ptl); in ssh_rtl_rx_command() 1032 status = ssh_ptl_init(&rtl->ptl, serdev, &ptl_ops); in ssh_rtl_init() 1066 ssh_ptl_destroy(&rtl->ptl); in ssh_rtl_destroy() 1079 status = ssh_ptl_tx_start(&rtl->ptl); in ssh_rtl_start() 1085 status = ssh_ptl_rx_start(&rtl->ptl); in ssh_rtl_start() 1088 ssh_ptl_tx_stop(&rtl->ptl); in ssh_rtl_start() [all …]
|
/linux-6.1.9/drivers/media/platform/allegro-dvt/ |
D | nal-hevc.c | 101 struct nal_hevc_profile_tier_level *ptl) in nal_hevc_rbsp_profile_tier_level() argument 106 rbsp_bits(rbsp, 2, &ptl->general_profile_space); in nal_hevc_rbsp_profile_tier_level() 107 rbsp_bit(rbsp, &ptl->general_tier_flag); in nal_hevc_rbsp_profile_tier_level() 108 rbsp_bits(rbsp, 5, &ptl->general_profile_idc); in nal_hevc_rbsp_profile_tier_level() 110 rbsp_bit(rbsp, &ptl->general_profile_compatibility_flag[i]); in nal_hevc_rbsp_profile_tier_level() 111 rbsp_bit(rbsp, &ptl->general_progressive_source_flag); in nal_hevc_rbsp_profile_tier_level() 112 rbsp_bit(rbsp, &ptl->general_interlaced_source_flag); in nal_hevc_rbsp_profile_tier_level() 113 rbsp_bit(rbsp, &ptl->general_non_packed_constraint_flag); in nal_hevc_rbsp_profile_tier_level() 114 rbsp_bit(rbsp, &ptl->general_frame_only_constraint_flag); in nal_hevc_rbsp_profile_tier_level() 115 if (ptl->general_profile_idc == 4 || in nal_hevc_rbsp_profile_tier_level() [all …]
|
/linux-6.1.9/arch/arm/lib/ |
D | uaccess_with_memcpy.c | 31 spinlock_t *ptl; in pin_page_for_write() local 60 ptl = ¤t->mm->page_table_lock; in pin_page_for_write() 61 spin_lock(ptl); in pin_page_for_write() 64 spin_unlock(ptl); in pin_page_for_write() 69 *ptlp = ptl; in pin_page_for_write() 76 pte = pte_offset_map_lock(current->mm, pmd, addr, &ptl); in pin_page_for_write() 79 pte_unmap_unlock(pte, ptl); in pin_page_for_write() 84 *ptlp = ptl; in pin_page_for_write() 102 spinlock_t *ptl; in __copy_to_user_memcpy() local 105 while (!pin_page_for_write(to, &pte, &ptl)) { in __copy_to_user_memcpy() [all …]
|
/linux-6.1.9/mm/ |
D | page_vma_mapped.c | 51 pvmw->ptl = pte_lockptr(pvmw->vma->vm_mm, pvmw->pmd); in map_pte() 52 spin_lock(pvmw->ptl); in map_pte() 177 pvmw->ptl = huge_pte_lock(hstate, mm, pvmw->pte); in page_vma_mapped_walk() 214 pvmw->ptl = pmd_lock(mm, pvmw->pmd); in page_vma_mapped_walk() 236 spin_unlock(pvmw->ptl); in page_vma_mapped_walk() 237 pvmw->ptl = NULL; in page_vma_mapped_walk() 247 spinlock_t *ptl = pmd_lock(mm, pvmw->pmd); in page_vma_mapped_walk() local 249 spin_unlock(ptl); in page_vma_mapped_walk() 266 if (pvmw->ptl) { in page_vma_mapped_walk() 267 spin_unlock(pvmw->ptl); in page_vma_mapped_walk() [all …]
|
D | huge_memory.c | 684 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in __do_huge_pmd_anonymous_page() 696 spin_unlock(vmf->ptl); in __do_huge_pmd_anonymous_page() 713 spin_unlock(vmf->ptl); in __do_huge_pmd_anonymous_page() 720 spin_unlock(vmf->ptl); in __do_huge_pmd_anonymous_page() 807 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_anonymous_page() 812 spin_unlock(vmf->ptl); in do_huge_pmd_anonymous_page() 815 spin_unlock(vmf->ptl); in do_huge_pmd_anonymous_page() 823 spin_unlock(vmf->ptl); in do_huge_pmd_anonymous_page() 826 spin_unlock(vmf->ptl); in do_huge_pmd_anonymous_page() 846 spinlock_t *ptl; in insert_pfn_pmd() local [all …]
|
D | memory.c | 442 spinlock_t *ptl = pmd_lock(mm, pmd); in pmd_install() local 463 spin_unlock(ptl); in pmd_install() 1403 spinlock_t *ptl; in zap_pte_range() local 1411 start_pte = pte_offset_map_lock(mm, pmd, addr, &ptl); in zap_pte_range() 1509 pte_unmap_unlock(start_pte, ptl); in zap_pte_range() 1550 spinlock_t *ptl = pmd_lock(tlb->mm, pmd); in zap_pmd_range() local 1556 spin_unlock(ptl); in zap_pmd_range() 1838 spinlock_t **ptl) in __get_locked_pte() argument 1844 return pte_alloc_map_lock(mm, pmd, addr, ptl); in __get_locked_pte() 1880 spinlock_t *ptl; in insert_page() local [all …]
|
D | mincore.c | 103 spinlock_t *ptl; in mincore_pte_range() local 109 ptl = pmd_trans_huge_lock(pmd, vma); in mincore_pte_range() 110 if (ptl) { in mincore_pte_range() 112 spin_unlock(ptl); in mincore_pte_range() 121 ptep = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in mincore_pte_range() 152 pte_unmap_unlock(ptep - 1, ptl); in mincore_pte_range()
|
D | madvise.c | 210 spinlock_t *ptl; in swapin_walk_pmd_entry() local 213 ptep = pte_offset_map_lock(vma->vm_mm, pmd, index, &ptl); in swapin_walk_pmd_entry() 215 pte_unmap_unlock(ptep, ptl); in swapin_walk_pmd_entry() 334 spinlock_t *ptl; in madvise_cold_or_pageout_pte_range() local 347 ptl = pmd_trans_huge_lock(pmd, vma); in madvise_cold_or_pageout_pte_range() 348 if (!ptl) in madvise_cold_or_pageout_pte_range() 371 spin_unlock(ptl); in madvise_cold_or_pageout_pte_range() 401 spin_unlock(ptl); in madvise_cold_or_pageout_pte_range() 412 orig_pte = pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in madvise_cold_or_pageout_pte_range() 440 pte_unmap_unlock(orig_pte, ptl); in madvise_cold_or_pageout_pte_range() [all …]
|
D | hmm.c | 417 spinlock_t *ptl = pud_trans_huge_lock(pudp, walk->vma); in hmm_vma_walk_pud() local 419 if (!ptl) in hmm_vma_walk_pud() 427 spin_unlock(ptl); in hmm_vma_walk_pud() 438 spin_unlock(ptl); in hmm_vma_walk_pud() 450 spin_unlock(ptl); in hmm_vma_walk_pud() 464 spin_unlock(ptl); in hmm_vma_walk_pud() 483 spinlock_t *ptl; in hmm_vma_walk_hugetlb_entry() local 486 ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte); in hmm_vma_walk_hugetlb_entry() 496 spin_unlock(ptl); in hmm_vma_walk_hugetlb_entry() 504 spin_unlock(ptl); in hmm_vma_walk_hugetlb_entry()
|
D | migrate_device.c | 66 spinlock_t *ptl; in migrate_vma_collect_pmd() local 76 ptl = pmd_lock(mm, pmdp); in migrate_vma_collect_pmd() 78 spin_unlock(ptl); in migrate_vma_collect_pmd() 84 spin_unlock(ptl); in migrate_vma_collect_pmd() 93 spin_unlock(ptl); in migrate_vma_collect_pmd() 112 ptep = pte_offset_map_lock(mm, pmdp, addr, &ptl); in migrate_vma_collect_pmd() 281 pte_unmap_unlock(ptep - 1, ptl); in migrate_vma_collect_pmd() 576 spinlock_t *ptl; in migrate_vma_insert_page() local 652 ptep = pte_offset_map_lock(mm, pmdp, addr, &ptl); in migrate_vma_insert_page() 690 pte_unmap_unlock(ptep, ptl); in migrate_vma_insert_page() [all …]
|
D | hugetlb.c | 5232 spinlock_t *ptl; in __unmap_hugepage_range() local 5267 ptl = huge_pte_lock(h, mm, ptep); in __unmap_hugepage_range() 5269 spin_unlock(ptl); in __unmap_hugepage_range() 5278 spin_unlock(ptl); in __unmap_hugepage_range() 5301 spin_unlock(ptl); in __unmap_hugepage_range() 5313 spin_unlock(ptl); in __unmap_hugepage_range() 5338 spin_unlock(ptl); in __unmap_hugepage_range() 5469 struct page *pagecache_page, spinlock_t *ptl) in hugetlb_wp() argument 5539 spin_unlock(ptl); in hugetlb_wp() 5574 spin_lock(ptl); in hugetlb_wp() [all …]
|
D | migrate.c | 294 spinlock_t *ptl) in __migration_entry_wait() argument 299 spin_lock(ptl); in __migration_entry_wait() 308 migration_entry_wait_on_locked(entry, ptep, ptl); in __migration_entry_wait() 311 pte_unmap_unlock(ptep, ptl); in __migration_entry_wait() 317 spinlock_t *ptl = pte_lockptr(mm, pmd); in migration_entry_wait() local 319 __migration_entry_wait(mm, ptep, ptl); in migration_entry_wait() 323 void __migration_entry_wait_huge(pte_t *ptep, spinlock_t *ptl) in __migration_entry_wait_huge() argument 327 spin_lock(ptl); in __migration_entry_wait_huge() 331 spin_unlock(ptl); in __migration_entry_wait_huge() 333 migration_entry_wait_on_locked(pte_to_swp_entry(pte), NULL, ptl); in __migration_entry_wait_huge() [all …]
|
/linux-6.1.9/arch/arm/mm/ |
D | fault-armv.c | 70 static inline void do_pte_lock(spinlock_t *ptl) in do_pte_lock() argument 76 spin_lock_nested(ptl, SINGLE_DEPTH_NESTING); in do_pte_lock() 79 static inline void do_pte_unlock(spinlock_t *ptl) in do_pte_unlock() argument 81 spin_unlock(ptl); in do_pte_unlock() 84 static inline void do_pte_lock(spinlock_t *ptl) {} in do_pte_lock() argument 85 static inline void do_pte_unlock(spinlock_t *ptl) {} in do_pte_unlock() argument 91 spinlock_t *ptl; in adjust_pte() local 120 ptl = pte_lockptr(vma->vm_mm, pmd); in adjust_pte() 122 do_pte_lock(ptl); in adjust_pte() 126 do_pte_unlock(ptl); in adjust_pte()
|
/linux-6.1.9/mm/damon/ |
D | vaddr.c | 304 spinlock_t *ptl; in damon_mkold_pmd_entry() local 307 ptl = pmd_lock(walk->mm, pmd); in damon_mkold_pmd_entry() 309 spin_unlock(ptl); in damon_mkold_pmd_entry() 315 spin_unlock(ptl); in damon_mkold_pmd_entry() 318 spin_unlock(ptl); in damon_mkold_pmd_entry() 323 pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in damon_mkold_pmd_entry() 328 pte_unmap_unlock(pte, ptl); in damon_mkold_pmd_entry() 366 spinlock_t *ptl; in damon_mkold_hugetlb_entry() local 369 ptl = huge_pte_lock(h, walk->mm, pte); in damon_mkold_hugetlb_entry() 377 spin_unlock(ptl); in damon_mkold_hugetlb_entry() [all …]
|
/linux-6.1.9/arch/powerpc/mm/ |
D | hugetlbpage.c | 47 unsigned int pshift, spinlock_t *ptl) in __hugepte_alloc() argument 82 spin_lock(ptl); in __hugepte_alloc() 102 spin_unlock(ptl); in __hugepte_alloc() 120 spinlock_t *ptl; in huge_pte_alloc() local 134 ptl = &mm->page_table_lock; in huge_pte_alloc() 144 ptl = pud_lockptr(mm, pu); in huge_pte_alloc() 155 ptl = pmd_lockptr(mm, pm); in huge_pte_alloc() 162 ptl = &mm->page_table_lock; in huge_pte_alloc() 170 ptl = pud_lockptr(mm, pu); in huge_pte_alloc() 177 ptl = pmd_lockptr(mm, pm); in huge_pte_alloc() [all …]
|
/linux-6.1.9/Documentation/translations/zh_CN/mm/ |
D | split_page_table_lock.rst | 63 面。这个区域与page->ptl共享存储。 79 page->ptl 82 page->ptl用于访问分割页表锁,其中'page'是包含该表的页面struct page。它 89 - 如果spinlock_t的大小大于long的大小,我们使用page->ptl作为spinlock_t 96 请不要直接访问page->ptl - -使用适当的辅助函数。
|
/linux-6.1.9/arch/s390/mm/ |
D | pgtable.c | 790 spinlock_t *ptl; in set_guest_storage_key() local 808 ptl = pmd_lock(mm, pmdp); in set_guest_storage_key() 810 spin_unlock(ptl); in set_guest_storage_key() 822 spin_unlock(ptl); in set_guest_storage_key() 825 spin_unlock(ptl); in set_guest_storage_key() 827 ptep = pte_offset_map_lock(mm, pmdp, addr, &ptl); in set_guest_storage_key() 852 pte_unmap_unlock(ptep, ptl); in set_guest_storage_key() 897 spinlock_t *ptl; in reset_guest_reference_bit() local 917 ptl = pmd_lock(mm, pmdp); in reset_guest_reference_bit() 919 spin_unlock(ptl); in reset_guest_reference_bit() [all …]
|
D | gmap.c | 543 spinlock_t *ptl; in __gmap_link() local 599 ptl = pmd_lock(mm, pmd); in __gmap_link() 621 spin_unlock(ptl); in __gmap_link() 677 spinlock_t *ptl; in __gmap_zap() local 691 ptep = get_locked_pte(gmap->mm, vmaddr, &ptl); in __gmap_zap() 694 pte_unmap_unlock(ptep, ptl); in __gmap_zap() 856 spinlock_t **ptl) in gmap_pte_op_walk() argument 865 return pte_alloc_map_lock(gmap->mm, (pmd_t *) table, gaddr, ptl); in gmap_pte_op_walk() 901 static void gmap_pte_op_end(spinlock_t *ptl) in gmap_pte_op_end() argument 903 if (ptl) in gmap_pte_op_end() [all …]
|
/linux-6.1.9/arch/s390/pci/ |
D | pci_mmio.c | 126 spinlock_t *ptl; in SYSCALL_DEFINE3() local 172 ret = follow_pte(vma->vm_mm, mmio_addr, &ptep, &ptl); in SYSCALL_DEFINE3() 184 pte_unmap_unlock(ptep, ptl); in SYSCALL_DEFINE3() 268 spinlock_t *ptl; in SYSCALL_DEFINE3() local 311 ret = follow_pte(vma->vm_mm, mmio_addr, &ptep, &ptl); in SYSCALL_DEFINE3() 325 pte_unmap_unlock(ptep, ptl); in SYSCALL_DEFINE3()
|
/linux-6.1.9/fs/proc/ |
D | task_mmu.c | 618 spinlock_t *ptl; in smaps_pte_range() local 620 ptl = pmd_trans_huge_lock(pmd, vma); in smaps_pte_range() 621 if (ptl) { in smaps_pte_range() 623 spin_unlock(ptl); in smaps_pte_range() 634 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in smaps_pte_range() 637 pte_unmap_unlock(pte - 1, ptl); in smaps_pte_range() 1166 spinlock_t *ptl; in clear_refs_pte_range() local 1169 ptl = pmd_trans_huge_lock(pmd, vma); in clear_refs_pte_range() 1170 if (ptl) { in clear_refs_pte_range() 1186 spin_unlock(ptl); in clear_refs_pte_range() [all …]
|
/linux-6.1.9/Documentation/mm/ |
D | split_page_table_lock.rst | 63 This field shares storage with page->ptl. 80 page->ptl 83 page->ptl is used to access split page table lock, where 'page' is struct 92 - if size of spinlock_t is bigger then size of long, we use page->ptl as 100 Please, never access page->ptl directly -- use appropriate helper.
|
/linux-6.1.9/arch/x86/kernel/ |
D | ldt.c | 292 spinlock_t *ptl; in map_ldt_struct() local 326 ptep = get_locked_pte(mm, va, &ptl); in map_ldt_struct() 339 pte_unmap_unlock(ptep, ptl); in map_ldt_struct() 365 spinlock_t *ptl; in unmap_ldt_struct() local 369 ptep = get_locked_pte(mm, va, &ptl); in unmap_ldt_struct() 371 pte_unmap_unlock(ptep, ptl); in unmap_ldt_struct()
|