Lines Matching refs:t_page

815 	struct nullb_page *t_page;  in null_alloc_page()  local
817 t_page = kmalloc(sizeof(struct nullb_page), GFP_NOIO); in null_alloc_page()
818 if (!t_page) in null_alloc_page()
821 t_page->page = alloc_pages(GFP_NOIO, 0); in null_alloc_page()
822 if (!t_page->page) { in null_alloc_page()
823 kfree(t_page); in null_alloc_page()
827 memset(t_page->bitmap, 0, sizeof(t_page->bitmap)); in null_alloc_page()
828 return t_page; in null_alloc_page()
831 static void null_free_page(struct nullb_page *t_page) in null_free_page() argument
833 __set_bit(NULLB_PAGE_FREE, t_page->bitmap); in null_free_page()
834 if (test_bit(NULLB_PAGE_LOCK, t_page->bitmap)) in null_free_page()
836 __free_page(t_page->page); in null_free_page()
837 kfree(t_page); in null_free_page()
852 struct nullb_page *t_page, *ret; in null_free_sector() local
859 t_page = radix_tree_lookup(root, idx); in null_free_sector()
860 if (t_page) { in null_free_sector()
861 __clear_bit(sector_bit, t_page->bitmap); in null_free_sector()
863 if (null_page_empty(t_page)) { in null_free_sector()
864 ret = radix_tree_delete_item(root, idx, t_page); in null_free_sector()
865 WARN_ON(ret != t_page); in null_free_sector()
874 struct nullb_page *t_page, bool is_cache) in null_radix_tree_insert() argument
880 if (radix_tree_insert(root, idx, t_page)) { in null_radix_tree_insert()
881 null_free_page(t_page); in null_radix_tree_insert()
882 t_page = radix_tree_lookup(root, idx); in null_radix_tree_insert()
883 WARN_ON(!t_page || t_page->page->index != idx); in null_radix_tree_insert()
887 return t_page; in null_radix_tree_insert()
924 struct nullb_page *t_page; in __null_lookup_page() local
931 t_page = radix_tree_lookup(root, idx); in __null_lookup_page()
932 WARN_ON(t_page && t_page->page->index != idx); in __null_lookup_page()
934 if (t_page && (for_write || test_bit(sector_bit, t_page->bitmap))) in __null_lookup_page()
935 return t_page; in __null_lookup_page()
958 struct nullb_page *t_page; in null_insert_page() local
960 t_page = null_lookup_page(nullb, sector, true, ignore_cache); in null_insert_page()
961 if (t_page) in null_insert_page()
962 return t_page; in null_insert_page()
966 t_page = null_alloc_page(); in null_insert_page()
967 if (!t_page) in null_insert_page()
975 t_page->page->index = idx; in null_insert_page()
976 t_page = null_radix_tree_insert(nullb, idx, t_page, !ignore_cache); in null_insert_page()
979 return t_page; in null_insert_page()
981 null_free_page(t_page); in null_insert_page()
992 struct nullb_page *t_page, *ret; in null_flush_cache_page() local
997 t_page = null_insert_page(nullb, idx << PAGE_SECTORS_SHIFT, true); in null_flush_cache_page()
1002 if (t_page && null_page_empty(t_page)) { in null_flush_cache_page()
1004 idx, t_page); in null_flush_cache_page()
1005 null_free_page(t_page); in null_flush_cache_page()
1010 if (!t_page) in null_flush_cache_page()
1014 dst = kmap_atomic(t_page->page); in null_flush_cache_page()
1022 __set_bit(i, t_page->bitmap); in null_flush_cache_page()
1094 struct nullb_page *t_page; in copy_to_nullb() local
1104 t_page = null_insert_page(nullb, sector, in copy_to_nullb()
1106 if (!t_page) in copy_to_nullb()
1110 dst = kmap_atomic(t_page->page); in copy_to_nullb()
1115 __set_bit(sector & SECTOR_MASK, t_page->bitmap); in copy_to_nullb()
1131 struct nullb_page *t_page; in copy_from_nullb() local
1138 t_page = null_lookup_page(nullb, sector, false, in copy_from_nullb()
1142 if (!t_page) { in copy_from_nullb()
1146 src = kmap_atomic(t_page->page); in copy_from_nullb()