/linux-3.4.99/drivers/media/video/ivtv/ ! |
D | ivtv-udma.c | 33 dma_page->page_count = dma_page->last - dma_page->first + 1; in ivtv_udma_get_page_info() 34 if (dma_page->page_count == 1) dma_page->tail -= dma_page->offset; in ivtv_udma_get_page_info() 48 for (i = 0; i < dma_page->page_count; i++) { in ivtv_udma_fill_sg_list() 49 unsigned int len = (i == dma_page->page_count - 1) ? in ivtv_udma_fill_sg_list() 112 if (dma->SG_length || dma->page_count) { in ivtv_udma_setup() 114 dma->SG_length, dma->page_count); in ivtv_udma_setup() 120 if (user_dma.page_count <= 0) { in ivtv_udma_setup() 122 user_dma.page_count, size_in_bytes, user_dma.offset); in ivtv_udma_setup() 129 user_dma.uaddr, user_dma.page_count, 0, 1, dma->map, NULL); in ivtv_udma_setup() 132 if (user_dma.page_count != err) { in ivtv_udma_setup() [all …]
|
D | ivtv-yuv.c | 67 if (dma->SG_length || dma->page_count) { in ivtv_yuv_prep_user_dma() 70 dma->SG_length, dma->page_count); in ivtv_yuv_prep_user_dma() 79 …y_pages = get_user_pages(current, current->mm, y_dma.uaddr, y_dma.page_count, 0, 1, &dma->map[0], … in ivtv_yuv_prep_user_dma() 81 if (y_pages == y_dma.page_count) { in ivtv_yuv_prep_user_dma() 83 uv_dma.uaddr, uv_dma.page_count, 0, 1, in ivtv_yuv_prep_user_dma() 88 if (y_pages != y_dma.page_count || uv_pages != uv_dma.page_count) { in ivtv_yuv_prep_user_dma() 91 if (y_pages == y_dma.page_count) { in ivtv_yuv_prep_user_dma() 94 "expecting %d\n", uv_pages, uv_dma.page_count); in ivtv_yuv_prep_user_dma() 106 "expecting %d\n", y_pages, y_dma.page_count); in ivtv_yuv_prep_user_dma() 123 dma->page_count = y_pages + uv_pages; in ivtv_yuv_prep_user_dma() [all …]
|
/linux-3.4.99/drivers/char/agp/ ! |
D | generic.c | 198 if (curr->page_count != 0) { in agp_free_memory() 203 for (i = 0; i < curr->page_count; i++) { in agp_free_memory() 208 for (i = 0; i < curr->page_count; i++) { in agp_free_memory() 235 size_t page_count, u32 type) in agp_allocate_memory() argument 246 if ((cur_memory + page_count > bridge->max_memory_agp) || in agp_allocate_memory() 247 (cur_memory + page_count < page_count)) in agp_allocate_memory() 251 new = agp_generic_alloc_user(page_count, type); in agp_allocate_memory() 258 new = bridge->driver->alloc_by_type(page_count, type); in agp_allocate_memory() 264 scratch_pages = (page_count + ENTRIES_PER_PAGE - 1) / ENTRIES_PER_PAGE; in agp_allocate_memory() 272 if (bridge->driver->agp_alloc_pages(bridge, new, page_count)) { in agp_allocate_memory() [all …]
|
D | i460-agp.c | 311 if ((io_pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count) > num_entries) { in i460_insert_memory_small_io_page() 317 while (j < (io_pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count)) { in i460_insert_memory_small_io_page() 327 for (i = 0, j = io_pg_start; i < mem->page_count; i++) { in i460_insert_memory_small_io_page() 346 for (i = pg_start; i < (pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count); i++) in i460_remove_memory_small_io_page() 415 end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; in i460_insert_memory_large_io_page() 417 end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; in i460_insert_memory_large_io_page() 473 end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; in i460_remove_memory_large_io_page() 475 end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; in i460_remove_memory_large_io_page()
|
D | sgi-agp.c | 175 if ((pg_start + mem->page_count) > num_entries) in sgi_tioca_insert_memory() 180 while (j < (pg_start + mem->page_count)) { in sgi_tioca_insert_memory() 191 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in sgi_tioca_insert_memory() 219 for (i = pg_start; i < (mem->page_count + pg_start); i++) { in sgi_tioca_remove_memory()
|
D | nvidia-agp.c | 209 if (mem->page_count == 0) in nvidia_insert_memory() 212 if ((pg_start + mem->page_count) > in nvidia_insert_memory() 216 for (j = pg_start; j < (pg_start + mem->page_count); j++) { in nvidia_insert_memory() 225 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in nvidia_insert_memory() 249 if (mem->page_count == 0) in nvidia_remove_memory() 252 for (i = pg_start; i < (mem->page_count + pg_start); i++) in nvidia_remove_memory()
|
D | ati-agp.c | 280 if (mem->page_count == 0) in ati_insert_memory() 283 if ((pg_start + mem->page_count) > num_entries) in ati_insert_memory() 287 while (j < (pg_start + mem->page_count)) { in ati_insert_memory() 301 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { in ati_insert_memory() 326 if (mem->page_count == 0) in ati_remove_memory() 329 for (i = pg_start; i < (mem->page_count + pg_start); i++) { in ati_remove_memory()
|
D | intel-gtt.c | 123 DBG("try unmapping %lu pages\n", (unsigned long)mem->page_count); in intel_gtt_unmap_memory() 215 if ((pg_start + mem->page_count) in i810_insert_dcache_entries() 222 for (i = pg_start; i < (pg_start + mem->page_count); i++) { in i810_insert_dcache_entries() 267 new->page_count = pg_count; in alloc_agpphysmem_i8xx() 278 if (curr->page_count == 4) in intel_i810_free_by_type() 939 if (mem->page_count == 0) in intel_fake_agp_insert_entries() 942 if (pg_start + mem->page_count > intel_private.base.gtt_total_entries) in intel_fake_agp_insert_entries() 955 ret = intel_gtt_map_memory(mem->pages, mem->page_count, in intel_fake_agp_insert_entries() 963 intel_gtt_insert_pages(pg_start, mem->page_count, mem->pages, in intel_fake_agp_insert_entries() 988 if (mem->page_count == 0) in intel_fake_agp_remove_entries() [all …]
|
D | uninorth-agp.c | 165 if (mem->page_count == 0) in uninorth_insert_memory() 171 if ((pg_start + mem->page_count) > num_entries) in uninorth_insert_memory() 175 for (i = 0; i < mem->page_count; ++i) { in uninorth_insert_memory() 184 for (i = 0; i < mem->page_count; i++) { in uninorth_insert_memory() 214 if (mem->page_count == 0) in uninorth_remove_memory() 218 for (i = 0; i < mem->page_count; ++i) { in uninorth_remove_memory()
|
D | agp.h | 199 struct agp_memory *agp_generic_alloc_by_type(size_t page_count, int type); 203 struct agp_memory *memory, size_t page_count); 220 struct agp_memory *agp_generic_alloc_user(size_t page_count, int type);
|
D | efficeon-agp.c | 241 int i, count = mem->page_count, num_entries; in efficeon_insert_memory() 249 if ((pg_start + mem->page_count) > num_entries) in efficeon_insert_memory() 290 int i, count = mem->page_count, num_entries; in efficeon_remove_memory() 296 if ((pg_start + mem->page_count) > num_entries) in efficeon_remove_memory()
|
D | ali-agp.c | 128 int i, page_count; in m1541_cache_flush() local 133 page_count = 1 << A_SIZE_32(agp_bridge->current_size)->page_order; in m1541_cache_flush() 134 for (i = 0; i < PAGE_SIZE * page_count; i += PAGE_SIZE) { in m1541_cache_flush()
|
D | parisc-agp.c | 137 io_pg_count = info->io_pages_per_kpage * mem->page_count; in parisc_agp_insert_memory() 154 for (i = 0, j = io_pg_start; i < mem->page_count; i++) { in parisc_agp_insert_memory() 183 io_pg_count = info->io_pages_per_kpage * mem->page_count; in parisc_agp_remove_memory()
|
/linux-3.4.99/drivers/gpu/drm/udl/ ! |
D | udl_gem.c | 125 int page_count, i; in udl_gem_get_pages() local 133 page_count = obj->base.size / PAGE_SIZE; in udl_gem_get_pages() 135 obj->pages = drm_malloc_ab(page_count, sizeof(struct page *)); in udl_gem_get_pages() 143 for (i = 0; i < page_count; i++) { in udl_gem_get_pages() 161 int page_count = obj->base.size / PAGE_SIZE; in udl_gem_put_pages() local 164 for (i = 0; i < page_count; i++) in udl_gem_put_pages() 173 int page_count = obj->base.size / PAGE_SIZE; in udl_gem_vmap() local 180 obj->vmapping = vmap(obj->pages, page_count, 0, PAGE_KERNEL); in udl_gem_vmap()
|
/linux-3.4.99/lib/ ! |
D | show_mem.c | 45 else if (page_count(page) == 1) in show_mem() 47 else if (page_count(page) > 1) in show_mem() 48 shared += page_count(page) - 1; in show_mem()
|
/linux-3.4.99/drivers/firewire/ ! |
D | core-iso.c | 43 int page_count, enum dma_data_direction direction) in fw_iso_buffer_init() argument 48 buffer->page_count = page_count; in fw_iso_buffer_init() 51 buffer->pages = kmalloc(page_count * sizeof(buffer->pages[0]), in fw_iso_buffer_init() 56 for (i = 0; i < buffer->page_count; i++) { in fw_iso_buffer_init() 93 for (i = 0; i < buffer->page_count; i++) { in fw_iso_buffer_map() 110 for (i = 0; i < buffer->page_count; i++) { in fw_iso_buffer_destroy() 129 for (i = 0; i < buffer->page_count; i++) { in fw_iso_buffer_lookup()
|
/linux-3.4.99/fs/btrfs/ ! |
D | scrub.c | 74 int page_count; member 81 int page_count; member 196 for (i = 0; i < sbio->page_count; i++) { in scrub_free_dev() 242 sbio->page_count = 0; in scrub_setup_dev() 365 BUG_ON(sblock->page_count < 1); in scrub_print_warning() 629 BUG_ON(sblock_to_check->page_count < 1); in scrub_handle_errored_block() 631 length = sblock_to_check->page_count * PAGE_SIZE; in scrub_handle_errored_block() 790 sblocks_for_recheck[mirror_index].page_count > 0; in scrub_handle_errored_block() 819 sblocks_for_recheck[mirror_index].page_count > 0; in scrub_handle_errored_block() 867 for (page_num = 0; page_num < sblock_bad->page_count; page_num++) { in scrub_handle_errored_block() [all …]
|
/linux-3.4.99/drivers/gpu/drm/i915/ ! |
D | i915_gem_tiling.c | 461 int page_count = obj->base.size >> PAGE_SHIFT; in i915_gem_object_do_bit_17_swizzle() local 467 for (i = 0; i < page_count; i++) { in i915_gem_object_do_bit_17_swizzle() 480 int page_count = obj->base.size >> PAGE_SHIFT; in i915_gem_object_save_bit_17_swizzle() local 484 obj->bit_17 = kmalloc(BITS_TO_LONGS(page_count) * in i915_gem_object_save_bit_17_swizzle() 493 for (i = 0; i < page_count; i++) { in i915_gem_object_save_bit_17_swizzle()
|
/linux-3.4.99/drivers/staging/android/ ! |
D | persistent_ram.c | 326 unsigned int page_count; in persistent_ram_buffer_map() local 331 page_count = DIV_ROUND_UP(size + offset_in_page(start), PAGE_SIZE); in persistent_ram_buffer_map() 335 pages = kmalloc(sizeof(struct page *) * page_count, GFP_KERNEL); in persistent_ram_buffer_map() 338 page_count); in persistent_ram_buffer_map() 342 for (i = 0; i < page_count; i++) { in persistent_ram_buffer_map() 346 prz->vaddr = vmap(pages, page_count, VM_MAP, prot); in persistent_ram_buffer_map() 349 pr_err("%s: Failed to map %u pages\n", __func__, page_count); in persistent_ram_buffer_map()
|
/linux-3.4.99/drivers/iommu/ ! |
D | tegra-gart.c | 56 u32 page_count; /* total remappable size */ member 78 iova < gart->iovmm_base + GART_PAGE_SIZE * gart->page_count; \ 145 gart_end = gart_start + gart->page_count * GART_PAGE_SIZE - 1; in gart_iova_range_valid() 381 gart->page_count = (resource_size(res_remap) >> GART_PAGE_SHIFT); in tegra_gart_probe() 383 gart->savedata = vmalloc(sizeof(u32) * gart->page_count); in tegra_gart_probe()
|
/linux-3.4.99/fs/xfs/ ! |
D | xfs_buf.c | 217 int page_count, in _xfs_buf_get_pages() argument 223 bp->b_page_count = page_count; in _xfs_buf_get_pages() 224 if (page_count <= XB_PAGES) { in _xfs_buf_get_pages() 228 page_count, xb_to_km(flags)); in _xfs_buf_get_pages() 232 memset(bp->b_pages, 0, sizeof(struct page *) * page_count); in _xfs_buf_get_pages() 294 unsigned short page_count, i; in xfs_buf_allocate_memory() local 328 page_count = xfs_buf_btoc(end) - xfs_buf_btoct(bp->b_file_offset); in xfs_buf_allocate_memory() 329 error = _xfs_buf_get_pages(bp, page_count, flags); in xfs_buf_allocate_memory() 738 int page_count; in xfs_buf_associate_memory() local 743 page_count = buflen >> PAGE_SHIFT; in xfs_buf_associate_memory() [all …]
|
/linux-3.4.99/drivers/gpu/drm/nouveau/ ! |
D | nouveau_bo.c | 493 u32 page_count = new_mem->num_pages; in nvc0_bo_move_m2mf() local 496 page_count = new_mem->num_pages; in nvc0_bo_move_m2mf() 497 while (page_count) { in nvc0_bo_move_m2mf() 498 int line_count = (page_count > 2047) ? 2047 : page_count; in nvc0_bo_move_m2mf() 517 page_count -= line_count; in nvc0_bo_move_m2mf() 630 u32 page_count = new_mem->num_pages; in nv04_bo_move_m2mf() local 641 page_count = new_mem->num_pages; in nv04_bo_move_m2mf() 642 while (page_count) { in nv04_bo_move_m2mf() 643 int line_count = (page_count > 2047) ? 2047 : page_count; in nv04_bo_move_m2mf() 662 page_count -= line_count; in nv04_bo_move_m2mf()
|
/linux-3.4.99/drivers/gpu/drm/ttm/ ! |
D | ttm_agp_backend.c | 63 mem->page_count = 0; in ttm_agp_bind() 70 mem->pages[mem->page_count++] = page; in ttm_agp_bind()
|
/linux-3.4.99/drivers/gpu/drm/ ! |
D | drm_bufs.c | 748 dma->page_count += byte_count >> PAGE_SHIFT; in drm_addbufs_agp() 782 int page_count; in drm_addbufs_pci() local 852 temp_pagelist = kmalloc((dma->page_count + (count << page_order)) * in drm_addbufs_pci() 862 dma->pagelist, dma->page_count * sizeof(*dma->pagelist)); in drm_addbufs_pci() 864 dma->page_count + (count << page_order)); in drm_addbufs_pci() 869 page_count = 0; in drm_addbufs_pci() 888 dma->page_count + page_count, in drm_addbufs_pci() 890 temp_pagelist[dma->page_count + page_count++] in drm_addbufs_pci() 950 if (dma->page_count) { in drm_addbufs_pci() 957 dma->page_count += entry->seg_count << page_order; in drm_addbufs_pci() [all …]
|
/linux-3.4.99/drivers/net/ethernet/brocade/bna/ ! |
D | bna_tx_rx.c | 1352 (bfi_q)->pages = htons((u16)(bna_qpt)->page_count); \ 1911 u32 page_count, in bna_rxq_qpt_setup() argument 1922 rxq->qpt.page_count = page_count; in bna_rxq_qpt_setup() 1927 for (i = 0; i < rxq->qpt.page_count; i++) { in bna_rxq_qpt_setup() 1938 u32 page_count, in bna_rxp_cqpt_setup() argument 1949 rxp->cq.qpt.page_count = page_count; in bna_rxp_cqpt_setup() 1954 for (i = 0; i < rxp->cq.qpt.page_count; i++) { in bna_rxp_cqpt_setup() 2328 u32 page_count; in bna_rx_create() local 2361 page_count = res_info[BNA_RX_RES_MEM_T_CQPT_PAGE].res_u.mem_info.num / in bna_rx_create() 2457 q0->rcb->page_count = dpage_count; in bna_rx_create() [all …]
|