Searched refs:pg_chunk (Results 1 – 4 of 4) sorted by relevance
119 struct fl_pg_chunk pg_chunk; member354 if (q->use_pages && d->pg_chunk.page) { in clear_rx_desc()355 (*d->pg_chunk.p_cnt)--; in clear_rx_desc()356 if (!*d->pg_chunk.p_cnt) in clear_rx_desc()358 d->pg_chunk.mapping, in clear_rx_desc()361 put_page(d->pg_chunk.page); in clear_rx_desc()362 d->pg_chunk.page = NULL; in clear_rx_desc()392 if (q->pg_chunk.page) { in free_rx_bufs()393 __free_pages(q->pg_chunk.page, q->order); in free_rx_bufs()394 q->pg_chunk.page = NULL; in free_rx_bufs()[all …]
119 struct fl_pg_chunk pg_chunk;/* page chunk cache */ member
1086 if ((lbq_desc->p.pg_chunk.offset + rx_ring->lbq_buf_size) in ql_get_curr_lchunk()1089 lbq_desc->p.pg_chunk.map, in ql_get_curr_lchunk()1125 if (!rx_ring->pg_chunk.page) { in ql_get_next_chunk()1127 rx_ring->pg_chunk.page = alloc_pages(__GFP_COLD | __GFP_COMP | in ql_get_next_chunk()1130 if (unlikely(!rx_ring->pg_chunk.page)) { in ql_get_next_chunk()1135 rx_ring->pg_chunk.offset = 0; in ql_get_next_chunk()1136 map = pci_map_page(qdev->pdev, rx_ring->pg_chunk.page, in ql_get_next_chunk()1140 __free_pages(rx_ring->pg_chunk.page, in ql_get_next_chunk()1146 rx_ring->pg_chunk.map = map; in ql_get_next_chunk()1147 rx_ring->pg_chunk.va = page_address(rx_ring->pg_chunk.page); in ql_get_next_chunk()[all …]
1369 struct page_chunk pg_chunk; member1441 struct page_chunk pg_chunk; /* current page for chunks */ member