page 15 arch/x86/include/asm/agp.h #define map_page_into_agp(page) set_pages_uc(page, 1) page 16 arch/x86/include/asm/agp.h #define unmap_page_from_agp(page) set_pages_wb(page, 1) page 25 arch/x86/include/asm/cacheflush.h static inline unsigned long get_page_memtype(struct page *pg) page 39 arch/x86/include/asm/cacheflush.h static inline void set_page_memtype(struct page *pg, unsigned long memtype) page 63 arch/x86/include/asm/cacheflush.h static inline unsigned long get_page_memtype(struct page *pg) { return -1; } page 64 arch/x86/include/asm/cacheflush.h static inline void set_page_memtype(struct page *pg, unsigned long memtype) { } page 112 arch/x86/include/asm/cacheflush.h int set_pages_array_uc(struct page **pages, int addrinarray); page 113 arch/x86/include/asm/cacheflush.h int set_pages_array_wc(struct page **pages, int addrinarray); page 114 arch/x86/include/asm/cacheflush.h int set_pages_array_wb(struct page **pages, int addrinarray); page 136 arch/x86/include/asm/cacheflush.h int set_pages_uc(struct page *page, int numpages); page 137 arch/x86/include/asm/cacheflush.h int set_pages_wb(struct page *page, int numpages); page 138 arch/x86/include/asm/cacheflush.h int set_pages_x(struct page *page, int numpages); page 139 arch/x86/include/asm/cacheflush.h int set_pages_nx(struct page *page, int numpages); page 140 arch/x86/include/asm/cacheflush.h int set_pages_ro(struct page *page, int numpages); page 141 arch/x86/include/asm/cacheflush.h int set_pages_rw(struct page *page, int numpages); page 57 arch/x86/include/asm/highmem.h extern void *kmap_high(struct page *page); page 58 arch/x86/include/asm/highmem.h extern void kunmap_high(struct page *page); page 60 arch/x86/include/asm/highmem.h void *kmap(struct page *page); page 61 arch/x86/include/asm/highmem.h void kunmap(struct page *page); page 63 arch/x86/include/asm/highmem.h void *kmap_atomic_prot(struct page *page, pgprot_t prot); page 64 arch/x86/include/asm/highmem.h void *kmap_atomic(struct page *page); page 68 arch/x86/include/asm/highmem.h struct page *kmap_atomic_to_page(void *ptr); page 86 arch/x86/include/asm/hugetlb.h static inline int arch_prepare_hugepage(struct page *page) page 91 arch/x86/include/asm/hugetlb.h static inline void arch_release_hugepage(struct page *page) page 95 arch/x86/include/asm/hugetlb.h static inline void arch_clear_hugepage_flags(struct page *page) page 138 arch/x86/include/asm/io.h #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT) page 149 arch/x86/include/asm/io.h #define isa_page_to_bus(page) ((unsigned int)page_to_phys(page)) page 955 arch/x86/include/asm/kvm_host.h struct page *page = pfn_to_page(shadow_page >> PAGE_SHIFT); page 957 arch/x86/include/asm/kvm_host.h return (struct kvm_mmu_page *)page_private(page); page 11 arch/x86/include/asm/mmx.h extern void mmx_clear_page(void *page); page 18 arch/x86/include/asm/page.h struct page; page 24 arch/x86/include/asm/page.h static inline void clear_user_page(void *page, unsigned long vaddr, page 25 arch/x86/include/asm/page.h struct page *pg) page 27 arch/x86/include/asm/page.h clear_page(page); page 31 arch/x86/include/asm/page.h struct page *topage) page 24 arch/x86/include/asm/page_32.h static inline void clear_page(void *page) page 26 arch/x86/include/asm/page_32.h mmx_clear_page(page); page 36 arch/x86/include/asm/page_32.h static inline void clear_page(void *page) page 38 arch/x86/include/asm/page_32.h memset(page, 0, PAGE_SIZE); page 37 arch/x86/include/asm/page_64.h void clear_page(void *page); page 46 arch/x86/include/asm/paravirt_types.h struct page; page 48 arch/x86/include/asm/pgalloc.h static inline void pte_free(struct mm_struct *mm, struct page *pte) page 54 arch/x86/include/asm/pgalloc.h extern void ___pte_free_tlb(struct mmu_gather *tlb, struct page *pte); page 56 arch/x86/include/asm/pgalloc.h static inline void __pte_free_tlb(struct mmu_gather *tlb, struct page *pte, page 70 arch/x86/include/asm/pgalloc.h struct page *pte) page 83 arch/x86/include/asm/pgalloc.h struct page *page; page 84 arch/x86/include/asm/pgalloc.h page = alloc_pages(GFP_KERNEL | __GFP_REPEAT | __GFP_ZERO, 0); page 85 arch/x86/include/asm/pgalloc.h if (!page) page 87 arch/x86/include/asm/pgalloc.h if (!pgtable_pmd_page_ctor(page)) { page 88 arch/x86/include/asm/pgalloc.h __free_pages(page, 0); page 91 arch/x86/include/asm/pgalloc.h return (pmd_t *)page_address(page); page 33 arch/x86/include/asm/pgtable.h extern struct mm_struct *pgd_page_get_mm(struct page *page); page 536 arch/x86/include/asm/pgtable.h #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot)) page 791 arch/x86/include/asm/pgtable.h #define mk_pmd(page, pgprot) pfn_pmd(page_to_pfn(page), (pgprot)) page 185 arch/x86/include/asm/pgtable_64.h #define vmemmap ((struct page *)VMEMMAP_START) page 345 arch/x86/include/asm/pgtable_types.h typedef struct page *pgtable_t; page 24 arch/x86/include/asm/xen/page-coherent.h static inline void xen_dma_map_page(struct device *hwdev, struct page *page, page 54 arch/x86/include/asm/xen/page.h struct page **pages, unsigned int count); page 55 arch/x86/include/asm/xen/page.h extern int m2p_add_override(unsigned long mfn, struct page *page, page 59 arch/x86/include/asm/xen/page.h struct page **pages, unsigned int count); page 60 arch/x86/include/asm/xen/page.h extern int m2p_remove_override(struct page *page, page 63 arch/x86/include/asm/xen/page.h extern struct page *m2p_find_override(unsigned long mfn); page 17 include/asm-generic/cacheflush.h #define flush_dcache_page(page) do { } while (0) page 26 include/asm-generic/cacheflush.h #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ page 29 include/asm-generic/cacheflush.h flush_icache_user_range(vma, page, vaddr, len); \ page 31 include/asm-generic/cacheflush.h #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ page 31 include/asm-generic/crash-driver.h map_virtual(u64 offset, struct page **pp) page 33 include/asm-generic/crash-driver.h struct page *page; page 51 include/asm-generic/crash-driver.h page = pfn_to_page(pfn); page 53 include/asm-generic/crash-driver.h vaddr = kmap(page); page 57 include/asm-generic/crash-driver.h pfn, (unsigned long)page); page 61 include/asm-generic/crash-driver.h *pp = page; page 65 include/asm-generic/crash-driver.h static inline void unmap_virtual(struct page *page) page 67 include/asm-generic/crash-driver.h kunmap(page); page 55 include/asm-generic/dma-mapping-broken.h dma_map_page(struct device *dev, struct page *page, unsigned long offset, page 71 include/asm-generic/dma-mapping-common.h static inline dma_addr_t dma_map_page(struct device *dev, struct page *page, page 78 include/asm-generic/dma-mapping-common.h kmemcheck_mark_initialized(page_address(page) + offset, size); page 80 include/asm-generic/dma-mapping-common.h addr = ops->map_page(dev, page, offset, size, dir, NULL); page 81 include/asm-generic/dma-mapping-common.h debug_dma_map_page(dev, page, offset, size, dir, addr, false); page 182 include/asm-generic/dma-mapping-common.h void *dma_common_contiguous_remap(struct page *page, size_t size, page 186 include/asm-generic/dma-mapping-common.h void *dma_common_pages_remap(struct page **pages, size_t size, page 4 include/asm-generic/hugetlb.h static inline pte_t mk_huge_pte(struct page *page, pgprot_t pgprot) page 6 include/asm-generic/hugetlb.h return mk_pte(page, pgprot); page 31 include/asm-generic/memory_model.h #define __page_to_pfn(page) ((unsigned long)((page) - mem_map) + \ page 42 include/asm-generic/memory_model.h ({ const struct page *__pg = (pg); \ page 52 include/asm-generic/memory_model.h #define __page_to_pfn(page) (unsigned long)((page) - vmemmap) page 60 include/asm-generic/memory_model.h ({ const struct page *__pg = (pg); \ page 28 include/asm-generic/page.h #define free_user_page(page, addr) free_page(addr) page 30 include/asm-generic/page.h #define clear_page(page) memset((page), 0, PAGE_SIZE) page 33 include/asm-generic/page.h #define clear_user_page(page, vaddr, pg) clear_page(page) page 51 include/asm-generic/page.h typedef struct page *pgtable_t; page 87 include/asm-generic/page.h #define page_to_virt(page) pfn_to_virt(page_to_pfn(page)) page 90 include/asm-generic/page.h #define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT) page 51 include/asm-generic/pci-dma-compat.h pci_map_page(struct pci_dev *hwdev, struct page *page, page 54 include/asm-generic/pci-dma-compat.h return dma_map_page(hwdev == NULL ? NULL : &hwdev->dev, page, offset, size, (enum dma_data_direction)direction); page 75 include/asm-generic/tlb.h struct page *pages[0]; page 109 include/asm-generic/tlb.h struct page *__pages[MMU_GATHER_BUNDLE]; page 119 include/asm-generic/tlb.h int __tlb_remove_page(struct mmu_gather *tlb, struct page *page); page 125 include/asm-generic/tlb.h static inline void tlb_remove_page(struct mmu_gather *tlb, struct page *page) page 127 include/asm-generic/tlb.h if (!__tlb_remove_page(tlb, page)) page 84 include/crypto/algapi.h struct page *page; page 89 include/crypto/algapi.h u8 *page; page 100 include/crypto/algapi.h void *page; page 113 include/crypto/algapi.h struct page *page; page 60 include/crypto/if_alg.h struct page *pages[ALG_MAX_PAGES]; page 28 include/crypto/internal/hash.h struct page *pg; page 92 include/crypto/scatterwalk.h static inline struct page *scatterwalk_page(struct scatter_walk *walk) page 887 include/drm/drmP.h void drm_clflush_pages(struct page *pages[], unsigned long num_pages); page 987 include/drm/drmP.h extern int drm_prime_sg_to_page_addr_arrays(struct sg_table *sgt, struct page **pages, page 989 include/drm/drmP.h extern struct sg_table *drm_prime_pages_to_sg(struct page **pages, int nr_pages); page 37 include/drm/drm_agpsupport.h struct page **pages, page 86 include/drm/drm_agpsupport.h struct page **pages, page 36 include/drm/drm_cache.h void drm_clflush_pages(struct page *pages[], unsigned long num_pages); page 172 include/drm/drm_gem.h struct page **drm_gem_get_pages(struct drm_gem_object *obj); page 173 include/drm/drm_gem.h void drm_gem_put_pages(struct drm_gem_object *obj, struct page **pages, page 124 include/drm/drm_legacy.h struct page **pagelist; page 278 include/drm/ttm/ttm_bo_api.h struct page *page; page 117 include/drm/ttm/ttm_bo_driver.h struct page *dummy_read_page; page 118 include/drm/ttm/ttm_bo_driver.h struct page **pages; page 334 include/drm/ttm/ttm_bo_driver.h struct page *dummy_read_page); page 470 include/drm/ttm/ttm_bo_driver.h struct page *dummy_read_page; page 581 include/drm/ttm/ttm_bo_driver.h struct page *dummy_read_page); page 584 include/drm/ttm/ttm_bo_driver.h struct page *dummy_read_page); page 1054 include/drm/ttm/ttm_bo_driver.h struct page *dummy_read_page); page 153 include/drm/ttm/ttm_memory.h struct page *page, page 156 include/drm/ttm/ttm_memory.h struct page *page); page 73 include/linux/agp_backend.h struct page **pages; page 129 include/linux/async_tx.h enum dma_transaction_type tx_type, struct page **dst, page 130 include/linux/async_tx.h int dst_count, struct page **src, int src_count, page 151 include/linux/async_tx.h struct page *page; page 172 include/linux/async_tx.h async_xor(struct page *dest, struct page **src_list, unsigned int offset, page 176 include/linux/async_tx.h async_xor_val(struct page *dest, struct page **src_list, unsigned int offset, page 181 include/linux/async_tx.h async_memcpy(struct page *dest, struct page *src, unsigned int dest_offset, page 188 include/linux/async_tx.h async_gen_syndrome(struct page **blocks, unsigned int offset, int src_cnt, page 192 include/linux/async_tx.h async_syndrome_val(struct page **blocks, unsigned int offset, int src_cnt, page 193 include/linux/async_tx.h size_t len, enum sum_check_flags *pqres, struct page *spare, page 198 include/linux/async_tx.h struct page **ptrs, struct async_submit_ctl *submit); page 202 include/linux/async_tx.h struct page **ptrs, struct async_submit_ctl *submit); page 201 include/linux/atmdev.h int (*proc_read)(struct atm_dev *dev,loff_t *pos,char *page); page 23 include/linux/backing-dev.h struct page; page 63 include/linux/balloon_compaction.h int (*migratepage)(struct balloon_dev_info *, struct page *newpage, page 64 include/linux/balloon_compaction.h struct page *page, enum migrate_mode mode); page 67 include/linux/balloon_compaction.h extern struct page *balloon_page_enqueue(struct balloon_dev_info *b_dev_info); page 68 include/linux/balloon_compaction.h extern struct page *balloon_page_dequeue(struct balloon_dev_info *b_dev_info); page 79 include/linux/balloon_compaction.h extern bool balloon_page_isolate(struct page *page); page 80 include/linux/balloon_compaction.h extern void balloon_page_putback(struct page *page); page 81 include/linux/balloon_compaction.h extern int balloon_page_migrate(struct page *newpage, page 82 include/linux/balloon_compaction.h struct page *page, enum migrate_mode mode); page 87 include/linux/balloon_compaction.h static inline bool __is_movable_balloon_page(struct page *page) page 89 include/linux/balloon_compaction.h return PageBalloon(page); page 100 include/linux/balloon_compaction.h static inline bool balloon_page_movable(struct page *page) page 102 include/linux/balloon_compaction.h return PageBalloon(page) && PagePrivate(page); page 109 include/linux/balloon_compaction.h static inline bool isolated_balloon_page(struct page *page) page 111 include/linux/balloon_compaction.h return PageBalloon(page); page 124 include/linux/balloon_compaction.h struct page *page) page 126 include/linux/balloon_compaction.h __SetPageBalloon(page); page 127 include/linux/balloon_compaction.h SetPagePrivate(page); page 128 include/linux/balloon_compaction.h set_page_private(page, (unsigned long)balloon); page 129 include/linux/balloon_compaction.h list_add(&page->lru, &balloon->pages); page 140 include/linux/balloon_compaction.h static inline void balloon_page_delete(struct page *page) page 142 include/linux/balloon_compaction.h __ClearPageBalloon(page); page 143 include/linux/balloon_compaction.h set_page_private(page, 0); page 144 include/linux/balloon_compaction.h if (PagePrivate(page)) { page 145 include/linux/balloon_compaction.h ClearPagePrivate(page); page 146 include/linux/balloon_compaction.h list_del(&page->lru); page 154 include/linux/balloon_compaction.h static inline struct balloon_dev_info *balloon_page_device(struct page *page) page 156 include/linux/balloon_compaction.h return (struct balloon_dev_info *)page_private(page); page 167 include/linux/balloon_compaction.h struct page *page) page 169 include/linux/balloon_compaction.h __SetPageBalloon(page); page 170 include/linux/balloon_compaction.h list_add(&page->lru, &balloon->pages); page 173 include/linux/balloon_compaction.h static inline void balloon_page_delete(struct page *page) page 175 include/linux/balloon_compaction.h __ClearPageBalloon(page); page 176 include/linux/balloon_compaction.h list_del(&page->lru); page 179 include/linux/balloon_compaction.h static inline bool __is_movable_balloon_page(struct page *page) page 184 include/linux/balloon_compaction.h static inline bool balloon_page_movable(struct page *page) page 189 include/linux/balloon_compaction.h static inline bool isolated_balloon_page(struct page *page) page 194 include/linux/balloon_compaction.h static inline bool balloon_page_isolate(struct page *page) page 199 include/linux/balloon_compaction.h static inline void balloon_page_putback(struct page *page) page 204 include/linux/balloon_compaction.h static inline int balloon_page_migrate(struct page *newpage, page 205 include/linux/balloon_compaction.h struct page *page, enum migrate_mode mode) page 21 include/linux/binfmts.h struct page *page[MAX_ARG_PAGES]; page 427 include/linux/bio.h extern int bio_add_page(struct bio *, struct page *, unsigned int,unsigned int); page 428 include/linux/bio.h extern int bio_add_pc_page(struct request_queue *, struct bio *, struct page *, page 702 include/linux/bio.h extern int bio_integrity_add_page(struct bio *, struct page *, unsigned int, unsigned int); page 13 include/linux/blk_types.h struct page; page 24 include/linux/blk_types.h struct page *bv_page; page 748 include/linux/blkdev.h struct page **pages; page 798 include/linux/blkdev.h extern void blk_add_request_payload(struct request *rq, struct page *page, page 1164 include/linux/blkdev.h sector_t nr_sects, gfp_t gfp_mask, struct page *page); page 1396 include/linux/blkdev.h typedef struct {struct page *v;} Sector; page 1603 include/linux/blkdev.h int (*rw_page)(struct block_device *, sector_t, struct page *, int rw); page 1622 include/linux/blkdev.h extern int bdev_read_page(struct block_device *, sector_t, struct page *); page 1623 include/linux/blkdev.h extern int bdev_write_page(struct block_device *, sector_t, struct page *, page 48 include/linux/buffer_head.h struct page; page 65 include/linux/buffer_head.h struct page *b_page; /* the page this bh is mapped to */ page 137 include/linux/buffer_head.h #define page_buffers(page) \ page 139 include/linux/buffer_head.h BUG_ON(!PagePrivate(page)); \ page 140 include/linux/buffer_head.h ((struct buffer_head *)page_private(page)); \ page 142 include/linux/buffer_head.h #define page_has_buffers(page) PagePrivate(page) page 144 include/linux/buffer_head.h void buffer_check_dirty_writeback(struct page *page, page 155 include/linux/buffer_head.h struct page *page, unsigned long offset); page 156 include/linux/buffer_head.h int try_to_free_buffers(struct page *); page 157 include/linux/buffer_head.h struct buffer_head *alloc_page_buffers(struct page *page, unsigned long size, page 159 include/linux/buffer_head.h void create_empty_buffers(struct page *, unsigned long, page 207 include/linux/buffer_head.h void block_invalidatepage(struct page *page, unsigned int offset, page 209 include/linux/buffer_head.h int block_write_full_page(struct page *page, get_block_t *get_block, page 211 include/linux/buffer_head.h int block_read_full_page(struct page*, get_block_t*); page 212 include/linux/buffer_head.h int block_is_partially_uptodate(struct page *page, unsigned long from, page 215 include/linux/buffer_head.h unsigned flags, struct page **pagep, get_block_t *get_block); page 216 include/linux/buffer_head.h int __block_write_begin(struct page *page, loff_t pos, unsigned len, page 220 include/linux/buffer_head.h struct page *, void *); page 223 include/linux/buffer_head.h struct page *, void *); page 224 include/linux/buffer_head.h void page_zero_new_buffers(struct page *page, unsigned from, unsigned to); page 226 include/linux/buffer_head.h unsigned, unsigned, struct page **, void **, page 229 include/linux/buffer_head.h int block_commit_write(struct page *page, unsigned from, unsigned to); page 251 include/linux/buffer_head.h struct page **, void **, get_block_t*); page 254 include/linux/buffer_head.h struct page *, void *); page 256 include/linux/buffer_head.h int nobh_writepage(struct page *page, get_block_t *get_block, page 265 include/linux/buffer_head.h static inline void attach_page_buffers(struct page *page, page 268 include/linux/buffer_head.h page_cache_get(page); page 269 include/linux/buffer_head.h SetPagePrivate(page); page 270 include/linux/buffer_head.h set_page_private(page, (unsigned long)head); page 384 include/linux/buffer_head.h extern int __set_page_dirty_buffers(struct page *page); page 389 include/linux/buffer_head.h static inline int try_to_free_buffers(struct page *page) { return 1; } page 207 include/linux/ceph/libceph.h extern void ceph_release_page_vector(struct page **pages, int num_pages); page 209 include/linux/ceph/libceph.h extern struct page **ceph_get_direct_page_vector(const void __user *data, page 212 include/linux/ceph/libceph.h extern void ceph_put_page_vector(struct page **pages, int num_pages, page 214 include/linux/ceph/libceph.h extern struct page **ceph_alloc_page_vector(int num_pages, gfp_t flags); page 215 include/linux/ceph/libceph.h extern int ceph_copy_user_to_page_vector(struct page **pages, page 218 include/linux/ceph/libceph.h extern void ceph_copy_to_page_vector(struct page **pages, page 221 include/linux/ceph/libceph.h extern void ceph_copy_from_page_vector(struct page **pages, page 224 include/linux/ceph/libceph.h extern void ceph_zero_page_vector_range(int off, int len, struct page **pages); page 103 include/linux/ceph/messenger.h struct page **pages; /* NOT OWNER. */ page 132 include/linux/ceph/messenger.h struct page *page; /* page from list */ page 277 include/linux/ceph/messenger.h extern void ceph_msg_data_add_pages(struct ceph_msg *msg, struct page **pages, page 61 include/linux/ceph/osd_client.h struct page **pages; page 249 include/linux/ceph/osd_client.h struct page **pages, u64 length, page 269 include/linux/ceph/osd_client.h struct page **pages, u64 length, page 286 include/linux/ceph/osd_client.h struct page **pages, u64 length, page 291 include/linux/ceph/osd_client.h struct page **pages, u64 length, page 347 include/linux/ceph/osd_client.h struct page **pages, int nr_pages, page 357 include/linux/ceph/osd_client.h struct page **pages, int nr_pages); page 28 include/linux/cleancache.h pgoff_t, struct page *); page 30 include/linux/cleancache.h pgoff_t, struct page *); page 40 include/linux/cleancache.h extern int __cleancache_get_page(struct page *); page 41 include/linux/cleancache.h extern void __cleancache_put_page(struct page *); page 42 include/linux/cleancache.h extern void __cleancache_invalidate_page(struct address_space *, struct page *); page 48 include/linux/cleancache.h static inline bool cleancache_fs_enabled(struct page *page) page 50 include/linux/cleancache.h return page->mapping->host->i_sb->cleancache_poolid >= 0; page 87 include/linux/cleancache.h static inline int cleancache_get_page(struct page *page) page 91 include/linux/cleancache.h if (cleancache_enabled && cleancache_fs_enabled(page)) page 92 include/linux/cleancache.h ret = __cleancache_get_page(page); page 96 include/linux/cleancache.h static inline void cleancache_put_page(struct page *page) page 98 include/linux/cleancache.h if (cleancache_enabled && cleancache_fs_enabled(page)) page 99 include/linux/cleancache.h __cleancache_put_page(page); page 103 include/linux/cleancache.h struct page *page) page 107 include/linux/cleancache.h __cleancache_invalidate_page(mapping, page); page 28 include/linux/cma.h extern struct page *cma_alloc(struct cma *cma, int count, unsigned int align); page 29 include/linux/cma.h extern bool cma_release(struct cma *cma, struct page *pages, int count); page 185 include/linux/configfs.h char *page) \ page 193 include/linux/configfs.h ret = _item##_attr->show(_item, page); \ page 198 include/linux/configfs.h const char *page, size_t count) \ page 206 include/linux/configfs.h ret = _item##_attr->store(_item, page, count); \ page 25 include/linux/dm-io.h struct page *page; page 59 include/linux/dma-contiguous.h struct page; page 114 include/linux/dma-contiguous.h struct page *dma_alloc_from_contiguous(struct device *dev, int count, page 116 include/linux/dma-contiguous.h bool dma_release_from_contiguous(struct device *dev, struct page *pages, page 147 include/linux/dma-contiguous.h struct page *dma_alloc_from_contiguous(struct device *dev, int count, page 154 include/linux/dma-contiguous.h bool dma_release_from_contiguous(struct device *dev, struct page *pages, page 37 include/linux/dma-debug.h extern void debug_dma_map_page(struct device *dev, struct page *page, page 88 include/linux/dma-debug.h extern void debug_dma_assert_idle(struct page *page); page 105 include/linux/dma-debug.h static inline void debug_dma_map_page(struct device *dev, struct page *page, page 188 include/linux/dma-debug.h static inline void debug_dma_assert_idle(struct page *page) page 30 include/linux/dma-mapping.h dma_addr_t (*map_page)(struct device *dev, struct page *page, page 1128 include/linux/dmaengine.h struct page **pages; page 1142 include/linux/dmaengine.h struct dma_pinned_list *pinned_list, struct page *page, page 699 include/linux/edac.h unsigned long page); page 162 include/linux/f2fs_fs.h #define ADDRS_PER_PAGE(page, fi) \ page 163 include/linux/f2fs_fs.h (IS_INODE(page) ? ADDRS_PER_INODE(fi) : ADDRS_PER_BLOCK) page 422 include/linux/firewire.h struct page **pages; page 14 include/linux/firmware.h struct page **pages; page 10 include/linux/frontswap.h int (*store)(unsigned, pgoff_t, struct page *); page 11 include/linux/frontswap.h int (*load)(unsigned, pgoff_t, struct page *); page 27 include/linux/frontswap.h extern int __frontswap_store(struct page *page); page 28 include/linux/frontswap.h extern int __frontswap_load(struct page *page); page 71 include/linux/frontswap.h static inline int frontswap_store(struct page *page) page 76 include/linux/frontswap.h ret = __frontswap_store(page); page 80 include/linux/frontswap.h static inline int frontswap_load(struct page *page) page 85 include/linux/frontswap.h ret = __frontswap_load(page); page 310 include/linux/fs.h struct page; page 333 include/linux/fs.h typedef int (*read_actor_t)(read_descriptor_t *, struct page *, page 337 include/linux/fs.h int (*writepage)(struct page *page, struct writeback_control *wbc); page 338 include/linux/fs.h int (*readpage)(struct file *, struct page *); page 344 include/linux/fs.h int (*set_page_dirty)(struct page *page); page 351 include/linux/fs.h struct page **pagep, void **fsdata); page 354 include/linux/fs.h struct page *page, void *fsdata); page 358 include/linux/fs.h void (*invalidatepage) (struct page *, unsigned int, unsigned int); page 359 include/linux/fs.h int (*releasepage) (struct page *, gfp_t); page 360 include/linux/fs.h void (*freepage)(struct page *); page 369 include/linux/fs.h struct page *, struct page *, enum migrate_mode); page 370 include/linux/fs.h int (*launder_page) (struct page *); page 371 include/linux/fs.h int (*is_partially_uptodate) (struct page *, unsigned long, page 373 include/linux/fs.h void (*is_dirty_writeback) (struct page *, bool *, bool *); page 374 include/linux/fs.h int (*error_remove_page)(struct address_space *, struct page *); page 390 include/linux/fs.h struct page **pagep, void **fsdata); page 394 include/linux/fs.h struct page *page, void *fsdata); page 1507 include/linux/fs.h ssize_t (*sendpage) (struct file *, struct page *, int, size_t, loff_t *, int); page 1594 include/linux/fs.h int (*bdev_try_to_free_page)(struct super_block*, struct page*, gfp_t); page 2476 include/linux/fs.h extern void block_sync_page(struct page *page); page 2623 include/linux/fs.h extern int simple_readpage(struct file *file, struct page *page); page 2626 include/linux/fs.h struct page **pagep, void **fsdata); page 2629 include/linux/fs.h struct page *page, void *fsdata); page 2657 include/linux/fs.h struct page *, struct page *, page 158 include/linux/fscache-cache.h struct page *page, page 298 include/linux/fscache-cache.h int (*write_page)(struct fscache_storage *op, struct page *page); page 305 include/linux/fscache-cache.h struct page *page); page 480 include/linux/fscache-cache.h struct page *page, int error) page 482 include/linux/fscache-cache.h op->end_io_func(page, op->context, error); page 543 include/linux/fscache-cache.h struct page *page); page 39 include/linux/fscache.h #define PageFsCache(page) PagePrivate2((page)) page 40 include/linux/fscache.h #define SetPageFsCache(page) SetPagePrivate2((page)) page 41 include/linux/fscache.h #define ClearPageFsCache(page) ClearPagePrivate2((page)) page 42 include/linux/fscache.h #define TestSetPageFsCache(page) TestSetPagePrivate2((page)) page 43 include/linux/fscache.h #define TestClearPageFsCache(page) TestClearPagePrivate2((page)) page 53 include/linux/fscache.h typedef void (*fscache_rw_complete_t)(struct page *page, page 145 include/linux/fscache.h struct page *page); page 228 include/linux/fscache.h struct page *, page 239 include/linux/fscache.h extern int __fscache_alloc_page(struct fscache_cookie *, struct page *, gfp_t); page 240 include/linux/fscache.h extern int __fscache_write_page(struct fscache_cookie *, struct page *, gfp_t); page 241 include/linux/fscache.h extern void __fscache_uncache_page(struct fscache_cookie *, struct page *); page 242 include/linux/fscache.h extern bool __fscache_check_page_write(struct fscache_cookie *, struct page *); page 243 include/linux/fscache.h extern void __fscache_wait_on_page_write(struct fscache_cookie *, struct page *); page 244 include/linux/fscache.h extern bool __fscache_maybe_release_page(struct fscache_cookie *, struct page *, page 543 include/linux/fscache.h struct page *page, page 549 include/linux/fscache.h return __fscache_read_or_alloc_page(cookie, page, end_io_func, page 627 include/linux/fscache.h struct page *page, page 631 include/linux/fscache.h return __fscache_alloc_page(cookie, page, gfp); page 676 include/linux/fscache.h struct page *page, page 680 include/linux/fscache.h return __fscache_write_page(cookie, page, gfp); page 701 include/linux/fscache.h struct page *page) page 704 include/linux/fscache.h __fscache_uncache_page(cookie, page); page 719 include/linux/fscache.h struct page *page) page 722 include/linux/fscache.h return __fscache_check_page_write(cookie, page); page 739 include/linux/fscache.h struct page *page) page 742 include/linux/fscache.h __fscache_wait_on_page_write(cookie, page); page 762 include/linux/fscache.h struct page *page, page 765 include/linux/fscache.h if (fscache_cookie_valid(cookie) && PageFsCache(page)) page 766 include/linux/fscache.h return __fscache_maybe_release_page(cookie, page, gfp); page 295 include/linux/gfp.h static inline void arch_free_page(struct page *page, int order) { } page 298 include/linux/gfp.h static inline void arch_alloc_page(struct page *page, int order) { } page 301 include/linux/gfp.h struct page * page 305 include/linux/gfp.h static inline struct page * page 312 include/linux/gfp.h static inline struct page *alloc_pages_node(int nid, gfp_t gfp_mask, page 322 include/linux/gfp.h static inline struct page *alloc_pages_exact_node(int nid, gfp_t gfp_mask, page 331 include/linux/gfp.h extern struct page *alloc_pages_current(gfp_t gfp_mask, unsigned order); page 333 include/linux/gfp.h static inline struct page * page 338 include/linux/gfp.h extern struct page *alloc_pages_vma(gfp_t gfp_mask, int order, page 353 include/linux/gfp.h extern struct page *alloc_kmem_pages(gfp_t gfp_mask, unsigned int order); page 354 include/linux/gfp.h extern struct page *alloc_kmem_pages_node(int nid, gfp_t gfp_mask, page 371 include/linux/gfp.h extern void __free_pages(struct page *page, unsigned int order); page 373 include/linux/gfp.h extern void free_hot_cold_page(struct page *page, bool cold); page 376 include/linux/gfp.h extern void __free_kmem_pages(struct page *page, unsigned int order); page 379 include/linux/gfp.h #define __free_page(page) __free_pages((page), 0) page 419 include/linux/gfp.h extern void init_cma_reserved_pageblock(struct page *page); page 14 include/linux/highmem.h static inline void flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr) page 20 include/linux/highmem.h static inline void flush_kernel_dcache_page(struct page *page) page 42 include/linux/highmem.h struct page *kmap_to_page(void *addr); page 48 include/linux/highmem.h static inline struct page *kmap_to_page(void *addr) page 56 include/linux/highmem.h static inline void *kmap(struct page *page) page 59 include/linux/highmem.h return page_address(page); page 62 include/linux/highmem.h static inline void kunmap(struct page *page) page 66 include/linux/highmem.h static inline void *kmap_atomic(struct page *page) page 69 include/linux/highmem.h return page_address(page); page 71 include/linux/highmem.h #define kmap_atomic_prot(page, prot) kmap_atomic(page) page 125 include/linux/highmem.h BUILD_BUG_ON(__same_type((addr), struct page *)); \ page 132 include/linux/highmem.h static inline void clear_user_highpage(struct page *page, unsigned long vaddr) page 134 include/linux/highmem.h void *addr = kmap_atomic(page); page 135 include/linux/highmem.h clear_user_page(addr, vaddr, page); page 155 include/linux/highmem.h static inline struct page * page 160 include/linux/highmem.h struct page *page = alloc_page_vma(GFP_HIGHUSER | movableflags, page 163 include/linux/highmem.h if (page) page 164 include/linux/highmem.h clear_user_highpage(page, vaddr); page 166 include/linux/highmem.h return page; page 178 include/linux/highmem.h static inline struct page * page 185 include/linux/highmem.h static inline void clear_highpage(struct page *page) page 187 include/linux/highmem.h void *kaddr = kmap_atomic(page); page 192 include/linux/highmem.h static inline void zero_user_segments(struct page *page, page 196 include/linux/highmem.h void *kaddr = kmap_atomic(page); page 207 include/linux/highmem.h flush_dcache_page(page); page 210 include/linux/highmem.h static inline void zero_user_segment(struct page *page, page 213 include/linux/highmem.h zero_user_segments(page, start, end, 0, 0); page 216 include/linux/highmem.h static inline void zero_user(struct page *page, page 219 include/linux/highmem.h zero_user_segments(page, start, start + size, 0, 0); page 224 include/linux/highmem.h static inline void copy_user_highpage(struct page *to, struct page *from, page 238 include/linux/highmem.h static inline void copy_highpage(struct page *to, struct page *from) page 18 include/linux/huge_mm.h extern struct page *follow_trans_huge_pmd(struct vm_area_struct *vma, page 54 include/linux/huge_mm.h extern pmd_t *page_check_address_pmd(struct page *page, page 96 include/linux/huge_mm.h extern int split_huge_page_to_list(struct page *page, struct list_head *list); page 97 include/linux/huge_mm.h static inline int split_huge_page(struct page *page) page 99 include/linux/huge_mm.h return split_huge_page_to_list(page, NULL); page 150 include/linux/huge_mm.h static inline int hpage_nr_pages(struct page *page) page 152 include/linux/huge_mm.h if (unlikely(PageTransHuge(page))) page 171 include/linux/huge_mm.h split_huge_page_to_list(struct page *page, struct list_head *list) page 175 include/linux/huge_mm.h static inline int split_huge_page(struct page *page) page 44 include/linux/hugetlb.h int PageHuge(struct page *page); page 58 include/linux/hugetlb.h struct page **, struct vm_area_struct **, page 61 include/linux/hugetlb.h unsigned long, unsigned long, struct page *); page 65 include/linux/hugetlb.h struct page *ref_page); page 68 include/linux/hugetlb.h struct page *ref_page); page 79 include/linux/hugetlb.h int dequeue_hwpoisoned_huge_page(struct page *page); page 80 include/linux/hugetlb.h bool isolate_huge_page(struct page *page, struct list_head *list); page 81 include/linux/hugetlb.h void putback_active_hugepage(struct page *page); page 82 include/linux/hugetlb.h bool is_hugepage_active(struct page *page); page 83 include/linux/hugetlb.h void free_huge_page(struct page *page); page 99 include/linux/hugetlb.h struct page *follow_huge_addr(struct mm_struct *mm, unsigned long address, page 101 include/linux/hugetlb.h struct page *follow_huge_pmd(struct mm_struct *mm, unsigned long address, page 103 include/linux/hugetlb.h struct page *follow_huge_pud(struct mm_struct *mm, unsigned long address, page 112 include/linux/hugetlb.h static inline int PageHuge(struct page *page) page 145 include/linux/hugetlb.h static inline int dequeue_hwpoisoned_huge_page(struct page *page) page 150 include/linux/hugetlb.h static inline bool isolate_huge_page(struct page *page, struct list_head *list) page 165 include/linux/hugetlb.h unsigned long end, struct page *ref_page) page 172 include/linux/hugetlb.h unsigned long end, struct page *ref_page) page 279 include/linux/hugetlb.h struct page *alloc_huge_page_node(struct hstate *h, int nid); page 280 include/linux/hugetlb.h struct page *alloc_huge_page_noerr(struct vm_area_struct *vma, page 365 include/linux/hugetlb.h struct page *page, int writable) page 371 include/linux/hugetlb.h static inline struct hstate *page_hstate(struct page *page) page 373 include/linux/hugetlb.h VM_BUG_ON_PAGE(!PageHuge(page), page); page 374 include/linux/hugetlb.h return size_to_hstate(PAGE_SIZE << compound_order(page)); page 387 include/linux/hugetlb.h pgoff_t __basepage_index(struct page *page); page 390 include/linux/hugetlb.h static inline pgoff_t basepage_index(struct page *page) page 392 include/linux/hugetlb.h if (!PageCompound(page)) page 393 include/linux/hugetlb.h return page->index; page 395 include/linux/hugetlb.h return __basepage_index(page); page 437 include/linux/hugetlb.h #define page_hstate(page) NULL page 451 include/linux/hugetlb.h static inline pgoff_t basepage_index(struct page *page) page 453 include/linux/hugetlb.h return page->index; page 30 include/linux/hugetlb_cgroup.h static inline struct hugetlb_cgroup *hugetlb_cgroup_from_page(struct page *page) page 32 include/linux/hugetlb_cgroup.h VM_BUG_ON_PAGE(!PageHuge(page), page); page 34 include/linux/hugetlb_cgroup.h if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) page 36 include/linux/hugetlb_cgroup.h return (struct hugetlb_cgroup *)page[2].lru.next; page 40 include/linux/hugetlb_cgroup.h int set_hugetlb_cgroup(struct page *page, struct hugetlb_cgroup *h_cg) page 42 include/linux/hugetlb_cgroup.h VM_BUG_ON_PAGE(!PageHuge(page), page); page 44 include/linux/hugetlb_cgroup.h if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) page 46 include/linux/hugetlb_cgroup.h page[2].lru.next = (void *)h_cg; page 61 include/linux/hugetlb_cgroup.h struct page *page); page 63 include/linux/hugetlb_cgroup.h struct page *page); page 67 include/linux/hugetlb_cgroup.h extern void hugetlb_cgroup_migrate(struct page *oldhpage, page 68 include/linux/hugetlb_cgroup.h struct page *newhpage); page 71 include/linux/hugetlb_cgroup.h static inline struct hugetlb_cgroup *hugetlb_cgroup_from_page(struct page *page) page 77 include/linux/hugetlb_cgroup.h int set_hugetlb_cgroup(struct page *page, struct hugetlb_cgroup *h_cg) page 97 include/linux/hugetlb_cgroup.h struct page *page) page 103 include/linux/hugetlb_cgroup.h hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, struct page *page) page 119 include/linux/hugetlb_cgroup.h static inline void hugetlb_cgroup_migrate(struct page *oldhpage, page 120 include/linux/hugetlb_cgroup.h struct page *newhpage) page 864 include/linux/jbd.h struct page *, unsigned int, unsigned int); page 865 include/linux/jbd.h extern int journal_try_to_free_buffers(journal_t *, struct page *, gfp_t); page 1134 include/linux/jbd2.h struct page *, unsigned int, unsigned int); page 1135 include/linux/jbd2.h extern int jbd2_journal_try_to_free_buffers(journal_t *, struct page *, gfp_t); page 128 include/linux/kexec.h struct page *control_code_page; page 129 include/linux/kexec.h struct page *swap_page; page 219 include/linux/kexec.h extern struct page *kimage_alloc_control_pages(struct kimage *image, page 267 include/linux/kexec.h #define kexec_flush_icache_page(page) page 11 include/linux/kmemcheck.h void kmemcheck_alloc_shadow(struct page *page, int order, gfp_t flags, int node); page 12 include/linux/kmemcheck.h void kmemcheck_free_shadow(struct page *page, int order); page 17 include/linux/kmemcheck.h void kmemcheck_pagealloc_alloc(struct page *p, unsigned int order, page 20 include/linux/kmemcheck.h void kmemcheck_show_pages(struct page *p, unsigned int n); page 21 include/linux/kmemcheck.h void kmemcheck_hide_pages(struct page *p, unsigned int n); page 23 include/linux/kmemcheck.h bool kmemcheck_page_is_tracked(struct page *p); page 30 include/linux/kmemcheck.h void kmemcheck_mark_unallocated_pages(struct page *p, unsigned int n); page 31 include/linux/kmemcheck.h void kmemcheck_mark_uninitialized_pages(struct page *p, unsigned int n); page 32 include/linux/kmemcheck.h void kmemcheck_mark_initialized_pages(struct page *p, unsigned int n); page 93 include/linux/kmemcheck.h kmemcheck_alloc_shadow(struct page *page, int order, gfp_t flags, int node) page 98 include/linux/kmemcheck.h kmemcheck_free_shadow(struct page *page, int order) page 113 include/linux/kmemcheck.h static inline void kmemcheck_pagealloc_alloc(struct page *p, page 118 include/linux/kmemcheck.h static inline bool kmemcheck_page_is_tracked(struct page *p) page 139 include/linux/kmemcheck.h static inline void kmemcheck_mark_unallocated_pages(struct page *p, page 144 include/linux/kmemcheck.h static inline void kmemcheck_mark_uninitialized_pages(struct page *p, page 149 include/linux/kmemcheck.h static inline void kmemcheck_mark_initialized_pages(struct page *p, page 44 include/linux/ks0108.h extern void ks0108_page(unsigned char page); page 44 include/linux/ksm.h static inline int PageKsm(struct page *page) page 46 include/linux/ksm.h return ((unsigned long)page->mapping & PAGE_MAPPING_FLAGS) == page 50 include/linux/ksm.h static inline struct stable_node *page_stable_node(struct page *page) page 52 include/linux/ksm.h return PageKsm(page) ? page_rmapping(page) : NULL; page 55 include/linux/ksm.h static inline void set_page_stable_node(struct page *page, page 58 include/linux/ksm.h page->mapping = (void *)stable_node + page 73 include/linux/ksm.h struct page *ksm_might_need_to_copy(struct page *page, page 76 include/linux/ksm.h int rmap_walk_ksm(struct page *page, struct rmap_walk_control *rwc); page 77 include/linux/ksm.h void ksm_migrate_page(struct page *newpage, struct page *oldpage); page 90 include/linux/ksm.h static inline int PageKsm(struct page *page) page 102 include/linux/ksm.h static inline struct page *ksm_might_need_to_copy(struct page *page, page 105 include/linux/ksm.h return page; page 108 include/linux/ksm.h static inline int page_referenced_ksm(struct page *page, page 114 include/linux/ksm.h static inline int rmap_walk_ksm(struct page *page, page 120 include/linux/ksm.h static inline void ksm_migrate_page(struct page *newpage, struct page *oldpage) page 106 include/linux/kvm_host.h static inline bool is_error_page(struct page *page) page 108 include/linux/kvm_host.h return IS_ERR(page); page 211 include/linux/kvm_host.h struct page **pagep); page 532 include/linux/kvm_host.h int gfn_to_page_many_atomic(struct kvm *kvm, gfn_t gfn, struct page **pages, page 535 include/linux/kvm_host.h struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn); page 541 include/linux/kvm_host.h void kvm_release_page_clean(struct page *page); page 542 include/linux/kvm_host.h void kvm_release_page_dirty(struct page *page); page 543 include/linux/kvm_host.h void kvm_set_page_accessed(struct page *page); page 29 include/linux/memcontrol.h struct page; page 57 include/linux/memcontrol.h int mem_cgroup_try_charge(struct page *page, struct mm_struct *mm, page 59 include/linux/memcontrol.h void mem_cgroup_commit_charge(struct page *page, struct mem_cgroup *memcg, page 61 include/linux/memcontrol.h void mem_cgroup_cancel_charge(struct page *page, struct mem_cgroup *memcg); page 62 include/linux/memcontrol.h void mem_cgroup_uncharge(struct page *page); page 65 include/linux/memcontrol.h void mem_cgroup_migrate(struct page *oldpage, struct page *newpage, page 69 include/linux/memcontrol.h struct lruvec *mem_cgroup_page_lruvec(struct page *, struct zone *); page 76 include/linux/memcontrol.h extern struct mem_cgroup *try_get_mem_cgroup_from_page(struct page *page); page 142 include/linux/memcontrol.h struct mem_cgroup *mem_cgroup_begin_page_stat(struct page *page, bool *locked, page 174 include/linux/memcontrol.h void mem_cgroup_split_huge_fixup(struct page *head); page 178 include/linux/memcontrol.h bool mem_cgroup_bad_page_check(struct page *page); page 179 include/linux/memcontrol.h void mem_cgroup_print_bad_page(struct page *page); page 184 include/linux/memcontrol.h static inline int mem_cgroup_try_charge(struct page *page, struct mm_struct *mm, page 192 include/linux/memcontrol.h static inline void mem_cgroup_commit_charge(struct page *page, page 198 include/linux/memcontrol.h static inline void mem_cgroup_cancel_charge(struct page *page, page 203 include/linux/memcontrol.h static inline void mem_cgroup_uncharge(struct page *page) page 211 include/linux/memcontrol.h static inline void mem_cgroup_migrate(struct page *oldpage, page 212 include/linux/memcontrol.h struct page *newpage, page 223 include/linux/memcontrol.h static inline struct lruvec *mem_cgroup_page_lruvec(struct page *page, page 229 include/linux/memcontrol.h static inline struct mem_cgroup *try_get_mem_cgroup_from_page(struct page *page) page 293 include/linux/memcontrol.h static inline struct mem_cgroup *mem_cgroup_begin_page_stat(struct page *page, page 340 include/linux/memcontrol.h static inline void mem_cgroup_split_huge_fixup(struct page *head) page 352 include/linux/memcontrol.h mem_cgroup_bad_page_check(struct page *page) page 358 include/linux/memcontrol.h mem_cgroup_print_bad_page(struct page *page) page 413 include/linux/memcontrol.h void __memcg_kmem_commit_charge(struct page *page, page 415 include/linux/memcontrol.h void __memcg_kmem_uncharge_pages(struct page *page, int order); page 474 include/linux/memcontrol.h memcg_kmem_uncharge_pages(struct page *page, int order) page 477 include/linux/memcontrol.h __memcg_kmem_uncharge_pages(page, order); page 492 include/linux/memcontrol.h memcg_kmem_commit_charge(struct page *page, struct mem_cgroup *memcg, int order) page 495 include/linux/memcontrol.h __memcg_kmem_commit_charge(page, memcg, order); page 534 include/linux/memcontrol.h static inline void memcg_kmem_uncharge_pages(struct page *page, int order) page 539 include/linux/memcontrol.h memcg_kmem_commit_charge(struct page *page, struct mem_cgroup *memcg, int order) page 9 include/linux/memory_hotplug.h struct page; page 84 include/linux/memory_hotplug.h extern int add_one_highpage(struct page *page, int pfn, int bad_ppro); page 90 include/linux/memory_hotplug.h typedef void (*online_page_callback_t)(struct page *page); page 95 include/linux/memory_hotplug.h extern void __online_page_set_limits(struct page *page); page 96 include/linux/memory_hotplug.h extern void __online_page_increment_counters(struct page *page); page 97 include/linux/memory_hotplug.h extern void __online_page_free(struct page *page); page 102 include/linux/memory_hotplug.h extern bool is_pageblock_removable_nolock(struct page *page); page 188 include/linux/memory_hotplug.h extern void put_page_bootmem(struct page *page); page 189 include/linux/memory_hotplug.h extern void get_page_bootmem(unsigned long ingo, struct page *page, page 270 include/linux/memory_hotplug.h extern struct page *sparse_decode_mem_map(unsigned long coded_mem_map, page 197 include/linux/mempolicy.h extern int mpol_misplaced(struct page *, struct vm_area_struct *, unsigned long); page 288 include/linux/mempolicy.h static inline int mpol_misplaced(struct page *page, struct vm_area_struct *vma, page 8 include/linux/migrate.h typedef struct page *new_page_t(struct page *page, unsigned long private, page 10 include/linux/migrate.h typedef void free_page_t(struct page *page, unsigned long private); page 33 include/linux/migrate.h struct page *, struct page *, enum migrate_mode); page 42 include/linux/migrate.h extern void migrate_page_copy(struct page *newpage, struct page *page); page 44 include/linux/migrate.h struct page *newpage, struct page *page); page 46 include/linux/migrate.h struct page *newpage, struct page *page, page 67 include/linux/migrate.h static inline void migrate_page_copy(struct page *newpage, page 68 include/linux/migrate.h struct page *page) {} page 71 include/linux/migrate.h struct page *newpage, struct page *page) page 81 include/linux/migrate.h extern int migrate_misplaced_page(struct page *page, page 92 include/linux/migrate.h static inline int migrate_misplaced_page(struct page *page, page 108 include/linux/migrate.h struct page *page, int node); page 114 include/linux/migrate.h struct page *page, int node) page 71 include/linux/mm.h #define nth_page(page,n) pfn_to_page(page_to_pfn((page)) + (n)) page 212 include/linux/mm.h struct page *page; /* ->fault handlers should return a page 282 include/linux/mm.h #define page_private(page) ((page)->private) page 283 include/linux/mm.h #define set_page_private(page, v) ((page)->private = (v)) page 286 include/linux/mm.h static inline void set_freepage_migratetype(struct page *page, int migratetype) page 288 include/linux/mm.h page->index = migratetype; page 292 include/linux/mm.h static inline int get_freepage_migratetype(struct page *page) page 294 include/linux/mm.h return page->index; page 320 include/linux/mm.h static inline int put_page_testzero(struct page *page) page 322 include/linux/mm.h VM_BUG_ON_PAGE(atomic_read(&page->_count) == 0, page); page 323 include/linux/mm.h return atomic_dec_and_test(&page->_count); page 332 include/linux/mm.h static inline int get_page_unless_zero(struct page *page) page 334 include/linux/mm.h return atomic_inc_not_zero(&page->_count); page 344 include/linux/mm.h static inline int put_page_unless_one(struct page *page) page 346 include/linux/mm.h return atomic_add_unless(&page->_count, -1, 1); page 353 include/linux/mm.h struct page *vmalloc_to_page(const void *addr); page 383 include/linux/mm.h static inline void compound_lock(struct page *page) page 386 include/linux/mm.h VM_BUG_ON_PAGE(PageSlab(page), page); page 387 include/linux/mm.h bit_spin_lock(PG_compound_lock, &page->flags); page 391 include/linux/mm.h static inline void compound_unlock(struct page *page) page 394 include/linux/mm.h VM_BUG_ON_PAGE(PageSlab(page), page); page 395 include/linux/mm.h bit_spin_unlock(PG_compound_lock, &page->flags); page 399 include/linux/mm.h static inline unsigned long compound_lock_irqsave(struct page *page) page 404 include/linux/mm.h compound_lock(page); page 409 include/linux/mm.h static inline void compound_unlock_irqrestore(struct page *page, page 413 include/linux/mm.h compound_unlock(page); page 418 include/linux/mm.h static inline struct page *compound_head_by_tail(struct page *tail) page 420 include/linux/mm.h struct page *head = tail->first_page; page 433 include/linux/mm.h static inline struct page *compound_head(struct page *page) page 435 include/linux/mm.h if (unlikely(PageTail(page))) page 436 include/linux/mm.h return compound_head_by_tail(page); page 437 include/linux/mm.h return page; page 445 include/linux/mm.h static inline void page_mapcount_reset(struct page *page) page 447 include/linux/mm.h atomic_set(&(page)->_mapcount, -1); page 450 include/linux/mm.h static inline int page_mapcount(struct page *page) page 452 include/linux/mm.h return atomic_read(&(page)->_mapcount) + 1; page 455 include/linux/mm.h static inline int page_count(struct page *page) page 457 include/linux/mm.h return atomic_read(&compound_head(page)->_count); page 461 include/linux/mm.h extern int PageHeadHuge(struct page *page_head); page 463 include/linux/mm.h static inline int PageHeadHuge(struct page *page_head) page 469 include/linux/mm.h static inline bool __compound_tail_refcounted(struct page *page) page 471 include/linux/mm.h return !PageSlab(page) && !PageHeadHuge(page); page 482 include/linux/mm.h static inline bool compound_tail_refcounted(struct page *page) page 484 include/linux/mm.h VM_BUG_ON_PAGE(!PageHead(page), page); page 485 include/linux/mm.h return __compound_tail_refcounted(page); page 488 include/linux/mm.h static inline void get_huge_page_tail(struct page *page) page 493 include/linux/mm.h VM_BUG_ON_PAGE(!PageTail(page), page); page 494 include/linux/mm.h VM_BUG_ON_PAGE(page_mapcount(page) < 0, page); page 495 include/linux/mm.h VM_BUG_ON_PAGE(atomic_read(&page->_count) != 0, page); page 496 include/linux/mm.h if (compound_tail_refcounted(page->first_page)) page 497 include/linux/mm.h atomic_inc(&page->_mapcount); page 500 include/linux/mm.h extern bool __get_page_tail(struct page *page); page 502 include/linux/mm.h static inline void get_page(struct page *page) page 504 include/linux/mm.h if (unlikely(PageTail(page))) page 505 include/linux/mm.h if (likely(__get_page_tail(page))) page 511 include/linux/mm.h VM_BUG_ON_PAGE(atomic_read(&page->_count) <= 0, page); page 512 include/linux/mm.h atomic_inc(&page->_count); page 515 include/linux/mm.h static inline struct page *virt_to_head_page(const void *x) page 517 include/linux/mm.h struct page *page = virt_to_page(x); page 518 include/linux/mm.h return compound_head(page); page 525 include/linux/mm.h static inline void init_page_count(struct page *page) page 527 include/linux/mm.h atomic_set(&page->_count, 1); page 541 include/linux/mm.h static inline int PageBuddy(struct page *page) page 543 include/linux/mm.h return atomic_read(&page->_mapcount) == PAGE_BUDDY_MAPCOUNT_VALUE; page 546 include/linux/mm.h static inline void __SetPageBuddy(struct page *page) page 548 include/linux/mm.h VM_BUG_ON_PAGE(atomic_read(&page->_mapcount) != -1, page); page 549 include/linux/mm.h atomic_set(&page->_mapcount, PAGE_BUDDY_MAPCOUNT_VALUE); page 552 include/linux/mm.h static inline void __ClearPageBuddy(struct page *page) page 554 include/linux/mm.h VM_BUG_ON_PAGE(!PageBuddy(page), page); page 555 include/linux/mm.h atomic_set(&page->_mapcount, -1); page 560 include/linux/mm.h static inline int PageBalloon(struct page *page) page 562 include/linux/mm.h return atomic_read(&page->_mapcount) == PAGE_BALLOON_MAPCOUNT_VALUE; page 565 include/linux/mm.h static inline void __SetPageBalloon(struct page *page) page 567 include/linux/mm.h VM_BUG_ON_PAGE(atomic_read(&page->_mapcount) != -1, page); page 568 include/linux/mm.h atomic_set(&page->_mapcount, PAGE_BALLOON_MAPCOUNT_VALUE); page 571 include/linux/mm.h static inline void __ClearPageBalloon(struct page *page) page 573 include/linux/mm.h VM_BUG_ON_PAGE(!PageBalloon(page), page); page 574 include/linux/mm.h atomic_set(&page->_mapcount, -1); page 577 include/linux/mm.h void put_page(struct page *page); page 580 include/linux/mm.h void split_page(struct page *page, unsigned int order); page 581 include/linux/mm.h int split_free_page(struct page *page); page 588 include/linux/mm.h typedef void compound_page_dtor(struct page *); page 590 include/linux/mm.h static inline void set_compound_page_dtor(struct page *page, page 593 include/linux/mm.h page[1].lru.next = (void *)dtor; page 596 include/linux/mm.h static inline compound_page_dtor *get_compound_page_dtor(struct page *page) page 598 include/linux/mm.h return (compound_page_dtor *)page[1].lru.next; page 601 include/linux/mm.h static inline int compound_order(struct page *page) page 603 include/linux/mm.h if (!PageHead(page)) page 605 include/linux/mm.h return (unsigned long)page[1].lru.prev; page 608 include/linux/mm.h static inline void set_compound_order(struct page *page, unsigned long order) page 610 include/linux/mm.h page[1].lru.prev = (void *)order; page 628 include/linux/mm.h struct page *page, pte_t *pte, bool write, bool anon); page 735 include/linux/mm.h static inline enum zone_type page_zonenum(const struct page *page) page 737 include/linux/mm.h return (page->flags >> ZONES_PGSHIFT) & ZONES_MASK; page 752 include/linux/mm.h static inline int page_zone_id(struct page *page) page 754 include/linux/mm.h return (page->flags >> ZONEID_PGSHIFT) & ZONEID_MASK; page 767 include/linux/mm.h extern int page_to_nid(const struct page *page); page 769 include/linux/mm.h static inline int page_to_nid(const struct page *page) page 771 include/linux/mm.h return (page->flags >> NODES_PGSHIFT) & NODES_MASK; page 813 include/linux/mm.h static inline int page_cpupid_xchg_last(struct page *page, int cpupid) page 815 include/linux/mm.h return xchg(&page->_last_cpupid, cpupid & LAST_CPUPID_MASK); page 818 include/linux/mm.h static inline int page_cpupid_last(struct page *page) page 820 include/linux/mm.h return page->_last_cpupid; page 822 include/linux/mm.h static inline void page_cpupid_reset_last(struct page *page) page 824 include/linux/mm.h page->_last_cpupid = -1 & LAST_CPUPID_MASK; page 827 include/linux/mm.h static inline int page_cpupid_last(struct page *page) page 829 include/linux/mm.h return (page->flags >> LAST_CPUPID_PGSHIFT) & LAST_CPUPID_MASK; page 832 include/linux/mm.h extern int page_cpupid_xchg_last(struct page *page, int cpupid); page 834 include/linux/mm.h static inline void page_cpupid_reset_last(struct page *page) page 838 include/linux/mm.h page->flags &= ~(LAST_CPUPID_MASK << LAST_CPUPID_PGSHIFT); page 839 include/linux/mm.h page->flags |= (cpupid & LAST_CPUPID_MASK) << LAST_CPUPID_PGSHIFT; page 843 include/linux/mm.h static inline int page_cpupid_xchg_last(struct page *page, int cpupid) page 845 include/linux/mm.h return page_to_nid(page); /* XXX */ page 848 include/linux/mm.h static inline int page_cpupid_last(struct page *page) page 850 include/linux/mm.h return page_to_nid(page); /* XXX */ page 878 include/linux/mm.h static inline void page_cpupid_reset_last(struct page *page) page 888 include/linux/mm.h static inline struct zone *page_zone(const struct page *page) page 890 include/linux/mm.h return &NODE_DATA(page_to_nid(page))->node_zones[page_zonenum(page)]; page 894 include/linux/mm.h static inline void set_page_section(struct page *page, unsigned long section) page 896 include/linux/mm.h page->flags &= ~(SECTIONS_MASK << SECTIONS_PGSHIFT); page 897 include/linux/mm.h page->flags |= (section & SECTIONS_MASK) << SECTIONS_PGSHIFT; page 900 include/linux/mm.h static inline unsigned long page_to_section(const struct page *page) page 902 include/linux/mm.h return (page->flags >> SECTIONS_PGSHIFT) & SECTIONS_MASK; page 906 include/linux/mm.h static inline void set_page_zone(struct page *page, enum zone_type zone) page 908 include/linux/mm.h page->flags &= ~(ZONES_MASK << ZONES_PGSHIFT); page 909 include/linux/mm.h page->flags |= (zone & ZONES_MASK) << ZONES_PGSHIFT; page 912 include/linux/mm.h static inline void set_page_node(struct page *page, unsigned long node) page 914 include/linux/mm.h page->flags &= ~(NODES_MASK << NODES_PGSHIFT); page 915 include/linux/mm.h page->flags |= (node & NODES_MASK) << NODES_PGSHIFT; page 918 include/linux/mm.h static inline void set_page_links(struct page *page, enum zone_type zone, page 921 include/linux/mm.h set_page_zone(page, zone); page 922 include/linux/mm.h set_page_node(page, node); page 924 include/linux/mm.h set_page_section(page, pfn_to_section_nr(pfn)); page 933 include/linux/mm.h static __always_inline void *lowmem_page_address(const struct page *page) page 935 include/linux/mm.h return __va(PFN_PHYS(page_to_pfn(page))); page 943 include/linux/mm.h static inline void *page_address(const struct page *page) page 945 include/linux/mm.h return page->virtual; page 947 include/linux/mm.h static inline void set_page_address(struct page *page, void *address) page 949 include/linux/mm.h page->virtual = address; page 955 include/linux/mm.h void *page_address(const struct page *page); page 956 include/linux/mm.h void set_page_address(struct page *page, void *virtual); page 961 include/linux/mm.h #define page_address(page) lowmem_page_address(page) page 962 include/linux/mm.h #define set_page_address(page, address) do { } while(0) page 986 include/linux/mm.h extern struct address_space *page_mapping(struct page *page); page 989 include/linux/mm.h static inline void *page_rmapping(struct page *page) page 991 include/linux/mm.h return (void *)((unsigned long)page->mapping & ~PAGE_MAPPING_FLAGS); page 994 include/linux/mm.h extern struct address_space *__page_file_mapping(struct page *); page 997 include/linux/mm.h struct address_space *page_file_mapping(struct page *page) page 999 include/linux/mm.h if (unlikely(PageSwapCache(page))) page 1000 include/linux/mm.h return __page_file_mapping(page); page 1002 include/linux/mm.h return page->mapping; page 1005 include/linux/mm.h static inline int PageAnon(struct page *page) page 1007 include/linux/mm.h return ((unsigned long)page->mapping & PAGE_MAPPING_ANON) != 0; page 1014 include/linux/mm.h static inline pgoff_t page_index(struct page *page) page 1016 include/linux/mm.h if (unlikely(PageSwapCache(page))) page 1017 include/linux/mm.h return page_private(page); page 1018 include/linux/mm.h return page->index; page 1021 include/linux/mm.h extern pgoff_t __page_file_index(struct page *page); page 1027 include/linux/mm.h static inline pgoff_t page_file_index(struct page *page) page 1029 include/linux/mm.h if (unlikely(PageSwapCache(page))) page 1030 include/linux/mm.h return __page_file_index(page); page 1032 include/linux/mm.h return page->index; page 1038 include/linux/mm.h static inline int page_mapped(struct page *page) page 1040 include/linux/mm.h return atomic_read(&(page)->_mapcount) >= 0; page 1114 include/linux/mm.h struct page *vm_normal_page(struct vm_area_struct *vma, unsigned long addr, page 1183 include/linux/mm.h int truncate_inode_page(struct address_space *mapping, struct page *page); page 1184 include/linux/mm.h int generic_error_remove_page(struct address_space *mapping, struct page *page); page 1185 include/linux/mm.h int invalidate_inode_page(struct page *page); page 1217 include/linux/mm.h unsigned int foll_flags, struct page **pages, page 1221 include/linux/mm.h int write, int force, struct page **pages, page 1224 include/linux/mm.h struct page **pages); page 1227 include/linux/mm.h struct page **pages); page 1228 include/linux/mm.h int get_kernel_page(unsigned long start, int write, struct page **pages); page 1229 include/linux/mm.h struct page *get_dump_page(unsigned long addr); page 1231 include/linux/mm.h extern int try_to_release_page(struct page * page, gfp_t gfp_mask); page 1232 include/linux/mm.h extern void do_invalidatepage(struct page *page, unsigned int offset, page 1235 include/linux/mm.h int __set_page_dirty_nobuffers(struct page *page); page 1236 include/linux/mm.h int __set_page_dirty_no_writeback(struct page *page); page 1238 include/linux/mm.h struct page *page); page 1239 include/linux/mm.h void account_page_dirtied(struct page *page, struct address_space *mapping); page 1240 include/linux/mm.h int set_page_dirty(struct page *page); page 1241 include/linux/mm.h int set_page_dirty_lock(struct page *page); page 1242 include/linux/mm.h int clear_page_dirty_for_io(struct page *page); page 1291 include/linux/mm.h struct page **pages); page 1429 include/linux/mm.h extern bool ptlock_alloc(struct page *page); page 1430 include/linux/mm.h extern void ptlock_free(struct page *page); page 1432 include/linux/mm.h static inline spinlock_t *ptlock_ptr(struct page *page) page 1434 include/linux/mm.h return page->ptl; page 1441 include/linux/mm.h static inline bool ptlock_alloc(struct page *page) page 1446 include/linux/mm.h static inline void ptlock_free(struct page *page) page 1450 include/linux/mm.h static inline spinlock_t *ptlock_ptr(struct page *page) page 1452 include/linux/mm.h return &page->ptl; page 1461 include/linux/mm.h static inline bool ptlock_init(struct page *page) page 1471 include/linux/mm.h VM_BUG_ON_PAGE(*(unsigned long *)&page->ptl, page); page 1472 include/linux/mm.h if (!ptlock_alloc(page)) page 1474 include/linux/mm.h spin_lock_init(ptlock_ptr(page)); page 1479 include/linux/mm.h static inline void pte_lock_deinit(struct page *page) page 1481 include/linux/mm.h page->mapping = NULL; page 1482 include/linux/mm.h ptlock_free(page); page 1494 include/linux/mm.h static inline bool ptlock_init(struct page *page) { return true; } page 1495 include/linux/mm.h static inline void pte_lock_deinit(struct page *page) {} page 1504 include/linux/mm.h static inline bool pgtable_page_ctor(struct page *page) page 1506 include/linux/mm.h inc_zone_page_state(page, NR_PAGETABLE); page 1507 include/linux/mm.h return ptlock_init(page); page 1510 include/linux/mm.h static inline void pgtable_page_dtor(struct page *page) page 1512 include/linux/mm.h pte_lock_deinit(page); page 1513 include/linux/mm.h dec_zone_page_state(page, NR_PAGETABLE); page 1546 include/linux/mm.h static struct page *pmd_to_page(pmd_t *pmd) page 1557 include/linux/mm.h static inline bool pgtable_pmd_page_ctor(struct page *page) page 1560 include/linux/mm.h page->pmd_huge_pte = NULL; page 1562 include/linux/mm.h return ptlock_init(page); page 1565 include/linux/mm.h static inline void pgtable_pmd_page_dtor(struct page *page) page 1568 include/linux/mm.h VM_BUG_ON_PAGE(page->pmd_huge_pte, page); page 1570 include/linux/mm.h ptlock_free(page); page 1582 include/linux/mm.h static inline bool pgtable_pmd_page_ctor(struct page *page) { return true; } page 1583 include/linux/mm.h static inline void pgtable_pmd_page_dtor(struct page *page) {} page 1615 include/linux/mm.h extern void free_highmem_page(struct page *page); page 1618 include/linux/mm.h extern void adjust_managed_page_count(struct page *page, long count); page 1622 include/linux/mm.h static inline void __free_reserved_page(struct page *page) page 1624 include/linux/mm.h ClearPageReserved(page); page 1625 include/linux/mm.h init_page_count(page); page 1626 include/linux/mm.h __free_page(page); page 1629 include/linux/mm.h static inline void free_reserved_page(struct page *page) page 1631 include/linux/mm.h __free_reserved_page(page); page 1632 include/linux/mm.h adjust_managed_page_count(page, 1); page 1635 include/linux/mm.h static inline void mark_page_reserved(struct page *page) page 1637 include/linux/mm.h SetPageReserved(page); page 1638 include/linux/mm.h adjust_managed_page_count(page, -1); page 1834 include/linux/mm.h unsigned long flags, struct page **pages); page 1907 include/linux/mm.h int write_one_page(struct page *page, int wait); page 1926 include/linux/mm.h struct page *pg, page 1999 include/linux/mm.h int vm_insert_page(struct vm_area_struct *, unsigned long addr, struct page *); page 2007 include/linux/mm.h struct page *follow_page_mask(struct vm_area_struct *vma, page 2011 include/linux/mm.h static inline struct page *follow_page(struct vm_area_struct *vma, page 2048 include/linux/mm.h extern void kernel_map_pages(struct page *page, int numpages, int enable); page 2050 include/linux/mm.h extern bool kernel_page_present(struct page *page); page 2054 include/linux/mm.h kernel_map_pages(struct page *page, int numpages, int enable) {} page 2056 include/linux/mm.h static inline bool kernel_page_present(struct page *page) { return true; } page 2095 include/linux/mm.h void sparse_mem_maps_populate_node(struct page **map_map, page 2101 include/linux/mm.h struct page *sparse_mem_map_populate(unsigned long pnum, int nid); page 2116 include/linux/mm.h void register_page_bootmem_memmap(unsigned long section_nr, struct page *map, page 2130 include/linux/mm.h extern void shake_page(struct page *p, int access); page 2132 include/linux/mm.h extern int soft_offline_page(struct page *page, int flags); page 2135 include/linux/mm.h extern void clear_huge_page(struct page *page, page 2138 include/linux/mm.h extern void copy_user_huge_page(struct page *dst, struct page *src, page 2151 include/linux/mm.h static inline bool page_is_guard(struct page *page) page 2153 include/linux/mm.h return test_bit(PAGE_DEBUG_FLAG_GUARD, &page->debug_flags); page 2157 include/linux/mm.h static inline bool page_is_guard(struct page *page) { return false; } page 20 include/linux/mm_inline.h static inline int page_is_file_cache(struct page *page) page 22 include/linux/mm_inline.h return !PageSwapBacked(page); page 25 include/linux/mm_inline.h static __always_inline void add_page_to_lru_list(struct page *page, page 28 include/linux/mm_inline.h int nr_pages = hpage_nr_pages(page); page 30 include/linux/mm_inline.h list_add(&page->lru, &lruvec->lists[lru]); page 34 include/linux/mm_inline.h static __always_inline void del_page_from_lru_list(struct page *page, page 37 include/linux/mm_inline.h int nr_pages = hpage_nr_pages(page); page 39 include/linux/mm_inline.h list_del(&page->lru); page 51 include/linux/mm_inline.h static inline enum lru_list page_lru_base_type(struct page *page) page 53 include/linux/mm_inline.h if (page_is_file_cache(page)) page 65 include/linux/mm_inline.h static __always_inline enum lru_list page_off_lru(struct page *page) page 69 include/linux/mm_inline.h if (PageUnevictable(page)) { page 70 include/linux/mm_inline.h __ClearPageUnevictable(page); page 73 include/linux/mm_inline.h lru = page_lru_base_type(page); page 74 include/linux/mm_inline.h if (PageActive(page)) { page 75 include/linux/mm_inline.h __ClearPageActive(page); page 89 include/linux/mm_inline.h static __always_inline enum lru_list page_lru(struct page *page) page 93 include/linux/mm_inline.h if (PageUnevictable(page)) page 96 include/linux/mm_inline.h lru = page_lru_base_type(page); page 97 include/linux/mm_inline.h if (PageActive(page)) page 131 include/linux/mm_types.h struct page *next; /* Next partial slab */ page 167 include/linux/mm_types.h struct page *first_page; /* Compound tail pages */ page 210 include/linux/mm_types.h struct page *page; page 517 include/linux/mm_types.h struct page **pages; page 6 include/linux/mmdebug.h struct page; page 10 include/linux/mmdebug.h extern void dump_page(struct page *page, const char *reason); page 11 include/linux/mmdebug.h extern void dump_page_badflags(struct page *page, const char *reason, page 18 include/linux/mmdebug.h #define VM_BUG_ON_PAGE(cond, page) \ page 21 include/linux/mmdebug.h dump_page(page, "VM_BUG_ON_PAGE(" __stringify(cond)")");\ page 44 include/linux/mmdebug.h #define VM_BUG_ON_PAGE(cond, page) VM_BUG_ON(cond) page 81 include/linux/mmzone.h #define get_pageblock_migratetype(page) \ page 82 include/linux/mmzone.h get_pfnblock_flags_mask(page, page_to_pfn(page), \ page 85 include/linux/mmzone.h static inline int get_pfnblock_migratetype(struct page *page, unsigned long pfn) page 88 include/linux/mmzone.h return get_pfnblock_flags_mask(page, pfn, PB_migrate_end, page 704 include/linux/mmzone.h extern struct page *mem_map; page 724 include/linux/mmzone.h struct page *node_mem_map; page 1080 include/linux/mmzone.h struct page; page 1149 include/linux/mmzone.h static inline struct page *__section_mem_map_addr(struct mem_section *section) page 1153 include/linux/mmzone.h return (struct page *)map; page 1261 include/linux/mmzone.h struct page *page, struct zone *zone); page 1264 include/linux/mmzone.h struct page *page, struct zone *zone) page 18 include/linux/mpage.h int mpage_readpage(struct page *page, get_block_t get_block); page 21 include/linux/mpage.h int mpage_writepage(struct page *page, get_block_t *get_block, page 490 include/linux/mtd/nand.h uint8_t *buf, int oob_required, int page); page 494 include/linux/mtd/nand.h uint8_t *buf, int oob_required, int page); page 496 include/linux/mtd/nand.h uint32_t offs, uint32_t len, uint8_t *buf, int page); page 503 include/linux/mtd/nand.h int page); page 505 include/linux/mtd/nand.h int page); page 506 include/linux/mtd/nand.h int (*read_oob)(struct mtd_info *mtd, struct nand_chip *chip, int page); page 508 include/linux/mtd/nand.h int page); page 649 include/linux/mtd/nand.h int (*erase)(struct mtd_info *mtd, int page); page 652 include/linux/mtd/nand.h int status, int page); page 655 include/linux/mtd/nand.h int oob_required, int page, int cached, int raw); page 122 include/linux/net.h struct page; page 180 include/linux/net.h ssize_t (*sendpage) (struct socket *sock, struct page *page, page 283 include/linux/net.h int kernel_sendpage(struct socket *sock, struct page *page, int offset, page 504 include/linux/nfs_fs.h extern int nfs_writepage(struct page *page, struct writeback_control *wbc); page 506 include/linux/nfs_fs.h extern int nfs_flush_incompatible(struct file *file, struct page *page); page 507 include/linux/nfs_fs.h extern int nfs_updatepage(struct file *, struct page *, unsigned int, unsigned int); page 514 include/linux/nfs_fs.h extern int nfs_wb_page(struct inode *inode, struct page* page); page 515 include/linux/nfs_fs.h extern int nfs_wb_page_cancel(struct inode *inode, struct page* page); page 529 include/linux/nfs_fs.h extern int nfs_readpage(struct file *, struct page *); page 533 include/linux/nfs_fs.h struct page *); page 41 include/linux/nfs_page.h struct page *wb_page; /* page to read in/write out */ page 99 include/linux/nfs_page.h struct page *page, page 215 include/linux/nfs_xdr.h struct page **pages; page 272 include/linux/nfs_xdr.h struct page *layoutupdate_page; page 273 include/linux/nfs_xdr.h struct page **layoutupdate_pages; page 507 include/linux/nfs_xdr.h struct page ** pages; page 636 include/linux/nfs_xdr.h struct page ** acl_pages; page 648 include/linux/nfs_xdr.h struct page ** acl_pages; page 658 include/linux/nfs_xdr.h struct page * acl_scratch; page 679 include/linux/nfs_xdr.h struct page ** pages; page 688 include/linux/nfs_xdr.h struct page ** pages; page 694 include/linux/nfs_xdr.h struct page ** pages; page 704 include/linux/nfs_xdr.h struct page ** pages; page 716 include/linux/nfs_xdr.h struct page ** pages; page 757 include/linux/nfs_xdr.h struct page ** pages; page 784 include/linux/nfs_xdr.h struct page ** pages; page 802 include/linux/nfs_xdr.h struct page ** pages; page 849 include/linux/nfs_xdr.h struct page ** pages; page 953 include/linux/nfs_xdr.h struct page ** pages; /* zero-copy data */ page 970 include/linux/nfs_xdr.h struct page ** pages; /* zero-copy data */ page 1048 include/linux/nfs_xdr.h struct page *page; page 1269 include/linux/nfs_xdr.h struct page **pagevec; page 1271 include/linux/nfs_xdr.h struct page *page_array[NFS_PAGEVEC_SIZE]; page 1421 include/linux/nfs_xdr.h int (*readlink)(struct inode *, struct page *, unsigned int, page 1433 include/linux/nfs_xdr.h int (*symlink) (struct inode *, struct dentry *, struct page *, page 1438 include/linux/nfs_xdr.h u64, struct page **, unsigned int, int); page 137 include/linux/page-flags.h static inline int Page##uname(const struct page *page) \ page 138 include/linux/page-flags.h { return test_bit(PG_##lname, &page->flags); } page 141 include/linux/page-flags.h static inline void SetPage##uname(struct page *page) \ page 142 include/linux/page-flags.h { set_bit(PG_##lname, &page->flags); } page 145 include/linux/page-flags.h static inline void ClearPage##uname(struct page *page) \ page 146 include/linux/page-flags.h { clear_bit(PG_##lname, &page->flags); } page 149 include/linux/page-flags.h static inline void __SetPage##uname(struct page *page) \ page 150 include/linux/page-flags.h { __set_bit(PG_##lname, &page->flags); } page 153 include/linux/page-flags.h static inline void __ClearPage##uname(struct page *page) \ page 154 include/linux/page-flags.h { __clear_bit(PG_##lname, &page->flags); } page 157 include/linux/page-flags.h static inline int TestSetPage##uname(struct page *page) \ page 158 include/linux/page-flags.h { return test_and_set_bit(PG_##lname, &page->flags); } page 161 include/linux/page-flags.h static inline int TestClearPage##uname(struct page *page) \ page 162 include/linux/page-flags.h { return test_and_clear_bit(PG_##lname, &page->flags); } page 165 include/linux/page-flags.h static inline int __TestClearPage##uname(struct page *page) \ page 166 include/linux/page-flags.h { return __test_and_clear_bit(PG_##lname, &page->flags); } page 178 include/linux/page-flags.h static inline int Page##uname(const struct page *page) { return 0; } page 181 include/linux/page-flags.h static inline void SetPage##uname(struct page *page) { } page 184 include/linux/page-flags.h static inline void ClearPage##uname(struct page *page) { } page 187 include/linux/page-flags.h static inline void __ClearPage##uname(struct page *page) { } page 190 include/linux/page-flags.h static inline int TestSetPage##uname(struct page *page) { return 0; } page 193 include/linux/page-flags.h static inline int TestClearPage##uname(struct page *page) { return 0; } page 196 include/linux/page-flags.h static inline int __TestClearPage##uname(struct page *page) { return 0; } page 204 include/linux/page-flags.h struct page; /* forward declaration */ page 287 include/linux/page-flags.h u64 stable_page_flags(struct page *page); page 289 include/linux/page-flags.h static inline int PageUptodate(struct page *page) page 291 include/linux/page-flags.h int ret = test_bit(PG_uptodate, &(page)->flags); page 307 include/linux/page-flags.h static inline void __SetPageUptodate(struct page *page) page 310 include/linux/page-flags.h __set_bit(PG_uptodate, &(page)->flags); page 313 include/linux/page-flags.h static inline void SetPageUptodate(struct page *page) page 321 include/linux/page-flags.h set_bit(PG_uptodate, &(page)->flags); page 326 include/linux/page-flags.h extern void cancel_dirty_page(struct page *page, unsigned int account_size); page 328 include/linux/page-flags.h int test_clear_page_writeback(struct page *page); page 329 include/linux/page-flags.h int __test_set_page_writeback(struct page *page, bool keep_write); page 331 include/linux/page-flags.h #define test_set_page_writeback(page) \ page 332 include/linux/page-flags.h __test_set_page_writeback(page, false) page 333 include/linux/page-flags.h #define test_set_page_writeback_keepwrite(page) \ page 334 include/linux/page-flags.h __test_set_page_writeback(page, true) page 336 include/linux/page-flags.h static inline void set_page_writeback(struct page *page) page 338 include/linux/page-flags.h test_set_page_writeback(page); page 341 include/linux/page-flags.h static inline void set_page_writeback_keepwrite(struct page *page) page 343 include/linux/page-flags.h test_set_page_writeback_keepwrite(page); page 358 include/linux/page-flags.h static inline int PageCompound(struct page *page) page 360 include/linux/page-flags.h return page->flags & ((1L << PG_head) | (1L << PG_tail)); page 364 include/linux/page-flags.h static inline void ClearPageCompound(struct page *page) page 366 include/linux/page-flags.h BUG_ON(!PageHead(page)); page 367 include/linux/page-flags.h ClearPageHead(page); page 396 include/linux/page-flags.h static inline int PageHead(struct page *page) page 398 include/linux/page-flags.h return ((page->flags & PG_head_tail_mask) == PG_head_mask); page 401 include/linux/page-flags.h static inline int PageTail(struct page *page) page 403 include/linux/page-flags.h return ((page->flags & PG_head_tail_mask) == PG_head_tail_mask); page 406 include/linux/page-flags.h static inline void __SetPageTail(struct page *page) page 408 include/linux/page-flags.h page->flags |= PG_head_tail_mask; page 411 include/linux/page-flags.h static inline void __ClearPageTail(struct page *page) page 413 include/linux/page-flags.h page->flags &= ~PG_head_tail_mask; page 417 include/linux/page-flags.h static inline void ClearPageCompound(struct page *page) page 419 include/linux/page-flags.h BUG_ON((page->flags & PG_head_tail_mask) != (1 << PG_compound)); page 420 include/linux/page-flags.h clear_bit(PG_compound, &page->flags); page 435 include/linux/page-flags.h static inline int PageTransHuge(struct page *page) page 437 include/linux/page-flags.h VM_BUG_ON_PAGE(PageTail(page), page); page 438 include/linux/page-flags.h return PageHead(page); page 446 include/linux/page-flags.h static inline int PageTransCompound(struct page *page) page 448 include/linux/page-flags.h return PageCompound(page); page 456 include/linux/page-flags.h static inline int PageTransTail(struct page *page) page 458 include/linux/page-flags.h return PageTail(page); page 463 include/linux/page-flags.h static inline int PageTransHuge(struct page *page) page 468 include/linux/page-flags.h static inline int PageTransCompound(struct page *page) page 473 include/linux/page-flags.h static inline int PageTransTail(struct page *page) page 483 include/linux/page-flags.h static inline int PageSlabPfmemalloc(struct page *page) page 485 include/linux/page-flags.h VM_BUG_ON_PAGE(!PageSlab(page), page); page 486 include/linux/page-flags.h return PageActive(page); page 489 include/linux/page-flags.h static inline void SetPageSlabPfmemalloc(struct page *page) page 491 include/linux/page-flags.h VM_BUG_ON_PAGE(!PageSlab(page), page); page 492 include/linux/page-flags.h SetPageActive(page); page 495 include/linux/page-flags.h static inline void __ClearPageSlabPfmemalloc(struct page *page) page 497 include/linux/page-flags.h VM_BUG_ON_PAGE(!PageSlab(page), page); page 498 include/linux/page-flags.h __ClearPageActive(page); page 501 include/linux/page-flags.h static inline void ClearPageSlabPfmemalloc(struct page *page) page 503 include/linux/page-flags.h VM_BUG_ON_PAGE(!PageSlab(page), page); page 504 include/linux/page-flags.h ClearPageActive(page); page 547 include/linux/page-flags.h static inline int page_has_private(struct page *page) page 549 include/linux/page-flags.h return !!(page->flags & PAGE_FLAGS_PRIVATE); page 9 include/linux/page-isolation.h static inline bool is_migrate_isolate_page(struct page *page) page 11 include/linux/page-isolation.h return get_pageblock_migratetype(page) == MIGRATE_ISOLATE; page 22 include/linux/page-isolation.h static inline bool is_migrate_isolate_page(struct page *page) page 32 include/linux/page-isolation.h bool has_unmovable_pages(struct zone *zone, struct page *page, int count, page 34 include/linux/page-isolation.h void set_pageblock_migratetype(struct page *page, int migratetype); page 35 include/linux/page-isolation.h int move_freepages_block(struct zone *zone, struct page *page, page 38 include/linux/page-isolation.h struct page *start_page, struct page *end_page, page 71 include/linux/page-isolation.h int set_migratetype_isolate(struct page *page, bool skip_hwpoisoned_pages); page 72 include/linux/page-isolation.h void unset_migratetype_isolate(struct page *page, unsigned migratetype); page 73 include/linux/page-isolation.h struct page *alloc_migrate_target(struct page *page, unsigned long private, page 42 include/linux/page_cgroup.h struct page_cgroup *lookup_page_cgroup(struct page *page); page 55 include/linux/page_cgroup.h static inline struct page_cgroup *lookup_page_cgroup(struct page *page) page 66 include/linux/pageblock-flags.h struct page; page 68 include/linux/pageblock-flags.h unsigned long get_pfnblock_flags_mask(struct page *page, page 73 include/linux/pageblock-flags.h void set_pfnblock_flags_mask(struct page *page, page 80 include/linux/pageblock-flags.h #define get_pageblock_flags_group(page, start_bitidx, end_bitidx) \ page 81 include/linux/pageblock-flags.h get_pfnblock_flags_mask(page, page_to_pfn(page), \ page 84 include/linux/pageblock-flags.h #define set_pageblock_flags_group(page, flags, start_bitidx, end_bitidx) \ page 85 include/linux/pageblock-flags.h set_pfnblock_flags_mask(page, flags, page_to_pfn(page), \ page 90 include/linux/pageblock-flags.h #define get_pageblock_skip(page) \ page 91 include/linux/pageblock-flags.h get_pageblock_flags_group(page, PB_migrate_skip, \ page 93 include/linux/pageblock-flags.h #define clear_pageblock_skip(page) \ page 94 include/linux/pageblock-flags.h set_pageblock_flags_group(page, 0, PB_migrate_skip, \ page 96 include/linux/pageblock-flags.h #define set_pageblock_skip(page) \ page 97 include/linux/pageblock-flags.h set_pageblock_flags_group(page, 1, PB_migrate_skip, \ page 95 include/linux/pagemap.h #define page_cache_get(page) get_page(page) page 96 include/linux/pagemap.h #define page_cache_release(page) put_page(page) page 97 include/linux/pagemap.h void release_pages(struct page **pages, int nr, bool cold); page 143 include/linux/pagemap.h static inline int page_cache_get_speculative(struct page *page) page 160 include/linux/pagemap.h VM_BUG_ON_PAGE(page_count(page) == 0, page); page 161 include/linux/pagemap.h atomic_inc(&page->_count); page 164 include/linux/pagemap.h if (unlikely(!get_page_unless_zero(page))) { page 173 include/linux/pagemap.h VM_BUG_ON_PAGE(PageTail(page), page); page 181 include/linux/pagemap.h static inline int page_cache_add_speculative(struct page *page, int count) page 189 include/linux/pagemap.h VM_BUG_ON_PAGE(page_count(page) == 0, page); page 190 include/linux/pagemap.h atomic_add(count, &page->_count); page 193 include/linux/pagemap.h if (unlikely(!atomic_add_unless(&page->_count, count, 0))) page 196 include/linux/pagemap.h VM_BUG_ON_PAGE(PageCompound(page) && page != compound_head(page), page); page 201 include/linux/pagemap.h static inline int page_freeze_refs(struct page *page, int count) page 203 include/linux/pagemap.h return likely(atomic_cmpxchg(&page->_count, count, 0) == count); page 206 include/linux/pagemap.h static inline void page_unfreeze_refs(struct page *page, int count) page 208 include/linux/pagemap.h VM_BUG_ON_PAGE(page_count(page) != 0, page); page 211 include/linux/pagemap.h atomic_set(&page->_count, count); page 215 include/linux/pagemap.h extern struct page *__page_cache_alloc(gfp_t gfp); page 217 include/linux/pagemap.h static inline struct page *__page_cache_alloc(gfp_t gfp) page 223 include/linux/pagemap.h static inline struct page *page_cache_alloc(struct address_space *x) page 228 include/linux/pagemap.h static inline struct page *page_cache_alloc_cold(struct address_space *x) page 233 include/linux/pagemap.h static inline struct page *page_cache_alloc_readahead(struct address_space *x) page 239 include/linux/pagemap.h typedef int filler_t(void *, struct page *); page 253 include/linux/pagemap.h struct page *pagecache_get_page(struct address_space *mapping, pgoff_t offset, page 266 include/linux/pagemap.h static inline struct page *find_get_page(struct address_space *mapping, page 272 include/linux/pagemap.h static inline struct page *find_get_page_flags(struct address_space *mapping, page 292 include/linux/pagemap.h static inline struct page *find_lock_page(struct address_space *mapping, page 317 include/linux/pagemap.h static inline struct page *find_or_create_page(struct address_space *mapping, page 338 include/linux/pagemap.h static inline struct page *grab_cache_page_nowait(struct address_space *mapping, page 346 include/linux/pagemap.h struct page *find_get_entry(struct address_space *mapping, pgoff_t offset); page 347 include/linux/pagemap.h struct page *find_lock_entry(struct address_space *mapping, pgoff_t offset); page 349 include/linux/pagemap.h unsigned int nr_entries, struct page **entries, page 352 include/linux/pagemap.h unsigned int nr_pages, struct page **pages); page 354 include/linux/pagemap.h unsigned int nr_pages, struct page **pages); page 356 include/linux/pagemap.h int tag, unsigned int nr_pages, struct page **pages); page 358 include/linux/pagemap.h struct page *grab_cache_page_write_begin(struct address_space *mapping, page 364 include/linux/pagemap.h static inline struct page *grab_cache_page(struct address_space *mapping, page 370 include/linux/pagemap.h extern struct page * read_cache_page(struct address_space *mapping, page 372 include/linux/pagemap.h extern struct page * read_cache_page_gfp(struct address_space *mapping, page 377 include/linux/pagemap.h static inline struct page *read_mapping_page(struct address_space *mapping, page 388 include/linux/pagemap.h static inline pgoff_t page_to_pgoff(struct page *page) page 390 include/linux/pagemap.h if (unlikely(PageHeadHuge(page))) page 391 include/linux/pagemap.h return page->index << compound_order(page); page 393 include/linux/pagemap.h return page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT); page 399 include/linux/pagemap.h static inline loff_t page_offset(struct page *page) page 401 include/linux/pagemap.h return ((loff_t)page->index) << PAGE_CACHE_SHIFT; page 404 include/linux/pagemap.h static inline loff_t page_file_offset(struct page *page) page 406 include/linux/pagemap.h return ((loff_t)page_file_index(page)) << PAGE_CACHE_SHIFT; page 423 include/linux/pagemap.h extern void __lock_page(struct page *page); page 424 include/linux/pagemap.h extern int __lock_page_killable(struct page *page); page 425 include/linux/pagemap.h extern int __lock_page_or_retry(struct page *page, struct mm_struct *mm, page 427 include/linux/pagemap.h extern void unlock_page(struct page *page); page 429 include/linux/pagemap.h static inline void __set_page_locked(struct page *page) page 431 include/linux/pagemap.h __set_bit(PG_locked, &page->flags); page 434 include/linux/pagemap.h static inline void __clear_page_locked(struct page *page) page 436 include/linux/pagemap.h __clear_bit(PG_locked, &page->flags); page 439 include/linux/pagemap.h static inline int trylock_page(struct page *page) page 441 include/linux/pagemap.h return (likely(!test_and_set_bit_lock(PG_locked, &page->flags))); page 447 include/linux/pagemap.h static inline void lock_page(struct page *page) page 450 include/linux/pagemap.h if (!trylock_page(page)) page 451 include/linux/pagemap.h __lock_page(page); page 459 include/linux/pagemap.h static inline int lock_page_killable(struct page *page) page 462 include/linux/pagemap.h if (!trylock_page(page)) page 463 include/linux/pagemap.h return __lock_page_killable(page); page 474 include/linux/pagemap.h static inline int lock_page_or_retry(struct page *page, struct mm_struct *mm, page 478 include/linux/pagemap.h return trylock_page(page) || __lock_page_or_retry(page, mm, flags); page 485 include/linux/pagemap.h extern void wait_on_page_bit(struct page *page, int bit_nr); page 487 include/linux/pagemap.h extern int wait_on_page_bit_killable(struct page *page, int bit_nr); page 488 include/linux/pagemap.h extern int wait_on_page_bit_killable_timeout(struct page *page, page 491 include/linux/pagemap.h static inline int wait_on_page_locked_killable(struct page *page) page 493 include/linux/pagemap.h if (PageLocked(page)) page 494 include/linux/pagemap.h return wait_on_page_bit_killable(page, PG_locked); page 498 include/linux/pagemap.h extern wait_queue_head_t *page_waitqueue(struct page *page); page 499 include/linux/pagemap.h static inline void wake_up_page(struct page *page, int bit) page 501 include/linux/pagemap.h __wake_up_bit(page_waitqueue(page), &page->flags, bit); page 511 include/linux/pagemap.h static inline void wait_on_page_locked(struct page *page) page 513 include/linux/pagemap.h if (PageLocked(page)) page 514 include/linux/pagemap.h wait_on_page_bit(page, PG_locked); page 520 include/linux/pagemap.h static inline void wait_on_page_writeback(struct page *page) page 522 include/linux/pagemap.h if (PageWriteback(page)) page 523 include/linux/pagemap.h wait_on_page_bit(page, PG_writeback); page 526 include/linux/pagemap.h extern void end_page_writeback(struct page *page); page 527 include/linux/pagemap.h void wait_for_stable_page(struct page *page); page 529 include/linux/pagemap.h void page_endio(struct page *page, int rw, int err); page 534 include/linux/pagemap.h extern void add_page_wait_queue(struct page *page, wait_queue_t *waiter); page 649 include/linux/pagemap.h int add_to_page_cache_locked(struct page *page, struct address_space *mapping, page 651 include/linux/pagemap.h int add_to_page_cache_lru(struct page *page, struct address_space *mapping, page 653 include/linux/pagemap.h extern void delete_from_page_cache(struct page *page); page 654 include/linux/pagemap.h extern void __delete_from_page_cache(struct page *page, void *shadow); page 655 include/linux/pagemap.h int replace_page_cache_page(struct page *old, struct page *new, gfp_t gfp_mask); page 661 include/linux/pagemap.h static inline int add_to_page_cache(struct page *page, page 666 include/linux/pagemap.h __set_page_locked(page); page 667 include/linux/pagemap.h error = add_to_page_cache_locked(page, mapping, offset, gfp_mask); page 669 include/linux/pagemap.h __clear_page_locked(page); page 14 include/linux/pagevec.h struct page; page 20 include/linux/pagevec.h struct page *pages[PAGEVEC_SIZE]; page 60 include/linux/pagevec.h static inline unsigned pagevec_add(struct pagevec *pvec, struct page *page) page 62 include/linux/pagevec.h pvec->pages[pvec->nr++] = page; page 544 include/linux/perf_event.h int page; page 907 include/linux/perf_event.h char *page) \ page 910 include/linux/perf_event.h return sprintf(page, _format "\n"); \ page 14 include/linux/pid_namespace.h void *page; page 21 include/linux/pipe_fs_i.h struct page *page; page 56 include/linux/pipe_fs_i.h struct page *tmp_page; page 133 include/linux/pktcdvd.h struct page *pages[PACKET_MAX_SIZE / FRAMES_PER_PAGE]; page 17 include/linux/quicklist.h void *page; page 39 include/linux/quicklist.h p = q->page; page 41 include/linux/quicklist.h q->page = p[0]; page 56 include/linux/quicklist.h struct page *page) page 61 include/linux/quicklist.h *(void **)p = q->page; page 62 include/linux/quicklist.h q->page = p; page 73 include/linux/quicklist.h struct page *page) page 75 include/linux/quicklist.h __quicklist_free(nr, dtor, page_address(page), page); page 43 include/linux/relay.h struct page **page_array; /* array of current buffer pages */ page 184 include/linux/ring_buffer.h size_t ring_buffer_page_len(void *page); page 98 include/linux/rmap.h static inline struct anon_vma *page_anon_vma(struct page *page) page 100 include/linux/rmap.h if (((unsigned long)page->mapping & PAGE_MAPPING_FLAGS) != page 103 include/linux/rmap.h return page_rmapping(page); page 157 include/linux/rmap.h struct anon_vma *page_get_anon_vma(struct page *page); page 162 include/linux/rmap.h void page_move_anon_rmap(struct page *, struct vm_area_struct *, unsigned long); page 163 include/linux/rmap.h void page_add_anon_rmap(struct page *, struct vm_area_struct *, unsigned long); page 164 include/linux/rmap.h void do_page_add_anon_rmap(struct page *, struct vm_area_struct *, page 166 include/linux/rmap.h void page_add_new_anon_rmap(struct page *, struct vm_area_struct *, unsigned long); page 167 include/linux/rmap.h void page_add_file_rmap(struct page *); page 168 include/linux/rmap.h void page_remove_rmap(struct page *); page 170 include/linux/rmap.h void hugepage_add_anon_rmap(struct page *, struct vm_area_struct *, page 172 include/linux/rmap.h void hugepage_add_new_anon_rmap(struct page *, struct vm_area_struct *, page 175 include/linux/rmap.h static inline void page_dup_rmap(struct page *page) page 177 include/linux/rmap.h atomic_inc(&page->_mapcount); page 183 include/linux/rmap.h int page_referenced(struct page *, int is_locked, page 188 include/linux/rmap.h int try_to_unmap(struct page *, enum ttu_flags flags); page 193 include/linux/rmap.h pte_t *__page_check_address(struct page *, struct mm_struct *, page 196 include/linux/rmap.h static inline pte_t *page_check_address(struct page *page, struct mm_struct *mm, page 202 include/linux/rmap.h __cond_lock(*ptlp, ptep = __page_check_address(page, mm, address, page 210 include/linux/rmap.h unsigned long page_address_in_vma(struct page *, struct vm_area_struct *); page 218 include/linux/rmap.h int page_mkclean(struct page *); page 224 include/linux/rmap.h int try_to_munlock(struct page *); page 229 include/linux/rmap.h struct anon_vma *page_lock_anon_vma_read(struct page *page); page 231 include/linux/rmap.h int page_mapped_in_vma(struct page *page, struct vm_area_struct *vma); page 245 include/linux/rmap.h int (*rmap_one)(struct page *page, struct vm_area_struct *vma, page 247 include/linux/rmap.h int (*done)(struct page *page); page 248 include/linux/rmap.h int (*file_nonlinear)(struct page *, struct address_space *, void *arg); page 249 include/linux/rmap.h struct anon_vma *(*anon_lock)(struct page *page); page 253 include/linux/rmap.h int rmap_walk(struct page *page, struct rmap_walk_control *rwc); page 261 include/linux/rmap.h static inline int page_referenced(struct page *page, int is_locked, page 269 include/linux/rmap.h #define try_to_unmap(page, refs) SWAP_FAIL page 271 include/linux/rmap.h static inline int page_mkclean(struct page *page) page 57 include/linux/scatterlist.h static inline void sg_assign_page(struct scatterlist *sg, struct page *page) page 65 include/linux/scatterlist.h BUG_ON((unsigned long) page & 0x03); page 70 include/linux/scatterlist.h sg->page_link = page_link | (unsigned long) page; page 87 include/linux/scatterlist.h static inline void sg_set_page(struct scatterlist *sg, struct page *page, page 90 include/linux/scatterlist.h sg_assign_page(sg, page); page 95 include/linux/scatterlist.h static inline struct page *sg_page(struct scatterlist *sg) page 101 include/linux/scatterlist.h return (struct page *)((sg)->page_link & ~0x3); page 238 include/linux/scatterlist.h struct page **pages, unsigned int n_pages, page 286 include/linux/scatterlist.h static inline struct page *sg_page_iter_page(struct sg_page_iter *piter) page 334 include/linux/scatterlist.h struct page *page; /* currently mapped page */ page 1679 include/linux/sched.h extern bool should_numa_migrate_memory(struct task_struct *p, struct page *page, page 1697 include/linux/sched.h struct page *page, int src_nid, int dst_cpu) page 58 include/linux/shmem_fs.h extern struct page *shmem_read_mapping_page_gfp(struct address_space *mapping, page 61 include/linux/shmem_fs.h extern int shmem_unuse(swp_entry_t entry, struct page *page); page 63 include/linux/shmem_fs.h static inline struct page *shmem_read_mapping_page( page 196 include/linux/skbuff.h struct page *p; page 197 include/linux/skbuff.h } page; page 1579 include/linux/skbuff.h struct page *page, int off, int size) page 1592 include/linux/skbuff.h frag->page.p = page; page 1596 include/linux/skbuff.h page = compound_head(page); page 1597 include/linux/skbuff.h if (page->pfmemalloc && !page->mapping) page 1616 include/linux/skbuff.h struct page *page, int off, int size) page 1618 include/linux/skbuff.h __skb_fill_page_desc(skb, i, page, off, size); page 1622 include/linux/skbuff.h void skb_add_rx_frag(struct sk_buff *skb, int i, struct page *page, int off, page 2189 include/linux/skbuff.h static inline struct page *__skb_alloc_pages(gfp_t gfp_mask, page 2193 include/linux/skbuff.h struct page *page; page 2200 include/linux/skbuff.h page = alloc_pages_node(NUMA_NO_NODE, gfp_mask, order); page 2201 include/linux/skbuff.h if (skb && page && page->pfmemalloc) page 2204 include/linux/skbuff.h return page; page 2216 include/linux/skbuff.h static inline struct page *__skb_alloc_page(gfp_t gfp_mask, page 2227 include/linux/skbuff.h static inline void skb_propagate_pfmemalloc(struct page *page, page 2230 include/linux/skbuff.h if (page && page->pfmemalloc) page 2240 include/linux/skbuff.h static inline struct page *skb_frag_page(const skb_frag_t *frag) page 2242 include/linux/skbuff.h return frag->page.p; page 2326 include/linux/skbuff.h static inline void __skb_frag_set_page(skb_frag_t *frag, struct page *page) page 2328 include/linux/skbuff.h frag->page.p = page; page 2340 include/linux/skbuff.h struct page *page) page 2342 include/linux/skbuff.h __skb_frag_set_page(&skb_shinfo(skb)->frags[f], page); page 2481 include/linux/skbuff.h const struct page *page, int off) page 2486 include/linux/skbuff.h return page == skb_frag_page(frag) && page 43 include/linux/slub_def.h struct page *page; /* The slab from which we are allocating */ page 44 include/linux/slub_def.h struct page *partial; /* Partially allocated frozen slabs */ page 54 include/linux/splice.h struct page **pages; /* page map */ page 65 include/linux/sunrpc/gss_api.h struct page **inpages); page 123 include/linux/sunrpc/gss_api.h struct page **inpages); page 85 include/linux/sunrpc/gss_krb5.h struct page **pages); /* v2 encryption function */ page 257 include/linux/sunrpc/gss_krb5.h struct xdr_buf *outbuf, struct page **pages); page 274 include/linux/sunrpc/gss_krb5.h int offset, struct page **pages); page 314 include/linux/sunrpc/gss_krb5.h struct page **pages); page 244 include/linux/sunrpc/svc.h struct page * rq_pages[RPCSVC_MAXPAGES]; page 245 include/linux/sunrpc/svc.h struct page * *rq_respages; /* points into rq_pages */ page 246 include/linux/sunrpc/svc.h struct page * *rq_next_page; /* next reply page to use */ page 247 include/linux/sunrpc/svc.h struct page * *rq_page_end; /* one past the last page */ page 347 include/linux/sunrpc/svc.h struct page **pp = --rqstp->rq_next_page; page 87 include/linux/sunrpc/svc_rdma.h struct page *pages[RPCSVC_MAXPAGES]; page 208 include/linux/sunrpc/svc_rdma.h struct page *svc_rdma_get_page(void); page 38 include/linux/sunrpc/svcsock.h struct page * sk_pages[RPCSVC_MAXPAGES]; /* received data */ page 59 include/linux/sunrpc/xdr.h struct page ** pages; /* Array of pages */ page 108 include/linux/sunrpc/xdr.h struct page **, unsigned int, unsigned int); page 193 include/linux/sunrpc/xdr.h extern void _copy_from_pages(char *p, struct page **pages, size_t pgbase, page 206 include/linux/sunrpc/xdr.h struct page **page_ptr; /* pointer to the current page */ page 221 include/linux/sunrpc/xdr.h extern void xdr_write_pages(struct xdr_stream *xdr, struct page **pages, page 226 include/linux/sunrpc/xdr.h struct page **pages, unsigned int len); page 78 include/linux/sunrpc/xprt.h struct page **rq_enc_pages; /* scratch pages for use by page 324 include/linux/suspend.h extern int swsusp_page_is_forbidden(struct page *); page 325 include/linux/suspend.h extern void swsusp_set_page_free(struct page *); page 326 include/linux/suspend.h extern void swsusp_unset_page_free(struct page *); page 338 include/linux/suspend.h static inline int swsusp_page_is_forbidden(struct page *p) { return 0; } page 339 include/linux/suspend.h static inline void swsusp_set_page_free(struct page *p) {} page 340 include/linux/suspend.h static inline void swsusp_unset_page_free(struct page *p) {} page 260 include/linux/swap.h void *workingset_eviction(struct address_space *mapping, struct page *page); page 262 include/linux/swap.h void workingset_activation(struct page *page); page 307 include/linux/swap.h extern void lru_cache_add(struct page *); page 308 include/linux/swap.h extern void lru_cache_add_anon(struct page *page); page 309 include/linux/swap.h extern void lru_cache_add_file(struct page *page); page 310 include/linux/swap.h extern void lru_add_page_tail(struct page *page, struct page *page_tail, page 312 include/linux/swap.h extern void activate_page(struct page *); page 313 include/linux/swap.h extern void mark_page_accessed(struct page *); page 317 include/linux/swap.h extern void rotate_reclaimable_page(struct page *page); page 318 include/linux/swap.h extern void deactivate_page(struct page *page); page 321 include/linux/swap.h extern void add_page_to_unevictable_list(struct page *page); page 323 include/linux/swap.h extern void lru_cache_add_active_or_unevictable(struct page *page, page 329 include/linux/swap.h extern int __isolate_lru_page(struct page *page, isolate_mode_t mode); page 340 include/linux/swap.h extern int remove_mapping(struct address_space *mapping, struct page *page); page 356 include/linux/swap.h extern int page_evictable(struct page *page); page 357 include/linux/swap.h extern void check_move_unevictable_pages(struct page **, int nr_pages); page 370 include/linux/swap.h extern void mem_cgroup_swapout(struct page *page, swp_entry_t entry); page 373 include/linux/swap.h static inline void mem_cgroup_swapout(struct page *page, swp_entry_t entry) page 382 include/linux/swap.h extern int swap_readpage(struct page *); page 383 include/linux/swap.h extern int swap_writepage(struct page *page, struct writeback_control *wbc); page 385 include/linux/swap.h extern int __swap_writepage(struct page *page, struct writeback_control *wbc, page 387 include/linux/swap.h extern int swap_set_page_dirty(struct page *page); page 400 include/linux/swap.h extern int add_to_swap(struct page *, struct list_head *list); page 401 include/linux/swap.h extern int add_to_swap_cache(struct page *, swp_entry_t, gfp_t); page 402 include/linux/swap.h extern int __add_to_swap_cache(struct page *page, swp_entry_t entry); page 403 include/linux/swap.h extern void __delete_from_swap_cache(struct page *); page 404 include/linux/swap.h extern void delete_from_swap_cache(struct page *); page 405 include/linux/swap.h extern void free_page_and_swap_cache(struct page *); page 406 include/linux/swap.h extern void free_pages_and_swap_cache(struct page **, int); page 407 include/linux/swap.h extern struct page *lookup_swap_cache(swp_entry_t); page 408 include/linux/swap.h extern struct page *read_swap_cache_async(swp_entry_t, gfp_t, page 410 include/linux/swap.h extern struct page *swapin_readahead(swp_entry_t, gfp_t, page 440 include/linux/swap.h extern sector_t map_swap_page(struct page *, struct block_device **); page 442 include/linux/swap.h extern int page_swapcount(struct page *); page 443 include/linux/swap.h extern struct swap_info_struct *page_swap_info(struct page *); page 444 include/linux/swap.h extern int reuse_swap_page(struct page *); page 445 include/linux/swap.h extern int try_to_free_swap(struct page *); page 450 include/linux/swap.h mem_cgroup_uncharge_swapcache(struct page *page, swp_entry_t ent, bool swapout); page 453 include/linux/swap.h mem_cgroup_uncharge_swapcache(struct page *page, swp_entry_t ent, bool swapout) page 470 include/linux/swap.h #define free_page_and_swap_cache(page) \ page 471 include/linux/swap.h page_cache_release(page) page 504 include/linux/swap.h static inline struct page *swapin_readahead(swp_entry_t swp, gfp_t gfp_mask, page 510 include/linux/swap.h static inline int swap_writepage(struct page *p, struct writeback_control *wbc) page 515 include/linux/swap.h static inline struct page *lookup_swap_cache(swp_entry_t swp) page 520 include/linux/swap.h static inline int add_to_swap(struct page *page, struct list_head *list) page 525 include/linux/swap.h static inline int add_to_swap_cache(struct page *page, swp_entry_t entry, page 531 include/linux/swap.h static inline void __delete_from_swap_cache(struct page *page) page 535 include/linux/swap.h static inline void delete_from_swap_cache(struct page *page) page 539 include/linux/swap.h static inline int page_swapcount(struct page *page) page 544 include/linux/swap.h #define reuse_swap_page(page) (page_mapcount(page) == 1) page 546 include/linux/swap.h static inline int try_to_free_swap(struct page *page) page 559 include/linux/swap.h mem_cgroup_uncharge_swapcache(struct page *page, swp_entry_t ent) page 106 include/linux/swapops.h static inline swp_entry_t make_migration_entry(struct page *page, int write) page 108 include/linux/swapops.h BUG_ON(!PageLocked(page)); page 110 include/linux/swapops.h page_to_pfn(page)); page 124 include/linux/swapops.h static inline struct page *migration_entry_to_page(swp_entry_t entry) page 126 include/linux/swapops.h struct page *p = pfn_to_page(swp_offset(entry)); page 146 include/linux/swapops.h #define make_migration_entry(page, write) swp_entry(0, 0) page 168 include/linux/swapops.h static inline swp_entry_t make_hwpoison_entry(struct page *page) page 170 include/linux/swapops.h BUG_ON(!PageLocked(page)); page 171 include/linux/swapops.h return swp_entry(SWP_HWPOISON, page_to_pfn(page)); page 180 include/linux/swapops.h static inline swp_entry_t make_hwpoison_entry(struct page *page) page 65 include/linux/swiotlb.h extern dma_addr_t swiotlb_map_page(struct device *dev, struct page *page, page 15 include/linux/uio.h struct page; page 74 include/linux/uio.h size_t iov_iter_copy_from_user_atomic(struct page *page, page 79 include/linux/uio.h size_t copy_page_to_iter(struct page *page, size_t offset, size_t bytes, page 81 include/linux/uio.h size_t copy_page_from_iter(struct page *page, size_t offset, size_t bytes, page 89 include/linux/uio.h ssize_t iov_iter_get_pages(struct iov_iter *i, struct page **pages, page 91 include/linux/uio.h ssize_t iov_iter_get_pages_alloc(struct iov_iter *i, struct page ***pages, page 35 include/linux/uprobes.h struct page; page 132 include/linux/uprobes.h extern void arch_uprobe_copy_ixol(struct page *page, unsigned long vaddr, page 11 include/linux/usb/gadget_configfs.h const char *page, size_t len) \ page 15 include/linux/usb/gadget_configfs.h ret = usb_string_copy(page, &gs->__name); \ page 23 include/linux/usb/gadget_configfs.h char *page) \ page 25 include/linux/usb/gadget_configfs.h return sprintf(page, "%s\n", gs->__name ?: ""); \ page 34 include/linux/vmalloc.h struct page **pages; page 55 include/linux/vmalloc.h extern void *vm_map_ram(struct page **pages, unsigned int count, page 81 include/linux/vmalloc.h extern void *vmap(struct page **pages, unsigned int count, page 116 include/linux/vmalloc.h struct page **pages); page 119 include/linux/vmalloc.h pgprot_t prot, struct page **pages); page 125 include/linux/vmalloc.h pgprot_t prot, struct page **pages) page 202 include/linux/vmstat.h void __inc_zone_page_state(struct page *, enum zone_stat_item); page 203 include/linux/vmstat.h void __dec_zone_page_state(struct page *, enum zone_stat_item); page 206 include/linux/vmstat.h void inc_zone_page_state(struct page *, enum zone_stat_item); page 207 include/linux/vmstat.h void dec_zone_page_state(struct page *, enum zone_stat_item); page 247 include/linux/vmstat.h static inline void __inc_zone_page_state(struct page *page, page 250 include/linux/vmstat.h __inc_zone_state(page_zone(page), item); page 253 include/linux/vmstat.h static inline void __dec_zone_page_state(struct page *page, page 256 include/linux/vmstat.h __dec_zone_state(page_zone(page), item); page 169 include/linux/writeback.h typedef int (*writepage_t)(struct page *page, struct writeback_control *wbc, page 184 include/linux/writeback.h void account_page_redirty(struct page *page); page 52 include/media/videobuf-dma-sg.h struct page **pages; page 56 include/media/videobuf-dma-sg.h struct page **vaddr_pages; page 560 include/net/bluetooth/hci.h __u8 page; page 1046 include/net/bluetooth/hci.h __u8 page; page 1050 include/net/bluetooth/hci.h __u8 page; page 1552 include/net/bluetooth/hci.h __u8 page; page 407 include/net/ieee802154_netdev.h u8 channel, u8 page, u8 cap); page 416 include/net/ieee802154_netdev.h u8 channel, u8 page, u8 bcn_ord, u8 sf_ord, page 419 include/net/ieee802154_netdev.h u8 type, u32 channels, u8 page, u8 duration); page 26 include/net/inet_common.h ssize_t inet_sendpage(struct socket *sock, struct page *page, int offset, page 132 include/net/ip.h ssize_t ip_append_page(struct sock *sk, struct flowi4 *fl4, struct page *page, page 171 include/net/mac802154.h int page, page 98 include/net/nl802154.h u8 status, u8 scan_type, u32 unscanned, u8 page, page 976 include/net/sock.h int (*sendpage)(struct sock *sk, struct page *page, page 1614 include/net/sock.h ssize_t sock_no_sendpage(struct socket *sock, struct page *page, int offset, page 1857 include/net/sock.h struct page *page, page 1862 include/net/sock.h err = skb_do_copy_data_nocache(sk, skb, from, page_address(page) + off, page 1876 include/net/sock.h struct sk_buff *skb, struct page *page, page 1882 include/net/sock.h page_address(page) + off, page 1887 include/net/sock.h } else if (copy_from_user(page_address(page) + off, from, copy)) page 357 include/net/tcp.h int tcp_sendpage(struct sock *sk, struct page *page, int offset, size_t size, page 1401 include/rdma/ib_verbs.h struct page *page, unsigned long offset, page 2165 include/rdma/ib_verbs.h struct page *page, page 2171 include/rdma/ib_verbs.h return dev->dma_ops->map_page(dev, page, offset, size, direction); page 2172 include/rdma/ib_verbs.h return dma_map_page(dev->dma_device, page, offset, size, direction); page 331 include/scsi/libfcoe.h struct page *crc_eof_page; page 112 include/scsi/osd_ore.h struct page * (*get_page)(void *priv, u64 page_index, bool *uptodate); page 113 include/scsi/osd_ore.h void (*put_page)(void *priv, struct page *page); page 131 include/scsi/osd_ore.h struct page **pages; page 146 include/scsi/osd_ore.h struct page **parity_pages; page 398 include/scsi/scsi_device.h extern int scsi_get_vpd_page(struct scsi_device *, u8 page, unsigned char *buf, page 91 include/sound/memalloc.h struct page **page_table; /* page table (for vmap/vunmap) */ page 81 include/sound/pcm.h struct page *(*page)(struct snd_pcm_substream *substream, page 945 include/sound/pcm.h struct page *snd_pcm_lib_get_vmalloc_page(struct snd_pcm_substream *substream, page 992 include/sound/pcm.h struct page *snd_pcm_sgbuf_ops_page(struct snd_pcm_substream *substream, page 111 include/target/configfs_macros.h char *page) \ page 119 include/target/configfs_macros.h ret = _name##_attr->show(_item, page); \ page 126 include/target/configfs_macros.h const char *page, size_t count) \ page 134 include/target/configfs_macros.h ret = _name##_attr->store(_item, page, count); \ page 307 include/trace/events/btrfs.h TP_PROTO(struct page *page, struct inode *inode, page 310 include/trace/events/btrfs.h TP_ARGS(page, inode, wbc), page 328 include/trace/events/btrfs.h __entry->index = page->index; page 356 include/trace/events/btrfs.h TP_PROTO(struct page *page, struct inode *inode, page 359 include/trace/events/btrfs.h TP_ARGS(page, inode, wbc) page 364 include/trace/events/btrfs.h TP_PROTO(struct page *page, u64 start, u64 end, int uptodate), page 366 include/trace/events/btrfs.h TP_ARGS(page, start, end, uptodate), page 378 include/trace/events/btrfs.h __entry->ino = page->mapping->host->i_ino; page 379 include/trace/events/btrfs.h __entry->index = page->index; page 384 include/trace/events/btrfs.h BTRFS_I(page->mapping->host)->root->root_key.objectid; page 235 include/trace/events/ext3.h TP_PROTO(struct page *page), page 237 include/trace/events/ext3.h TP_ARGS(page), page 247 include/trace/events/ext3.h __entry->index = page->index; page 248 include/trace/events/ext3.h __entry->ino = page->mapping->host->i_ino; page 249 include/trace/events/ext3.h __entry->dev = page->mapping->host->i_sb->s_dev; page 259 include/trace/events/ext3.h TP_PROTO(struct page *page), page 261 include/trace/events/ext3.h TP_ARGS(page) page 266 include/trace/events/ext3.h TP_PROTO(struct page *page), page 268 include/trace/events/ext3.h TP_ARGS(page) page 273 include/trace/events/ext3.h TP_PROTO(struct page *page), page 275 include/trace/events/ext3.h TP_ARGS(page) page 280 include/trace/events/ext3.h TP_PROTO(struct page *page), page 282 include/trace/events/ext3.h TP_ARGS(page) page 287 include/trace/events/ext3.h TP_PROTO(struct page *page), page 289 include/trace/events/ext3.h TP_ARGS(page) page 293 include/trace/events/ext3.h TP_PROTO(struct page *page, unsigned int offset, unsigned int length), page 295 include/trace/events/ext3.h TP_ARGS(page, offset, length), page 307 include/trace/events/ext3.h __entry->index = page->index; page 310 include/trace/events/ext3.h __entry->ino = page->mapping->host->i_ino; page 311 include/trace/events/ext3.h __entry->dev = page->mapping->host->i_sb->s_dev; page 474 include/trace/events/ext4.h TP_PROTO(struct page *page), page 476 include/trace/events/ext4.h TP_ARGS(page), page 486 include/trace/events/ext4.h __entry->dev = page->mapping->host->i_sb->s_dev; page 487 include/trace/events/ext4.h __entry->ino = page->mapping->host->i_ino; page 488 include/trace/events/ext4.h __entry->index = page->index; page 499 include/trace/events/ext4.h TP_PROTO(struct page *page), page 501 include/trace/events/ext4.h TP_ARGS(page) page 506 include/trace/events/ext4.h TP_PROTO(struct page *page), page 508 include/trace/events/ext4.h TP_ARGS(page) page 513 include/trace/events/ext4.h TP_PROTO(struct page *page), page 515 include/trace/events/ext4.h TP_ARGS(page) page 519 include/trace/events/ext4.h TP_PROTO(struct page *page, unsigned int offset, unsigned int length), page 521 include/trace/events/ext4.h TP_ARGS(page, offset, length), page 532 include/trace/events/ext4.h __entry->dev = page->mapping->host->i_sb->s_dev; page 533 include/trace/events/ext4.h __entry->ino = page->mapping->host->i_ino; page 534 include/trace/events/ext4.h __entry->index = page->index; page 547 include/trace/events/ext4.h TP_PROTO(struct page *page, unsigned int offset, unsigned int length), page 549 include/trace/events/ext4.h TP_ARGS(page, offset, length) page 553 include/trace/events/ext4.h TP_PROTO(struct page *page, unsigned int offset, unsigned int length), page 555 include/trace/events/ext4.h TP_ARGS(page, offset, length) page 445 include/trace/events/f2fs.h TP_PROTO(struct page *page, sector_t blkaddr, int type), page 447 include/trace/events/f2fs.h TP_ARGS(page, blkaddr, type), page 449 include/trace/events/f2fs.h TP_CONDITION(page->mapping), page 460 include/trace/events/f2fs.h __entry->dev = page->mapping->host->i_sb->s_dev; page 461 include/trace/events/f2fs.h __entry->ino = page->mapping->host->i_ino; page 462 include/trace/events/f2fs.h __entry->index = page->index; page 793 include/trace/events/f2fs.h TP_PROTO(struct page *page, int type), page 795 include/trace/events/f2fs.h TP_ARGS(page, type), page 808 include/trace/events/f2fs.h __entry->dev = page->mapping->host->i_sb->s_dev; page 809 include/trace/events/f2fs.h __entry->ino = page->mapping->host->i_ino; page 811 include/trace/events/f2fs.h __entry->dir = S_ISDIR(page->mapping->host->i_mode); page 812 include/trace/events/f2fs.h __entry->index = page->index; page 813 include/trace/events/f2fs.h __entry->dirty = PageDirty(page); page 814 include/trace/events/f2fs.h __entry->uptodate = PageUptodate(page); page 829 include/trace/events/f2fs.h TP_PROTO(struct page *page, int type), page 831 include/trace/events/f2fs.h TP_ARGS(page, type) page 836 include/trace/events/f2fs.h TP_PROTO(struct page *page, int type), page 838 include/trace/events/f2fs.h TP_ARGS(page, type) page 843 include/trace/events/f2fs.h TP_PROTO(struct page *page, int type), page 845 include/trace/events/f2fs.h TP_ARGS(page, type) page 850 include/trace/events/f2fs.h TP_PROTO(struct page *page, int type), page 852 include/trace/events/f2fs.h TP_ARGS(page, type) page 921 include/trace/events/f2fs.h TP_PROTO(struct page *page, int rw, int type, block_t blk_addr), page 923 include/trace/events/f2fs.h TP_ARGS(page, rw, type, blk_addr), page 935 include/trace/events/f2fs.h __entry->dev = page->mapping->host->i_sb->s_dev; page 936 include/trace/events/f2fs.h __entry->ino = page->mapping->host->i_ino; page 939 include/trace/events/f2fs.h __entry->index = page->index; page 16 include/trace/events/filemap.h TP_PROTO(struct page *page), page 18 include/trace/events/filemap.h TP_ARGS(page), page 21 include/trace/events/filemap.h __field(struct page *, page) page 28 include/trace/events/filemap.h __entry->page = page; page 29 include/trace/events/filemap.h __entry->i_ino = page->mapping->host->i_ino; page 30 include/trace/events/filemap.h __entry->index = page->index; page 31 include/trace/events/filemap.h if (page->mapping->host->i_sb) page 32 include/trace/events/filemap.h __entry->s_dev = page->mapping->host->i_sb->s_dev; page 34 include/trace/events/filemap.h __entry->s_dev = page->mapping->host->i_rdev; page 40 include/trace/events/filemap.h __entry->page, page 41 include/trace/events/filemap.h page_to_pfn(__entry->page), page 46 include/trace/events/filemap.h TP_PROTO(struct page *page), page 47 include/trace/events/filemap.h TP_ARGS(page) page 51 include/trace/events/filemap.h TP_PROTO(struct page *page), page 52 include/trace/events/filemap.h TP_ARGS(page) page 152 include/trace/events/kmem.h TP_PROTO(struct page *page, unsigned int order), page 154 include/trace/events/kmem.h TP_ARGS(page, order), page 157 include/trace/events/kmem.h __field( struct page *, page ) page 162 include/trace/events/kmem.h __entry->page = page; page 167 include/trace/events/kmem.h __entry->page, page 168 include/trace/events/kmem.h page_to_pfn(__entry->page), page 174 include/trace/events/kmem.h TP_PROTO(struct page *page, int cold), page 176 include/trace/events/kmem.h TP_ARGS(page, cold), page 179 include/trace/events/kmem.h __field( struct page *, page ) page 184 include/trace/events/kmem.h __entry->page = page; page 189 include/trace/events/kmem.h __entry->page, page 190 include/trace/events/kmem.h page_to_pfn(__entry->page), page 196 include/trace/events/kmem.h TP_PROTO(struct page *page, unsigned int order, page 199 include/trace/events/kmem.h TP_ARGS(page, order, gfp_flags, migratetype), page 202 include/trace/events/kmem.h __field( struct page *, page ) page 209 include/trace/events/kmem.h __entry->page = page; page 216 include/trace/events/kmem.h __entry->page, page 217 include/trace/events/kmem.h __entry->page ? page_to_pfn(__entry->page) : 0, page 225 include/trace/events/kmem.h TP_PROTO(struct page *page, unsigned int order, int migratetype), page 227 include/trace/events/kmem.h TP_ARGS(page, order, migratetype), page 230 include/trace/events/kmem.h __field( struct page *, page ) page 236 include/trace/events/kmem.h __entry->page = page; page 242 include/trace/events/kmem.h __entry->page, page 243 include/trace/events/kmem.h __entry->page ? page_to_pfn(__entry->page) : 0, page 251 include/trace/events/kmem.h TP_PROTO(struct page *page, unsigned int order, int migratetype), page 253 include/trace/events/kmem.h TP_ARGS(page, order, migratetype) page 258 include/trace/events/kmem.h TP_PROTO(struct page *page, unsigned int order, int migratetype), page 260 include/trace/events/kmem.h TP_ARGS(page, order, migratetype), page 263 include/trace/events/kmem.h __entry->page, page_to_pfn(__entry->page), page 269 include/trace/events/kmem.h TP_PROTO(struct page *page, page 273 include/trace/events/kmem.h TP_ARGS(page, page 278 include/trace/events/kmem.h __field( struct page *, page ) page 287 include/trace/events/kmem.h __entry->page = page; page 296 include/trace/events/kmem.h __entry->page, page 297 include/trace/events/kmem.h page_to_pfn(__entry->page), page 18 include/trace/events/pagemap.h #define trace_pagemap_flags(page) ( \ page 19 include/trace/events/pagemap.h (PageAnon(page) ? PAGEMAP_ANONYMOUS : PAGEMAP_FILE) | \ page 20 include/trace/events/pagemap.h (page_mapped(page) ? PAGEMAP_MAPPED : 0) | \ page 21 include/trace/events/pagemap.h (PageSwapCache(page) ? PAGEMAP_SWAPCACHE : 0) | \ page 22 include/trace/events/pagemap.h (PageSwapBacked(page) ? PAGEMAP_SWAPBACKED : 0) | \ page 23 include/trace/events/pagemap.h (PageMappedToDisk(page) ? PAGEMAP_MAPPEDDISK : 0) | \ page 24 include/trace/events/pagemap.h (page_has_private(page) ? PAGEMAP_BUFFERS : 0) \ page 30 include/trace/events/pagemap.h struct page *page, page 34 include/trace/events/pagemap.h TP_ARGS(page, lru), page 37 include/trace/events/pagemap.h __field(struct page *, page ) page 44 include/trace/events/pagemap.h __entry->page = page; page 45 include/trace/events/pagemap.h __entry->pfn = page_to_pfn(page); page 47 include/trace/events/pagemap.h __entry->flags = trace_pagemap_flags(page); page 52 include/trace/events/pagemap.h __entry->page, page 65 include/trace/events/pagemap.h TP_PROTO(struct page *page), page 67 include/trace/events/pagemap.h TP_ARGS(page), page 70 include/trace/events/pagemap.h __field(struct page *, page ) page 75 include/trace/events/pagemap.h __entry->page = page; page 76 include/trace/events/pagemap.h __entry->pfn = page_to_pfn(page); page 80 include/trace/events/pagemap.h TP_printk("page=%p pfn=%lu", __entry->page, __entry->pfn) page 28 include/trace/events/vmscan.h #define trace_reclaim_flags(page) ( \ page 29 include/trace/events/vmscan.h (page_is_file_cache(page) ? RECLAIM_WB_FILE : RECLAIM_WB_ANON) | \ page 333 include/trace/events/vmscan.h TP_PROTO(struct page *page, page 336 include/trace/events/vmscan.h TP_ARGS(page, reclaim_flags), page 339 include/trace/events/vmscan.h __field(struct page *, page) page 344 include/trace/events/vmscan.h __entry->page = page; page 349 include/trace/events/vmscan.h __entry->page, page 350 include/trace/events/vmscan.h page_to_pfn(__entry->page), page 38 include/trace/events/writeback.h TP_PROTO(struct page *page, struct address_space *mapping), page 40 include/trace/events/writeback.h TP_ARGS(page, mapping), page 52 include/trace/events/writeback.h __entry->index = page->index; page 87 include/uapi/linux/i2o-dev.h unsigned int page; /* HTML page */ page 28 include/xen/balloon.h int alloc_xenballooned_pages(int nr_pages, struct page **pages, page 30 include/xen/balloon.h void free_xenballooned_pages(int nr_pages, struct page **pages); page 32 include/xen/balloon.h struct page *get_balloon_scratch_page(void); page 82 include/xen/grant_table.h unsigned long page); page 168 include/xen/grant_table.h struct page **pages, unsigned int count); page 171 include/xen/grant_table.h struct page **pages, unsigned int count); page 95 include/xen/interface/io/fbif.h #define XENFB_IN_RING(page) \ page 96 include/xen/interface/io/fbif.h ((union xenfb_in_event *)((char *)(page) + XENFB_IN_RING_OFFS)) page 97 include/xen/interface/io/fbif.h #define XENFB_IN_RING_REF(page, idx) \ page 98 include/xen/interface/io/fbif.h (XENFB_IN_RING((page))[(idx) % XENFB_IN_RING_LEN]) page 103 include/xen/interface/io/fbif.h #define XENFB_OUT_RING(page) \ page 104 include/xen/interface/io/fbif.h ((union xenfb_out_event *)((char *)(page) + XENFB_OUT_RING_OFFS)) page 105 include/xen/interface/io/fbif.h #define XENFB_OUT_RING_REF(page, idx) \ page 106 include/xen/interface/io/fbif.h (XENFB_OUT_RING((page))[(idx) % XENFB_OUT_RING_LEN]) page 98 include/xen/interface/io/kbdif.h #define XENKBD_IN_RING(page) \ page 99 include/xen/interface/io/kbdif.h ((union xenkbd_in_event *)((char *)(page) + XENKBD_IN_RING_OFFS)) page 100 include/xen/interface/io/kbdif.h #define XENKBD_IN_RING_REF(page, idx) \ page 101 include/xen/interface/io/kbdif.h (XENKBD_IN_RING((page))[(idx) % XENKBD_IN_RING_LEN]) page 106 include/xen/interface/io/kbdif.h #define XENKBD_OUT_RING(page) \ page 107 include/xen/interface/io/kbdif.h ((union xenkbd_out_event *)((char *)(page) + XENKBD_OUT_RING_OFFS)) page 108 include/xen/interface/io/kbdif.h #define XENKBD_OUT_RING_REF(page, idx) \ page 109 include/xen/interface/io/kbdif.h (XENKBD_OUT_RING((page))[(idx) % XENKBD_OUT_RING_LEN]) page 19 include/xen/swiotlb-xen.h extern dma_addr_t xen_swiotlb_map_page(struct device *dev, struct page *page, page 34 include/xen/xen-ops.h struct page **pages); page 36 include/xen/xen-ops.h int numpgs, struct page **pages); page 34 scripts/kconfig/lxdialog/textbox.c static char *page; page 70 scripts/kconfig/lxdialog/textbox.c page = buf; /* page is pointer to start of page to be displayed */ page 156 scripts/kconfig/lxdialog/textbox.c page = buf; page 167 scripts/kconfig/lxdialog/textbox.c page = buf + strlen(buf); page 266 scripts/kconfig/lxdialog/textbox.c while (s < page && (s = strchr(s, '\n'))) { page 287 scripts/kconfig/lxdialog/textbox.c if (*page == '\0') { page 293 scripts/kconfig/lxdialog/textbox.c if (page == buf) { page 297 scripts/kconfig/lxdialog/textbox.c page--; page 299 scripts/kconfig/lxdialog/textbox.c if (page == buf) { page 303 scripts/kconfig/lxdialog/textbox.c page--; page 304 scripts/kconfig/lxdialog/textbox.c } while (*page != '\n'); page 305 scripts/kconfig/lxdialog/textbox.c page++; page 322 scripts/kconfig/lxdialog/textbox.c end = page; page 324 scripts/kconfig/lxdialog/textbox.c update_text(buf, page - buf, end - buf, data); page 375 scripts/kconfig/lxdialog/textbox.c while (*page != '\n') { page 376 scripts/kconfig/lxdialog/textbox.c if (*page == '\0') { page 380 scripts/kconfig/lxdialog/textbox.c line[i++] = *(page++); page 385 scripts/kconfig/lxdialog/textbox.c page++; page 391 scripts/kconfig/lxdialog/textbox.c page++; /* move past '\n' */ page 405 scripts/kconfig/lxdialog/textbox.c percent = (page - buf) * 100 / strlen(buf);