Lines Matching refs:size_t

17 	void *(*alloc)(struct device *dev, size_t size,
20 void (*free)(struct device *dev, size_t size, void *vaddr,
22 struct page *(*alloc_pages_op)(struct device *dev, size_t size,
25 void (*free_pages)(struct device *dev, size_t size, struct page *vaddr,
28 void *, dma_addr_t, size_t, unsigned long attrs);
31 void *cpu_addr, dma_addr_t dma_addr, size_t size,
35 unsigned long offset, size_t size,
38 size_t size, enum dma_data_direction dir,
50 size_t size, enum dma_data_direction dir,
53 size_t size, enum dma_data_direction dir,
56 size_t size, enum dma_data_direction dir);
58 dma_addr_t dma_handle, size_t size,
64 void (*cache_sync)(struct device *dev, void *vaddr, size_t size,
68 size_t (*max_mapping_size)(struct device *dev);
69 size_t (*opt_mapping_size)(void);
113 struct page *dma_alloc_from_contiguous(struct device *dev, size_t count,
117 struct page *dma_alloc_contiguous(struct device *dev, size_t size, gfp_t gfp);
118 void dma_free_contiguous(struct device *dev, struct page *page, size_t size);
136 size_t count, unsigned int order, bool no_warn) in dma_alloc_from_contiguous()
146 static inline struct page *dma_alloc_contiguous(struct device *dev, size_t size, in dma_alloc_contiguous()
152 size_t size) in dma_free_contiguous()
160 dma_addr_t device_addr, size_t size);
166 void *cpu_addr, size_t size, int *ret);
169 phys_addr_t phys_addr, dma_addr_t device_addr, size_t size) in dma_declare_coherent_memory()
185 size_t size, int *ret);
186 int dma_init_global_coherent(phys_addr_t phys_addr, size_t size);
198 void *cpu_addr, size_t size, int *ret) in dma_mmap_from_global_coherent()
205 void *cpu_addr, dma_addr_t dma_addr, size_t size,
208 void *cpu_addr, dma_addr_t dma_addr, size_t size,
210 struct page *dma_common_alloc_pages(struct device *dev, size_t size,
212 void dma_common_free_pages(struct device *dev, size_t size, struct page *vaddr,
216 void *dma_common_contiguous_remap(struct page *page, size_t size, pgprot_t prot,
218 void *dma_common_pages_remap(struct page **pages, size_t size, pgprot_t prot,
220 void dma_common_free_remap(void *cpu_addr, size_t size);
222 struct page *dma_alloc_from_pool(struct device *dev, size_t size,
224 bool (*phys_addr_ok)(struct device *, phys_addr_t, size_t));
225 bool dma_free_from_pool(struct device *dev, void *start, size_t size);
284 static inline bool dma_kmalloc_size_aligned(size_t size) in dma_kmalloc_size_aligned()
310 static inline bool dma_kmalloc_needs_bounce(struct device *dev, size_t size, in dma_kmalloc_needs_bounce()
316 void *arch_dma_alloc(struct device *dev, size_t size, dma_addr_t *dma_handle,
318 void arch_dma_free(struct device *dev, size_t size, void *cpu_addr,
348 void arch_sync_dma_for_device(phys_addr_t paddr, size_t size,
351 static inline void arch_sync_dma_for_device(phys_addr_t paddr, size_t size, in arch_sync_dma_for_device()
358 void arch_sync_dma_for_cpu(phys_addr_t paddr, size_t size,
361 static inline void arch_sync_dma_for_cpu(phys_addr_t paddr, size_t size, in arch_sync_dma_for_cpu()
376 void arch_dma_prep_coherent(struct page *page, size_t size);
378 static inline void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent()
384 void arch_dma_mark_clean(phys_addr_t paddr, size_t size);
386 static inline void arch_dma_mark_clean(phys_addr_t paddr, size_t size) in arch_dma_mark_clean()
391 void *arch_dma_set_uncached(void *addr, size_t size);
392 void arch_dma_clear_uncached(void *addr, size_t size);