Lines Matching refs:size
22 * @size: number of bytes to write back
25 * instruction. Note that @size is internally rounded up to be cache
26 * line size aligned.
28 static void clean_cache_range(void *addr, size_t size)
32 void *vend = addr + size;
40 void arch_wb_cache_pmem(void *addr, size_t size)
42 clean_cache_range(addr, size);
46 long __copy_user_flushcache(void *dst, const void __user *src, unsigned size)
52 rc = __copy_user_nocache(dst, src, size);
59 * destination or size is not naturally aligned. That is:
60 * - Require 8-byte alignment when size is 8 bytes or larger.
61 * - Require 4-byte alignment when size is 4 bytes.
63 if (size < 8) {
64 if (!IS_ALIGNED(dest, 4) || size != 4)
65 clean_cache_range(dst, size);
73 if (size > flushed && !IS_ALIGNED(size - flushed, 8))
74 clean_cache_range(dst + size - 1, 1);
80 void __memcpy_flushcache(void *_dst, const void *_src, size_t size)
87 size_t len = min_t(size_t, size, ALIGN(dest, 8) - dest);
93 size -= len;
94 if (!size)
99 while (size >= 32) {
112 size -= 32;
116 while (size >= 8) {
123 size -= 8;
127 while (size >= 4) {
134 size -= 4;
138 if (size) {
139 memcpy((void *) dest, (void *) source, size);
140 clean_cache_range((void *) dest, size);