Lines Matching refs:buff
125 * Algorithm accounts for buff being misaligned.
126 * If buff is not aligned, will over-read bytes but not use the bytes that it
130 do_csum_with_alignment(const unsigned char *buff, int len)
138 * buff. This should always be in the same page and cache line.
141 offset = (unsigned long)buff & OFFSET_MASK;
142 kasan_check_read(buff, len);
143 ptr = (const unsigned long *)(buff - offset);
146 * Clear the most significant bytes that were over-read if buff was not
156 end = (const unsigned long *)(buff + len);
231 * misaligned accesses, or when buff is known to be aligned.
234 do_csum_no_alignment(const unsigned char *buff, int len)
239 ptr = (const unsigned long *)(buff);
242 kasan_check_read(buff, len);
244 end = (const unsigned long *)(buff + len);
303 * Will do a light-weight address alignment if buff is misaligned, unless
306 unsigned int do_csum(const unsigned char *buff, int len)
321 if (has_fast_unaligned_accesses() || (((unsigned long)buff & OFFSET_MASK) == 0))
322 return do_csum_no_alignment(buff, len);
324 return do_csum_with_alignment(buff, len);