Lines Matching refs:buff
42 * it's best to have buff aligned on a 64-bit boundary
44 __wsum csum_partial(const void *buff, int len, __wsum sum)
52 temp64 = update_csum_40b(temp64, buff);
53 temp64_2 = update_csum_40b(temp64_2, buff + 40);
54 buff += 80;
68 temp64 = update_csum_40b(temp64, buff);
72 buff += 40;
82 : [src] "r"(buff), "m"(*(const char(*)[32])buff));
83 buff += 32;
90 : [src] "r"(buff), "m"(*(const char(*)[16])buff));
91 buff += 16;
97 : [src] "r"(buff), "m"(*(const char(*)[8])buff));
98 buff += 8;
104 trail = (load_unaligned_zeropad(buff) << shift) >> shift;
119 __sum16 ip_compute_csum(const void *buff, int len)
121 return csum_fold(csum_partial(buff, len, 0));