Lines Matching refs:src_cnt

32 	int src_cnt = unmap->to_cnt;
37 while (src_cnt) {
41 xor_src_cnt = min(src_cnt, (int)dma->max_xor);
45 if (src_cnt > xor_src_cnt) {
86 if (src_cnt > xor_src_cnt) {
88 src_cnt -= xor_src_cnt;
90 src_cnt++;
102 int src_cnt, size_t len, struct async_submit_ctl *submit)
116 for (i = 0; i < src_cnt; i++)
120 src_cnt = xor_src_cnt;
127 while (src_cnt > 0) {
129 xor_src_cnt = min(src_cnt, MAX_XOR_BLOCKS);
133 src_cnt -= xor_src_cnt;
142 unsigned int *src_offs, int src_cnt, int len)
152 for (i = 0; i < src_cnt; i++) {
165 * @src_cnt: number of source pages
184 int src_cnt, size_t len, struct async_submit_ctl *submit)
188 src_cnt, len);
192 BUG_ON(src_cnt <= 1);
195 unmap = dmaengine_get_unmap_data(device->dev, src_cnt+1, GFP_NOWAIT);
198 src_offs, src_cnt, len)) {
206 for (i = 0, j = 0; i < src_cnt; i++) {
234 src_cnt--;
244 src_cnt, len, submit);
256 * @src_cnt: number of source pages
274 int src_cnt, size_t len, struct async_submit_ctl *submit)
277 src_cnt, len, submit);
288 struct page **src_list, int src_cnt, size_t len)
294 src_cnt, len);
303 * @src_cnt: number of source pages
317 int src_cnt, size_t len, enum sum_check_flags *result,
320 struct dma_chan *chan = xor_val_chan(submit, dest, src_list, src_cnt, len);
325 BUG_ON(src_cnt <= 1);
328 unmap = dmaengine_get_unmap_data(device->dev, src_cnt, GFP_NOWAIT);
330 if (unmap && src_cnt <= device->max_xor &&
331 dma_xor_aligned_offsets(device, offset, src_offs, src_cnt, len)) {
342 for (i = 0; i < src_cnt; i++) {
350 tx = device->device_prep_dma_xor_val(chan, unmap->addr, src_cnt,
359 unmap->addr, src_cnt, len, result,
369 WARN_ONCE(device && src_cnt <= device->max_xor,
377 src_cnt, len, submit);
397 * @src_cnt: number of source pages
410 int src_cnt, size_t len, enum sum_check_flags *result,
413 return async_xor_val_offs(dest, offset, src_list, NULL, src_cnt,