Lines Matching refs:buf_size

62 				      u32 buf_size, enum dma_data_direction dir,
69 dma_sync(stat->isp->dev, buf->dma_addr + (buf_size & PAGE_MASK),
70 buf_size & ~PAGE_MASK, MAGIC_SIZE, dir);
75 u32 buf_size,
81 __isp_stat_buf_sync_magic(stat, buf, buf_size, dir,
87 u32 buf_size,
93 __isp_stat_buf_sync_magic(stat, buf, buf_size, dir,
100 const u32 buf_size = IS_H3A_AF(stat) ?
101 buf->buf_size + AF_EXTRA_DATA : buf->buf_size;
106 isp_stat_buf_sync_magic_for_cpu(stat, buf, buf_size, DMA_FROM_DEVICE);
121 for (w = buf->virt_addr + buf_size, end = w + MAGIC_SIZE;
131 isp_stat_buf_sync_magic_for_device(stat, buf, buf_size,
140 const u32 buf_size = IS_H3A_AF(stat) ?
141 stat->buf_size + AF_EXTRA_DATA : stat->buf_size;
143 isp_stat_buf_sync_magic_for_cpu(stat, buf, buf_size, DMA_FROM_DEVICE);
147 * buf->buf_size is set only after the buffer is queued. For now the
148 * right buf_size for the current configuration is pointed by
149 * stat->buf_size.
152 memset(buf->virt_addr + buf_size, MAGIC_NUM, MAGIC_SIZE);
154 isp_stat_buf_sync_magic_for_device(stat, buf, buf_size,
240 stat->active_buf->buf_size = stat->buf_size;
310 if (buf->buf_size > data->buf_size) {
322 buf->buf_size);
503 data->buf_size = buf->buf_size;
543 u32 buf_size = user_cfg->buf_size;
549 stat->subdev.name, (unsigned long)buf_size);
559 if (buf_size != user_cfg->buf_size)
563 (unsigned long)user_cfg->buf_size);
569 * Hack 2: MAGIC_SIZE is added to buf_size so a magic word can be
577 buf_size = user_cfg->buf_size * 2 + MAGIC_SIZE;
583 buf_size += AF_EXTRA_DATA * (NUM_H3A_RECOVER_BUFS + 2);
587 buf_size += recover_cfg->buf_size *
590 buf_size = PAGE_ALIGN(buf_size);
592 buf_size = PAGE_ALIGN(user_cfg->buf_size + MAGIC_SIZE);
595 ret = isp_stat_bufs_alloc(stat, buf_size);