• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /macosx-10.5.8/xnu-1228.15.4/bsd/vfs/

Lines Matching refs:cl

2436 	struct cl_extent cl;
2518 cl.b_addr = (daddr64_t)(upl_f_offset / PAGE_SIZE_64);
2783 cl.e_addr = (daddr64_t)((upl_f_offset + (off_t)upl_size) / PAGE_SIZE_64);
2806 sparse_cluster_add(wbp, vp, &cl, newEOF, callback, callback_arg);
2845 if (cl.b_addr >= wbp->cl_clusters[cl_index].b_addr) {
2849 if (cl.e_addr <= (wbp->cl_clusters[cl_index].b_addr + max_cluster_pgcount)) {
2854 if (cl.e_addr > wbp->cl_clusters[cl_index].e_addr)
2858 wbp->cl_clusters[cl_index].e_addr = cl.e_addr;
2861 if (cl.b_addr < (wbp->cl_clusters[cl_index].b_addr + max_cluster_pgcount)) {
2874 cl.b_addr = wbp->cl_clusters[cl_index].e_addr;
2888 if ((wbp->cl_clusters[cl_index].e_addr - cl.b_addr) <= max_cluster_pgcount) {
2895 wbp->cl_clusters[cl_index].b_addr = cl.b_addr;
2897 if (cl.e_addr > wbp->cl_clusters[cl_index].e_addr) {
2905 wbp->cl_clusters[cl_index].e_addr = cl.e_addr;
2919 if (cl.e_addr > wbp->cl_clusters[cl_index].e_addr - max_cluster_pgcount) {
2927 cl.e_addr = wbp->cl_clusters[cl_index].b_addr;
2983 sparse_cluster_add(wbp, vp, &cl, newEOF, callback, callback_arg);
3002 wbp->cl_clusters[wbp->cl_number].b_addr = cl.b_addr;
3003 wbp->cl_clusters[wbp->cl_number].e_addr = cl.e_addr;
3028 retval = cluster_push_now(vp, &cl, newEOF, flags, callback, callback_arg);
4607 struct cl_extent cl;
4623 cl.b_addr = l_clusters[cl_index].b_addr;
4624 cl.e_addr = l_clusters[cl_index].e_addr;
4626 cluster_push_now(vp, &cl, EOF, flags, callback, callback_arg);
4702 cluster_push_now(vnode_t vp, struct cl_extent *cl, off_t EOF, int flags, int (*callback)(buf_t, void *), void *callback_arg)
4727 (int)cl->b_addr, (int)cl->e_addr, (int)EOF, flags, 0);
4729 if ((pages_in_upl = (int)(cl->e_addr - cl->b_addr)) == 0) {
4735 upl_f_offset = (off_t)(cl->b_addr * PAGE_SIZE_64);
4880 struct cl_extent cl;
4882 for (cl.b_addr = wbp->cl_clusters[cl_index].b_addr; cl.b_addr < wbp->cl_clusters[cl_index].e_addr; cl.b_addr++) {
4884 if (ubc_page_op(vp, (off_t)(cl.b_addr * PAGE_SIZE_64), 0, NULL, &flags) == KERN_SUCCESS) {
4886 cl.e_addr = cl.b_addr + 1;
4888 sparse_cluster_add(wbp, vp, &cl, EOF, callback, callback_arg);
4905 struct cl_extent cl;
4918 cl.b_addr = (daddr64_t)(offset / PAGE_SIZE_64);
4919 cl.e_addr = (daddr64_t)((offset + length) / PAGE_SIZE_64);
4921 wbp->cl_scdirty -= (int)(cl.e_addr - cl.b_addr);
4923 cluster_push_now(vp, &cl, EOF, push_flag & IO_PASSIVE, callback, callback_arg);
4936 sparse_cluster_add(struct cl_writebehind *wbp, vnode_t vp, struct cl_extent *cl, off_t EOF, int (*callback)(buf_t, void *), void *callback_arg)
4942 KERNEL_DEBUG((FSDBG_CODE(DBG_FSRW, 80)) | DBG_FUNC_START, (int)wbp->cl_scmap, wbp->cl_scdirty, (int)cl->b_addr, (int)cl->e_addr, 0);
4944 offset = (off_t)(cl->b_addr * PAGE_SIZE_64);
4945 length = ((u_int)(cl->e_addr - cl->b_addr)) * PAGE_SIZE;