• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /macosx-10.5.8/xnu-1228.15.4/bsd/vfs/

Lines Matching refs:b_addr

1517 		     (int)extent->b_addr, (int)extent->e_addr, (int)rap->cl_lastr, 0, 0);
1519 if (extent->b_addr == rap->cl_lastr && extent->b_addr == extent->e_addr) {
1524 if (rap->cl_lastr == -1 || (extent->b_addr != rap->cl_lastr && extent->b_addr != (rap->cl_lastr + 1))) {
1560 read_size = (extent->e_addr + 1) - extent->b_addr;
2518 cl.b_addr = (daddr64_t)(upl_f_offset / PAGE_SIZE_64);
2845 if (cl.b_addr >= wbp->cl_clusters[cl_index].b_addr) {
2849 if (cl.e_addr <= (wbp->cl_clusters[cl_index].b_addr + max_cluster_pgcount)) {
2861 if (cl.b_addr < (wbp->cl_clusters[cl_index].b_addr + max_cluster_pgcount)) {
2867 * and update the b_addr for the current write to reflect that
2872 wbp->cl_clusters[cl_index].e_addr = wbp->cl_clusters[cl_index].b_addr + max_cluster_pgcount;
2874 cl.b_addr = wbp->cl_clusters[cl_index].e_addr;
2888 if ((wbp->cl_clusters[cl_index].e_addr - cl.b_addr) <= max_cluster_pgcount) {
2895 wbp->cl_clusters[cl_index].b_addr = cl.b_addr;
2925 wbp->cl_clusters[cl_index].b_addr = wbp->cl_clusters[cl_index].e_addr - max_cluster_pgcount;
2927 cl.e_addr = wbp->cl_clusters[cl_index].b_addr;
3002 wbp->cl_clusters[wbp->cl_number].b_addr = cl.b_addr;
3197 extent.b_addr = uio->uio_offset / PAGE_SIZE_64;
3200 if (rap != NULL && rap->cl_ralen && (rap->cl_lastr == extent.b_addr || (rap->cl_lastr + 1) == extent.b_addr)) {
4557 if (wbp->cl_clusters[cl_index1].b_addr == wbp->cl_clusters[cl_index1].e_addr)
4561 else if (wbp->cl_clusters[cl_index1].b_addr < wbp->cl_clusters[min_index].b_addr)
4566 l_clusters[cl_index].b_addr = wbp->cl_clusters[min_index].b_addr;
4570 wbp->cl_clusters[min_index].b_addr = wbp->cl_clusters[min_index].e_addr;
4591 * note that e_addr is not inclusive, so it will be equal to the b_addr of the next cluster if they
4599 if ((l_clusters[i].e_addr - l_clusters[i].b_addr) != max_cluster_pgcount)
4601 if (l_clusters[i].e_addr != l_clusters[i+1].b_addr)
4623 cl.b_addr = l_clusters[cl_index].b_addr;
4628 l_clusters[cl_index].b_addr = 0;
4655 if (l_clusters[cl_index].b_addr == l_clusters[cl_index].e_addr)
4657 wbp->cl_clusters[cl_index1].b_addr = l_clusters[cl_index].b_addr;
4681 if (l_clusters[cl_index].b_addr == l_clusters[cl_index].e_addr)
4684 wbp->cl_clusters[cl_index1].b_addr = l_clusters[cl_index].b_addr;
4727 (int)cl->b_addr, (int)cl->e_addr, (int)EOF, flags, 0);
4729 if ((pages_in_upl = (int)(cl->e_addr - cl->b_addr)) == 0) {
4735 upl_f_offset = (off_t)(cl->b_addr * PAGE_SIZE_64);
4882 for (cl.b_addr = wbp->cl_clusters[cl_index].b_addr; cl.b_addr < wbp->cl_clusters[cl_index].e_addr; cl.b_addr++) {
4884 if (ubc_page_op(vp, (off_t)(cl.b_addr * PAGE_SIZE_64), 0, NULL, &flags) == KERN_SUCCESS) {
4886 cl.e_addr = cl.b_addr + 1;
4918 cl.b_addr = (daddr64_t)(offset / PAGE_SIZE_64);
4921 wbp->cl_scdirty -= (int)(cl.e_addr - cl.b_addr);
4942 KERNEL_DEBUG((FSDBG_CODE(DBG_FSRW, 80)) | DBG_FUNC_START, (int)wbp->cl_scmap, wbp->cl_scdirty, (int)cl->b_addr, (int)cl->e_addr, 0);
4944 offset = (off_t)(cl->b_addr * PAGE_SIZE_64);
4945 length = ((u_int)(cl->e_addr - cl->b_addr)) * PAGE_SIZE;