Lines Matching defs:frame
79 * store the lower bits (the offset into the physical frame) because we can
190 assert(extract_paddr(r) != 0 && "a region includes physical frame 0");
419 /* We've hit a physical contiguity break (== frame
452 for (dma_frame_t **frame = __start__dma_frames;
453 frame < __stop__dma_frames; frame++) {
454 uintptr_t base = (uintptr_t)ptr & ~MASK(ffs((*frame)->size) - 1);
455 if (base == (*frame)->vaddr) {
456 return *frame;
465 dma_frame_t *frame = get_frame_desc(ptr);
466 uintptr_t offset = (uintptr_t)ptr & MASK(ffs(frame->size) - 1);
467 if (frame) {
468 seL4_ARCH_Page_GetAddress_t res = seL4_ARCH_Page_GetAddress(frame->cap);
472 .description = "failed to reverse virtual mapping to a DMA frame",
488 for (dma_frame_t **frame = __start__dma_frames;
489 frame < __stop__dma_frames; frame++) {
490 uintptr_t base = (uintptr_t)ptr & ~MASK(ffs((*frame)->size) - 1);
491 if (base == (*frame)->vaddr) {
492 return (*frame)->cap;
751 * an uncached frame we simply lose some performance in (1) and make (2) a
782 dma_frame_t *frame = get_frame_desc(addr);
783 if (frame == NULL) {
788 /* If the frame is uncached then the cache op isn't required. This assumes
793 if (frame->cached == 0) {
796 seL4_CPtr frame_cap = frame->cap;
803 size_t page_size_of_region = frame->size;
805 if ((frame_start_offset + size) > frame->size) {