Lines Matching refs:a_out
63 const struct exec *a_out = (const struct exec *) imgp->image_header;
71 if (((a_out->a_magic >> 16) & 0xff) != 0x64)
77 switch ((int)(a_out->a_magic & 0xffff)) {
89 bss_size = round_page(a_out->a_bss);
92 (u_long)a_out->a_text, (u_long)a_out->a_data, bss_size);
98 if (a_out->a_entry < virtual_offset ||
99 a_out->a_entry >= virtual_offset + a_out->a_text ||
100 a_out->a_text & PAGE_MASK || a_out->a_data & PAGE_MASK)
104 if (a_out->a_data + a_out->a_text > imgp->attr->va_size)
110 if (a_out->a_text > maxtsiz ||
111 a_out->a_data + bss_size > lim_cur_proc(imgp->proc, RLIMIT_DATA) ||
112 racct_set(imgp->proc, RACCT_DATA, a_out->a_data + bss_size) != 0) {
142 a_out->a_text + a_out->a_data + bss_size, 0, VMFS_NO_SPACE,
148 a_out->a_text + a_out->a_data, UIO_USERSPACE, 0,
161 vmaddr + a_out->a_text, VM_PROT_EXECUTE|VM_PROT_READ, TRUE);
173 a_out->a_text + a_out->a_data,
182 (u_long)a_out->a_text + (u_long)a_out->a_data);
187 error = vm_map_protect(&vmspace->vm_map, vmaddr + a_out->a_text,
188 vmaddr + a_out->a_text + a_out->a_data, VM_PROT_ALL, FALSE);
196 vmaddr = virtual_offset + a_out->a_text + a_out->a_data;
208 vmspace->vm_tsize = round_page(a_out->a_text) >> PAGE_SHIFT;
209 vmspace->vm_dsize = round_page(a_out->a_data + bss_size) >> PAGE_SHIFT;
212 (caddr_t)(void *)(uintptr_t)(virtual_offset + a_out->a_text);
216 imgp->entry_addr = a_out->a_entry;