94 95#ifdef __i386__ 96#include <machine/cputypes.h> 97#endif 98 99#define BSD_TO_LINUX_SIGNAL(sig) \ 100 (((sig) <= LINUX_SIGTBLSZ) ? bsd_to_linux_signal[_SIG_IDX(sig)] : sig) 101 102static unsigned int linux_to_bsd_resource[LINUX_RLIM_NLIMITS] = { 103 RLIMIT_CPU, RLIMIT_FSIZE, RLIMIT_DATA, RLIMIT_STACK, 104 RLIMIT_CORE, RLIMIT_RSS, RLIMIT_NPROC, RLIMIT_NOFILE, 105 RLIMIT_MEMLOCK, RLIMIT_AS 106}; 107 108struct l_sysinfo { 109 l_long uptime; /* Seconds since boot */ 110 l_ulong loads[3]; /* 1, 5, and 15 minute load averages */ 111#define LINUX_SYSINFO_LOADS_SCALE 65536 112 l_ulong totalram; /* Total usable main memory size */ 113 l_ulong freeram; /* Available memory size */ 114 l_ulong sharedram; /* Amount of shared memory */ 115 l_ulong bufferram; /* Memory used by buffers */ 116 l_ulong totalswap; /* Total swap space size */ 117 l_ulong freeswap; /* swap space still available */ 118 l_ushort procs; /* Number of current processes */ 119 l_ushort pads; 120 l_ulong totalbig; 121 l_ulong freebig; 122 l_uint mem_unit; 123 char _f[20-2*sizeof(l_long)-sizeof(l_int)]; /* padding */ 124}; 125int 126linux_sysinfo(struct thread *td, struct linux_sysinfo_args *args) 127{ 128 struct l_sysinfo sysinfo; 129 vm_object_t object; 130 int i, j; 131 struct timespec ts; 132 133 getnanouptime(&ts); 134 if (ts.tv_nsec != 0) 135 ts.tv_sec++; 136 sysinfo.uptime = ts.tv_sec; 137 138 /* Use the information from the mib to get our load averages */ 139 for (i = 0; i < 3; i++) 140 sysinfo.loads[i] = averunnable.ldavg[i] * 141 LINUX_SYSINFO_LOADS_SCALE / averunnable.fscale; 142 143 sysinfo.totalram = physmem * PAGE_SIZE; 144 sysinfo.freeram = sysinfo.totalram - cnt.v_wire_count * PAGE_SIZE; 145 146 sysinfo.sharedram = 0; 147 mtx_lock(&vm_object_list_mtx); 148 TAILQ_FOREACH(object, &vm_object_list, object_list) 149 if (object->shadow_count > 1) 150 sysinfo.sharedram += object->resident_page_count; 151 mtx_unlock(&vm_object_list_mtx); 152 153 sysinfo.sharedram *= PAGE_SIZE; 154 sysinfo.bufferram = 0; 155 156 swap_pager_status(&i, &j); 157 sysinfo.totalswap = i * PAGE_SIZE; 158 sysinfo.freeswap = (i - j) * PAGE_SIZE; 159 160 sysinfo.procs = nprocs; 161 162 /* The following are only present in newer Linux kernels. */ 163 sysinfo.totalbig = 0; 164 sysinfo.freebig = 0; 165 sysinfo.mem_unit = 1; 166 167 return copyout(&sysinfo, args->info, sizeof(sysinfo)); 168} 169 170int 171linux_alarm(struct thread *td, struct linux_alarm_args *args) 172{ 173 struct itimerval it, old_it; 174 int error; 175 176#ifdef DEBUG 177 if (ldebug(alarm)) 178 printf(ARGS(alarm, "%u"), args->secs); 179#endif 180 181 if (args->secs > 100000000) 182 return (EINVAL); 183 184 it.it_value.tv_sec = (long)args->secs; 185 it.it_value.tv_usec = 0; 186 it.it_interval.tv_sec = 0; 187 it.it_interval.tv_usec = 0; 188 error = kern_setitimer(td, ITIMER_REAL, &it, &old_it); 189 if (error) 190 return (error); 191 if (timevalisset(&old_it.it_value)) { 192 if (old_it.it_value.tv_usec != 0) 193 old_it.it_value.tv_sec++; 194 td->td_retval[0] = old_it.it_value.tv_sec; 195 } 196 return (0); 197} 198 199int 200linux_brk(struct thread *td, struct linux_brk_args *args) 201{ 202 struct vmspace *vm = td->td_proc->p_vmspace; 203 vm_offset_t new, old; 204 struct obreak_args /* { 205 char * nsize; 206 } */ tmp; 207 208#ifdef DEBUG 209 if (ldebug(brk)) 210 printf(ARGS(brk, "%p"), (void *)(uintptr_t)args->dsend); 211#endif 212 old = (vm_offset_t)vm->vm_daddr + ctob(vm->vm_dsize); 213 new = (vm_offset_t)args->dsend; 214 tmp.nsize = (char *)new; 215 if (((caddr_t)new > vm->vm_daddr) && !obreak(td, &tmp)) 216 td->td_retval[0] = (long)new; 217 else 218 td->td_retval[0] = (long)old; 219 220 return 0; 221} 222 223#if defined(__i386__) 224/* XXX: what about amd64/linux32? */ 225 226int 227linux_uselib(struct thread *td, struct linux_uselib_args *args) 228{ 229 struct nameidata ni; 230 struct vnode *vp; 231 struct exec *a_out; 232 struct vattr attr; 233 vm_offset_t vmaddr; 234 unsigned long file_offset; 235 vm_offset_t buffer; 236 unsigned long bss_size; 237 char *library; 238 int error; 239 int locked, vfslocked; 240 241 LCONVPATHEXIST(td, args->library, &library); 242 243#ifdef DEBUG 244 if (ldebug(uselib)) 245 printf(ARGS(uselib, "%s"), library); 246#endif 247 248 a_out = NULL; 249 vfslocked = 0; 250 locked = 0; 251 vp = NULL; 252 253 NDINIT(&ni, LOOKUP, ISOPEN | FOLLOW | LOCKLEAF | MPSAFE | AUDITVNODE1, 254 UIO_SYSSPACE, library, td); 255 error = namei(&ni); 256 LFREEPATH(library); 257 if (error) 258 goto cleanup; 259 260 vp = ni.ni_vp; 261 vfslocked = NDHASGIANT(&ni); 262 NDFREE(&ni, NDF_ONLY_PNBUF); 263 264 /* 265 * From here on down, we have a locked vnode that must be unlocked. 266 * XXX: The code below largely duplicates exec_check_permissions(). 267 */ 268 locked = 1; 269 270 /* Writable? */ 271 if (vp->v_writecount) { 272 error = ETXTBSY; 273 goto cleanup; 274 } 275 276 /* Executable? */ 277 error = VOP_GETATTR(vp, &attr, td->td_ucred, td); 278 if (error) 279 goto cleanup; 280 281 if ((vp->v_mount->mnt_flag & MNT_NOEXEC) || 282 ((attr.va_mode & 0111) == 0) || (attr.va_type != VREG)) { 283 /* EACCESS is what exec(2) returns. */ 284 error = ENOEXEC; 285 goto cleanup; 286 } 287 288 /* Sensible size? */ 289 if (attr.va_size == 0) { 290 error = ENOEXEC; 291 goto cleanup; 292 } 293 294 /* Can we access it? */ 295 error = VOP_ACCESS(vp, VEXEC, td->td_ucred, td); 296 if (error) 297 goto cleanup; 298 299 /* 300 * XXX: This should use vn_open() so that it is properly authorized, 301 * and to reduce code redundancy all over the place here. 302 * XXX: Not really, it duplicates far more of exec_check_permissions() 303 * than vn_open(). 304 */ 305#ifdef MAC 306 error = mac_vnode_check_open(td->td_ucred, vp, FREAD); 307 if (error) 308 goto cleanup; 309#endif 310 error = VOP_OPEN(vp, FREAD, td->td_ucred, td, NULL); 311 if (error) 312 goto cleanup; 313 314 /* Pull in executable header into kernel_map */ 315 error = vm_mmap(kernel_map, (vm_offset_t *)&a_out, PAGE_SIZE, 316 VM_PROT_READ, VM_PROT_READ, 0, OBJT_VNODE, vp, 0); 317 if (error) 318 goto cleanup; 319 320 /* Is it a Linux binary ? */ 321 if (((a_out->a_magic >> 16) & 0xff) != 0x64) { 322 error = ENOEXEC; 323 goto cleanup; 324 } 325 326 /* 327 * While we are here, we should REALLY do some more checks 328 */ 329 330 /* Set file/virtual offset based on a.out variant. */ 331 switch ((int)(a_out->a_magic & 0xffff)) { 332 case 0413: /* ZMAGIC */ 333 file_offset = 1024; 334 break; 335 case 0314: /* QMAGIC */ 336 file_offset = 0; 337 break; 338 default: 339 error = ENOEXEC; 340 goto cleanup; 341 } 342 343 bss_size = round_page(a_out->a_bss); 344 345 /* Check various fields in header for validity/bounds. */ 346 if (a_out->a_text & PAGE_MASK || a_out->a_data & PAGE_MASK) { 347 error = ENOEXEC; 348 goto cleanup; 349 } 350 351 /* text + data can't exceed file size */ 352 if (a_out->a_data + a_out->a_text > attr.va_size) { 353 error = EFAULT; 354 goto cleanup; 355 } 356 357 /* 358 * text/data/bss must not exceed limits 359 * XXX - this is not complete. it should check current usage PLUS 360 * the resources needed by this library. 361 */ 362 PROC_LOCK(td->td_proc); 363 if (a_out->a_text > maxtsiz || 364 a_out->a_data + bss_size > lim_cur(td->td_proc, RLIMIT_DATA)) { 365 PROC_UNLOCK(td->td_proc); 366 error = ENOMEM; 367 goto cleanup; 368 } 369 PROC_UNLOCK(td->td_proc); 370 371 /* 372 * Prevent more writers. 373 * XXX: Note that if any of the VM operations fail below we don't 374 * clear this flag. 375 */ 376 vp->v_vflag |= VV_TEXT; 377 378 /* 379 * Lock no longer needed 380 */ 381 locked = 0; 382 VOP_UNLOCK(vp, 0); 383 VFS_UNLOCK_GIANT(vfslocked); 384 385 /* 386 * Check if file_offset page aligned. Currently we cannot handle 387 * misalinged file offsets, and so we read in the entire image 388 * (what a waste). 389 */ 390 if (file_offset & PAGE_MASK) { 391#ifdef DEBUG 392 printf("uselib: Non page aligned binary %lu\n", file_offset); 393#endif 394 /* Map text+data read/write/execute */ 395 396 /* a_entry is the load address and is page aligned */ 397 vmaddr = trunc_page(a_out->a_entry); 398 399 /* get anon user mapping, read+write+execute */ 400 error = vm_map_find(&td->td_proc->p_vmspace->vm_map, NULL, 0, 401 &vmaddr, a_out->a_text + a_out->a_data, FALSE, VM_PROT_ALL, 402 VM_PROT_ALL, 0); 403 if (error) 404 goto cleanup; 405 406 /* map file into kernel_map */ 407 error = vm_mmap(kernel_map, &buffer, 408 round_page(a_out->a_text + a_out->a_data + file_offset), 409 VM_PROT_READ, VM_PROT_READ, 0, OBJT_VNODE, vp, 410 trunc_page(file_offset)); 411 if (error) 412 goto cleanup; 413 414 /* copy from kernel VM space to user space */ 415 error = copyout(PTRIN(buffer + file_offset), 416 (void *)vmaddr, a_out->a_text + a_out->a_data); 417 418 /* release temporary kernel space */ 419 vm_map_remove(kernel_map, buffer, buffer + 420 round_page(a_out->a_text + a_out->a_data + file_offset)); 421 422 if (error) 423 goto cleanup; 424 } else { 425#ifdef DEBUG 426 printf("uselib: Page aligned binary %lu\n", file_offset); 427#endif 428 /* 429 * for QMAGIC, a_entry is 20 bytes beyond the load address 430 * to skip the executable header 431 */ 432 vmaddr = trunc_page(a_out->a_entry); 433 434 /* 435 * Map it all into the process's space as a single 436 * copy-on-write "data" segment. 437 */ 438 error = vm_mmap(&td->td_proc->p_vmspace->vm_map, &vmaddr, 439 a_out->a_text + a_out->a_data, VM_PROT_ALL, VM_PROT_ALL, 440 MAP_PRIVATE | MAP_FIXED, OBJT_VNODE, vp, file_offset); 441 if (error) 442 goto cleanup; 443 } 444#ifdef DEBUG 445 printf("mem=%08lx = %08lx %08lx\n", (long)vmaddr, ((long *)vmaddr)[0], 446 ((long *)vmaddr)[1]); 447#endif 448 if (bss_size != 0) { 449 /* Calculate BSS start address */ 450 vmaddr = trunc_page(a_out->a_entry) + a_out->a_text + 451 a_out->a_data; 452 453 /* allocate some 'anon' space */ 454 error = vm_map_find(&td->td_proc->p_vmspace->vm_map, NULL, 0, 455 &vmaddr, bss_size, FALSE, VM_PROT_ALL, VM_PROT_ALL, 0); 456 if (error) 457 goto cleanup; 458 } 459 460cleanup: 461 /* Unlock vnode if needed */ 462 if (locked) { 463 VOP_UNLOCK(vp, 0); 464 VFS_UNLOCK_GIANT(vfslocked); 465 } 466 467 /* Release the kernel mapping. */ 468 if (a_out) 469 vm_map_remove(kernel_map, (vm_offset_t)a_out, 470 (vm_offset_t)a_out + PAGE_SIZE); 471 472 return error; 473} 474 475#endif /* __i386__ */ 476 477int 478linux_select(struct thread *td, struct linux_select_args *args) 479{ 480 l_timeval ltv; 481 struct timeval tv0, tv1, utv, *tvp; 482 int error; 483 484#ifdef DEBUG 485 if (ldebug(select)) 486 printf(ARGS(select, "%d, %p, %p, %p, %p"), args->nfds, 487 (void *)args->readfds, (void *)args->writefds, 488 (void *)args->exceptfds, (void *)args->timeout); 489#endif 490 491 /* 492 * Store current time for computation of the amount of 493 * time left. 494 */ 495 if (args->timeout) { 496 if ((error = copyin(args->timeout, <v, sizeof(ltv)))) 497 goto select_out; 498 utv.tv_sec = ltv.tv_sec; 499 utv.tv_usec = ltv.tv_usec; 500#ifdef DEBUG 501 if (ldebug(select)) 502 printf(LMSG("incoming timeout (%jd/%ld)"), 503 (intmax_t)utv.tv_sec, utv.tv_usec); 504#endif 505 506 if (itimerfix(&utv)) { 507 /* 508 * The timeval was invalid. Convert it to something 509 * valid that will act as it does under Linux. 510 */ 511 utv.tv_sec += utv.tv_usec / 1000000; 512 utv.tv_usec %= 1000000; 513 if (utv.tv_usec < 0) { 514 utv.tv_sec -= 1; 515 utv.tv_usec += 1000000; 516 } 517 if (utv.tv_sec < 0) 518 timevalclear(&utv); 519 } 520 microtime(&tv0); 521 tvp = &utv; 522 } else 523 tvp = NULL; 524 525 error = kern_select(td, args->nfds, args->readfds, args->writefds, 526 args->exceptfds, tvp); 527 528#ifdef DEBUG 529 if (ldebug(select)) 530 printf(LMSG("real select returns %d"), error); 531#endif 532 if (error) { 533 /* 534 * See fs/select.c in the Linux kernel. Without this, 535 * Maelstrom doesn't work. 536 */ 537 if (error == ERESTART) 538 error = EINTR; 539 goto select_out; 540 } 541 542 if (args->timeout) { 543 if (td->td_retval[0]) { 544 /* 545 * Compute how much time was left of the timeout, 546 * by subtracting the current time and the time 547 * before we started the call, and subtracting 548 * that result from the user-supplied value. 549 */ 550 microtime(&tv1); 551 timevalsub(&tv1, &tv0); 552 timevalsub(&utv, &tv1); 553 if (utv.tv_sec < 0) 554 timevalclear(&utv); 555 } else 556 timevalclear(&utv); 557#ifdef DEBUG 558 if (ldebug(select)) 559 printf(LMSG("outgoing timeout (%jd/%ld)"), 560 (intmax_t)utv.tv_sec, utv.tv_usec); 561#endif 562 ltv.tv_sec = utv.tv_sec; 563 ltv.tv_usec = utv.tv_usec; 564 if ((error = copyout(<v, args->timeout, sizeof(ltv)))) 565 goto select_out; 566 } 567 568select_out: 569#ifdef DEBUG 570 if (ldebug(select)) 571 printf(LMSG("select_out -> %d"), error); 572#endif 573 return error; 574} 575 576int 577linux_mremap(struct thread *td, struct linux_mremap_args *args) 578{ 579 struct munmap_args /* { 580 void *addr; 581 size_t len; 582 } */ bsd_args; 583 int error = 0; 584 585#ifdef DEBUG 586 if (ldebug(mremap)) 587 printf(ARGS(mremap, "%p, %08lx, %08lx, %08lx"), 588 (void *)(uintptr_t)args->addr, 589 (unsigned long)args->old_len, 590 (unsigned long)args->new_len, 591 (unsigned long)args->flags); 592#endif 593 594 if (args->flags & ~(LINUX_MREMAP_FIXED | LINUX_MREMAP_MAYMOVE)) { 595 td->td_retval[0] = 0; 596 return (EINVAL); 597 } 598 599 /* 600 * Check for the page alignment. 601 * Linux defines PAGE_MASK to be FreeBSD ~PAGE_MASK. 602 */ 603 if (args->addr & PAGE_MASK) { 604 td->td_retval[0] = 0; 605 return (EINVAL); 606 } 607 608 args->new_len = round_page(args->new_len); 609 args->old_len = round_page(args->old_len); 610 611 if (args->new_len > args->old_len) { 612 td->td_retval[0] = 0; 613 return ENOMEM; 614 } 615 616 if (args->new_len < args->old_len) { 617 bsd_args.addr = 618 (caddr_t)((uintptr_t)args->addr + args->new_len); 619 bsd_args.len = args->old_len - args->new_len; 620 error = munmap(td, &bsd_args); 621 } 622 623 td->td_retval[0] = error ? 0 : (uintptr_t)args->addr; 624 return error; 625} 626 627#define LINUX_MS_ASYNC 0x0001 628#define LINUX_MS_INVALIDATE 0x0002 629#define LINUX_MS_SYNC 0x0004 630 631int 632linux_msync(struct thread *td, struct linux_msync_args *args) 633{ 634 struct msync_args bsd_args; 635 636 bsd_args.addr = (caddr_t)(uintptr_t)args->addr; 637 bsd_args.len = (uintptr_t)args->len; 638 bsd_args.flags = args->fl & ~LINUX_MS_SYNC; 639 640 return msync(td, &bsd_args); 641} 642 643int 644linux_time(struct thread *td, struct linux_time_args *args) 645{ 646 struct timeval tv; 647 l_time_t tm; 648 int error; 649 650#ifdef DEBUG 651 if (ldebug(time)) 652 printf(ARGS(time, "*")); 653#endif 654 655 microtime(&tv); 656 tm = tv.tv_sec; 657 if (args->tm && (error = copyout(&tm, args->tm, sizeof(tm)))) 658 return error; 659 td->td_retval[0] = tm; 660 return 0; 661} 662 663struct l_times_argv { 664 l_long tms_utime; 665 l_long tms_stime; 666 l_long tms_cutime; 667 l_long tms_cstime; 668}; 669 670#define CLK_TCK 100 /* Linux uses 100 */ 671 672#define CONVTCK(r) (r.tv_sec * CLK_TCK + r.tv_usec / (1000000 / CLK_TCK)) 673 674int 675linux_times(struct thread *td, struct linux_times_args *args) 676{ 677 struct timeval tv, utime, stime, cutime, cstime; 678 struct l_times_argv tms; 679 struct proc *p; 680 int error; 681 682#ifdef DEBUG 683 if (ldebug(times)) 684 printf(ARGS(times, "*")); 685#endif 686 687 if (args->buf != NULL) { 688 p = td->td_proc; 689 PROC_LOCK(p); 690 PROC_SLOCK(p); 691 calcru(p, &utime, &stime); 692 PROC_SUNLOCK(p); 693 calccru(p, &cutime, &cstime); 694 PROC_UNLOCK(p); 695 696 tms.tms_utime = CONVTCK(utime); 697 tms.tms_stime = CONVTCK(stime); 698 699 tms.tms_cutime = CONVTCK(cutime); 700 tms.tms_cstime = CONVTCK(cstime); 701 702 if ((error = copyout(&tms, args->buf, sizeof(tms)))) 703 return error; 704 } 705 706 microuptime(&tv); 707 td->td_retval[0] = (int)CONVTCK(tv); 708 return 0; 709} 710 711int 712linux_newuname(struct thread *td, struct linux_newuname_args *args) 713{ 714 struct l_new_utsname utsname; 715 char osname[LINUX_MAX_UTSNAME]; 716 char osrelease[LINUX_MAX_UTSNAME]; 717 char *p; 718 719#ifdef DEBUG 720 if (ldebug(newuname)) 721 printf(ARGS(newuname, "*")); 722#endif 723 724 linux_get_osname(td, osname); 725 linux_get_osrelease(td, osrelease); 726 727 bzero(&utsname, sizeof(utsname)); 728 strlcpy(utsname.sysname, osname, LINUX_MAX_UTSNAME); 729 getcredhostname(td->td_ucred, utsname.nodename, LINUX_MAX_UTSNAME); 730 strlcpy(utsname.release, osrelease, LINUX_MAX_UTSNAME); 731 strlcpy(utsname.version, version, LINUX_MAX_UTSNAME); 732 for (p = utsname.version; *p != '\0'; ++p) 733 if (*p == '\n') { 734 *p = '\0'; 735 break; 736 } 737#ifdef __i386__ 738 { 739 const char *class; 740 741 switch (cpu_class) { 742 case CPUCLASS_686: 743 class = "i686"; 744 break; 745 case CPUCLASS_586: 746 class = "i586"; 747 break; 748 case CPUCLASS_486: 749 class = "i486"; 750 break; 751 default: 752 class = "i386"; 753 } 754 strlcpy(utsname.machine, class, LINUX_MAX_UTSNAME); 755 } 756#elif defined(__amd64__) /* XXX: Linux can change 'personality'. */ 757#ifdef COMPAT_LINUX32 758 strlcpy(utsname.machine, "i686", LINUX_MAX_UTSNAME); 759#else 760 strlcpy(utsname.machine, "x86_64", LINUX_MAX_UTSNAME); 761#endif /* COMPAT_LINUX32 */ 762#else /* something other than i386 or amd64 - assume we and Linux agree */ 763 strlcpy(utsname.machine, machine, LINUX_MAX_UTSNAME); 764#endif /* __i386__ */ 765 strlcpy(utsname.domainname, domainname, LINUX_MAX_UTSNAME); 766 767 return (copyout(&utsname, args->buf, sizeof(utsname))); 768} 769 770#if defined(__i386__) || (defined(__amd64__) && defined(COMPAT_LINUX32)) 771struct l_utimbuf { 772 l_time_t l_actime; 773 l_time_t l_modtime; 774}; 775 776int 777linux_utime(struct thread *td, struct linux_utime_args *args) 778{ 779 struct timeval tv[2], *tvp; 780 struct l_utimbuf lut; 781 char *fname; 782 int error; 783 784 LCONVPATHEXIST(td, args->fname, &fname); 785 786#ifdef DEBUG 787 if (ldebug(utime)) 788 printf(ARGS(utime, "%s, *"), fname); 789#endif 790 791 if (args->times) { 792 if ((error = copyin(args->times, &lut, sizeof lut))) { 793 LFREEPATH(fname); 794 return error; 795 } 796 tv[0].tv_sec = lut.l_actime; 797 tv[0].tv_usec = 0; 798 tv[1].tv_sec = lut.l_modtime; 799 tv[1].tv_usec = 0; 800 tvp = tv; 801 } else 802 tvp = NULL; 803 804 error = kern_utimes(td, fname, UIO_SYSSPACE, tvp, UIO_SYSSPACE); 805 LFREEPATH(fname); 806 return (error); 807} 808 809int 810linux_utimes(struct thread *td, struct linux_utimes_args *args) 811{ 812 l_timeval ltv[2]; 813 struct timeval tv[2], *tvp = NULL; 814 char *fname; 815 int error; 816 817 LCONVPATHEXIST(td, args->fname, &fname); 818 819#ifdef DEBUG 820 if (ldebug(utimes)) 821 printf(ARGS(utimes, "%s, *"), fname); 822#endif 823 824 if (args->tptr != NULL) { 825 if ((error = copyin(args->tptr, ltv, sizeof ltv))) { 826 LFREEPATH(fname); 827 return (error); 828 } 829 tv[0].tv_sec = ltv[0].tv_sec; 830 tv[0].tv_usec = ltv[0].tv_usec; 831 tv[1].tv_sec = ltv[1].tv_sec; 832 tv[1].tv_usec = ltv[1].tv_usec; 833 tvp = tv; 834 } 835 836 error = kern_utimes(td, fname, UIO_SYSSPACE, tvp, UIO_SYSSPACE); 837 LFREEPATH(fname); 838 return (error); 839} 840 841int 842linux_futimesat(struct thread *td, struct linux_futimesat_args *args) 843{ 844 l_timeval ltv[2]; 845 struct timeval tv[2], *tvp = NULL; 846 char *fname; 847 int error, dfd; 848 849 dfd = (args->dfd == LINUX_AT_FDCWD) ? AT_FDCWD : args->dfd; 850 LCONVPATHEXIST_AT(td, args->filename, &fname, dfd); 851 852#ifdef DEBUG 853 if (ldebug(futimesat)) 854 printf(ARGS(futimesat, "%s, *"), fname); 855#endif 856 857 if (args->utimes != NULL) { 858 if ((error = copyin(args->utimes, ltv, sizeof ltv))) { 859 LFREEPATH(fname); 860 return (error); 861 } 862 tv[0].tv_sec = ltv[0].tv_sec; 863 tv[0].tv_usec = ltv[0].tv_usec; 864 tv[1].tv_sec = ltv[1].tv_sec; 865 tv[1].tv_usec = ltv[1].tv_usec; 866 tvp = tv; 867 } 868 869 error = kern_utimesat(td, dfd, fname, UIO_SYSSPACE, tvp, UIO_SYSSPACE); 870 LFREEPATH(fname); 871 return (error); 872} 873#endif /* __i386__ || (__amd64__ && COMPAT_LINUX32) */ 874 875#define __WCLONE 0x80000000 876 877int 878linux_waitpid(struct thread *td, struct linux_waitpid_args *args) 879{ 880 int error, options, tmpstat; 881 882#ifdef DEBUG 883 if (ldebug(waitpid)) 884 printf(ARGS(waitpid, "%d, %p, %d"), 885 args->pid, (void *)args->status, args->options); 886#endif 887 /* 888 * this is necessary because the test in kern_wait doesn't work 889 * because we mess with the options here 890 */ 891 if (args->options & ~(WUNTRACED | WNOHANG | WCONTINUED | __WCLONE)) 892 return (EINVAL); 893 894 options = (args->options & (WNOHANG | WUNTRACED)); 895 /* WLINUXCLONE should be equal to __WCLONE, but we make sure */ 896 if (args->options & __WCLONE) 897 options |= WLINUXCLONE; 898 899 error = kern_wait(td, args->pid, &tmpstat, options, NULL); 900 if (error) 901 return error; 902 903 if (args->status) { 904 tmpstat &= 0xffff; 905 if (WIFSIGNALED(tmpstat)) 906 tmpstat = (tmpstat & 0xffffff80) | 907 BSD_TO_LINUX_SIGNAL(WTERMSIG(tmpstat)); 908 else if (WIFSTOPPED(tmpstat)) 909 tmpstat = (tmpstat & 0xffff00ff) | 910 (BSD_TO_LINUX_SIGNAL(WSTOPSIG(tmpstat)) << 8); 911 return copyout(&tmpstat, args->status, sizeof(int)); 912 } 913 914 return 0; 915} 916 917int 918linux_wait4(struct thread *td, struct linux_wait4_args *args) 919{ 920 int error, options, tmpstat; 921 struct rusage ru, *rup; 922 struct proc *p; 923 924#ifdef DEBUG 925 if (ldebug(wait4)) 926 printf(ARGS(wait4, "%d, %p, %d, %p"), 927 args->pid, (void *)args->status, args->options, 928 (void *)args->rusage); 929#endif 930 931 options = (args->options & (WNOHANG | WUNTRACED)); 932 /* WLINUXCLONE should be equal to __WCLONE, but we make sure */ 933 if (args->options & __WCLONE) 934 options |= WLINUXCLONE; 935 936 if (args->rusage != NULL) 937 rup = &ru; 938 else 939 rup = NULL; 940 error = kern_wait(td, args->pid, &tmpstat, options, rup); 941 if (error) 942 return error; 943 944 p = td->td_proc; 945 PROC_LOCK(p); 946 sigqueue_delete(&p->p_sigqueue, SIGCHLD); 947 PROC_UNLOCK(p); 948 949 if (args->status) { 950 tmpstat &= 0xffff; 951 if (WIFSIGNALED(tmpstat)) 952 tmpstat = (tmpstat & 0xffffff80) | 953 BSD_TO_LINUX_SIGNAL(WTERMSIG(tmpstat)); 954 else if (WIFSTOPPED(tmpstat)) 955 tmpstat = (tmpstat & 0xffff00ff) | 956 (BSD_TO_LINUX_SIGNAL(WSTOPSIG(tmpstat)) << 8); 957 error = copyout(&tmpstat, args->status, sizeof(int)); 958 } 959 if (args->rusage != NULL && error == 0) 960 error = copyout(&ru, args->rusage, sizeof(ru)); 961 962 return (error); 963} 964 965int 966linux_mknod(struct thread *td, struct linux_mknod_args *args) 967{ 968 char *path; 969 int error; 970 971 LCONVPATHCREAT(td, args->path, &path); 972 973#ifdef DEBUG 974 if (ldebug(mknod)) 975 printf(ARGS(mknod, "%s, %d, %d"), path, args->mode, args->dev); 976#endif 977 978 switch (args->mode & S_IFMT) { 979 case S_IFIFO: 980 case S_IFSOCK: 981 error = kern_mkfifo(td, path, UIO_SYSSPACE, args->mode); 982 break; 983 984 case S_IFCHR: 985 case S_IFBLK: 986 error = kern_mknod(td, path, UIO_SYSSPACE, args->mode, 987 args->dev); 988 break; 989 990 case S_IFDIR: 991 error = EPERM; 992 break; 993 994 case 0: 995 args->mode |= S_IFREG; 996 /* FALLTHROUGH */ 997 case S_IFREG: 998 error = kern_open(td, path, UIO_SYSSPACE, 999 O_WRONLY | O_CREAT | O_TRUNC, args->mode); 1000 if (error == 0) 1001 kern_close(td, td->td_retval[0]); 1002 break; 1003 1004 default: 1005 error = EINVAL; 1006 break; 1007 } 1008 LFREEPATH(path); 1009 return (error); 1010} 1011 1012int 1013linux_mknodat(struct thread *td, struct linux_mknodat_args *args) 1014{ 1015 char *path; 1016 int error, dfd; 1017 1018 dfd = (args->dfd == LINUX_AT_FDCWD) ? AT_FDCWD : args->dfd; 1019 LCONVPATHCREAT_AT(td, args->filename, &path, dfd); 1020 1021#ifdef DEBUG 1022 if (ldebug(mknodat)) 1023 printf(ARGS(mknodat, "%s, %d, %d"), path, args->mode, args->dev); 1024#endif 1025 1026 switch (args->mode & S_IFMT) { 1027 case S_IFIFO: 1028 case S_IFSOCK: 1029 error = kern_mkfifoat(td, dfd, path, UIO_SYSSPACE, args->mode); 1030 break; 1031 1032 case S_IFCHR: 1033 case S_IFBLK: 1034 error = kern_mknodat(td, dfd, path, UIO_SYSSPACE, args->mode, 1035 args->dev); 1036 break; 1037 1038 case S_IFDIR: 1039 error = EPERM; 1040 break; 1041 1042 case 0: 1043 args->mode |= S_IFREG; 1044 /* FALLTHROUGH */ 1045 case S_IFREG: 1046 error = kern_openat(td, dfd, path, UIO_SYSSPACE, 1047 O_WRONLY | O_CREAT | O_TRUNC, args->mode); 1048 if (error == 0) 1049 kern_close(td, td->td_retval[0]); 1050 break; 1051 1052 default: 1053 error = EINVAL; 1054 break; 1055 } 1056 LFREEPATH(path); 1057 return (error); 1058} 1059 1060/* 1061 * UGH! This is just about the dumbest idea I've ever heard!! 1062 */ 1063int 1064linux_personality(struct thread *td, struct linux_personality_args *args) 1065{ 1066#ifdef DEBUG 1067 if (ldebug(personality)) 1068 printf(ARGS(personality, "%lu"), (unsigned long)args->per); 1069#endif 1070 if (args->per != 0) 1071 return EINVAL; 1072 1073 /* Yes Jim, it's still a Linux... */ 1074 td->td_retval[0] = 0; 1075 return 0; 1076} 1077 1078struct l_itimerval { 1079 l_timeval it_interval; 1080 l_timeval it_value; 1081}; 1082 1083#define B2L_ITIMERVAL(bip, lip) \ 1084 (bip)->it_interval.tv_sec = (lip)->it_interval.tv_sec; \ 1085 (bip)->it_interval.tv_usec = (lip)->it_interval.tv_usec; \ 1086 (bip)->it_value.tv_sec = (lip)->it_value.tv_sec; \ 1087 (bip)->it_value.tv_usec = (lip)->it_value.tv_usec; 1088 1089int 1090linux_setitimer(struct thread *td, struct linux_setitimer_args *uap) 1091{ 1092 int error; 1093 struct l_itimerval ls; 1094 struct itimerval aitv, oitv; 1095 1096#ifdef DEBUG 1097 if (ldebug(setitimer)) 1098 printf(ARGS(setitimer, "%p, %p"), 1099 (void *)uap->itv, (void *)uap->oitv); 1100#endif 1101 1102 if (uap->itv == NULL) { 1103 uap->itv = uap->oitv; 1104 return (linux_getitimer(td, (struct linux_getitimer_args *)uap)); 1105 } 1106 1107 error = copyin(uap->itv, &ls, sizeof(ls)); 1108 if (error != 0) 1109 return (error); 1110 B2L_ITIMERVAL(&aitv, &ls); 1111#ifdef DEBUG 1112 if (ldebug(setitimer)) { 1113 printf("setitimer: value: sec: %jd, usec: %ld\n", 1114 (intmax_t)aitv.it_value.tv_sec, aitv.it_value.tv_usec); 1115 printf("setitimer: interval: sec: %jd, usec: %ld\n", 1116 (intmax_t)aitv.it_interval.tv_sec, aitv.it_interval.tv_usec); 1117 } 1118#endif 1119 error = kern_setitimer(td, uap->which, &aitv, &oitv); 1120 if (error != 0 || uap->oitv == NULL) 1121 return (error); 1122 B2L_ITIMERVAL(&ls, &oitv); 1123 1124 return (copyout(&ls, uap->oitv, sizeof(ls))); 1125} 1126 1127int 1128linux_getitimer(struct thread *td, struct linux_getitimer_args *uap) 1129{ 1130 int error; 1131 struct l_itimerval ls; 1132 struct itimerval aitv; 1133 1134#ifdef DEBUG 1135 if (ldebug(getitimer)) 1136 printf(ARGS(getitimer, "%p"), (void *)uap->itv); 1137#endif 1138 error = kern_getitimer(td, uap->which, &aitv); 1139 if (error != 0) 1140 return (error); 1141 B2L_ITIMERVAL(&ls, &aitv); 1142 return (copyout(&ls, uap->itv, sizeof(ls))); 1143} 1144 1145int 1146linux_nice(struct thread *td, struct linux_nice_args *args) 1147{ 1148 struct setpriority_args bsd_args; 1149 1150 bsd_args.which = PRIO_PROCESS; 1151 bsd_args.who = 0; /* current process */ 1152 bsd_args.prio = args->inc; 1153 return setpriority(td, &bsd_args); 1154} 1155 1156int 1157linux_setgroups(struct thread *td, struct linux_setgroups_args *args) 1158{ 1159 struct ucred *newcred, *oldcred; 1160 l_gid_t linux_gidset[NGROUPS]; 1161 gid_t *bsd_gidset; 1162 int ngrp, error; 1163 struct proc *p; 1164 1165 ngrp = args->gidsetsize; 1166 if (ngrp < 0 || ngrp >= NGROUPS) 1167 return (EINVAL); 1168 error = copyin(args->grouplist, linux_gidset, ngrp * sizeof(l_gid_t)); 1169 if (error) 1170 return (error); 1171 newcred = crget(); 1172 p = td->td_proc; 1173 PROC_LOCK(p); 1174 oldcred = p->p_ucred; 1175 1176 /* 1177 * cr_groups[0] holds egid. Setting the whole set from 1178 * the supplied set will cause egid to be changed too. 1179 * Keep cr_groups[0] unchanged to prevent that. 1180 */ 1181 1182 if ((error = priv_check_cred(oldcred, PRIV_CRED_SETGROUPS, 0)) != 0) { 1183 PROC_UNLOCK(p); 1184 crfree(newcred); 1185 return (error); 1186 } 1187 1188 crcopy(newcred, oldcred); 1189 if (ngrp > 0) { 1190 newcred->cr_ngroups = ngrp + 1; 1191 1192 bsd_gidset = newcred->cr_groups; 1193 ngrp--; 1194 while (ngrp >= 0) { 1195 bsd_gidset[ngrp + 1] = linux_gidset[ngrp]; 1196 ngrp--; 1197 } 1198 } else 1199 newcred->cr_ngroups = 1; 1200 1201 setsugid(p); 1202 p->p_ucred = newcred; 1203 PROC_UNLOCK(p); 1204 crfree(oldcred); 1205 return (0); 1206} 1207 1208int 1209linux_getgroups(struct thread *td, struct linux_getgroups_args *args) 1210{ 1211 struct ucred *cred; 1212 l_gid_t linux_gidset[NGROUPS]; 1213 gid_t *bsd_gidset; 1214 int bsd_gidsetsz, ngrp, error; 1215 1216 cred = td->td_ucred; 1217 bsd_gidset = cred->cr_groups; 1218 bsd_gidsetsz = cred->cr_ngroups - 1; 1219 1220 /* 1221 * cr_groups[0] holds egid. Returning the whole set 1222 * here will cause a duplicate. Exclude cr_groups[0] 1223 * to prevent that. 1224 */ 1225 1226 if ((ngrp = args->gidsetsize) == 0) { 1227 td->td_retval[0] = bsd_gidsetsz; 1228 return (0); 1229 } 1230 1231 if (ngrp < bsd_gidsetsz) 1232 return (EINVAL); 1233 1234 ngrp = 0; 1235 while (ngrp < bsd_gidsetsz) { 1236 linux_gidset[ngrp] = bsd_gidset[ngrp + 1]; 1237 ngrp++; 1238 } 1239 1240 if ((error = copyout(linux_gidset, args->grouplist, 1241 ngrp * sizeof(l_gid_t)))) 1242 return (error); 1243 1244 td->td_retval[0] = ngrp; 1245 return (0); 1246} 1247 1248int 1249linux_setrlimit(struct thread *td, struct linux_setrlimit_args *args) 1250{ 1251 struct rlimit bsd_rlim; 1252 struct l_rlimit rlim; 1253 u_int which; 1254 int error; 1255 1256#ifdef DEBUG 1257 if (ldebug(setrlimit)) 1258 printf(ARGS(setrlimit, "%d, %p"), 1259 args->resource, (void *)args->rlim); 1260#endif 1261 1262 if (args->resource >= LINUX_RLIM_NLIMITS) 1263 return (EINVAL); 1264 1265 which = linux_to_bsd_resource[args->resource]; 1266 if (which == -1) 1267 return (EINVAL); 1268 1269 error = copyin(args->rlim, &rlim, sizeof(rlim)); 1270 if (error) 1271 return (error); 1272 1273 bsd_rlim.rlim_cur = (rlim_t)rlim.rlim_cur; 1274 bsd_rlim.rlim_max = (rlim_t)rlim.rlim_max; 1275 return (kern_setrlimit(td, which, &bsd_rlim)); 1276} 1277 1278int 1279linux_old_getrlimit(struct thread *td, struct linux_old_getrlimit_args *args) 1280{ 1281 struct l_rlimit rlim; 1282 struct proc *p = td->td_proc; 1283 struct rlimit bsd_rlim; 1284 u_int which; 1285 1286#ifdef DEBUG 1287 if (ldebug(old_getrlimit)) 1288 printf(ARGS(old_getrlimit, "%d, %p"), 1289 args->resource, (void *)args->rlim); 1290#endif 1291 1292 if (args->resource >= LINUX_RLIM_NLIMITS) 1293 return (EINVAL); 1294 1295 which = linux_to_bsd_resource[args->resource]; 1296 if (which == -1) 1297 return (EINVAL); 1298 1299 PROC_LOCK(p); 1300 lim_rlimit(p, which, &bsd_rlim); 1301 PROC_UNLOCK(p); 1302 1303#ifdef COMPAT_LINUX32 1304 rlim.rlim_cur = (unsigned int)bsd_rlim.rlim_cur; 1305 if (rlim.rlim_cur == UINT_MAX) 1306 rlim.rlim_cur = INT_MAX; 1307 rlim.rlim_max = (unsigned int)bsd_rlim.rlim_max; 1308 if (rlim.rlim_max == UINT_MAX) 1309 rlim.rlim_max = INT_MAX; 1310#else 1311 rlim.rlim_cur = (unsigned long)bsd_rlim.rlim_cur; 1312 if (rlim.rlim_cur == ULONG_MAX) 1313 rlim.rlim_cur = LONG_MAX; 1314 rlim.rlim_max = (unsigned long)bsd_rlim.rlim_max; 1315 if (rlim.rlim_max == ULONG_MAX) 1316 rlim.rlim_max = LONG_MAX; 1317#endif 1318 return (copyout(&rlim, args->rlim, sizeof(rlim))); 1319} 1320 1321int 1322linux_getrlimit(struct thread *td, struct linux_getrlimit_args *args) 1323{ 1324 struct l_rlimit rlim; 1325 struct proc *p = td->td_proc; 1326 struct rlimit bsd_rlim; 1327 u_int which; 1328 1329#ifdef DEBUG 1330 if (ldebug(getrlimit)) 1331 printf(ARGS(getrlimit, "%d, %p"), 1332 args->resource, (void *)args->rlim); 1333#endif 1334 1335 if (args->resource >= LINUX_RLIM_NLIMITS) 1336 return (EINVAL); 1337 1338 which = linux_to_bsd_resource[args->resource]; 1339 if (which == -1) 1340 return (EINVAL); 1341 1342 PROC_LOCK(p); 1343 lim_rlimit(p, which, &bsd_rlim); 1344 PROC_UNLOCK(p); 1345 1346 rlim.rlim_cur = (l_ulong)bsd_rlim.rlim_cur; 1347 rlim.rlim_max = (l_ulong)bsd_rlim.rlim_max; 1348 return (copyout(&rlim, args->rlim, sizeof(rlim))); 1349} 1350 1351int 1352linux_sched_setscheduler(struct thread *td, 1353 struct linux_sched_setscheduler_args *args) 1354{ 1355 struct sched_setscheduler_args bsd; 1356 1357#ifdef DEBUG 1358 if (ldebug(sched_setscheduler)) 1359 printf(ARGS(sched_setscheduler, "%d, %d, %p"), 1360 args->pid, args->policy, (const void *)args->param); 1361#endif 1362 1363 switch (args->policy) { 1364 case LINUX_SCHED_OTHER: 1365 bsd.policy = SCHED_OTHER; 1366 break; 1367 case LINUX_SCHED_FIFO: 1368 bsd.policy = SCHED_FIFO; 1369 break; 1370 case LINUX_SCHED_RR: 1371 bsd.policy = SCHED_RR; 1372 break; 1373 default: 1374 return EINVAL; 1375 } 1376 1377 bsd.pid = args->pid; 1378 bsd.param = (struct sched_param *)args->param; 1379 return sched_setscheduler(td, &bsd); 1380} 1381 1382int 1383linux_sched_getscheduler(struct thread *td, 1384 struct linux_sched_getscheduler_args *args) 1385{ 1386 struct sched_getscheduler_args bsd; 1387 int error; 1388 1389#ifdef DEBUG 1390 if (ldebug(sched_getscheduler)) 1391 printf(ARGS(sched_getscheduler, "%d"), args->pid); 1392#endif 1393 1394 bsd.pid = args->pid; 1395 error = sched_getscheduler(td, &bsd); 1396 1397 switch (td->td_retval[0]) { 1398 case SCHED_OTHER: 1399 td->td_retval[0] = LINUX_SCHED_OTHER; 1400 break; 1401 case SCHED_FIFO: 1402 td->td_retval[0] = LINUX_SCHED_FIFO; 1403 break; 1404 case SCHED_RR: 1405 td->td_retval[0] = LINUX_SCHED_RR; 1406 break; 1407 } 1408 1409 return error; 1410} 1411 1412int 1413linux_sched_get_priority_max(struct thread *td, 1414 struct linux_sched_get_priority_max_args *args) 1415{ 1416 struct sched_get_priority_max_args bsd; 1417 1418#ifdef DEBUG 1419 if (ldebug(sched_get_priority_max)) 1420 printf(ARGS(sched_get_priority_max, "%d"), args->policy); 1421#endif 1422 1423 switch (args->policy) { 1424 case LINUX_SCHED_OTHER: 1425 bsd.policy = SCHED_OTHER; 1426 break; 1427 case LINUX_SCHED_FIFO: 1428 bsd.policy = SCHED_FIFO; 1429 break; 1430 case LINUX_SCHED_RR: 1431 bsd.policy = SCHED_RR; 1432 break; 1433 default: 1434 return EINVAL; 1435 } 1436 return sched_get_priority_max(td, &bsd); 1437} 1438 1439int 1440linux_sched_get_priority_min(struct thread *td, 1441 struct linux_sched_get_priority_min_args *args) 1442{ 1443 struct sched_get_priority_min_args bsd; 1444 1445#ifdef DEBUG 1446 if (ldebug(sched_get_priority_min)) 1447 printf(ARGS(sched_get_priority_min, "%d"), args->policy); 1448#endif 1449 1450 switch (args->policy) { 1451 case LINUX_SCHED_OTHER: 1452 bsd.policy = SCHED_OTHER; 1453 break; 1454 case LINUX_SCHED_FIFO: 1455 bsd.policy = SCHED_FIFO; 1456 break; 1457 case LINUX_SCHED_RR: 1458 bsd.policy = SCHED_RR; 1459 break; 1460 default: 1461 return EINVAL; 1462 } 1463 return sched_get_priority_min(td, &bsd); 1464} 1465 1466#define REBOOT_CAD_ON 0x89abcdef 1467#define REBOOT_CAD_OFF 0 1468#define REBOOT_HALT 0xcdef0123 1469#define REBOOT_RESTART 0x01234567 1470#define REBOOT_RESTART2 0xA1B2C3D4 1471#define REBOOT_POWEROFF 0x4321FEDC 1472#define REBOOT_MAGIC1 0xfee1dead 1473#define REBOOT_MAGIC2 0x28121969 1474#define REBOOT_MAGIC2A 0x05121996 1475#define REBOOT_MAGIC2B 0x16041998 1476 1477int 1478linux_reboot(struct thread *td, struct linux_reboot_args *args) 1479{ 1480 struct reboot_args bsd_args; 1481 1482#ifdef DEBUG 1483 if (ldebug(reboot)) 1484 printf(ARGS(reboot, "0x%x"), args->cmd); 1485#endif 1486 1487 if (args->magic1 != REBOOT_MAGIC1) 1488 return EINVAL; 1489 1490 switch (args->magic2) { 1491 case REBOOT_MAGIC2: 1492 case REBOOT_MAGIC2A: 1493 case REBOOT_MAGIC2B: 1494 break; 1495 default: 1496 return EINVAL; 1497 } 1498 1499 switch (args->cmd) { 1500 case REBOOT_CAD_ON: 1501 case REBOOT_CAD_OFF: 1502 return (priv_check(td, PRIV_REBOOT)); 1503 case REBOOT_HALT: 1504 bsd_args.opt = RB_HALT; 1505 break; 1506 case REBOOT_RESTART: 1507 case REBOOT_RESTART2: 1508 bsd_args.opt = 0; 1509 break; 1510 case REBOOT_POWEROFF: 1511 bsd_args.opt = RB_POWEROFF; 1512 break; 1513 default: 1514 return EINVAL; 1515 } 1516 return reboot(td, &bsd_args); 1517} 1518 1519 1520/* 1521 * The FreeBSD native getpid(2), getgid(2) and getuid(2) also modify 1522 * td->td_retval[1] when COMPAT_43 is defined. This clobbers registers that 1523 * are assumed to be preserved. The following lightweight syscalls fixes 1524 * this. See also linux_getgid16() and linux_getuid16() in linux_uid16.c 1525 * 1526 * linux_getpid() - MP SAFE 1527 * linux_getgid() - MP SAFE 1528 * linux_getuid() - MP SAFE 1529 */ 1530 1531int 1532linux_getpid(struct thread *td, struct linux_getpid_args *args) 1533{ 1534 struct linux_emuldata *em; 1535 1536#ifdef DEBUG 1537 if (ldebug(getpid)) 1538 printf(ARGS(getpid, "")); 1539#endif 1540 1541 if (linux_use26(td)) { 1542 em = em_find(td->td_proc, EMUL_DONTLOCK); 1543 KASSERT(em != NULL, ("getpid: emuldata not found.\n")); 1544 td->td_retval[0] = em->shared->group_pid; 1545 } else { 1546 td->td_retval[0] = td->td_proc->p_pid; 1547 } 1548 1549 return (0); 1550} 1551 1552int 1553linux_gettid(struct thread *td, struct linux_gettid_args *args) 1554{ 1555 1556#ifdef DEBUG 1557 if (ldebug(gettid)) 1558 printf(ARGS(gettid, "")); 1559#endif 1560 1561 td->td_retval[0] = td->td_proc->p_pid; 1562 return (0); 1563} 1564 1565 1566int 1567linux_getppid(struct thread *td, struct linux_getppid_args *args) 1568{ 1569 struct linux_emuldata *em; 1570 struct proc *p, *pp; 1571 1572#ifdef DEBUG 1573 if (ldebug(getppid)) 1574 printf(ARGS(getppid, "")); 1575#endif 1576 1577 if (!linux_use26(td)) { 1578 PROC_LOCK(td->td_proc); 1579 td->td_retval[0] = td->td_proc->p_pptr->p_pid; 1580 PROC_UNLOCK(td->td_proc); 1581 return (0); 1582 } 1583 1584 em = em_find(td->td_proc, EMUL_DONTLOCK); 1585 1586 KASSERT(em != NULL, ("getppid: process emuldata not found.\n")); 1587 1588 /* find the group leader */ 1589 p = pfind(em->shared->group_pid); 1590 1591 if (p == NULL) { 1592#ifdef DEBUG 1593 printf(LMSG("parent process not found.\n")); 1594#endif 1595 return (0); 1596 } 1597 1598 pp = p->p_pptr; /* switch to parent */ 1599 PROC_LOCK(pp); 1600 PROC_UNLOCK(p); 1601 1602 /* if its also linux process */ 1603 if (pp->p_sysent == &elf_linux_sysvec) { 1604 em = em_find(pp, EMUL_DONTLOCK); 1605 KASSERT(em != NULL, ("getppid: parent emuldata not found.\n")); 1606 1607 td->td_retval[0] = em->shared->group_pid; 1608 } else 1609 td->td_retval[0] = pp->p_pid; 1610 1611 PROC_UNLOCK(pp); 1612 1613 return (0); 1614} 1615 1616int 1617linux_getgid(struct thread *td, struct linux_getgid_args *args) 1618{ 1619 1620#ifdef DEBUG 1621 if (ldebug(getgid)) 1622 printf(ARGS(getgid, "")); 1623#endif 1624 1625 td->td_retval[0] = td->td_ucred->cr_rgid; 1626 return (0); 1627} 1628 1629int 1630linux_getuid(struct thread *td, struct linux_getuid_args *args) 1631{ 1632 1633#ifdef DEBUG 1634 if (ldebug(getuid)) 1635 printf(ARGS(getuid, "")); 1636#endif 1637 1638 td->td_retval[0] = td->td_ucred->cr_ruid; 1639 return (0); 1640} 1641 1642 1643int 1644linux_getsid(struct thread *td, struct linux_getsid_args *args) 1645{ 1646 struct getsid_args bsd; 1647 1648#ifdef DEBUG 1649 if (ldebug(getsid)) 1650 printf(ARGS(getsid, "%i"), args->pid); 1651#endif 1652 1653 bsd.pid = args->pid; 1654 return getsid(td, &bsd); 1655} 1656 1657int 1658linux_nosys(struct thread *td, struct nosys_args *ignore) 1659{ 1660 1661 return (ENOSYS); 1662} 1663 1664int 1665linux_getpriority(struct thread *td, struct linux_getpriority_args *args) 1666{ 1667 struct getpriority_args bsd_args; 1668 int error; 1669 1670#ifdef DEBUG 1671 if (ldebug(getpriority)) 1672 printf(ARGS(getpriority, "%i, %i"), args->which, args->who); 1673#endif 1674 1675 bsd_args.which = args->which; 1676 bsd_args.who = args->who; 1677 error = getpriority(td, &bsd_args); 1678 td->td_retval[0] = 20 - td->td_retval[0]; 1679 return error; 1680} 1681 1682int 1683linux_sethostname(struct thread *td, struct linux_sethostname_args *args) 1684{ 1685 int name[2]; 1686 1687#ifdef DEBUG 1688 if (ldebug(sethostname)) 1689 printf(ARGS(sethostname, "*, %i"), args->len); 1690#endif 1691 1692 name[0] = CTL_KERN; 1693 name[1] = KERN_HOSTNAME; 1694 return (userland_sysctl(td, name, 2, 0, 0, 0, args->hostname, 1695 args->len, 0, 0)); 1696} 1697 1698int 1699linux_exit_group(struct thread *td, struct linux_exit_group_args *args) 1700{ 1701 struct linux_emuldata *em, *td_em, *tmp_em; 1702 struct proc *sp; 1703 1704#ifdef DEBUG 1705 if (ldebug(exit_group)) 1706 printf(ARGS(exit_group, "%i"), args->error_code); 1707#endif 1708 1709 if (linux_use26(td)) { 1710 td_em = em_find(td->td_proc, EMUL_DONTLOCK); 1711 1712 KASSERT(td_em != NULL, ("exit_group: emuldata not found.\n")); 1713 1714 EMUL_SHARED_RLOCK(&emul_shared_lock); 1715 LIST_FOREACH_SAFE(em, &td_em->shared->threads, threads, tmp_em) { 1716 if (em->pid == td_em->pid) 1717 continue; 1718 1719 sp = pfind(em->pid); 1720 psignal(sp, SIGKILL); 1721 PROC_UNLOCK(sp); 1722#ifdef DEBUG 1723 printf(LMSG("linux_sys_exit_group: kill PID %d\n"), em->pid); 1724#endif 1725 } 1726 1727 EMUL_SHARED_RUNLOCK(&emul_shared_lock); 1728 } 1729 /* 1730 * XXX: we should send a signal to the parent if 1731 * SIGNAL_EXIT_GROUP is set. We ignore that (temporarily?) 1732 * as it doesnt occur often. 1733 */ 1734 exit1(td, W_EXITCODE(args->error_code, 0)); 1735 1736 return (0); 1737} 1738 1739int 1740linux_prctl(struct thread *td, struct linux_prctl_args *args) 1741{ 1742 int error = 0, max_size; 1743 struct proc *p = td->td_proc; 1744 char comm[LINUX_MAX_COMM_LEN]; 1745 struct linux_emuldata *em; 1746 int pdeath_signal; 1747 1748#ifdef DEBUG 1749 if (ldebug(prctl)) 1750 printf(ARGS(prctl, "%d, %d, %d, %d, %d"), args->option, 1751 args->arg2, args->arg3, args->arg4, args->arg5); 1752#endif 1753 1754 switch (args->option) { 1755 case LINUX_PR_SET_PDEATHSIG: 1756 if (!LINUX_SIG_VALID(args->arg2)) 1757 return (EINVAL); 1758 em = em_find(p, EMUL_DOLOCK); 1759 KASSERT(em != NULL, ("prctl: emuldata not found.\n")); 1760 em->pdeath_signal = args->arg2; 1761 EMUL_UNLOCK(&emul_lock); 1762 break; 1763 case LINUX_PR_GET_PDEATHSIG: 1764 em = em_find(p, EMUL_DOLOCK); 1765 KASSERT(em != NULL, ("prctl: emuldata not found.\n")); 1766 pdeath_signal = em->pdeath_signal; 1767 EMUL_UNLOCK(&emul_lock); 1768 error = copyout(&pdeath_signal, 1769 (void *)(register_t)args->arg2, 1770 sizeof(pdeath_signal)); 1771 break; 1772 case LINUX_PR_SET_NAME: 1773 /* 1774 * To be on the safe side we need to make sure to not 1775 * overflow the size a linux program expects. We already 1776 * do this here in the copyin, so that we don't need to 1777 * check on copyout. 1778 */ 1779 max_size = MIN(sizeof(comm), sizeof(p->p_comm)); 1780 error = copyinstr((void *)(register_t)args->arg2, comm, 1781 max_size, NULL); 1782 1783 /* Linux silently truncates the name if it is too long. */ 1784 if (error == ENAMETOOLONG) { 1785 /* 1786 * XXX: copyinstr() isn't documented to populate the 1787 * array completely, so do a copyin() to be on the 1788 * safe side. This should be changed in case 1789 * copyinstr() is changed to guarantee this. 1790 */ 1791 error = copyin((void *)(register_t)args->arg2, comm, 1792 max_size - 1); 1793 comm[max_size - 1] = '\0'; 1794 } 1795 if (error) 1796 return (error); 1797 1798 PROC_LOCK(p); 1799 strlcpy(p->p_comm, comm, sizeof(p->p_comm)); 1800 PROC_UNLOCK(p); 1801 break; 1802 case LINUX_PR_GET_NAME: 1803 PROC_LOCK(p); 1804 strlcpy(comm, p->p_comm, sizeof(comm)); 1805 PROC_UNLOCK(p); 1806 error = copyout(comm, (void *)(register_t)args->arg2, 1807 strlen(comm) + 1); 1808 break; 1809 default: 1810 error = EINVAL; 1811 break; 1812 } 1813 1814 return (error); 1815} 1816 1817/* 1818 * Get affinity of a process. 1819 */ 1820int 1821linux_sched_getaffinity(struct thread *td, 1822 struct linux_sched_getaffinity_args *args) 1823{ 1824 int error; 1825 struct cpuset_getaffinity_args cga; 1826 1827#ifdef DEBUG 1828 if (ldebug(sched_getaffinity)) 1829 printf(ARGS(sched_getaffinity, "%d, %d, *"), args->pid, 1830 args->len); 1831#endif 1832 1833 cga.level = CPU_LEVEL_WHICH; 1834 cga.which = CPU_WHICH_PID; 1835 cga.id = args->pid; 1836 cga.cpusetsize = sizeof(cpumask_t); 1837 cga.mask = (cpuset_t *) args->user_mask_ptr; 1838 1839 if ((error = cpuset_getaffinity(td, &cga)) == 0) 1840 td->td_retval[0] = sizeof(cpumask_t); 1841 1842 return (error); 1843} 1844 1845/* 1846 * Set affinity of a process. 1847 */ 1848int 1849linux_sched_setaffinity(struct thread *td, 1850 struct linux_sched_setaffinity_args *args) 1851{ 1852 struct cpuset_setaffinity_args csa; 1853 1854#ifdef DEBUG 1855 if (ldebug(sched_setaffinity)) 1856 printf(ARGS(sched_setaffinity, "%d, %d, *"), args->pid, 1857 args->len); 1858#endif 1859 csa.level = CPU_LEVEL_WHICH; 1860 csa.which = CPU_WHICH_PID; 1861 csa.id = args->pid; 1862 csa.cpusetsize = args->len; 1863 csa.mask = (cpuset_t *) args->user_mask_ptr; 1864 1865 return (cpuset_setaffinity(td, &csa)); 1866}
| 93 94#ifdef __i386__ 95#include <machine/cputypes.h> 96#endif 97 98#define BSD_TO_LINUX_SIGNAL(sig) \ 99 (((sig) <= LINUX_SIGTBLSZ) ? bsd_to_linux_signal[_SIG_IDX(sig)] : sig) 100 101static unsigned int linux_to_bsd_resource[LINUX_RLIM_NLIMITS] = { 102 RLIMIT_CPU, RLIMIT_FSIZE, RLIMIT_DATA, RLIMIT_STACK, 103 RLIMIT_CORE, RLIMIT_RSS, RLIMIT_NPROC, RLIMIT_NOFILE, 104 RLIMIT_MEMLOCK, RLIMIT_AS 105}; 106 107struct l_sysinfo { 108 l_long uptime; /* Seconds since boot */ 109 l_ulong loads[3]; /* 1, 5, and 15 minute load averages */ 110#define LINUX_SYSINFO_LOADS_SCALE 65536 111 l_ulong totalram; /* Total usable main memory size */ 112 l_ulong freeram; /* Available memory size */ 113 l_ulong sharedram; /* Amount of shared memory */ 114 l_ulong bufferram; /* Memory used by buffers */ 115 l_ulong totalswap; /* Total swap space size */ 116 l_ulong freeswap; /* swap space still available */ 117 l_ushort procs; /* Number of current processes */ 118 l_ushort pads; 119 l_ulong totalbig; 120 l_ulong freebig; 121 l_uint mem_unit; 122 char _f[20-2*sizeof(l_long)-sizeof(l_int)]; /* padding */ 123}; 124int 125linux_sysinfo(struct thread *td, struct linux_sysinfo_args *args) 126{ 127 struct l_sysinfo sysinfo; 128 vm_object_t object; 129 int i, j; 130 struct timespec ts; 131 132 getnanouptime(&ts); 133 if (ts.tv_nsec != 0) 134 ts.tv_sec++; 135 sysinfo.uptime = ts.tv_sec; 136 137 /* Use the information from the mib to get our load averages */ 138 for (i = 0; i < 3; i++) 139 sysinfo.loads[i] = averunnable.ldavg[i] * 140 LINUX_SYSINFO_LOADS_SCALE / averunnable.fscale; 141 142 sysinfo.totalram = physmem * PAGE_SIZE; 143 sysinfo.freeram = sysinfo.totalram - cnt.v_wire_count * PAGE_SIZE; 144 145 sysinfo.sharedram = 0; 146 mtx_lock(&vm_object_list_mtx); 147 TAILQ_FOREACH(object, &vm_object_list, object_list) 148 if (object->shadow_count > 1) 149 sysinfo.sharedram += object->resident_page_count; 150 mtx_unlock(&vm_object_list_mtx); 151 152 sysinfo.sharedram *= PAGE_SIZE; 153 sysinfo.bufferram = 0; 154 155 swap_pager_status(&i, &j); 156 sysinfo.totalswap = i * PAGE_SIZE; 157 sysinfo.freeswap = (i - j) * PAGE_SIZE; 158 159 sysinfo.procs = nprocs; 160 161 /* The following are only present in newer Linux kernels. */ 162 sysinfo.totalbig = 0; 163 sysinfo.freebig = 0; 164 sysinfo.mem_unit = 1; 165 166 return copyout(&sysinfo, args->info, sizeof(sysinfo)); 167} 168 169int 170linux_alarm(struct thread *td, struct linux_alarm_args *args) 171{ 172 struct itimerval it, old_it; 173 int error; 174 175#ifdef DEBUG 176 if (ldebug(alarm)) 177 printf(ARGS(alarm, "%u"), args->secs); 178#endif 179 180 if (args->secs > 100000000) 181 return (EINVAL); 182 183 it.it_value.tv_sec = (long)args->secs; 184 it.it_value.tv_usec = 0; 185 it.it_interval.tv_sec = 0; 186 it.it_interval.tv_usec = 0; 187 error = kern_setitimer(td, ITIMER_REAL, &it, &old_it); 188 if (error) 189 return (error); 190 if (timevalisset(&old_it.it_value)) { 191 if (old_it.it_value.tv_usec != 0) 192 old_it.it_value.tv_sec++; 193 td->td_retval[0] = old_it.it_value.tv_sec; 194 } 195 return (0); 196} 197 198int 199linux_brk(struct thread *td, struct linux_brk_args *args) 200{ 201 struct vmspace *vm = td->td_proc->p_vmspace; 202 vm_offset_t new, old; 203 struct obreak_args /* { 204 char * nsize; 205 } */ tmp; 206 207#ifdef DEBUG 208 if (ldebug(brk)) 209 printf(ARGS(brk, "%p"), (void *)(uintptr_t)args->dsend); 210#endif 211 old = (vm_offset_t)vm->vm_daddr + ctob(vm->vm_dsize); 212 new = (vm_offset_t)args->dsend; 213 tmp.nsize = (char *)new; 214 if (((caddr_t)new > vm->vm_daddr) && !obreak(td, &tmp)) 215 td->td_retval[0] = (long)new; 216 else 217 td->td_retval[0] = (long)old; 218 219 return 0; 220} 221 222#if defined(__i386__) 223/* XXX: what about amd64/linux32? */ 224 225int 226linux_uselib(struct thread *td, struct linux_uselib_args *args) 227{ 228 struct nameidata ni; 229 struct vnode *vp; 230 struct exec *a_out; 231 struct vattr attr; 232 vm_offset_t vmaddr; 233 unsigned long file_offset; 234 vm_offset_t buffer; 235 unsigned long bss_size; 236 char *library; 237 int error; 238 int locked, vfslocked; 239 240 LCONVPATHEXIST(td, args->library, &library); 241 242#ifdef DEBUG 243 if (ldebug(uselib)) 244 printf(ARGS(uselib, "%s"), library); 245#endif 246 247 a_out = NULL; 248 vfslocked = 0; 249 locked = 0; 250 vp = NULL; 251 252 NDINIT(&ni, LOOKUP, ISOPEN | FOLLOW | LOCKLEAF | MPSAFE | AUDITVNODE1, 253 UIO_SYSSPACE, library, td); 254 error = namei(&ni); 255 LFREEPATH(library); 256 if (error) 257 goto cleanup; 258 259 vp = ni.ni_vp; 260 vfslocked = NDHASGIANT(&ni); 261 NDFREE(&ni, NDF_ONLY_PNBUF); 262 263 /* 264 * From here on down, we have a locked vnode that must be unlocked. 265 * XXX: The code below largely duplicates exec_check_permissions(). 266 */ 267 locked = 1; 268 269 /* Writable? */ 270 if (vp->v_writecount) { 271 error = ETXTBSY; 272 goto cleanup; 273 } 274 275 /* Executable? */ 276 error = VOP_GETATTR(vp, &attr, td->td_ucred, td); 277 if (error) 278 goto cleanup; 279 280 if ((vp->v_mount->mnt_flag & MNT_NOEXEC) || 281 ((attr.va_mode & 0111) == 0) || (attr.va_type != VREG)) { 282 /* EACCESS is what exec(2) returns. */ 283 error = ENOEXEC; 284 goto cleanup; 285 } 286 287 /* Sensible size? */ 288 if (attr.va_size == 0) { 289 error = ENOEXEC; 290 goto cleanup; 291 } 292 293 /* Can we access it? */ 294 error = VOP_ACCESS(vp, VEXEC, td->td_ucred, td); 295 if (error) 296 goto cleanup; 297 298 /* 299 * XXX: This should use vn_open() so that it is properly authorized, 300 * and to reduce code redundancy all over the place here. 301 * XXX: Not really, it duplicates far more of exec_check_permissions() 302 * than vn_open(). 303 */ 304#ifdef MAC 305 error = mac_vnode_check_open(td->td_ucred, vp, FREAD); 306 if (error) 307 goto cleanup; 308#endif 309 error = VOP_OPEN(vp, FREAD, td->td_ucred, td, NULL); 310 if (error) 311 goto cleanup; 312 313 /* Pull in executable header into kernel_map */ 314 error = vm_mmap(kernel_map, (vm_offset_t *)&a_out, PAGE_SIZE, 315 VM_PROT_READ, VM_PROT_READ, 0, OBJT_VNODE, vp, 0); 316 if (error) 317 goto cleanup; 318 319 /* Is it a Linux binary ? */ 320 if (((a_out->a_magic >> 16) & 0xff) != 0x64) { 321 error = ENOEXEC; 322 goto cleanup; 323 } 324 325 /* 326 * While we are here, we should REALLY do some more checks 327 */ 328 329 /* Set file/virtual offset based on a.out variant. */ 330 switch ((int)(a_out->a_magic & 0xffff)) { 331 case 0413: /* ZMAGIC */ 332 file_offset = 1024; 333 break; 334 case 0314: /* QMAGIC */ 335 file_offset = 0; 336 break; 337 default: 338 error = ENOEXEC; 339 goto cleanup; 340 } 341 342 bss_size = round_page(a_out->a_bss); 343 344 /* Check various fields in header for validity/bounds. */ 345 if (a_out->a_text & PAGE_MASK || a_out->a_data & PAGE_MASK) { 346 error = ENOEXEC; 347 goto cleanup; 348 } 349 350 /* text + data can't exceed file size */ 351 if (a_out->a_data + a_out->a_text > attr.va_size) { 352 error = EFAULT; 353 goto cleanup; 354 } 355 356 /* 357 * text/data/bss must not exceed limits 358 * XXX - this is not complete. it should check current usage PLUS 359 * the resources needed by this library. 360 */ 361 PROC_LOCK(td->td_proc); 362 if (a_out->a_text > maxtsiz || 363 a_out->a_data + bss_size > lim_cur(td->td_proc, RLIMIT_DATA)) { 364 PROC_UNLOCK(td->td_proc); 365 error = ENOMEM; 366 goto cleanup; 367 } 368 PROC_UNLOCK(td->td_proc); 369 370 /* 371 * Prevent more writers. 372 * XXX: Note that if any of the VM operations fail below we don't 373 * clear this flag. 374 */ 375 vp->v_vflag |= VV_TEXT; 376 377 /* 378 * Lock no longer needed 379 */ 380 locked = 0; 381 VOP_UNLOCK(vp, 0); 382 VFS_UNLOCK_GIANT(vfslocked); 383 384 /* 385 * Check if file_offset page aligned. Currently we cannot handle 386 * misalinged file offsets, and so we read in the entire image 387 * (what a waste). 388 */ 389 if (file_offset & PAGE_MASK) { 390#ifdef DEBUG 391 printf("uselib: Non page aligned binary %lu\n", file_offset); 392#endif 393 /* Map text+data read/write/execute */ 394 395 /* a_entry is the load address and is page aligned */ 396 vmaddr = trunc_page(a_out->a_entry); 397 398 /* get anon user mapping, read+write+execute */ 399 error = vm_map_find(&td->td_proc->p_vmspace->vm_map, NULL, 0, 400 &vmaddr, a_out->a_text + a_out->a_data, FALSE, VM_PROT_ALL, 401 VM_PROT_ALL, 0); 402 if (error) 403 goto cleanup; 404 405 /* map file into kernel_map */ 406 error = vm_mmap(kernel_map, &buffer, 407 round_page(a_out->a_text + a_out->a_data + file_offset), 408 VM_PROT_READ, VM_PROT_READ, 0, OBJT_VNODE, vp, 409 trunc_page(file_offset)); 410 if (error) 411 goto cleanup; 412 413 /* copy from kernel VM space to user space */ 414 error = copyout(PTRIN(buffer + file_offset), 415 (void *)vmaddr, a_out->a_text + a_out->a_data); 416 417 /* release temporary kernel space */ 418 vm_map_remove(kernel_map, buffer, buffer + 419 round_page(a_out->a_text + a_out->a_data + file_offset)); 420 421 if (error) 422 goto cleanup; 423 } else { 424#ifdef DEBUG 425 printf("uselib: Page aligned binary %lu\n", file_offset); 426#endif 427 /* 428 * for QMAGIC, a_entry is 20 bytes beyond the load address 429 * to skip the executable header 430 */ 431 vmaddr = trunc_page(a_out->a_entry); 432 433 /* 434 * Map it all into the process's space as a single 435 * copy-on-write "data" segment. 436 */ 437 error = vm_mmap(&td->td_proc->p_vmspace->vm_map, &vmaddr, 438 a_out->a_text + a_out->a_data, VM_PROT_ALL, VM_PROT_ALL, 439 MAP_PRIVATE | MAP_FIXED, OBJT_VNODE, vp, file_offset); 440 if (error) 441 goto cleanup; 442 } 443#ifdef DEBUG 444 printf("mem=%08lx = %08lx %08lx\n", (long)vmaddr, ((long *)vmaddr)[0], 445 ((long *)vmaddr)[1]); 446#endif 447 if (bss_size != 0) { 448 /* Calculate BSS start address */ 449 vmaddr = trunc_page(a_out->a_entry) + a_out->a_text + 450 a_out->a_data; 451 452 /* allocate some 'anon' space */ 453 error = vm_map_find(&td->td_proc->p_vmspace->vm_map, NULL, 0, 454 &vmaddr, bss_size, FALSE, VM_PROT_ALL, VM_PROT_ALL, 0); 455 if (error) 456 goto cleanup; 457 } 458 459cleanup: 460 /* Unlock vnode if needed */ 461 if (locked) { 462 VOP_UNLOCK(vp, 0); 463 VFS_UNLOCK_GIANT(vfslocked); 464 } 465 466 /* Release the kernel mapping. */ 467 if (a_out) 468 vm_map_remove(kernel_map, (vm_offset_t)a_out, 469 (vm_offset_t)a_out + PAGE_SIZE); 470 471 return error; 472} 473 474#endif /* __i386__ */ 475 476int 477linux_select(struct thread *td, struct linux_select_args *args) 478{ 479 l_timeval ltv; 480 struct timeval tv0, tv1, utv, *tvp; 481 int error; 482 483#ifdef DEBUG 484 if (ldebug(select)) 485 printf(ARGS(select, "%d, %p, %p, %p, %p"), args->nfds, 486 (void *)args->readfds, (void *)args->writefds, 487 (void *)args->exceptfds, (void *)args->timeout); 488#endif 489 490 /* 491 * Store current time for computation of the amount of 492 * time left. 493 */ 494 if (args->timeout) { 495 if ((error = copyin(args->timeout, <v, sizeof(ltv)))) 496 goto select_out; 497 utv.tv_sec = ltv.tv_sec; 498 utv.tv_usec = ltv.tv_usec; 499#ifdef DEBUG 500 if (ldebug(select)) 501 printf(LMSG("incoming timeout (%jd/%ld)"), 502 (intmax_t)utv.tv_sec, utv.tv_usec); 503#endif 504 505 if (itimerfix(&utv)) { 506 /* 507 * The timeval was invalid. Convert it to something 508 * valid that will act as it does under Linux. 509 */ 510 utv.tv_sec += utv.tv_usec / 1000000; 511 utv.tv_usec %= 1000000; 512 if (utv.tv_usec < 0) { 513 utv.tv_sec -= 1; 514 utv.tv_usec += 1000000; 515 } 516 if (utv.tv_sec < 0) 517 timevalclear(&utv); 518 } 519 microtime(&tv0); 520 tvp = &utv; 521 } else 522 tvp = NULL; 523 524 error = kern_select(td, args->nfds, args->readfds, args->writefds, 525 args->exceptfds, tvp); 526 527#ifdef DEBUG 528 if (ldebug(select)) 529 printf(LMSG("real select returns %d"), error); 530#endif 531 if (error) { 532 /* 533 * See fs/select.c in the Linux kernel. Without this, 534 * Maelstrom doesn't work. 535 */ 536 if (error == ERESTART) 537 error = EINTR; 538 goto select_out; 539 } 540 541 if (args->timeout) { 542 if (td->td_retval[0]) { 543 /* 544 * Compute how much time was left of the timeout, 545 * by subtracting the current time and the time 546 * before we started the call, and subtracting 547 * that result from the user-supplied value. 548 */ 549 microtime(&tv1); 550 timevalsub(&tv1, &tv0); 551 timevalsub(&utv, &tv1); 552 if (utv.tv_sec < 0) 553 timevalclear(&utv); 554 } else 555 timevalclear(&utv); 556#ifdef DEBUG 557 if (ldebug(select)) 558 printf(LMSG("outgoing timeout (%jd/%ld)"), 559 (intmax_t)utv.tv_sec, utv.tv_usec); 560#endif 561 ltv.tv_sec = utv.tv_sec; 562 ltv.tv_usec = utv.tv_usec; 563 if ((error = copyout(<v, args->timeout, sizeof(ltv)))) 564 goto select_out; 565 } 566 567select_out: 568#ifdef DEBUG 569 if (ldebug(select)) 570 printf(LMSG("select_out -> %d"), error); 571#endif 572 return error; 573} 574 575int 576linux_mremap(struct thread *td, struct linux_mremap_args *args) 577{ 578 struct munmap_args /* { 579 void *addr; 580 size_t len; 581 } */ bsd_args; 582 int error = 0; 583 584#ifdef DEBUG 585 if (ldebug(mremap)) 586 printf(ARGS(mremap, "%p, %08lx, %08lx, %08lx"), 587 (void *)(uintptr_t)args->addr, 588 (unsigned long)args->old_len, 589 (unsigned long)args->new_len, 590 (unsigned long)args->flags); 591#endif 592 593 if (args->flags & ~(LINUX_MREMAP_FIXED | LINUX_MREMAP_MAYMOVE)) { 594 td->td_retval[0] = 0; 595 return (EINVAL); 596 } 597 598 /* 599 * Check for the page alignment. 600 * Linux defines PAGE_MASK to be FreeBSD ~PAGE_MASK. 601 */ 602 if (args->addr & PAGE_MASK) { 603 td->td_retval[0] = 0; 604 return (EINVAL); 605 } 606 607 args->new_len = round_page(args->new_len); 608 args->old_len = round_page(args->old_len); 609 610 if (args->new_len > args->old_len) { 611 td->td_retval[0] = 0; 612 return ENOMEM; 613 } 614 615 if (args->new_len < args->old_len) { 616 bsd_args.addr = 617 (caddr_t)((uintptr_t)args->addr + args->new_len); 618 bsd_args.len = args->old_len - args->new_len; 619 error = munmap(td, &bsd_args); 620 } 621 622 td->td_retval[0] = error ? 0 : (uintptr_t)args->addr; 623 return error; 624} 625 626#define LINUX_MS_ASYNC 0x0001 627#define LINUX_MS_INVALIDATE 0x0002 628#define LINUX_MS_SYNC 0x0004 629 630int 631linux_msync(struct thread *td, struct linux_msync_args *args) 632{ 633 struct msync_args bsd_args; 634 635 bsd_args.addr = (caddr_t)(uintptr_t)args->addr; 636 bsd_args.len = (uintptr_t)args->len; 637 bsd_args.flags = args->fl & ~LINUX_MS_SYNC; 638 639 return msync(td, &bsd_args); 640} 641 642int 643linux_time(struct thread *td, struct linux_time_args *args) 644{ 645 struct timeval tv; 646 l_time_t tm; 647 int error; 648 649#ifdef DEBUG 650 if (ldebug(time)) 651 printf(ARGS(time, "*")); 652#endif 653 654 microtime(&tv); 655 tm = tv.tv_sec; 656 if (args->tm && (error = copyout(&tm, args->tm, sizeof(tm)))) 657 return error; 658 td->td_retval[0] = tm; 659 return 0; 660} 661 662struct l_times_argv { 663 l_long tms_utime; 664 l_long tms_stime; 665 l_long tms_cutime; 666 l_long tms_cstime; 667}; 668 669#define CLK_TCK 100 /* Linux uses 100 */ 670 671#define CONVTCK(r) (r.tv_sec * CLK_TCK + r.tv_usec / (1000000 / CLK_TCK)) 672 673int 674linux_times(struct thread *td, struct linux_times_args *args) 675{ 676 struct timeval tv, utime, stime, cutime, cstime; 677 struct l_times_argv tms; 678 struct proc *p; 679 int error; 680 681#ifdef DEBUG 682 if (ldebug(times)) 683 printf(ARGS(times, "*")); 684#endif 685 686 if (args->buf != NULL) { 687 p = td->td_proc; 688 PROC_LOCK(p); 689 PROC_SLOCK(p); 690 calcru(p, &utime, &stime); 691 PROC_SUNLOCK(p); 692 calccru(p, &cutime, &cstime); 693 PROC_UNLOCK(p); 694 695 tms.tms_utime = CONVTCK(utime); 696 tms.tms_stime = CONVTCK(stime); 697 698 tms.tms_cutime = CONVTCK(cutime); 699 tms.tms_cstime = CONVTCK(cstime); 700 701 if ((error = copyout(&tms, args->buf, sizeof(tms)))) 702 return error; 703 } 704 705 microuptime(&tv); 706 td->td_retval[0] = (int)CONVTCK(tv); 707 return 0; 708} 709 710int 711linux_newuname(struct thread *td, struct linux_newuname_args *args) 712{ 713 struct l_new_utsname utsname; 714 char osname[LINUX_MAX_UTSNAME]; 715 char osrelease[LINUX_MAX_UTSNAME]; 716 char *p; 717 718#ifdef DEBUG 719 if (ldebug(newuname)) 720 printf(ARGS(newuname, "*")); 721#endif 722 723 linux_get_osname(td, osname); 724 linux_get_osrelease(td, osrelease); 725 726 bzero(&utsname, sizeof(utsname)); 727 strlcpy(utsname.sysname, osname, LINUX_MAX_UTSNAME); 728 getcredhostname(td->td_ucred, utsname.nodename, LINUX_MAX_UTSNAME); 729 strlcpy(utsname.release, osrelease, LINUX_MAX_UTSNAME); 730 strlcpy(utsname.version, version, LINUX_MAX_UTSNAME); 731 for (p = utsname.version; *p != '\0'; ++p) 732 if (*p == '\n') { 733 *p = '\0'; 734 break; 735 } 736#ifdef __i386__ 737 { 738 const char *class; 739 740 switch (cpu_class) { 741 case CPUCLASS_686: 742 class = "i686"; 743 break; 744 case CPUCLASS_586: 745 class = "i586"; 746 break; 747 case CPUCLASS_486: 748 class = "i486"; 749 break; 750 default: 751 class = "i386"; 752 } 753 strlcpy(utsname.machine, class, LINUX_MAX_UTSNAME); 754 } 755#elif defined(__amd64__) /* XXX: Linux can change 'personality'. */ 756#ifdef COMPAT_LINUX32 757 strlcpy(utsname.machine, "i686", LINUX_MAX_UTSNAME); 758#else 759 strlcpy(utsname.machine, "x86_64", LINUX_MAX_UTSNAME); 760#endif /* COMPAT_LINUX32 */ 761#else /* something other than i386 or amd64 - assume we and Linux agree */ 762 strlcpy(utsname.machine, machine, LINUX_MAX_UTSNAME); 763#endif /* __i386__ */ 764 strlcpy(utsname.domainname, domainname, LINUX_MAX_UTSNAME); 765 766 return (copyout(&utsname, args->buf, sizeof(utsname))); 767} 768 769#if defined(__i386__) || (defined(__amd64__) && defined(COMPAT_LINUX32)) 770struct l_utimbuf { 771 l_time_t l_actime; 772 l_time_t l_modtime; 773}; 774 775int 776linux_utime(struct thread *td, struct linux_utime_args *args) 777{ 778 struct timeval tv[2], *tvp; 779 struct l_utimbuf lut; 780 char *fname; 781 int error; 782 783 LCONVPATHEXIST(td, args->fname, &fname); 784 785#ifdef DEBUG 786 if (ldebug(utime)) 787 printf(ARGS(utime, "%s, *"), fname); 788#endif 789 790 if (args->times) { 791 if ((error = copyin(args->times, &lut, sizeof lut))) { 792 LFREEPATH(fname); 793 return error; 794 } 795 tv[0].tv_sec = lut.l_actime; 796 tv[0].tv_usec = 0; 797 tv[1].tv_sec = lut.l_modtime; 798 tv[1].tv_usec = 0; 799 tvp = tv; 800 } else 801 tvp = NULL; 802 803 error = kern_utimes(td, fname, UIO_SYSSPACE, tvp, UIO_SYSSPACE); 804 LFREEPATH(fname); 805 return (error); 806} 807 808int 809linux_utimes(struct thread *td, struct linux_utimes_args *args) 810{ 811 l_timeval ltv[2]; 812 struct timeval tv[2], *tvp = NULL; 813 char *fname; 814 int error; 815 816 LCONVPATHEXIST(td, args->fname, &fname); 817 818#ifdef DEBUG 819 if (ldebug(utimes)) 820 printf(ARGS(utimes, "%s, *"), fname); 821#endif 822 823 if (args->tptr != NULL) { 824 if ((error = copyin(args->tptr, ltv, sizeof ltv))) { 825 LFREEPATH(fname); 826 return (error); 827 } 828 tv[0].tv_sec = ltv[0].tv_sec; 829 tv[0].tv_usec = ltv[0].tv_usec; 830 tv[1].tv_sec = ltv[1].tv_sec; 831 tv[1].tv_usec = ltv[1].tv_usec; 832 tvp = tv; 833 } 834 835 error = kern_utimes(td, fname, UIO_SYSSPACE, tvp, UIO_SYSSPACE); 836 LFREEPATH(fname); 837 return (error); 838} 839 840int 841linux_futimesat(struct thread *td, struct linux_futimesat_args *args) 842{ 843 l_timeval ltv[2]; 844 struct timeval tv[2], *tvp = NULL; 845 char *fname; 846 int error, dfd; 847 848 dfd = (args->dfd == LINUX_AT_FDCWD) ? AT_FDCWD : args->dfd; 849 LCONVPATHEXIST_AT(td, args->filename, &fname, dfd); 850 851#ifdef DEBUG 852 if (ldebug(futimesat)) 853 printf(ARGS(futimesat, "%s, *"), fname); 854#endif 855 856 if (args->utimes != NULL) { 857 if ((error = copyin(args->utimes, ltv, sizeof ltv))) { 858 LFREEPATH(fname); 859 return (error); 860 } 861 tv[0].tv_sec = ltv[0].tv_sec; 862 tv[0].tv_usec = ltv[0].tv_usec; 863 tv[1].tv_sec = ltv[1].tv_sec; 864 tv[1].tv_usec = ltv[1].tv_usec; 865 tvp = tv; 866 } 867 868 error = kern_utimesat(td, dfd, fname, UIO_SYSSPACE, tvp, UIO_SYSSPACE); 869 LFREEPATH(fname); 870 return (error); 871} 872#endif /* __i386__ || (__amd64__ && COMPAT_LINUX32) */ 873 874#define __WCLONE 0x80000000 875 876int 877linux_waitpid(struct thread *td, struct linux_waitpid_args *args) 878{ 879 int error, options, tmpstat; 880 881#ifdef DEBUG 882 if (ldebug(waitpid)) 883 printf(ARGS(waitpid, "%d, %p, %d"), 884 args->pid, (void *)args->status, args->options); 885#endif 886 /* 887 * this is necessary because the test in kern_wait doesn't work 888 * because we mess with the options here 889 */ 890 if (args->options & ~(WUNTRACED | WNOHANG | WCONTINUED | __WCLONE)) 891 return (EINVAL); 892 893 options = (args->options & (WNOHANG | WUNTRACED)); 894 /* WLINUXCLONE should be equal to __WCLONE, but we make sure */ 895 if (args->options & __WCLONE) 896 options |= WLINUXCLONE; 897 898 error = kern_wait(td, args->pid, &tmpstat, options, NULL); 899 if (error) 900 return error; 901 902 if (args->status) { 903 tmpstat &= 0xffff; 904 if (WIFSIGNALED(tmpstat)) 905 tmpstat = (tmpstat & 0xffffff80) | 906 BSD_TO_LINUX_SIGNAL(WTERMSIG(tmpstat)); 907 else if (WIFSTOPPED(tmpstat)) 908 tmpstat = (tmpstat & 0xffff00ff) | 909 (BSD_TO_LINUX_SIGNAL(WSTOPSIG(tmpstat)) << 8); 910 return copyout(&tmpstat, args->status, sizeof(int)); 911 } 912 913 return 0; 914} 915 916int 917linux_wait4(struct thread *td, struct linux_wait4_args *args) 918{ 919 int error, options, tmpstat; 920 struct rusage ru, *rup; 921 struct proc *p; 922 923#ifdef DEBUG 924 if (ldebug(wait4)) 925 printf(ARGS(wait4, "%d, %p, %d, %p"), 926 args->pid, (void *)args->status, args->options, 927 (void *)args->rusage); 928#endif 929 930 options = (args->options & (WNOHANG | WUNTRACED)); 931 /* WLINUXCLONE should be equal to __WCLONE, but we make sure */ 932 if (args->options & __WCLONE) 933 options |= WLINUXCLONE; 934 935 if (args->rusage != NULL) 936 rup = &ru; 937 else 938 rup = NULL; 939 error = kern_wait(td, args->pid, &tmpstat, options, rup); 940 if (error) 941 return error; 942 943 p = td->td_proc; 944 PROC_LOCK(p); 945 sigqueue_delete(&p->p_sigqueue, SIGCHLD); 946 PROC_UNLOCK(p); 947 948 if (args->status) { 949 tmpstat &= 0xffff; 950 if (WIFSIGNALED(tmpstat)) 951 tmpstat = (tmpstat & 0xffffff80) | 952 BSD_TO_LINUX_SIGNAL(WTERMSIG(tmpstat)); 953 else if (WIFSTOPPED(tmpstat)) 954 tmpstat = (tmpstat & 0xffff00ff) | 955 (BSD_TO_LINUX_SIGNAL(WSTOPSIG(tmpstat)) << 8); 956 error = copyout(&tmpstat, args->status, sizeof(int)); 957 } 958 if (args->rusage != NULL && error == 0) 959 error = copyout(&ru, args->rusage, sizeof(ru)); 960 961 return (error); 962} 963 964int 965linux_mknod(struct thread *td, struct linux_mknod_args *args) 966{ 967 char *path; 968 int error; 969 970 LCONVPATHCREAT(td, args->path, &path); 971 972#ifdef DEBUG 973 if (ldebug(mknod)) 974 printf(ARGS(mknod, "%s, %d, %d"), path, args->mode, args->dev); 975#endif 976 977 switch (args->mode & S_IFMT) { 978 case S_IFIFO: 979 case S_IFSOCK: 980 error = kern_mkfifo(td, path, UIO_SYSSPACE, args->mode); 981 break; 982 983 case S_IFCHR: 984 case S_IFBLK: 985 error = kern_mknod(td, path, UIO_SYSSPACE, args->mode, 986 args->dev); 987 break; 988 989 case S_IFDIR: 990 error = EPERM; 991 break; 992 993 case 0: 994 args->mode |= S_IFREG; 995 /* FALLTHROUGH */ 996 case S_IFREG: 997 error = kern_open(td, path, UIO_SYSSPACE, 998 O_WRONLY | O_CREAT | O_TRUNC, args->mode); 999 if (error == 0) 1000 kern_close(td, td->td_retval[0]); 1001 break; 1002 1003 default: 1004 error = EINVAL; 1005 break; 1006 } 1007 LFREEPATH(path); 1008 return (error); 1009} 1010 1011int 1012linux_mknodat(struct thread *td, struct linux_mknodat_args *args) 1013{ 1014 char *path; 1015 int error, dfd; 1016 1017 dfd = (args->dfd == LINUX_AT_FDCWD) ? AT_FDCWD : args->dfd; 1018 LCONVPATHCREAT_AT(td, args->filename, &path, dfd); 1019 1020#ifdef DEBUG 1021 if (ldebug(mknodat)) 1022 printf(ARGS(mknodat, "%s, %d, %d"), path, args->mode, args->dev); 1023#endif 1024 1025 switch (args->mode & S_IFMT) { 1026 case S_IFIFO: 1027 case S_IFSOCK: 1028 error = kern_mkfifoat(td, dfd, path, UIO_SYSSPACE, args->mode); 1029 break; 1030 1031 case S_IFCHR: 1032 case S_IFBLK: 1033 error = kern_mknodat(td, dfd, path, UIO_SYSSPACE, args->mode, 1034 args->dev); 1035 break; 1036 1037 case S_IFDIR: 1038 error = EPERM; 1039 break; 1040 1041 case 0: 1042 args->mode |= S_IFREG; 1043 /* FALLTHROUGH */ 1044 case S_IFREG: 1045 error = kern_openat(td, dfd, path, UIO_SYSSPACE, 1046 O_WRONLY | O_CREAT | O_TRUNC, args->mode); 1047 if (error == 0) 1048 kern_close(td, td->td_retval[0]); 1049 break; 1050 1051 default: 1052 error = EINVAL; 1053 break; 1054 } 1055 LFREEPATH(path); 1056 return (error); 1057} 1058 1059/* 1060 * UGH! This is just about the dumbest idea I've ever heard!! 1061 */ 1062int 1063linux_personality(struct thread *td, struct linux_personality_args *args) 1064{ 1065#ifdef DEBUG 1066 if (ldebug(personality)) 1067 printf(ARGS(personality, "%lu"), (unsigned long)args->per); 1068#endif 1069 if (args->per != 0) 1070 return EINVAL; 1071 1072 /* Yes Jim, it's still a Linux... */ 1073 td->td_retval[0] = 0; 1074 return 0; 1075} 1076 1077struct l_itimerval { 1078 l_timeval it_interval; 1079 l_timeval it_value; 1080}; 1081 1082#define B2L_ITIMERVAL(bip, lip) \ 1083 (bip)->it_interval.tv_sec = (lip)->it_interval.tv_sec; \ 1084 (bip)->it_interval.tv_usec = (lip)->it_interval.tv_usec; \ 1085 (bip)->it_value.tv_sec = (lip)->it_value.tv_sec; \ 1086 (bip)->it_value.tv_usec = (lip)->it_value.tv_usec; 1087 1088int 1089linux_setitimer(struct thread *td, struct linux_setitimer_args *uap) 1090{ 1091 int error; 1092 struct l_itimerval ls; 1093 struct itimerval aitv, oitv; 1094 1095#ifdef DEBUG 1096 if (ldebug(setitimer)) 1097 printf(ARGS(setitimer, "%p, %p"), 1098 (void *)uap->itv, (void *)uap->oitv); 1099#endif 1100 1101 if (uap->itv == NULL) { 1102 uap->itv = uap->oitv; 1103 return (linux_getitimer(td, (struct linux_getitimer_args *)uap)); 1104 } 1105 1106 error = copyin(uap->itv, &ls, sizeof(ls)); 1107 if (error != 0) 1108 return (error); 1109 B2L_ITIMERVAL(&aitv, &ls); 1110#ifdef DEBUG 1111 if (ldebug(setitimer)) { 1112 printf("setitimer: value: sec: %jd, usec: %ld\n", 1113 (intmax_t)aitv.it_value.tv_sec, aitv.it_value.tv_usec); 1114 printf("setitimer: interval: sec: %jd, usec: %ld\n", 1115 (intmax_t)aitv.it_interval.tv_sec, aitv.it_interval.tv_usec); 1116 } 1117#endif 1118 error = kern_setitimer(td, uap->which, &aitv, &oitv); 1119 if (error != 0 || uap->oitv == NULL) 1120 return (error); 1121 B2L_ITIMERVAL(&ls, &oitv); 1122 1123 return (copyout(&ls, uap->oitv, sizeof(ls))); 1124} 1125 1126int 1127linux_getitimer(struct thread *td, struct linux_getitimer_args *uap) 1128{ 1129 int error; 1130 struct l_itimerval ls; 1131 struct itimerval aitv; 1132 1133#ifdef DEBUG 1134 if (ldebug(getitimer)) 1135 printf(ARGS(getitimer, "%p"), (void *)uap->itv); 1136#endif 1137 error = kern_getitimer(td, uap->which, &aitv); 1138 if (error != 0) 1139 return (error); 1140 B2L_ITIMERVAL(&ls, &aitv); 1141 return (copyout(&ls, uap->itv, sizeof(ls))); 1142} 1143 1144int 1145linux_nice(struct thread *td, struct linux_nice_args *args) 1146{ 1147 struct setpriority_args bsd_args; 1148 1149 bsd_args.which = PRIO_PROCESS; 1150 bsd_args.who = 0; /* current process */ 1151 bsd_args.prio = args->inc; 1152 return setpriority(td, &bsd_args); 1153} 1154 1155int 1156linux_setgroups(struct thread *td, struct linux_setgroups_args *args) 1157{ 1158 struct ucred *newcred, *oldcred; 1159 l_gid_t linux_gidset[NGROUPS]; 1160 gid_t *bsd_gidset; 1161 int ngrp, error; 1162 struct proc *p; 1163 1164 ngrp = args->gidsetsize; 1165 if (ngrp < 0 || ngrp >= NGROUPS) 1166 return (EINVAL); 1167 error = copyin(args->grouplist, linux_gidset, ngrp * sizeof(l_gid_t)); 1168 if (error) 1169 return (error); 1170 newcred = crget(); 1171 p = td->td_proc; 1172 PROC_LOCK(p); 1173 oldcred = p->p_ucred; 1174 1175 /* 1176 * cr_groups[0] holds egid. Setting the whole set from 1177 * the supplied set will cause egid to be changed too. 1178 * Keep cr_groups[0] unchanged to prevent that. 1179 */ 1180 1181 if ((error = priv_check_cred(oldcred, PRIV_CRED_SETGROUPS, 0)) != 0) { 1182 PROC_UNLOCK(p); 1183 crfree(newcred); 1184 return (error); 1185 } 1186 1187 crcopy(newcred, oldcred); 1188 if (ngrp > 0) { 1189 newcred->cr_ngroups = ngrp + 1; 1190 1191 bsd_gidset = newcred->cr_groups; 1192 ngrp--; 1193 while (ngrp >= 0) { 1194 bsd_gidset[ngrp + 1] = linux_gidset[ngrp]; 1195 ngrp--; 1196 } 1197 } else 1198 newcred->cr_ngroups = 1; 1199 1200 setsugid(p); 1201 p->p_ucred = newcred; 1202 PROC_UNLOCK(p); 1203 crfree(oldcred); 1204 return (0); 1205} 1206 1207int 1208linux_getgroups(struct thread *td, struct linux_getgroups_args *args) 1209{ 1210 struct ucred *cred; 1211 l_gid_t linux_gidset[NGROUPS]; 1212 gid_t *bsd_gidset; 1213 int bsd_gidsetsz, ngrp, error; 1214 1215 cred = td->td_ucred; 1216 bsd_gidset = cred->cr_groups; 1217 bsd_gidsetsz = cred->cr_ngroups - 1; 1218 1219 /* 1220 * cr_groups[0] holds egid. Returning the whole set 1221 * here will cause a duplicate. Exclude cr_groups[0] 1222 * to prevent that. 1223 */ 1224 1225 if ((ngrp = args->gidsetsize) == 0) { 1226 td->td_retval[0] = bsd_gidsetsz; 1227 return (0); 1228 } 1229 1230 if (ngrp < bsd_gidsetsz) 1231 return (EINVAL); 1232 1233 ngrp = 0; 1234 while (ngrp < bsd_gidsetsz) { 1235 linux_gidset[ngrp] = bsd_gidset[ngrp + 1]; 1236 ngrp++; 1237 } 1238 1239 if ((error = copyout(linux_gidset, args->grouplist, 1240 ngrp * sizeof(l_gid_t)))) 1241 return (error); 1242 1243 td->td_retval[0] = ngrp; 1244 return (0); 1245} 1246 1247int 1248linux_setrlimit(struct thread *td, struct linux_setrlimit_args *args) 1249{ 1250 struct rlimit bsd_rlim; 1251 struct l_rlimit rlim; 1252 u_int which; 1253 int error; 1254 1255#ifdef DEBUG 1256 if (ldebug(setrlimit)) 1257 printf(ARGS(setrlimit, "%d, %p"), 1258 args->resource, (void *)args->rlim); 1259#endif 1260 1261 if (args->resource >= LINUX_RLIM_NLIMITS) 1262 return (EINVAL); 1263 1264 which = linux_to_bsd_resource[args->resource]; 1265 if (which == -1) 1266 return (EINVAL); 1267 1268 error = copyin(args->rlim, &rlim, sizeof(rlim)); 1269 if (error) 1270 return (error); 1271 1272 bsd_rlim.rlim_cur = (rlim_t)rlim.rlim_cur; 1273 bsd_rlim.rlim_max = (rlim_t)rlim.rlim_max; 1274 return (kern_setrlimit(td, which, &bsd_rlim)); 1275} 1276 1277int 1278linux_old_getrlimit(struct thread *td, struct linux_old_getrlimit_args *args) 1279{ 1280 struct l_rlimit rlim; 1281 struct proc *p = td->td_proc; 1282 struct rlimit bsd_rlim; 1283 u_int which; 1284 1285#ifdef DEBUG 1286 if (ldebug(old_getrlimit)) 1287 printf(ARGS(old_getrlimit, "%d, %p"), 1288 args->resource, (void *)args->rlim); 1289#endif 1290 1291 if (args->resource >= LINUX_RLIM_NLIMITS) 1292 return (EINVAL); 1293 1294 which = linux_to_bsd_resource[args->resource]; 1295 if (which == -1) 1296 return (EINVAL); 1297 1298 PROC_LOCK(p); 1299 lim_rlimit(p, which, &bsd_rlim); 1300 PROC_UNLOCK(p); 1301 1302#ifdef COMPAT_LINUX32 1303 rlim.rlim_cur = (unsigned int)bsd_rlim.rlim_cur; 1304 if (rlim.rlim_cur == UINT_MAX) 1305 rlim.rlim_cur = INT_MAX; 1306 rlim.rlim_max = (unsigned int)bsd_rlim.rlim_max; 1307 if (rlim.rlim_max == UINT_MAX) 1308 rlim.rlim_max = INT_MAX; 1309#else 1310 rlim.rlim_cur = (unsigned long)bsd_rlim.rlim_cur; 1311 if (rlim.rlim_cur == ULONG_MAX) 1312 rlim.rlim_cur = LONG_MAX; 1313 rlim.rlim_max = (unsigned long)bsd_rlim.rlim_max; 1314 if (rlim.rlim_max == ULONG_MAX) 1315 rlim.rlim_max = LONG_MAX; 1316#endif 1317 return (copyout(&rlim, args->rlim, sizeof(rlim))); 1318} 1319 1320int 1321linux_getrlimit(struct thread *td, struct linux_getrlimit_args *args) 1322{ 1323 struct l_rlimit rlim; 1324 struct proc *p = td->td_proc; 1325 struct rlimit bsd_rlim; 1326 u_int which; 1327 1328#ifdef DEBUG 1329 if (ldebug(getrlimit)) 1330 printf(ARGS(getrlimit, "%d, %p"), 1331 args->resource, (void *)args->rlim); 1332#endif 1333 1334 if (args->resource >= LINUX_RLIM_NLIMITS) 1335 return (EINVAL); 1336 1337 which = linux_to_bsd_resource[args->resource]; 1338 if (which == -1) 1339 return (EINVAL); 1340 1341 PROC_LOCK(p); 1342 lim_rlimit(p, which, &bsd_rlim); 1343 PROC_UNLOCK(p); 1344 1345 rlim.rlim_cur = (l_ulong)bsd_rlim.rlim_cur; 1346 rlim.rlim_max = (l_ulong)bsd_rlim.rlim_max; 1347 return (copyout(&rlim, args->rlim, sizeof(rlim))); 1348} 1349 1350int 1351linux_sched_setscheduler(struct thread *td, 1352 struct linux_sched_setscheduler_args *args) 1353{ 1354 struct sched_setscheduler_args bsd; 1355 1356#ifdef DEBUG 1357 if (ldebug(sched_setscheduler)) 1358 printf(ARGS(sched_setscheduler, "%d, %d, %p"), 1359 args->pid, args->policy, (const void *)args->param); 1360#endif 1361 1362 switch (args->policy) { 1363 case LINUX_SCHED_OTHER: 1364 bsd.policy = SCHED_OTHER; 1365 break; 1366 case LINUX_SCHED_FIFO: 1367 bsd.policy = SCHED_FIFO; 1368 break; 1369 case LINUX_SCHED_RR: 1370 bsd.policy = SCHED_RR; 1371 break; 1372 default: 1373 return EINVAL; 1374 } 1375 1376 bsd.pid = args->pid; 1377 bsd.param = (struct sched_param *)args->param; 1378 return sched_setscheduler(td, &bsd); 1379} 1380 1381int 1382linux_sched_getscheduler(struct thread *td, 1383 struct linux_sched_getscheduler_args *args) 1384{ 1385 struct sched_getscheduler_args bsd; 1386 int error; 1387 1388#ifdef DEBUG 1389 if (ldebug(sched_getscheduler)) 1390 printf(ARGS(sched_getscheduler, "%d"), args->pid); 1391#endif 1392 1393 bsd.pid = args->pid; 1394 error = sched_getscheduler(td, &bsd); 1395 1396 switch (td->td_retval[0]) { 1397 case SCHED_OTHER: 1398 td->td_retval[0] = LINUX_SCHED_OTHER; 1399 break; 1400 case SCHED_FIFO: 1401 td->td_retval[0] = LINUX_SCHED_FIFO; 1402 break; 1403 case SCHED_RR: 1404 td->td_retval[0] = LINUX_SCHED_RR; 1405 break; 1406 } 1407 1408 return error; 1409} 1410 1411int 1412linux_sched_get_priority_max(struct thread *td, 1413 struct linux_sched_get_priority_max_args *args) 1414{ 1415 struct sched_get_priority_max_args bsd; 1416 1417#ifdef DEBUG 1418 if (ldebug(sched_get_priority_max)) 1419 printf(ARGS(sched_get_priority_max, "%d"), args->policy); 1420#endif 1421 1422 switch (args->policy) { 1423 case LINUX_SCHED_OTHER: 1424 bsd.policy = SCHED_OTHER; 1425 break; 1426 case LINUX_SCHED_FIFO: 1427 bsd.policy = SCHED_FIFO; 1428 break; 1429 case LINUX_SCHED_RR: 1430 bsd.policy = SCHED_RR; 1431 break; 1432 default: 1433 return EINVAL; 1434 } 1435 return sched_get_priority_max(td, &bsd); 1436} 1437 1438int 1439linux_sched_get_priority_min(struct thread *td, 1440 struct linux_sched_get_priority_min_args *args) 1441{ 1442 struct sched_get_priority_min_args bsd; 1443 1444#ifdef DEBUG 1445 if (ldebug(sched_get_priority_min)) 1446 printf(ARGS(sched_get_priority_min, "%d"), args->policy); 1447#endif 1448 1449 switch (args->policy) { 1450 case LINUX_SCHED_OTHER: 1451 bsd.policy = SCHED_OTHER; 1452 break; 1453 case LINUX_SCHED_FIFO: 1454 bsd.policy = SCHED_FIFO; 1455 break; 1456 case LINUX_SCHED_RR: 1457 bsd.policy = SCHED_RR; 1458 break; 1459 default: 1460 return EINVAL; 1461 } 1462 return sched_get_priority_min(td, &bsd); 1463} 1464 1465#define REBOOT_CAD_ON 0x89abcdef 1466#define REBOOT_CAD_OFF 0 1467#define REBOOT_HALT 0xcdef0123 1468#define REBOOT_RESTART 0x01234567 1469#define REBOOT_RESTART2 0xA1B2C3D4 1470#define REBOOT_POWEROFF 0x4321FEDC 1471#define REBOOT_MAGIC1 0xfee1dead 1472#define REBOOT_MAGIC2 0x28121969 1473#define REBOOT_MAGIC2A 0x05121996 1474#define REBOOT_MAGIC2B 0x16041998 1475 1476int 1477linux_reboot(struct thread *td, struct linux_reboot_args *args) 1478{ 1479 struct reboot_args bsd_args; 1480 1481#ifdef DEBUG 1482 if (ldebug(reboot)) 1483 printf(ARGS(reboot, "0x%x"), args->cmd); 1484#endif 1485 1486 if (args->magic1 != REBOOT_MAGIC1) 1487 return EINVAL; 1488 1489 switch (args->magic2) { 1490 case REBOOT_MAGIC2: 1491 case REBOOT_MAGIC2A: 1492 case REBOOT_MAGIC2B: 1493 break; 1494 default: 1495 return EINVAL; 1496 } 1497 1498 switch (args->cmd) { 1499 case REBOOT_CAD_ON: 1500 case REBOOT_CAD_OFF: 1501 return (priv_check(td, PRIV_REBOOT)); 1502 case REBOOT_HALT: 1503 bsd_args.opt = RB_HALT; 1504 break; 1505 case REBOOT_RESTART: 1506 case REBOOT_RESTART2: 1507 bsd_args.opt = 0; 1508 break; 1509 case REBOOT_POWEROFF: 1510 bsd_args.opt = RB_POWEROFF; 1511 break; 1512 default: 1513 return EINVAL; 1514 } 1515 return reboot(td, &bsd_args); 1516} 1517 1518 1519/* 1520 * The FreeBSD native getpid(2), getgid(2) and getuid(2) also modify 1521 * td->td_retval[1] when COMPAT_43 is defined. This clobbers registers that 1522 * are assumed to be preserved. The following lightweight syscalls fixes 1523 * this. See also linux_getgid16() and linux_getuid16() in linux_uid16.c 1524 * 1525 * linux_getpid() - MP SAFE 1526 * linux_getgid() - MP SAFE 1527 * linux_getuid() - MP SAFE 1528 */ 1529 1530int 1531linux_getpid(struct thread *td, struct linux_getpid_args *args) 1532{ 1533 struct linux_emuldata *em; 1534 1535#ifdef DEBUG 1536 if (ldebug(getpid)) 1537 printf(ARGS(getpid, "")); 1538#endif 1539 1540 if (linux_use26(td)) { 1541 em = em_find(td->td_proc, EMUL_DONTLOCK); 1542 KASSERT(em != NULL, ("getpid: emuldata not found.\n")); 1543 td->td_retval[0] = em->shared->group_pid; 1544 } else { 1545 td->td_retval[0] = td->td_proc->p_pid; 1546 } 1547 1548 return (0); 1549} 1550 1551int 1552linux_gettid(struct thread *td, struct linux_gettid_args *args) 1553{ 1554 1555#ifdef DEBUG 1556 if (ldebug(gettid)) 1557 printf(ARGS(gettid, "")); 1558#endif 1559 1560 td->td_retval[0] = td->td_proc->p_pid; 1561 return (0); 1562} 1563 1564 1565int 1566linux_getppid(struct thread *td, struct linux_getppid_args *args) 1567{ 1568 struct linux_emuldata *em; 1569 struct proc *p, *pp; 1570 1571#ifdef DEBUG 1572 if (ldebug(getppid)) 1573 printf(ARGS(getppid, "")); 1574#endif 1575 1576 if (!linux_use26(td)) { 1577 PROC_LOCK(td->td_proc); 1578 td->td_retval[0] = td->td_proc->p_pptr->p_pid; 1579 PROC_UNLOCK(td->td_proc); 1580 return (0); 1581 } 1582 1583 em = em_find(td->td_proc, EMUL_DONTLOCK); 1584 1585 KASSERT(em != NULL, ("getppid: process emuldata not found.\n")); 1586 1587 /* find the group leader */ 1588 p = pfind(em->shared->group_pid); 1589 1590 if (p == NULL) { 1591#ifdef DEBUG 1592 printf(LMSG("parent process not found.\n")); 1593#endif 1594 return (0); 1595 } 1596 1597 pp = p->p_pptr; /* switch to parent */ 1598 PROC_LOCK(pp); 1599 PROC_UNLOCK(p); 1600 1601 /* if its also linux process */ 1602 if (pp->p_sysent == &elf_linux_sysvec) { 1603 em = em_find(pp, EMUL_DONTLOCK); 1604 KASSERT(em != NULL, ("getppid: parent emuldata not found.\n")); 1605 1606 td->td_retval[0] = em->shared->group_pid; 1607 } else 1608 td->td_retval[0] = pp->p_pid; 1609 1610 PROC_UNLOCK(pp); 1611 1612 return (0); 1613} 1614 1615int 1616linux_getgid(struct thread *td, struct linux_getgid_args *args) 1617{ 1618 1619#ifdef DEBUG 1620 if (ldebug(getgid)) 1621 printf(ARGS(getgid, "")); 1622#endif 1623 1624 td->td_retval[0] = td->td_ucred->cr_rgid; 1625 return (0); 1626} 1627 1628int 1629linux_getuid(struct thread *td, struct linux_getuid_args *args) 1630{ 1631 1632#ifdef DEBUG 1633 if (ldebug(getuid)) 1634 printf(ARGS(getuid, "")); 1635#endif 1636 1637 td->td_retval[0] = td->td_ucred->cr_ruid; 1638 return (0); 1639} 1640 1641 1642int 1643linux_getsid(struct thread *td, struct linux_getsid_args *args) 1644{ 1645 struct getsid_args bsd; 1646 1647#ifdef DEBUG 1648 if (ldebug(getsid)) 1649 printf(ARGS(getsid, "%i"), args->pid); 1650#endif 1651 1652 bsd.pid = args->pid; 1653 return getsid(td, &bsd); 1654} 1655 1656int 1657linux_nosys(struct thread *td, struct nosys_args *ignore) 1658{ 1659 1660 return (ENOSYS); 1661} 1662 1663int 1664linux_getpriority(struct thread *td, struct linux_getpriority_args *args) 1665{ 1666 struct getpriority_args bsd_args; 1667 int error; 1668 1669#ifdef DEBUG 1670 if (ldebug(getpriority)) 1671 printf(ARGS(getpriority, "%i, %i"), args->which, args->who); 1672#endif 1673 1674 bsd_args.which = args->which; 1675 bsd_args.who = args->who; 1676 error = getpriority(td, &bsd_args); 1677 td->td_retval[0] = 20 - td->td_retval[0]; 1678 return error; 1679} 1680 1681int 1682linux_sethostname(struct thread *td, struct linux_sethostname_args *args) 1683{ 1684 int name[2]; 1685 1686#ifdef DEBUG 1687 if (ldebug(sethostname)) 1688 printf(ARGS(sethostname, "*, %i"), args->len); 1689#endif 1690 1691 name[0] = CTL_KERN; 1692 name[1] = KERN_HOSTNAME; 1693 return (userland_sysctl(td, name, 2, 0, 0, 0, args->hostname, 1694 args->len, 0, 0)); 1695} 1696 1697int 1698linux_exit_group(struct thread *td, struct linux_exit_group_args *args) 1699{ 1700 struct linux_emuldata *em, *td_em, *tmp_em; 1701 struct proc *sp; 1702 1703#ifdef DEBUG 1704 if (ldebug(exit_group)) 1705 printf(ARGS(exit_group, "%i"), args->error_code); 1706#endif 1707 1708 if (linux_use26(td)) { 1709 td_em = em_find(td->td_proc, EMUL_DONTLOCK); 1710 1711 KASSERT(td_em != NULL, ("exit_group: emuldata not found.\n")); 1712 1713 EMUL_SHARED_RLOCK(&emul_shared_lock); 1714 LIST_FOREACH_SAFE(em, &td_em->shared->threads, threads, tmp_em) { 1715 if (em->pid == td_em->pid) 1716 continue; 1717 1718 sp = pfind(em->pid); 1719 psignal(sp, SIGKILL); 1720 PROC_UNLOCK(sp); 1721#ifdef DEBUG 1722 printf(LMSG("linux_sys_exit_group: kill PID %d\n"), em->pid); 1723#endif 1724 } 1725 1726 EMUL_SHARED_RUNLOCK(&emul_shared_lock); 1727 } 1728 /* 1729 * XXX: we should send a signal to the parent if 1730 * SIGNAL_EXIT_GROUP is set. We ignore that (temporarily?) 1731 * as it doesnt occur often. 1732 */ 1733 exit1(td, W_EXITCODE(args->error_code, 0)); 1734 1735 return (0); 1736} 1737 1738int 1739linux_prctl(struct thread *td, struct linux_prctl_args *args) 1740{ 1741 int error = 0, max_size; 1742 struct proc *p = td->td_proc; 1743 char comm[LINUX_MAX_COMM_LEN]; 1744 struct linux_emuldata *em; 1745 int pdeath_signal; 1746 1747#ifdef DEBUG 1748 if (ldebug(prctl)) 1749 printf(ARGS(prctl, "%d, %d, %d, %d, %d"), args->option, 1750 args->arg2, args->arg3, args->arg4, args->arg5); 1751#endif 1752 1753 switch (args->option) { 1754 case LINUX_PR_SET_PDEATHSIG: 1755 if (!LINUX_SIG_VALID(args->arg2)) 1756 return (EINVAL); 1757 em = em_find(p, EMUL_DOLOCK); 1758 KASSERT(em != NULL, ("prctl: emuldata not found.\n")); 1759 em->pdeath_signal = args->arg2; 1760 EMUL_UNLOCK(&emul_lock); 1761 break; 1762 case LINUX_PR_GET_PDEATHSIG: 1763 em = em_find(p, EMUL_DOLOCK); 1764 KASSERT(em != NULL, ("prctl: emuldata not found.\n")); 1765 pdeath_signal = em->pdeath_signal; 1766 EMUL_UNLOCK(&emul_lock); 1767 error = copyout(&pdeath_signal, 1768 (void *)(register_t)args->arg2, 1769 sizeof(pdeath_signal)); 1770 break; 1771 case LINUX_PR_SET_NAME: 1772 /* 1773 * To be on the safe side we need to make sure to not 1774 * overflow the size a linux program expects. We already 1775 * do this here in the copyin, so that we don't need to 1776 * check on copyout. 1777 */ 1778 max_size = MIN(sizeof(comm), sizeof(p->p_comm)); 1779 error = copyinstr((void *)(register_t)args->arg2, comm, 1780 max_size, NULL); 1781 1782 /* Linux silently truncates the name if it is too long. */ 1783 if (error == ENAMETOOLONG) { 1784 /* 1785 * XXX: copyinstr() isn't documented to populate the 1786 * array completely, so do a copyin() to be on the 1787 * safe side. This should be changed in case 1788 * copyinstr() is changed to guarantee this. 1789 */ 1790 error = copyin((void *)(register_t)args->arg2, comm, 1791 max_size - 1); 1792 comm[max_size - 1] = '\0'; 1793 } 1794 if (error) 1795 return (error); 1796 1797 PROC_LOCK(p); 1798 strlcpy(p->p_comm, comm, sizeof(p->p_comm)); 1799 PROC_UNLOCK(p); 1800 break; 1801 case LINUX_PR_GET_NAME: 1802 PROC_LOCK(p); 1803 strlcpy(comm, p->p_comm, sizeof(comm)); 1804 PROC_UNLOCK(p); 1805 error = copyout(comm, (void *)(register_t)args->arg2, 1806 strlen(comm) + 1); 1807 break; 1808 default: 1809 error = EINVAL; 1810 break; 1811 } 1812 1813 return (error); 1814} 1815 1816/* 1817 * Get affinity of a process. 1818 */ 1819int 1820linux_sched_getaffinity(struct thread *td, 1821 struct linux_sched_getaffinity_args *args) 1822{ 1823 int error; 1824 struct cpuset_getaffinity_args cga; 1825 1826#ifdef DEBUG 1827 if (ldebug(sched_getaffinity)) 1828 printf(ARGS(sched_getaffinity, "%d, %d, *"), args->pid, 1829 args->len); 1830#endif 1831 1832 cga.level = CPU_LEVEL_WHICH; 1833 cga.which = CPU_WHICH_PID; 1834 cga.id = args->pid; 1835 cga.cpusetsize = sizeof(cpumask_t); 1836 cga.mask = (cpuset_t *) args->user_mask_ptr; 1837 1838 if ((error = cpuset_getaffinity(td, &cga)) == 0) 1839 td->td_retval[0] = sizeof(cpumask_t); 1840 1841 return (error); 1842} 1843 1844/* 1845 * Set affinity of a process. 1846 */ 1847int 1848linux_sched_setaffinity(struct thread *td, 1849 struct linux_sched_setaffinity_args *args) 1850{ 1851 struct cpuset_setaffinity_args csa; 1852 1853#ifdef DEBUG 1854 if (ldebug(sched_setaffinity)) 1855 printf(ARGS(sched_setaffinity, "%d, %d, *"), args->pid, 1856 args->len); 1857#endif 1858 csa.level = CPU_LEVEL_WHICH; 1859 csa.which = CPU_WHICH_PID; 1860 csa.id = args->pid; 1861 csa.cpusetsize = args->len; 1862 csa.mask = (cpuset_t *) args->user_mask_ptr; 1863 1864 return (cpuset_setaffinity(td, &csa)); 1865}
|