Lines Matching refs:NBYTES
44 #define NBYTES 8
51 #define NBYTES 4
55 #define UNIT(unit) ((unit)*NBYTES)
350 #define NBYTES 8
367 #define NBYTES 4
388 #define FIRST(unit) ((unit)*NBYTES)
389 #define REST(unit) (FIRST(unit)+NBYTES-1)
391 #define ADDRMASK (NBYTES-1)
418 * If len < NBYTES use byte operations.
420 sltu t2, len, NBYTES
434 beqz t0, .Lcleanup_both_aligned # len < 8*NBYTES
436 SUB len, 8*NBYTES # subtract here for bgez loop
447 SUB len, len, 8*NBYTES
448 ADD src, src, 8*NBYTES
466 ADD dst, dst, 8*NBYTES
469 ADD len, 8*NBYTES # revert len (see above)
472 * len == the number of bytes left to copy < 8*NBYTES
477 sltu t0, len, 4*NBYTES
479 and rem, len, (NBYTES-1) # rem = len % NBYTES
481 * len >= 4*NBYTES
487 SUB len, len, 4*NBYTES
488 ADD src, src, 4*NBYTES
498 ADD dst, dst, 4*NBYTES
503 * rem = len % NBYTES
509 ADD src, src, NBYTES
510 SUB len, len, NBYTES
514 ADD dst, dst, NBYTES
519 * src and dst are aligned, need to copy rem bytes (rem < NBYTES)
532 li bits, 8*NBYTES
548 * len >= NBYTES
555 ADD t2, zero, NBYTES
573 and rem, len, (4*NBYTES-1) # rem = len % 4*NBYTES
583 SUB len, len, 4*NBYTES
590 ADD src, src, 4*NBYTES
603 ADD dst, dst, 4*NBYTES
609 and rem, len, NBYTES-1 # rem = len % NBYTES
615 ADD src, src, NBYTES
616 SUB len, len, NBYTES
620 ADD dst, dst, NBYTES
628 /* 0 < len < NBYTES */
633 #define SHIFT_START 8*(NBYTES-1)
656 EXC( lbu t0, NBYTES-2(src), .Ll_exc_copy)
658 EXC( sb t0, NBYTES-2(dst), .Ls_exc)