/linux-master/tools/testing/selftests/bpf/verifier/ |
H A D | atomic_invalid.c | 15 __INVALID_ATOMIC_ACCESS_TEST(BPF_ADD | BPF_FETCH), 17 __INVALID_ATOMIC_ACCESS_TEST(BPF_ADD | BPF_FETCH), 19 __INVALID_ATOMIC_ACCESS_TEST(BPF_AND | BPF_FETCH), 21 __INVALID_ATOMIC_ACCESS_TEST(BPF_OR | BPF_FETCH), 23 __INVALID_ATOMIC_ACCESS_TEST(BPF_XOR | BPF_FETCH),
|
H A D | atomic_fetch.c | 10 BPF_ATOMIC_OP(BPF_DW, BPF_AND | BPF_FETCH, BPF_REG_2, BPF_REG_1, 0), 34 BPF_ATOMIC_OP(BPF_DW, BPF_AND | BPF_FETCH, BPF_REG_2, BPF_REG_1, 0), 58 BPF_ATOMIC_OP(BPF_W, BPF_AND | BPF_FETCH, BPF_REG_2, BPF_REG_1, 0), 81 BPF_ATOMIC_OP(BPF_W, BPF_AND | BPF_FETCH, BPF_REG_2, BPF_REG_1, 0), 121 __ATOMIC_FETCH_OP_TEST(BPF_REG_1, BPF_REG_2, 1, BPF_ADD | BPF_FETCH, 2, 3), 122 __ATOMIC_FETCH_OP_TEST(BPF_REG_0, BPF_REG_1, 1, BPF_ADD | BPF_FETCH, 2, 3), 123 __ATOMIC_FETCH_OP_TEST(BPF_REG_1, BPF_REG_0, 1, BPF_ADD | BPF_FETCH, 2, 3), 124 __ATOMIC_FETCH_OP_TEST(BPF_REG_2, BPF_REG_3, 1, BPF_ADD | BPF_FETCH, 2, 3), 125 __ATOMIC_FETCH_OP_TEST(BPF_REG_4, BPF_REG_5, 1, BPF_ADD | BPF_FETCH, 2, 3), 126 __ATOMIC_FETCH_OP_TEST(BPF_REG_9, BPF_REG_8, 1, BPF_ADD | BPF_FETCH, [all...] |
H A D | atomic_fetch_add.c | 9 BPF_ATOMIC_OP(BPF_DW, BPF_ADD | BPF_FETCH, BPF_REG_10, BPF_REG_1, -8), 31 BPF_ATOMIC_OP(BPF_W, BPF_ADD | BPF_FETCH, BPF_REG_10, BPF_REG_1, -4), 50 BPF_ATOMIC_OP(BPF_DW, BPF_ADD | BPF_FETCH, BPF_REG_10, BPF_REG_10, -8), 62 BPF_ATOMIC_OP(BPF_DW, BPF_ADD | BPF_FETCH, BPF_REG_10, BPF_REG_2, -8), 75 BPF_ATOMIC_OP(BPF_DW, BPF_ADD | BPF_FETCH, BPF_REG_2, BPF_REG_0, -8), 96 BPF_ATOMIC_OP(BPF_DW, BPF_ADD | BPF_FETCH, BPF_REG_2, BPF_REG_3, 0),
|
H A D | atomic_bounds.c | 18 BPF_ATOMIC_OP(BPF_DW, BPF_ADD | BPF_FETCH, BPF_REG_10, BPF_REG_1, -8),
|
H A D | atomic_and.c | 14 /* r1 should not be clobbered, no BPF_FETCH flag */ 30 BPF_ATOMIC_OP(BPF_DW, BPF_AND | BPF_FETCH, BPF_REG_10, BPF_REG_1, -8), 60 BPF_ATOMIC_OP(BPF_W, BPF_AND | BPF_FETCH, BPF_REG_10, BPF_REG_1, -4), 85 BPF_ATOMIC_OP(BPF_DW, BPF_AND | BPF_FETCH, BPF_REG_10, BPF_REG_0, -8),
|
H A D | atomic_or.c | 14 /* r1 should not be clobbered, no BPF_FETCH flag */ 30 BPF_ATOMIC_OP(BPF_DW, BPF_OR | BPF_FETCH, BPF_REG_10, BPF_REG_1, -8), 60 BPF_ATOMIC_OP(BPF_W, BPF_OR | BPF_FETCH, BPF_REG_10, BPF_REG_1, -4), 88 BPF_ATOMIC_OP(BPF_W, BPF_OR | BPF_FETCH, BPF_REG_10, BPF_REG_1, -8),
|
H A D | atomic_xor.c | 14 /* r1 should not be clobbered, no BPF_FETCH flag */ 30 BPF_ATOMIC_OP(BPF_DW, BPF_XOR | BPF_FETCH, BPF_REG_10, BPF_REG_1, -8), 60 BPF_ATOMIC_OP(BPF_W, BPF_XOR | BPF_FETCH, BPF_REG_10, BPF_REG_1, -4),
|
/linux-master/arch/mips/net/ |
H A D | bpf_jit_comp64.c | 391 case BPF_ADD | BPF_FETCH: 395 case BPF_AND | BPF_FETCH: 399 case BPF_OR | BPF_FETCH: 403 case BPF_XOR | BPF_FETCH: 414 if (code & BPF_FETCH) { 853 case BPF_ADD | BPF_FETCH: 855 case BPF_AND | BPF_FETCH: 857 case BPF_OR | BPF_FETCH: 859 case BPF_XOR | BPF_FETCH: 863 } else if (imm & BPF_FETCH) { [all...] |
H A D | bpf_jit_comp32.c | 741 case BPF_ADD | BPF_FETCH: 747 case BPF_SUB | BPF_FETCH: 753 case BPF_OR | BPF_FETCH: 759 case BPF_AND | BPF_FETCH: 765 case BPF_XOR | BPF_FETCH: 777 if (code & BPF_FETCH) { 822 case BPF_ADD | BPF_FETCH: 828 case BPF_SUB | BPF_FETCH: 834 case BPF_OR | BPF_FETCH: 840 case BPF_AND | BPF_FETCH [all...] |
H A D | bpf_jit_comp.c | 415 case BPF_ADD | BPF_FETCH: 419 case BPF_AND | BPF_FETCH: 423 case BPF_OR | BPF_FETCH: 427 case BPF_XOR | BPF_FETCH: 438 if (code & BPF_FETCH) {
|
/linux-master/kernel/bpf/ |
H A D | disasm.c | 231 (insn->imm == (BPF_ADD | BPF_FETCH) || 232 insn->imm == (BPF_AND | BPF_FETCH) || 233 insn->imm == (BPF_OR | BPF_FETCH) || 234 insn->imm == (BPF_XOR | BPF_FETCH))) {
|
H A D | core.c | 2141 case BOP | BPF_FETCH: \
|
H A D | verifier.c | 3207 (insn->imm & BPF_FETCH)) { 6982 case BPF_ADD | BPF_FETCH: 6984 case BPF_AND | BPF_FETCH: 6986 case BPF_OR | BPF_FETCH: 6988 case BPF_XOR | BPF_FETCH: 7042 if (insn->imm & BPF_FETCH) {
|
/linux-master/tools/testing/selftests/bpf/ |
H A D | disasm.c | 231 (insn->imm == (BPF_ADD | BPF_FETCH) || 232 insn->imm == (BPF_AND | BPF_FETCH) || 233 insn->imm == (BPF_OR | BPF_FETCH) || 234 insn->imm == (BPF_XOR | BPF_FETCH))) {
|
/linux-master/arch/arm64/net/ |
H A D | bpf_jit_comp.c | 520 case BPF_ADD | BPF_FETCH: 523 case BPF_AND | BPF_FETCH: 527 case BPF_OR | BPF_FETCH: 530 case BPF_XOR | BPF_FETCH: 594 } else if (imm == (BPF_ADD | BPF_FETCH) || 595 imm == (BPF_AND | BPF_FETCH) || 596 imm == (BPF_OR | BPF_FETCH) || 597 imm == (BPF_XOR | BPF_FETCH)) { 603 if (imm == (BPF_ADD | BPF_FETCH)) 605 else if (imm == (BPF_AND | BPF_FETCH)) [all...] |
/linux-master/arch/powerpc/net/ |
H A D | bpf_jit_comp64.c | 814 if (imm & BPF_FETCH) 819 case BPF_ADD | BPF_FETCH: 823 case BPF_AND | BPF_FETCH: 827 case BPF_OR | BPF_FETCH: 831 case BPF_XOR | BPF_FETCH: 867 if (imm & BPF_FETCH) {
|
H A D | bpf_jit_comp32.c | 860 if (imm & BPF_FETCH) 865 case BPF_ADD | BPF_FETCH: 869 case BPF_AND | BPF_FETCH: 873 case BPF_OR | BPF_FETCH: 877 case BPF_XOR | BPF_FETCH: 906 /* For the BPF_FETCH variant, get old data into src_reg */ 907 if (imm & BPF_FETCH) {
|
/linux-master/lib/ |
H A D | test_bpf.c | 1613 if (op & BPF_FETCH) 1660 if (op & BPF_FETCH) 1808 return __bpf_fill_atomic64(self, BPF_ADD | BPF_FETCH); 1813 return __bpf_fill_atomic64(self, BPF_AND | BPF_FETCH); 1818 return __bpf_fill_atomic64(self, BPF_OR | BPF_FETCH); 1823 return __bpf_fill_atomic64(self, BPF_XOR | BPF_FETCH); 1860 return __bpf_fill_atomic32(self, BPF_ADD | BPF_FETCH); 1865 return __bpf_fill_atomic32(self, BPF_AND | BPF_FETCH); 1870 return __bpf_fill_atomic32(self, BPF_OR | BPF_FETCH); 1875 return __bpf_fill_atomic32(self, BPF_XOR | BPF_FETCH); [all...] |
/linux-master/arch/loongarch/net/ |
H A D | bpf_jit.c | 325 case BPF_ADD | BPF_FETCH: 333 case BPF_AND | BPF_FETCH: 341 case BPF_OR | BPF_FETCH: 349 case BPF_XOR | BPF_FETCH:
|
/linux-master/arch/s390/net/ |
H A D | bpf_jit_comp.c | 1428 (insn->imm & BPF_FETCH) ? src_reg : REG_W0, \ 1430 if (is32 && (insn->imm & BPF_FETCH)) \ 1434 case BPF_ADD | BPF_FETCH: 1439 case BPF_AND | BPF_FETCH: 1444 case BPF_OR | BPF_FETCH: 1449 case BPF_XOR | BPF_FETCH:
|
/linux-master/include/uapi/linux/ |
H A D | bpf.h | 50 #define BPF_FETCH 0x01 /* not an opcode on its own, used to build others */ macro 51 #define BPF_XCHG (0xe0 | BPF_FETCH) /* atomic exchange */ 52 #define BPF_CMPXCHG (0xf0 | BPF_FETCH) /* atomic compare-and-write */
|
/linux-master/tools/include/uapi/linux/ |
H A D | bpf.h | 50 #define BPF_FETCH 0x01 /* not an opcode on its own, used to build others */ macro 51 #define BPF_XCHG (0xe0 | BPF_FETCH) /* atomic exchange */ 52 #define BPF_CMPXCHG (0xf0 | BPF_FETCH) /* atomic compare-and-write */
|
/linux-master/arch/riscv/net/ |
H A D | bpf_jit_comp64.c | 500 case BPF_ADD | BPF_FETCH: 506 case BPF_AND | BPF_FETCH: 512 case BPF_OR | BPF_FETCH: 518 case BPF_XOR | BPF_FETCH:
|
/linux-master/arch/x86/net/ |
H A D | bpf_jit_comp.c | 1149 case BPF_ADD | BPF_FETCH: 1916 if (insn->imm == (BPF_AND | BPF_FETCH) || 1917 insn->imm == (BPF_OR | BPF_FETCH) || 1918 insn->imm == (BPF_XOR | BPF_FETCH)) {
|