/linux-master/drivers/net/dsa/sja1105/ |
H A D | sja1105_static_config.c | 57 size_t len, enum packing_op op) 59 int rc = packing(buf, val, start, end, len, op, QUIRK_LSW32_IS_FIRST); 95 enum packing_op op) 100 sja1105_packing(buf, &entry->destmeta, 95, 48, size, op); 101 sja1105_packing(buf, &entry->srcmeta, 47, 0, size, op); 106 enum packing_op op) 111 sja1105_packing(buf, &entry->cas_master, 126, 126, size, op); 112 sja1105_packing(buf, &entry->destmeta, 125, 78, size, op); 113 sja1105_packing(buf, &entry->srcmeta, 77, 30, size, op); 118 enum packing_op op) 56 sja1105_packing(void *buf, u64 *val, int start, int end, size_t len, enum packing_op op) argument 94 sja1105et_avb_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 105 sja1105pqrs_avb_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 117 sja1105et_general_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 149 sja1105pqrs_general_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 183 sja1110_general_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 221 sja1105_l2_forwarding_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 235 sja1110_l2_forwarding_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 249 sja1105_l2_forwarding_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 265 sja1110_l2_forwarding_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 287 sja1105et_l2_lookup_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 302 sja1105pqrs_l2_lookup_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 324 sja1110_l2_lookup_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 346 sja1105et_l2_lookup_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 360 sja1105pqrs_l2_lookup_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 388 sja1110_l2_lookup_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 418 sja1105_l2_policing_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 432 sja1110_l2_policing_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 446 sja1105et_mac_config_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 480 sja1105pqrs_mac_config_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 514 sja1110_mac_config_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 549 sja1105_schedule_entry_points_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 561 sja1105_schedule_entry_points_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 574 sja1110_schedule_entry_points_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 586 sja1105_schedule_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 599 sja1110_schedule_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 612 sja1105_schedule_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 631 sja1110_schedule_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 651 sja1105_vl_forwarding_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 666 sja1110_vl_forwarding_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 680 sja1105_vl_forwarding_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 693 sja1110_vl_forwarding_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 706 sja1105_vl_lookup_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 740 sja1110_vl_lookup_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 774 sja1105_vl_policing_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 790 sja1110_vl_policing_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 806 sja1105_vlan_lookup_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 821 sja1110_vlan_lookup_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 837 sja1105_xmii_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 853 sja1110_xmii_params_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 871 sja1105_retagging_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 887 sja1110_retagging_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 903 sja1110_pcp_remapping_entry_packing(void *buf, void *entry_ptr, enum packing_op op) argument 917 sja1105_table_header_packing(void *buf, void *entry_ptr, enum packing_op op) argument [all...] |
/linux-master/drivers/acpi/acpica/ |
H A D | dswexec.c | 78 status = acpi_ds_create_operand(walk_state, walk_state->op, 0); 148 walk_state->op)); 172 * out_op - Where to return op if a new one is created 186 union acpi_parse_object *op; local 192 op = walk_state->op; 193 if (!op) { 199 op = *out_op; 200 walk_state->op = op; 329 union acpi_parse_object *op; local [all...] |
H A D | psparse.c | 91 * op - Op to complete 101 union acpi_parse_object *op) 109 ACPI_FUNCTION_TRACE_PTR(ps_complete_this_op, op); 113 if (!op) { 117 acpi_ex_stop_trace_opcode(op, walk_state); 119 /* Delete this op and the subtree below it if asked to */ 129 if (op->common.parent) { 130 prev = op->common.parent->common.value.arg; 140 * with a return value op (placeholder op) 100 acpi_ps_complete_this_op(struct acpi_walk_state *walk_state, union acpi_parse_object *op) argument 297 acpi_ps_next_parse_state(struct acpi_walk_state *walk_state, union acpi_parse_object *op, acpi_status callback_status) argument [all...] |
/linux-master/arch/powerpc/kvm/ |
H A D | emulate_loadstore.c | 76 struct instruction_op op; local 96 if (analyse_instr(&op, &vcpu->arch.regs, inst) == 0) { 97 int type = op.type & INSTR_TYPE_MASK; 98 int size = GETSIZE(op.type); 104 int instr_byte_swap = op.type & BYTEREV; 106 if (op.type & SIGNEXT) 108 op.reg, size, !instr_byte_swap); 111 op.reg, size, !instr_byte_swap); 113 if ((op.type & UPDATE) && (emulated != EMULATE_FAIL)) 114 kvmppc_set_gpr(vcpu, op [all...] |
/linux-master/tools/crypto/tcrypt/ |
H A D | tcrypt_speed_compare.py | 62 op = match.group(2) 63 return alg, op 94 alg, op = "", "" 101 alg, op = _alg, _op 104 if op not in result[alg]: 105 result[alg][op] = [] 109 result[alg][op].append(parsed_result) 117 for op in base[alg].keys(): 118 if op not in merged[alg]: 119 merged[alg][op] [all...] |
/linux-master/drivers/spi/ |
H A D | spi-wpcm-fiu.c | 136 bool (*match)(const struct spi_mem_op *op); 137 int (*exec)(struct spi_mem *mem, const struct spi_mem_op *op); 140 static bool wpcm_fiu_normal_match(const struct spi_mem_op *op) argument 143 if (op->cmd.opcode == 0x0b) 146 return (op->addr.nbytes == 0 || op->addr.nbytes == 3) && 147 op->dummy.nbytes == 0 && op->data.nbytes <= 4; 150 static int wpcm_fiu_normal_exec(struct spi_mem *mem, const struct spi_mem_op *op) argument 155 wpcm_fiu_set_opcode(fiu, op 169 wpcm_fiu_fast_read_match(const struct spi_mem_op *op) argument 177 wpcm_fiu_fast_read_exec(struct spi_mem *mem, const struct spi_mem_op *op) argument 190 wpcm_fiu_4ba_match(const struct spi_mem_op *op) argument 195 wpcm_fiu_4ba_exec(struct spi_mem *mem, const struct spi_mem_op *op) argument 233 wpcm_fiu_rdid_match(const struct spi_mem_op *op) argument 240 wpcm_fiu_rdid_exec(struct spi_mem *mem, const struct spi_mem_op *op) argument 266 wpcm_fiu_dummy_match(const struct spi_mem_op *op) argument 277 wpcm_fiu_dummy_exec(struct spi_mem *mem, const struct spi_mem_op *op) argument 308 wpcm_fiu_find_op_shape(const struct spi_mem_op *op) argument 322 wpcm_fiu_supports_op(struct spi_mem *mem, const struct spi_mem_op *op) argument 352 wpcm_fiu_exec_op(struct spi_mem *mem, const struct spi_mem_op *op) argument 367 wpcm_fiu_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op) argument [all...] |
H A D | spi-mem.c | 23 * @op: the memory operation containing the buffer to map 27 * Some controllers might want to do DMA on the data buffer embedded in @op. 32 * op->data.buf.{in,out} is DMA-able before calling this function. 37 const struct spi_mem_op *op, 42 if (!op->data.nbytes) 45 if (op->data.dir == SPI_MEM_DATA_OUT && ctlr->dma_tx) 47 else if (op->data.dir == SPI_MEM_DATA_IN && ctlr->dma_rx) 55 return spi_map_buf(ctlr, dmadev, sgt, op->data.buf.in, op->data.nbytes, 56 op 36 spi_controller_dma_map_mem_op_data(struct spi_controller *ctlr, const struct spi_mem_op *op, struct sg_table *sgt) argument 82 spi_controller_dma_unmap_mem_op_data(struct spi_controller *ctlr, const struct spi_mem_op *op, struct sg_table *sgt) argument 142 spi_mem_check_buswidth(struct spi_mem *mem, const struct spi_mem_op *op) argument 164 spi_mem_default_supports_op(struct spi_mem *mem, const struct spi_mem_op *op) argument 199 spi_mem_check_op(const struct spi_mem_op *op) argument 227 spi_mem_internal_supports_op(struct spi_mem *mem, const struct spi_mem_op *op) argument 253 spi_mem_supports_op(struct spi_mem *mem, const struct spi_mem_op *op) argument 300 spi_mem_add_op_stats(struct spi_statistics __percpu *pcpu_stats, const struct spi_mem_op *op, int exec_op_ret) argument 355 spi_mem_exec_op(struct spi_mem *mem, const struct spi_mem_op *op) argument 502 spi_mem_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op) argument 531 struct spi_mem_op op = desc->info.op_tmpl; local 551 struct spi_mem_op op = desc->info.op_tmpl; local 807 spi_mem_read_status(struct spi_mem *mem, const struct spi_mem_op *op, u16 *status) argument 842 spi_mem_poll_status(struct spi_mem *mem, const struct spi_mem_op *op, u16 mask, u16 match, unsigned long initial_delay_us, unsigned long polling_delay_us, u16 timeout_ms) argument [all...] |
/linux-master/arch/arm64/include/asm/ |
H A D | atomic.h | 19 #define ATOMIC_OP(op) \ 20 static __always_inline void arch_##op(int i, atomic_t *v) \ 22 __lse_ll_sc_body(op, i, v); \ 34 #define ATOMIC_FETCH_OP(name, op) \ 35 static __always_inline int arch_##op##name(int i, atomic_t *v) \ 37 return __lse_ll_sc_body(op##name, i, v); \ 40 #define ATOMIC_FETCH_OPS(op) \ 41 ATOMIC_FETCH_OP(_relaxed, op) \ 42 ATOMIC_FETCH_OP(_acquire, op) \ 43 ATOMIC_FETCH_OP(_release, op) \ [all...] |
/linux-master/arch/powerpc/kernel/ |
H A D | optprobes.c | 50 struct instruction_op op; local 88 analyse_instr(&op, ®s, ppc_inst_read(p->ainsn.insn)) == 1) { 89 emulate_update_regs(®s, &op); 96 static void optimized_callback(struct optimized_kprobe *op, argument 99 /* This is possible if op is under delayed unoptimizing */ 100 if (kprobe_disabled(&op->kp)) 106 kprobes_inc_nmissed_count(&op->kp); 108 __this_cpu_write(current_kprobe, &op->kp); 109 regs_set_return_ip(regs, (unsigned long)op->kp.addr); 111 opt_pre_handler(&op 119 arch_remove_optimized_kprobe(struct optimized_kprobe *op) argument 154 arch_prepare_optimized_kprobe(struct optimized_kprobe *op, struct kprobe *p) argument 261 arch_check_optimized_kprobe(struct optimized_kprobe *op) argument 269 struct optimized_kprobe *op; local 284 arch_unoptimize_kprobe(struct optimized_kprobe *op) argument 291 struct optimized_kprobe *op; local 300 arch_within_optimized_kprobe(struct optimized_kprobe *op, kprobe_opcode_t *addr) argument [all...] |
/linux-master/fs/afs/ |
H A D | yfsclient.c | 333 struct afs_operation *op = call->op; local 342 xdr_decode_YFSFetchStatus(&bp, call, &op->file[0].scb); 343 xdr_decode_YFSVolSync(&bp, &op->volsync); 354 struct afs_operation *op = call->op; local 355 struct afs_vnode_param *vp = &op->file[0]; 356 struct afs_read *req = op->fetch.req; 435 xdr_decode_YFSVolSync(&bp, &op->volsync); 456 .op 464 yfs_fs_fetch_data(struct afs_operation *op) argument 506 struct afs_operation *op = call->op; local 543 yfs_fs_create_file(struct afs_operation *op) argument 594 yfs_fs_make_dir(struct afs_operation *op) argument 638 struct afs_operation *op = call->op; local 684 yfs_fs_remove_file2(struct afs_operation *op) argument 723 struct afs_operation *op = call->op; local 753 yfs_fs_remove_file(struct afs_operation *op) argument 798 yfs_fs_remove_dir(struct afs_operation *op) argument 835 struct afs_operation *op = call->op; local 868 yfs_fs_link(struct afs_operation *op) argument 909 struct afs_operation *op = call->op; local 945 yfs_fs_symlink(struct afs_operation *op) argument 991 struct afs_operation *op = call->op; local 1027 yfs_fs_rename(struct afs_operation *op) argument 1079 yfs_fs_store_data(struct afs_operation *op) argument 1143 yfs_fs_setattr_size(struct afs_operation *op) argument 1183 yfs_fs_setattr(struct afs_operation *op) argument 1223 struct afs_operation *op = call->op; local 1359 yfs_fs_get_volume_status(struct afs_operation *op) argument 1424 yfs_fs_set_lock(struct afs_operation *op) argument 1457 yfs_fs_extend_lock(struct afs_operation *op) argument 1488 yfs_fs_release_lock(struct afs_operation *op) argument 1521 struct afs_operation *op = call->op; local 1553 yfs_fs_fetch_status(struct afs_operation *op) argument 1588 struct afs_operation *op = call->op; local 1728 yfs_fs_inline_bulk_status(struct afs_operation *op) argument 1769 struct afs_operation *op = call->op; local 1897 yfs_fs_fetch_opaque_acl(struct afs_operation *op) argument 1940 yfs_fs_store_opaque_acl2(struct afs_operation *op) argument [all...] |
H A D | xattr.c | 17 static void afs_acl_success(struct afs_operation *op) argument 19 afs_vnode_commit_status(op, &op->file[0]); 22 static void afs_acl_put(struct afs_operation *op) argument 24 kfree(op->acl); 41 struct afs_operation *op; local 46 op = afs_alloc_operation(NULL, vnode->volume); 47 if (IS_ERR(op)) 50 afs_op_set_vnode(op, 0, vnode); 51 op 73 afs_make_acl(struct afs_operation *op, const void *buffer, size_t size) argument 105 struct afs_operation *op; local 143 struct afs_operation *op; local 236 struct afs_operation *op; local [all...] |
/linux-master/drivers/xen/xen-pciback/ |
H A D | pciback_ops.c | 143 struct pci_dev *dev, struct xen_pci_op *op) 158 op->value = 0; 165 op->value = dev->irq ? xen_pirq_from_irq(dev->irq) : 0; 167 dev_dbg(&dev->dev, "MSI: %d\n", op->value); 178 struct pci_dev *dev, struct xen_pci_op *op) 189 op->value = dev->irq ? xen_pirq_from_irq(dev->irq) : 0; 191 dev_dbg(&dev->dev, "MSI: %d\n", op->value); 198 struct pci_dev *dev, struct xen_pci_op *op) 207 if (op->value > SH_INFO_MAX_VEC) 222 entries = kmalloc_array(op 142 xen_pcibk_enable_msi(struct xen_pcibk_device *pdev, struct pci_dev *dev, struct xen_pci_op *op) argument 177 xen_pcibk_disable_msi(struct xen_pcibk_device *pdev, struct pci_dev *dev, struct xen_pci_op *op) argument 197 xen_pcibk_enable_msix(struct xen_pcibk_device *pdev, struct pci_dev *dev, struct xen_pci_op *op) argument 256 xen_pcibk_disable_msix(struct xen_pcibk_device *pdev, struct pci_dev *dev, struct xen_pci_op *op) argument 325 struct xen_pci_op *op = &pdev->op; local [all...] |
/linux-master/arch/x86/kernel/kprobes/ |
H A D | opt.c | 39 struct optimized_kprobe *op; local 48 op = container_of(kp, struct optimized_kprobe, kp); 49 /* If op is optimized or under unoptimizing */ 50 if (list_empty(&op->list) || optprobe_queued_unopt(op)) 60 * bytes must be recovered from op->optinsn.copied_insn buffer. 68 memcpy(buf + 1, op->optinsn.copied_insn, DISP32_SIZE); 71 memcpy(buf, op->optinsn.copied_insn + offs, DISP32_SIZE - offs); 177 optimized_callback(struct optimized_kprobe *op, struct pt_regs *regs) argument 179 /* This is possible if op i 341 arch_check_optimized_kprobe(struct optimized_kprobe *op) argument 356 arch_within_optimized_kprobe(struct optimized_kprobe *op, kprobe_opcode_t *addr) argument 365 __arch_remove_optimized_kprobe(struct optimized_kprobe *op, int dirty) argument 381 arch_remove_optimized_kprobe(struct optimized_kprobe *op) argument 391 arch_prepare_optimized_kprobe(struct optimized_kprobe *op, struct kprobe *__unused) argument 475 struct optimized_kprobe *op, *tmp; local 504 arch_unoptimize_kprobe(struct optimized_kprobe *op) argument 532 struct optimized_kprobe *op, *tmp; local 542 struct optimized_kprobe *op; local [all...] |
/linux-master/tools/objtool/arch/loongarch/ |
H A D | decode.c | 73 #define ADD_OP(op) \ 74 if (!(op = calloc(1, sizeof(*op)))) \ 76 else for (*ops_list = op, ops_list = &op->next; op; op = NULL) 120 struct stack_op *op) 127 ADD_OP(op) { 128 op 117 decode_insn_reg2i12_fomat(union loongarch_instruction inst, struct instruction *insn, struct stack_op **ops_list, struct stack_op *op) argument 176 decode_insn_reg2i14_fomat(union loongarch_instruction inst, struct instruction *insn, struct stack_op **ops_list, struct stack_op *op) argument 281 struct stack_op *op = NULL; local [all...] |
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_uvmm.c | 62 enum vm_bind_op op; member in struct:bind_job_op 460 struct drm_gpuva_op *op = last; local 465 drm_gpuva_for_each_op_from_reverse(op, ops) { 466 switch (op->op) { 471 struct drm_gpuva_op_remap *r = &op->remap; 484 op_unmap_prepare_unwind(op->unmap.va); 497 drm_gpuva_for_each_op(op, ops) { 498 switch (op->op) { 581 op_map_prepare(struct nouveau_uvmm *uvmm, struct nouveau_uvma **puvma, struct drm_gpuva_op_map *op, struct uvmm_map_args *args) argument 620 struct drm_gpuva_op *op; local 759 op_gem_obj(struct drm_gpuva_op *op) argument 831 struct drm_gpuva_op *op; local 873 struct drm_gpuva_op *op; local 969 struct bind_job_op *op, *next; local 993 bind_validate_op(struct nouveau_job *job, struct bind_job_op *op) argument 1018 struct bind_job_op *op; local 1092 struct bind_job_op *op; local 1125 struct drm_gpuva_op *op; local 1156 struct bind_job_op *op; local 1198 struct bind_job_op *op; local 1445 struct bind_job_op *op; local 1479 struct bind_job_op *op; local 1548 struct bind_job_op *op; local 1580 struct bind_job_op *op, *next; local 1594 struct bind_job_op *op; local [all...] |
/linux-master/drivers/crypto/ccp/ |
H A D | ccp-dev-v3.c | 74 static int ccp_do_cmd(struct ccp_op *op, u32 *cr, unsigned int cr_count) argument 76 struct ccp_cmd_queue *cmd_q = op->cmd_q; 90 | (op->jobid << REQ0_JOBID_SHIFT) 93 if (op->soc) 97 if (op->ioc || !cmd_q->free_slots) 122 | op->jobid; 131 } else if (op->soc) { 135 | op->jobid; 148 static int ccp_perform_aes(struct ccp_op *op) argument 154 | (op 179 ccp_perform_xts_aes(struct ccp_op *op) argument 206 ccp_perform_sha(struct ccp_op *op) argument 232 ccp_perform_rsa(struct ccp_op *op) argument 253 ccp_perform_passthru(struct ccp_op *op) argument 294 ccp_perform_ecc(struct ccp_op *op) argument [all...] |
/linux-master/tools/build/feature/ |
H A D | test-dwarf_getlocations.c | 9 Dwarf_Op *op; local 12 return (int)dwarf_getlocations(&attr, offset, &base, &start, &end, &op, &nops);
|
/linux-master/include/linux/ |
H A D | atomic.h | 58 #define __atomic_op_acquire(op, args...) \ 60 typeof(op##_relaxed(args)) __ret = op##_relaxed(args); \ 65 #define __atomic_op_release(op, args...) \ 68 op##_relaxed(args); \ 71 #define __atomic_op_fence(op, args...) \ 73 typeof(op##_relaxed(args)) __ret; \ 75 __ret = op##_relaxed(args); \
|
/linux-master/arch/sparc/lib/ |
H A D | atomic_64.S | 21 #define ATOMIC_OP(op) \ 22 ENTRY(arch_atomic_##op) /* %o0 = increment, %o1 = atomic_ptr */ \ 25 op %g1, %o0, %g7; \ 33 ENDPROC(arch_atomic_##op); \ 34 EXPORT_SYMBOL(arch_atomic_##op); 36 #define ATOMIC_OP_RETURN(op) \ 37 ENTRY(arch_atomic_##op##_return) /* %o0 = increment, %o1 = atomic_ptr */\ 40 op %g1, %o0, %g7; \ 44 op %g1, %o0, %g1; \ 48 ENDPROC(arch_atomic_##op##_retur [all...] |
/linux-master/fs/bcachefs/ |
H A D | logged_ops.h | 11 static inline int bch2_logged_op_update(struct btree_trans *trans, struct bkey_i *op) argument 13 return bch2_btree_insert_nonextent(trans, BTREE_ID_logged_ops, op, 0);
|
/linux-master/arch/m68k/mac/ |
H A D | mac.h | 12 int mac_hwclk(int op, struct rtc_time *t);
|
/linux-master/security/selinux/include/ |
H A D | audit.h | 21 * @op: the operator the rule uses 29 int selinux_audit_rule_init(u32 field, u32 op, char *rulestr, void **rule); 44 * @op: the operator the rule uses 50 int selinux_audit_rule_match(u32 sid, u32 field, u32 op, void *rule);
|
/linux-master/drivers/gpu/drm/xe/instructions/ |
H A D | xe_gsc_commands.h | 26 #define __GSC_INSTR(op, dl) \ 28 REG_FIELD_PREP(GSC_OPCODE, op) | \
|
/linux-master/drivers/base/ |
H A D | trace.h | 20 TP_PROTO(struct device *dev, const char *op, void *node, const char *name, size_t size), 21 TP_ARGS(dev, op, node, name, size), 25 __field(const char *, op) 32 __entry->op = op; 38 __entry->op, __entry->node, __entry->name, __entry->size) 42 TP_PROTO(struct device *dev, const char *op, void *node, const char *name, size_t size), 43 TP_ARGS(dev, op, node, name, size)
|
/linux-master/arch/csky/include/asm/ |
H A D | atomic.h | 25 #define ATOMIC_OP(op) \ 27 void arch_atomic_##op(int i, atomic_t *v) \ 32 " " #op " %0, %1 \n" \ 48 #define ATOMIC_FETCH_OP(op) \ 50 int arch_atomic_fetch_##op##_relaxed(int i, atomic_t *v) \ 56 " " #op " %0, %2 \n" \ 65 #define ATOMIC_OP_RETURN(op, c_op) \ 67 int arch_atomic_##op##_return_relaxed(int i, atomic_t *v) \ 69 return arch_atomic_fetch_##op##_relaxed(i, v) c_op i; \ 72 #define ATOMIC_OPS(op, c_o [all...] |