/linux-master/sound/soc/ti/ |
H A D | davinci-mcasp.c | 1662 .access = (SNDRV_CTL_ELEM_ACCESS_READWRITE | 1670 .access = SNDRV_CTL_ELEM_ACCESS_READ,
|
/linux-master/sound/soc/sof/ |
H A D | sof-audio.h | 369 unsigned int access; member in struct:snd_sof_control 425 * setup_mutex against non stream domain races (kcontrol access for
|
/linux-master/sound/soc/codecs/ |
H A D | wsa881x.c | 203 .access = SNDRV_CTL_ELEM_ACCESS_TLV_READ |\
|
H A D | rt722-sdca.c | 688 .access = SNDRV_CTL_ELEM_ACCESS_TLV_READ | \
|
H A D | rt715-sdca.c | 441 .access = SNDRV_CTL_ELEM_ACCESS_TLV_READ | \ 451 .access = SNDRV_CTL_ELEM_ACCESS_TLV_READ | \
|
/linux-master/sound/pci/hda/ |
H A D | cs35l56_hda.c | 340 .access = SNDRV_CTL_ELEM_ACCESS_READWRITE, 375 ctl_template.access = (SNDRV_CTL_ELEM_ACCESS_READWRITE | SNDRV_CTL_ELEM_ACCESS_TLV_READ);
|
H A D | tas2781_hda_i2c.c | 42 .access = SNDRV_CTL_ELEM_ACCESS_TLV_READ |\
|
/linux-master/sound/pci/emu10k1/ |
H A D | emumixer.c | 673 .access = SNDRV_CTL_ELEM_ACCESS_READWRITE, 719 .access = SNDRV_CTL_ELEM_ACCESS_READWRITE, 790 .access = SNDRV_CTL_ELEM_ACCESS_READWRITE, 848 .access = SNDRV_CTL_ELEM_ACCESS_READWRITE, 1010 .access = SNDRV_CTL_ELEM_ACCESS_READWRITE, 1057 .access = SNDRV_CTL_ELEM_ACCESS_READWRITE, 1108 .access = SNDRV_CTL_ELEM_ACCESS_READWRITE, 1159 .access = SNDRV_CTL_ELEM_ACCESS_READWRITE, 1313 .access = SNDRV_CTL_ELEM_ACCESS_READWRITE | 1393 .access [all...] |
/linux-master/fs/nfsd/ |
H A D | nfs4xdr.c | 775 struct nfsd4_access *access = &u->access; local 776 if (xdr_stream_decode_u32(argp->xdr, &access->ac_req_access) < 0) 1100 /* Note: unlike access bits, deny bits may be zero. */ 2789 /* access mask */ 3925 struct nfsd4_access *access = &u->access; local 3930 status = nfsd4_encode_uint32_t(xdr, access->ac_supported); 3933 /* access */ 3934 return nfsd4_encode_uint32_t(xdr, access [all...] |
/linux-master/include/linux/ |
H A D | mm.h | 291 #define VM_SEQ_READ 0x00008000 /* App will access data sequentially */ 366 # define VM_SPARC_ADI VM_ARCH_1 /* Uses ADI tag for access control */ 376 # define VM_MTE VM_HIGH_ARCH_0 /* Use Tagged memory for access control */ 441 /* VMA basic access permission flags */ 605 int (*access)(struct vm_area_struct *vma, unsigned long addr, member in struct:vm_operations_struct 1146 * This can be called when MMU is off so it must not access 1358 * servicing faults for write access. In the normal case, do always want 1498 * allocation immediately. Do not access the memory or the struct folio 1518 * allocation immediately. Do not access the memory or the struct folio 1674 /* page access tim [all...] |
/linux-master/arch/arc/kernel/ |
H A D | head.S | 51 ; Unaligned access is disabled at reset, so re-enable early as 52 ; gcc 7.3.1 (ARC GNU 2018.03) onwards generates unaligned access
|
H A D | entry-arcv2.S | 121 ; Instruction fetch or Data access, under a single Exception Vector 163 ; ECR in pt_regs provides whether access was R/W/X
|
/linux-master/fs/smb/server/ |
H A D | smb2pdu.c | 2045 * @access: file access flags 2051 static int smb2_create_open_flags(bool file_present, __le32 access, argument 2057 if (access & FILE_READ_DESIRED_ACCESS_LE && 2058 access & FILE_WRITE_DESIRE_ACCESS_LE) { 2061 } else if (access & FILE_WRITE_DESIRE_ACCESS_LE) { 2069 if (access == FILE_READ_ATTRIBUTES_LE) 2311 * @get_write: get write access to a mount 2980 pr_err("Invalid desired access : 0x%x\n", 3021 "get query maximal access contex [all...] |
/linux-master/fs/smb/client/ |
H A D | smb2pdu.c | 2677 /* and one more ACE to allow access for authenticated users */ 3244 oparms->fid->access = oparms->desired_access;
|
H A D | smb2ops.c | 1405 cfile->fid.access = fid->access;
|
H A D | cifsglob.h | 1087 bool expired_pwd; /* track if access denied or expired pwd so can know if need to update */ 1412 __u32 access; member in struct:cifs_fid
|
/linux-master/arch/x86/kvm/ |
H A D | x86.c | 323 * Return true if we want to ignore/silent this failed msr access. 807 * else the access will fault indefinitely (and to emulate hardware). 7523 gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u64 access, argument 7532 access |= PFERR_USER_MASK; 7533 t_gpa = mmu->gva_to_gpa(vcpu, mmu, gpa, access, exception); 7543 u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0; local 7544 return mmu->gva_to_gpa(vcpu, mmu, gva, access, exception); 7553 u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0; local 7554 access |= PFERR_WRITE_MASK; 7555 return mmu->gva_to_gpa(vcpu, mmu, gva, access, exceptio 7568 kvm_read_guest_virt_helper(gva_t addr, void *val, unsigned int bytes, struct kvm_vcpu *vcpu, u64 access, struct x86_exception *exception) argument 7606 u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0; local 7631 u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0; local 7650 u64 access = 0; local 7660 kvm_write_guest_virt_helper(gva_t addr, void *val, unsigned int bytes, struct kvm_vcpu *vcpu, u64 access, struct x86_exception *exception) argument 7695 u64 access = PFERR_WRITE_MASK; local 7771 u64 access = ((static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0) local 13565 u64 access = error_code & local [all...] |
/linux-master/arch/x86/kvm/vmx/ |
H A D | vmx.h | 447 #define __BUILD_VMX_MSR_BITMAP_HELPER(rtype, action, bitop, access, base) \ 448 static inline rtype vmx_##action##_msr_bitmap_##access(unsigned long *bitmap, \
|
/linux-master/arch/x86/kvm/mmu/ |
H A D | mmu.c | 293 unsigned int access) 295 u64 spte = make_mmio_spte(vcpu, gfn, access); 628 * lost when the SPTE is marked for access tracking. 729 * For leaf SPTEs, fetch the *guest* access permissions being shadowed. Note 730 * that the SPTE itself may have a more constrained access permissions that 741 * KVM is not shadowing any guest page tables, so the "guest access 745 * is shadowing a guest huge page with small pages, the guest access 746 * permissions being shadowed are the access permissions of the huge 749 * In both cases, sp->role.access contains the correct access bit 292 mark_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, u64 gfn, unsigned int access) argument 754 kvm_mmu_page_set_translation(struct kvm_mmu_page *sp, int index, gfn_t gfn, unsigned int access) argument 773 kvm_mmu_page_set_access(struct kvm_mmu_page *sp, int index, unsigned int access) argument 1641 __rmap_add(struct kvm *kvm, struct kvm_mmu_memory_cache *cache, const struct kvm_memory_slot *slot, u64 *spte, gfn_t gfn, unsigned int access) argument 1665 rmap_add(struct kvm_vcpu *vcpu, const struct kvm_memory_slot *slot, u64 *spte, gfn_t gfn, unsigned int access) argument 2307 kvm_mmu_child_role(u64 *sptep, bool direct, unsigned int access) argument 2353 kvm_mmu_get_child_sp(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, bool direct, unsigned int access) argument 2983 unsigned int access = sp->role.access; local 3311 kvm_handle_noslot_fault(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault, unsigned int access) argument 4088 nonpaging_gva_to_gpa(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, gpa_t vaddr, u64 access, struct x86_exception *exception) argument 4201 unsigned int access = get_mmio_spte_access(spte); local 4413 kvm_faultin_pfn(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault, unsigned int access) argument 4812 sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, unsigned int access) argument 6495 unsigned int access; local [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | kvm_host.h | 341 unsigned access:3; member in struct:kvm_mmu_page_role::__anon6 453 gpa_t gva_or_gpa, u64 access, 559 * set in a single access, e.g. to reprogram all counters when the PMU 1421 /* Guest can access the SGX PROVISIONKEY. */ 2012 * access because doing so also zaps the translation for
|
/linux-master/tools/testing/selftests/net/tcp_ao/lib/ |
H A D | setup.c | 298 optmem_ns = !access(optmem_file, F_OK);
|
/linux-master/tools/perf/util/ |
H A D | annotate.c | 1356 * of 'struct disasm_line' to have an easy access. 1890 access(filename, R_OK)) { 1903 if (access(filename, R_OK) && errno == ENOENT && dso->nsinfo) { 3569 * Get register number and access offset from the given instruction. 3926 * Some instructions can be fused and the actual memory access came
|
/linux-master/tools/include/uapi/sound/ |
H A D | asound.h | 92 SNDRV_HWDEP_IFACE_USB_STREAM, /* direct access to usb stream */ 1089 /* bits 30 and 31 are obsoleted (for indirect access) */ 1122 unsigned int access; /* R: value access (bitmask) - SNDRV_CTL_ELEM_ACCESS_* */ member in struct:snd_ctl_elem_info 1150 unsigned int indirect: 1; /* W: indirect access - obsoleted */
|
/linux-master/fs/kernfs/ |
H A D | file.c | 425 if (of->vm_ops->access) 426 ret = of->vm_ops->access(vma, addr, buf, len, write); 436 .access = kernfs_vma_access, 681 * Always instantiate seq_file even if read access doesn't use 682 * seq_file or is not requested. This unifies private data access 705 /* nobody has access to @of yet, skip @of->mutex */
|
/linux-master/drivers/vhost/ |
H A D | vhost.c | 988 /* No one will access memory at this point */ 1090 struct iovec iov[], int iov_size, int access); 1102 * could be access through iotlb. So -EAGAIN should 1137 * could be access through iotlb. So -EAGAIN should 1184 vq_err(vq, "Non atomic userspace memory access: uaddr " 1195 * could be access through iotlb. So -EAGAIN should 1351 static bool umem_access_ok(u64 uaddr, u64 size, int access) argument 1359 if ((access & VHOST_ACCESS_RO) && 1362 if ((access & VHOST_ACCESS_WO) && 1564 static int vhost_iotlb_miss(struct vhost_virtqueue *vq, u64 iova, int access) argument 1611 int access = (type == VHOST_ADDR_USED) ? local 1618 iotlb_access_ok(struct vhost_virtqueue *vq, int access, u64 addr, u64 len, int type) argument 2332 translate_desc(struct vhost_virtqueue *vq, u64 addr, u32 len, struct iovec iov[], int iov_size, int access) argument 2403 int ret, access; local 2504 int ret, access; local [all...] |