/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | mxgpu_vi.c | 515 struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, flr_work); local 516 struct amdgpu_device *adev = container_of(virt, struct amdgpu_device, virt); 565 &adev->virt.flr_work), 585 adev->virt.ack_irq.num_types = 1; 586 adev->virt.ack_irq.funcs = &xgpu_vi_mailbox_ack_irq_funcs; 587 adev->virt.rcv_irq.num_types = 1; 588 adev->virt.rcv_irq.funcs = &xgpu_vi_mailbox_rcv_irq_funcs; 595 r = amdgpu_irq_add_id(adev, AMDGPU_IRQ_CLIENTID_LEGACY, 135, &adev->virt.rcv_irq); 599 r = amdgpu_irq_add_id(adev, AMDGPU_IRQ_CLIENTID_LEGACY, 138, &adev->virt [all...] |
/linux-master/drivers/platform/x86/dell/ |
H A D | dcdbas.c | 56 smi_buffer->virt = dma_alloc_coherent(&dcdbas_pdev->dev, size, 58 if (!smi_buffer->virt) { 75 if (!smi_buffer->virt) 81 smi_buffer->virt, smi_buffer->dma); 82 smi_buffer->virt = NULL; 93 if (!smi_buf.virt || wsmt_enabled) 119 if (smi_buf.virt) 120 memcpy(tmp.virt, smi_buf.virt, smi_buf.size); 171 ret = memory_read_from_buffer(buf, count, &pos, smi_buf.virt, [all...] |
/linux-master/drivers/scsi/elx/libefc/ |
H A D | efc_device.c | 539 struct fc_frame_header *hdr = cbdata->header->dma.virt; 542 efc_node_save_sparms(node, cbdata->payload->dma.virt); 544 cbdata->header->dma.virt, 568 struct fc_frame_header *hdr = cbdata->header->dma.virt; 573 cbdata->payload->dma.virt, 592 struct fc_frame_header *hdr = cbdata->header->dma.virt; 617 struct fc_frame_header *hdr = cbdata->header->dma.virt; 708 efc_node_save_sparms(node, cbdata->payload->dma.virt); 710 cbdata->header->dma.virt, 730 efc_process_prli_payload(node, cbdata->payload->dma.virt); [all...] |
/linux-master/crypto/ |
H A D | crypto_null.c | 84 if (walk.src.virt.addr != walk.dst.virt.addr) 85 memcpy(walk.dst.virt.addr, walk.src.virt.addr,
|
/linux-master/arch/arm/include/debug/ |
H A D | sti.S | 19 ldr \rv, =CONFIG_DEBUG_UART_VIRT @ virt base
|
/linux-master/drivers/scsi/elx/include/ |
H A D | efc_common.h | 13 void *virt; member in struct:efc_dma
|
/linux-master/arch/arm/include/asm/mach/ |
H A D | map.h | 58 extern int ioremap_page(unsigned long virt, unsigned long phys,
|
/linux-master/arch/arm/kernel/ |
H A D | reboot.c | 12 #include <asm/virt.h>
|
/linux-master/drivers/soc/ux500/ |
H A D | ux500-soc-id.c | 39 void __iomem *virt = ioremap(addr, 4); local 42 if (!virt) 45 asicid = readl(virt); 46 iounmap(virt);
|
/linux-master/drivers/mtd/maps/ |
H A D | pxa2xx-flash.c | 63 info->map.virt = ioremap(info->map.phys, info->map.size); 64 if (!info->map.virt) { 86 iounmap((void *)info->map.virt); 108 iounmap(info->map.virt);
|
H A D | plat-ram.c | 124 info->map.virt = devm_platform_get_and_ioremap_resource(pdev, 0, &res); 125 if (IS_ERR(info->map.virt)) { 126 err = PTR_ERR(info->map.virt); 141 dev_dbg(&pdev->dev, "virt %p, %lu bytes\n", info->map.virt, info->map.size);
|
H A D | esb2rom.c | 98 void __iomem* virt; member in struct:esb2rom_window 139 if (window->virt) { 140 iounmap(window->virt); 141 window->virt = NULL; 252 window->virt = ioremap(window->phys, window->size); 253 if (!window->virt) { 290 map->map.virt = (void __iomem *) 291 (((unsigned long)(window->virt)) + offset); 347 map->map.virt = window->virt; [all...] |
/linux-master/arch/s390/crypto/ |
H A D | chacha-glue.c | 52 chacha_crypt_generic(state, walk.dst.virt.addr, 53 walk.src.virt.addr, nbytes, 56 chacha20_crypt_s390(state, walk.dst.virt.addr, 57 walk.src.virt.addr, nbytes,
|
/linux-master/arch/powerpc/crypto/ |
H A D | aes_ctr.c | 74 u8 *src = walk->src.virt.addr; 75 u8 *dst = walk->dst.virt.addr; 111 aes_p8_ctr32_encrypt_blocks(walk.src.virt.addr, 112 walk.dst.virt.addr,
|
H A D | aes_xts.c | 118 aes_p8_xts_encrypt(walk.src.virt.addr, 119 walk.dst.virt.addr, 123 aes_p8_xts_decrypt(walk.src.virt.addr, 124 walk.dst.virt.addr,
|
H A D | aes-spe-glue.c | 196 ppc_encrypt_ecb(walk.dst.virt.addr, walk.src.virt.addr, 199 ppc_decrypt_ecb(walk.dst.virt.addr, walk.src.virt.addr, 235 ppc_encrypt_cbc(walk.dst.virt.addr, walk.src.virt.addr, 239 ppc_decrypt_cbc(walk.dst.virt.addr, walk.src.virt.addr, 276 ppc_crypt_ctr(walk.dst.virt.addr, walk.src.virt [all...] |
/linux-master/drivers/scsi/elx/efct/ |
H A D | efct_unsol.c | 46 hdr = seq->header->dma.virt; 147 struct fc_frame_header *fchdr = seq->header->dma.virt; 229 struct fc_frame_header *req_hdr = seq->header->dma.virt; 235 u8 *heap_virt_base = seq->payload->dma.virt; 276 ctx->payload.virt = heap_virt_base + heap_offset; 287 memcpy(ctx->payload.virt, payload, payload_len); 317 struct fcp_cmnd *fcpcmd = seq->payload->dma.virt; 342 struct fc_frame_header *fchdr = seq->header->dma.virt; 352 cmnd = seq->payload->dma.virt; 466 struct fc_frame_header *hdr = seq->header->dma.virt; [all...] |
/linux-master/arch/x86/pci/ |
H A D | numachip.c | 23 if (cfg && cfg->virt) 24 return cfg->virt + (PCI_MMCFG_BUS_OFFSET(bus) | (devfn << 12));
|
/linux-master/arch/x86/crypto/ |
H A D | sm4_aesni_avx_glue.c | 48 const u8 *src = walk.src.virt.addr; 49 u8 *dst = walk.dst.virt.addr; 103 const u8 *src = walk.src.virt.addr; 104 u8 *dst = walk.dst.virt.addr; 136 const u8 *src = walk.src.virt.addr; 137 u8 *dst = walk.dst.virt.addr; 201 const u8 *src = walk.src.virt.addr; 202 u8 *dst = walk.dst.virt.addr;
|
/linux-master/arch/powerpc/mm/book3s32/ |
H A D | mmu.c | 116 static void setibat(int index, unsigned long virt, phys_addr_t phys, argument 128 bat[0].batu = virt | (bl << 2) | 2; /* Vs=1, Vp=0 */ 130 if (!is_kernel_addr(virt)) 258 void __init setbat(int index, unsigned long virt, phys_addr_t phys, argument 284 bat[1].batu = virt | (bl << 2) | 2; /* Vs=1, Vp=0 */ 286 if (!is_kernel_addr(virt)) 293 bat_addrs[index].start = virt; 294 bat_addrs[index].limit = virt + ((bl + 1) << 17) - 1;
|
/linux-master/arch/arm/crypto/ |
H A D | aes-ce-glue.c | 179 ce_aes_ecb_encrypt(walk.dst.virt.addr, walk.src.virt.addr, 199 ce_aes_ecb_decrypt(walk.dst.virt.addr, walk.src.virt.addr, 217 ce_aes_cbc_encrypt(walk->dst.virt.addr, walk->src.virt.addr, 247 ce_aes_cbc_decrypt(walk->dst.virt.addr, walk->src.virt.addr, 317 ce_aes_cbc_cts_encrypt(walk.dst.virt.addr, walk.src.virt [all...] |
/linux-master/arch/sparc/crypto/ |
H A D | camellia_glue.c | 112 op(walk.src.virt.addr, walk.dst.virt.addr, 159 op(walk.src.virt.addr, walk.dst.virt.addr, 188 op(walk.src.virt.addr, walk.dst.virt.addr,
|
/linux-master/arch/arm64/crypto/ |
H A D | aes-glue.c | 191 aes_ecb_encrypt(walk.dst.virt.addr, walk.src.virt.addr, 211 aes_ecb_decrypt(walk.dst.virt.addr, walk.src.virt.addr, 229 aes_cbc_encrypt(walk->dst.virt.addr, walk->src.virt.addr, 258 aes_cbc_decrypt(walk->dst.virt.addr, walk->src.virt.addr, 327 aes_cbc_cts_encrypt(walk.dst.virt.addr, walk.src.virt [all...] |
/linux-master/drivers/gpu/drm/tegra/ |
H A D | nvdec.c | 228 nvdec->falcon.firmware.virt, 232 nvdec->falcon.firmware.virt, 250 void *virt; local 253 if (nvdec->falcon.firmware.virt) 263 virt = dma_alloc_coherent(nvdec->dev, size, &iova, GFP_KERNEL); 269 virt = tegra_drm_alloc(tegra, size, &iova); 270 if (IS_ERR(virt)) 271 return PTR_ERR(virt); 274 nvdec->falcon.firmware.virt = virt; [all...] |
/linux-master/drivers/staging/media/atomisp/pci/mmu/ |
H A D | isp_mmu.c | 101 void *virt; local 103 virt = (void *)__get_free_page(GFP_KERNEL | GFP_DMA32); 105 if (!virt) 112 set_memory_uc((unsigned long)virt, 1); 115 page = virt_to_phys(virt); 127 void *virt; local 133 virt = phys_to_virt(page); 136 set_memory_wb((unsigned long)virt, 1); 139 free_page((unsigned long)virt); 149 "\tL1 PT: virt [all...] |