/linux-master/drivers/vfio/pci/mlx5/ |
H A D | cmd.h | 123 int nent; member in struct:mlx5_vhca_cq_buf
|
H A D | cmd.c | 992 struct mlx5_vhca_cq_buf *buf, int nent, 1000 err = mlx5_frag_buf_alloc_node(mdev, nent * cqe_size, frag_buf, 1007 buf->nent = nent; 1017 for (i = 0; i < buf->nent; i++) { 1621 u32 nent = size / entry_size; local 1628 (nent > qp->max_msg_size / entry_size))) 1633 for (i = 0; i < nent; i++) { 991 alloc_cq_frag_buf(struct mlx5_core_dev *mdev, struct mlx5_vhca_cq_buf *buf, int nent, int cqe_size) argument
|
/linux-master/drivers/rapidio/ |
H A D | rio_cm.c | 327 * @nent: max number of entries to fill 331 static void riocm_rx_fill(struct cm_dev *cm, int nent) argument 338 for (i = 0; i < RIOCM_RX_RING_SIZE && cm->rx_slots && nent; i++) { 345 nent--; 1568 u32 nent; local 1589 nent = min(info[0], cm->npeers); 1590 buf = kcalloc(nent + 2, sizeof(u32), GFP_KERNEL); 1601 if (++i == nent)
|
/linux-master/drivers/infiniband/hw/mthca/ |
H A D | mthca_provider.c | 583 int nent; local 619 for (nent = 1; nent <= entries; nent <<= 1) 622 err = mthca_init_cq(to_mdev(ibdev), nent, context,
|
H A D | mthca_cmd.c | 664 int nent = 0; local 694 pages[nent * 2] = cpu_to_be64(virt); 698 pages[nent * 2 + 1] = 704 if (++nent == MTHCA_MAILBOX_SIZE / 16) { 705 err = mthca_cmd(dev, mailbox->dma, nent, 0, op, 709 nent = 0; 714 if (nent) 715 err = mthca_cmd(dev, mailbox->dma, nent, 0, op,
|
/linux-master/drivers/mailbox/ |
H A D | bcm-pdc-mailbox.c | 1201 int nent; local 1210 nent = dma_map_sg(dev, mssg->spu.src, src_nent, DMA_TO_DEVICE); 1211 if (unlikely(nent == 0)) 1217 nent = dma_map_sg(dev, mssg->spu.dst, dst_nent, 1219 if (unlikely(nent == 0)) {
|
/linux-master/arch/x86/kvm/ |
H A D | hyperv.c | 2776 int i, nent = ARRAY_SIZE(cpuid_entries); local 2781 if (cpuid->nent < nent) 2784 if (cpuid->nent > nent) 2785 cpuid->nent = nent; 2787 for (i = 0; i < nent; i++) { 2912 nent * sizeof(struct kvm_cpuid_entry2)))
|
/linux-master/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 240 __u32 nent; member in struct:kvm_cpuid 262 __u32 nent; member in struct:kvm_cpuid2
|
/linux-master/tools/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 240 __u32 nent; member in struct:kvm_cpuid 262 __u32 nent; member in struct:kvm_cpuid2
|
/linux-master/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | pci.c | 1362 int nent = 0; local 1384 mlxsw_cmd_mbox_map_fa_pa_set(mbox, nent, mem_item->mapaddr); 1385 mlxsw_cmd_mbox_map_fa_log2size_set(mbox, nent, 0); /* 1 page */ 1386 if (++nent == MLXSW_CMD_MAP_FA_VPM_ENTRIES_MAX) { 1387 err = mlxsw_cmd_map_fa(mlxsw_pci->core, mbox, nent); 1390 nent = 0; 1395 if (nent) { 1396 err = mlxsw_cmd_map_fa(mlxsw_pci->core, mbox, nent);
|
/linux-master/include/xen/interface/ |
H A D | platform.h | 245 uint32_t nent; member in struct:xenpf_firmware_info::__anon859::xenpf_efi_info::__anon863
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | gen8_ppgtt.c | 526 u16 index, max, nent, i; local 529 nent = 1; 570 nent = 16; 580 for (i = 0; i < nent; i++) {
|
/linux-master/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 101 static int mlx4_ib_alloc_cq_buf(struct mlx4_ib_dev *dev, struct mlx4_ib_cq_buf *buf, int nent) argument 105 err = mlx4_buf_alloc(dev->dev, nent * dev->dev->caps.cqe_size, 127 mlx4_buf_free(dev->dev, nent * buf->entry_size, &buf->buf);
|
/linux-master/arch/alpha/kernel/ |
H A D | pci_iommu.c | 110 long i, p, nent; 120 nent = arena->size >> PAGE_SHIFT; 125 while (i < n && p+i < nent) { 109 long i, p, nent; local
|
/linux-master/drivers/vdpa/mlx5/core/ |
H A D | mr.c | 44 for_each_sg(mr->sg_head.sgl, sg, mr->nent, i) { 276 mr->nent = dma_map_sg_attrs(dma, mr->sg_head.sgl, mr->nsg, DMA_BIDIRECTIONAL, 0); 277 if (!mr->nent) {
|
/linux-master/drivers/spi/ |
H A D | spi-topcliff-pch.c | 117 int nent; member in struct:pch_spi_dma_ctrl 768 dma_sync_sg_for_cpu(&data->host->dev, dma->sg_rx_p, dma->nent, 771 dma_sync_sg_for_cpu(&data->host->dev, dma->sg_tx_p, dma->nent, 1026 dma->nent = num; 1088 dma->nent = num;
|
/linux-master/drivers/tty/serial/ |
H A D | atmel_serial.c | 1017 int ret, nent; local 1039 nent = dma_map_sg(port->dev, 1044 if (!nent) { 1196 int ret, nent; local 1220 nent = dma_map_sg(port->dev, 1225 if (!nent) {
|
/linux-master/drivers/net/ethernet/mellanox/mlx4/ |
H A D | fw.c | 1525 int nent = 0; local 1555 pages[nent * 2] = cpu_to_be64(virt); 1559 pages[nent * 2 + 1] = 1565 if (++nent == MLX4_MAILBOX_SIZE / 16) { 1566 err = mlx4_cmd(dev, mailbox->dma, nent, 0, op, 1571 nent = 0; 1576 if (nent) 1577 err = mlx4_cmd(dev, mailbox->dma, nent, 0, op,
|
H A D | mlx4.h | 400 int nent; member in struct:mlx4_eq
|
/linux-master/drivers/nvme/target/ |
H A D | fc.c | 2098 unsigned int nent; local 2100 sg = sgl_alloc(fod->req.transfer_len, GFP_KERNEL, &nent); 2105 fod->data_sg_cnt = nent; 2106 fod->data_sg_cnt = fc_dma_map_sg(fod->tgtport->dev, sg, nent,
|
/linux-master/drivers/vdpa/mlx5/net/ |
H A D | mlx5_vnet.c | 55 int nent; member in struct:mlx5_vdpa_cq_buf 292 static int cq_frag_buf_alloc(struct mlx5_vdpa_net *ndev, struct mlx5_vdpa_cq_buf *buf, int nent) argument 299 err = mlx5_frag_buf_alloc_node(ndev->mvdev.mdev, nent * MLX5_VDPA_CQE_SIZE, frag_buf, 307 buf->nent = nent; 336 for (i = 0; i < buf->nent; i++) {
|
/linux-master/drivers/infiniband/hw/qib/ |
H A D | qib.h | 1382 int qib_pcie_params(struct qib_devdata *dd, u32 minw, u32 *nent);
|
/linux-master/include/linux/mlx4/ |
H A D | device.h | 1146 int mlx4_cq_alloc(struct mlx4_dev *dev, int nent, struct mlx4_mtt *mtt,
|
/linux-master/drivers/scsi/pm8001/ |
H A D | pm80xx_hwi.h | 958 __le32 nent; member in struct:dek_mgmt_req
|
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | mlx5_ib.h | 552 int nent; member in struct:mlx5_ib_cq_buf
|