/linux-master/drivers/infiniband/sw/rxe/ |
H A D | rxe_mr.c | 27 int mr_check_range(struct rxe_mr *mr, u64 iova, size_t length) argument 29 switch (mr->ibmr.type) { 35 if (iova < mr->ibmr.iova || 36 iova + length > mr->ibmr.iova + mr->ibmr.length) { 37 rxe_dbg_mr(mr, "iova/length out of range\n"); 43 rxe_dbg_mr(mr, "mr type not supported\n"); 48 static void rxe_mr_init(int access, struct rxe_mr *mr) argument 50 u32 key = mr 67 rxe_mr_init_dma(int access, struct rxe_mr *mr) argument 75 rxe_mr_iova_to_index(struct rxe_mr *mr, u64 iova) argument 80 rxe_mr_iova_to_page_offset(struct rxe_mr *mr, u64 iova) argument 94 rxe_mr_fill_pages_from_sgt(struct rxe_mr *mr, struct sg_table *sgt) argument 129 rxe_mr_init_user(struct rxe_dev *rxe, u64 start, u64 length, int access, struct rxe_mr *mr) argument 159 rxe_mr_alloc(struct rxe_mr *mr, int num_buf) argument 188 rxe_mr_init_fast(int max_pages, struct rxe_mr *mr) argument 210 struct rxe_mr *mr = to_rmr(ibmr); local 234 struct rxe_mr *mr = to_rmr(ibmr); local 245 rxe_mr_copy_xarray(struct rxe_mr *mr, u64 iova, void *addr, unsigned int length, enum rxe_mr_copy_dir dir) argument 277 rxe_mr_copy_dma(struct rxe_mr *mr, u64 dma_addr, void *addr, unsigned int length, enum rxe_mr_copy_dir dir) argument 304 rxe_mr_copy(struct rxe_mr *mr, u64 iova, void *addr, unsigned int length, enum rxe_mr_copy_dir dir) argument 344 struct rxe_mr *mr = NULL; local 424 rxe_flush_pmem_iova(struct rxe_mr *mr, u64 iova, unsigned int length) argument 471 rxe_mr_do_atomic_op(struct rxe_mr *mr, u64 iova, int opcode, u64 compare, u64 swap_add, u64 *orig_val) argument 529 rxe_mr_do_atomic_write(struct rxe_mr *mr, u64 iova, u64 value) argument 577 rxe_mr_do_atomic_write(struct rxe_mr *mr, u64 iova, u64 value) argument 619 struct rxe_mr *mr; local 641 struct rxe_mr *mr; local 690 struct rxe_mr *mr = to_rmr(wqe->wr.wr.reg.mr); local 724 struct rxe_mr *mr = container_of(elem, typeof(*mr), elem); local [all...] |
H A D | rxe_mw.c | 51 struct rxe_mw *mw, struct rxe_mr *mr, int access) 83 if (unlikely(!mr || wqe->wr.wr.mw.length == 0)) { 91 if (!mr) 94 if (unlikely(mr->access & IB_ZERO_BASED)) { 100 if (unlikely(!(mr->access & IB_ACCESS_MW_BIND))) { 109 !(mr->access & IB_ACCESS_LOCAL_WRITE))) { 117 if (unlikely(wqe->wr.wr.mw.length > mr->ibmr.length)) { 123 if (unlikely((wqe->wr.wr.mw.addr < mr->ibmr.iova) || 125 (mr->ibmr.iova + mr 50 rxe_check_bind_mw(struct rxe_qp *qp, struct rxe_send_wqe *wqe, struct rxe_mw *mw, struct rxe_mr *mr, int access) argument 135 rxe_do_bind_mw(struct rxe_qp *qp, struct rxe_send_wqe *wqe, struct rxe_mw *mw, struct rxe_mr *mr, int access) argument 168 struct rxe_mr *mr; local 239 struct rxe_mr *mr; local 320 struct rxe_mr *mr = mw->mr; local [all...] |
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
H A D | gddr5.c | 75 ram->mr[0] &= ~0xf7f; 76 ram->mr[0] |= (WR & 0x0f) << 8; 77 ram->mr[0] |= (CL & 0x0f) << 3; 78 ram->mr[0] |= (WL & 0x07) << 0; 80 ram->mr[1] &= ~0x0bf; 81 ram->mr[1] |= (xd & 0x01) << 7; 82 ram->mr[1] |= (at[0] & 0x03) << 4; 83 ram->mr[1] |= (dt & 0x03) << 2; 84 ram->mr[1] |= (ds & 0x03) << 0; 89 ram->mr1_nuts = ram->mr[ [all...] |
H A D | gddr3.c | 89 DLL = !(ram->mr[1] & 0x1); 90 RON = !((ram->mr[1] & 0x300) >> 8); 98 ODT = (ram->mr[1] & 0xc) >> 2; 101 hi = ram->mr[2] & 0x1; 107 ram->mr[0] &= ~0xf74; 108 ram->mr[0] |= (CWL & 0x07) << 9; 109 ram->mr[0] |= (CL & 0x07) << 4; 110 ram->mr[0] |= (CL & 0x08) >> 1; 112 ram->mr[1] &= ~0x3fc; 113 ram->mr[ [all...] |
H A D | sddr3.c | 92 ODT = (ram->mr[1] & 0x004) >> 2 | 93 (ram->mr[1] & 0x040) >> 5 | 94 (ram->mr[1] & 0x200) >> 7; 106 ram->mr[0] &= ~0xf74; 107 ram->mr[0] |= (WR & 0x07) << 9; 108 ram->mr[0] |= (CL & 0x0e) << 3; 109 ram->mr[0] |= (CL & 0x01) << 2; 111 ram->mr[1] &= ~0x245; 112 ram->mr[1] |= (ODT & 0x1) << 2; 113 ram->mr[ [all...] |
H A D | sddr2.c | 82 ODT = (ram->mr[1] & 0x004) >> 2 | 83 (ram->mr[1] & 0x040) >> 5; 91 ram->mr[0] &= ~0xf70; 92 ram->mr[0] |= (WR & 0x07) << 9; 93 ram->mr[0] |= (CL & 0x07) << 4; 95 ram->mr[1] &= ~0x045; 96 ram->mr[1] |= (ODT & 0x1) << 2; 97 ram->mr[1] |= (ODT & 0x2) << 5; 98 ram->mr[1] |= !DLL;
|
/linux-master/drivers/infiniband/sw/rdmavt/ |
H A D | mr.c | 11 #include "mr.h" 75 static void rvt_deinit_mregion(struct rvt_mregion *mr) argument 77 int i = mr->mapsz; 79 mr->mapsz = 0; 81 kfree(mr->map[--i]); 82 percpu_ref_exit(&mr->refcount); 87 struct rvt_mregion *mr = container_of(ref, struct rvt_mregion, local 90 complete(&mr->comp); 93 static int rvt_init_mregion(struct rvt_mregion *mr, struct ib_pd *pd, argument 99 mr 135 rvt_alloc_lkey(struct rvt_mregion *mr, int dma_region) argument 205 rvt_free_lkey(struct rvt_mregion *mr) argument 239 struct rvt_mr *mr; local 272 __rvt_free_mr(struct rvt_mr *mr) argument 288 struct rvt_mr *mr; local 340 struct rvt_mr *mr; local 409 struct rvt_mregion *mr = (struct rvt_mregion *)v; local 424 rvt_dereg_clean_qps(struct rvt_mregion *mr) argument 442 rvt_check_refs(struct rvt_mregion *mr, const char *t) argument 471 rvt_mr_has_lkey(struct rvt_mregion *mr, u32 lkey) argument 511 struct rvt_mr *mr = to_imr(ibmr); local 538 struct rvt_mr *mr; local 559 struct rvt_mr *mr = to_imr(ibmr); local 591 struct rvt_mr *mr = to_imr(ibmr); local 617 struct rvt_mr *mr = to_imr(ibmr); local 651 struct rvt_mregion *mr; local 720 struct rvt_mregion *mr; local 831 struct rvt_mregion *mr; local [all...] |
H A D | trace_mr.h | 15 #include "mr.h" 21 TP_PROTO(struct rvt_mregion *mr, u16 m, u16 n, void *v, size_t len), 22 TP_ARGS(mr, m, n, v, len), 24 RDI_DEV_ENTRY(ib_to_rvt(mr->pd->device)) 37 RDI_DEV_ASSIGN(ib_to_rvt(mr->pd->device)); 40 __entry->iova = mr->iova; 41 __entry->user_base = mr->user_base; 42 __entry->lkey = mr->lkey; 46 __entry->length = mr->length; 47 __entry->offset = mr [all...] |
/linux-master/drivers/infiniband/core/ |
H A D | mr_pool.c | 10 struct ib_mr *mr; local 14 mr = list_first_entry_or_null(list, struct ib_mr, qp_entry); 15 if (mr) { 16 list_del(&mr->qp_entry); 21 return mr; 25 void ib_mr_pool_put(struct ib_qp *qp, struct list_head *list, struct ib_mr *mr) argument 30 list_add(&mr->qp_entry, list); 39 struct ib_mr *mr; local 45 mr = ib_alloc_mr_integrity(qp->pd, max_num_sg, 48 mr 68 struct ib_mr *mr; local [all...] |
H A D | uverbs_std_types_mr.c | 95 struct ib_mr *mr; local 127 mr = pd->device->ops.reg_dm_mr(pd, dm, &attr, attrs); 128 if (IS_ERR(mr)) 129 return PTR_ERR(mr); 131 mr->device = pd->device; 132 mr->pd = pd; 133 mr->type = IB_MR_TYPE_DM; 134 mr->dm = dm; 135 mr->uobject = uobj; 139 rdma_restrack_new(&mr 159 struct ib_mr *mr = local 197 struct ib_mr *mr; local [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_ttm_pm.h | 17 int i915_ttm_backup_region(struct intel_memory_region *mr, u32 flags); 19 void i915_ttm_recover_region(struct intel_memory_region *mr); 21 int i915_ttm_restore_region(struct intel_memory_region *mr, u32 flags);
|
H A D | i915_gem_lmem.c | 44 struct intel_memory_region *mr = READ_ONCE(obj->mm.region); local 51 return mr && (mr->type == INTEL_MEMORY_LOCAL || 52 mr->type == INTEL_MEMORY_STOLEN_LOCAL); 68 struct intel_memory_region *mr = READ_ONCE(obj->mm.region); local 74 return mr && (mr->type == INTEL_MEMORY_LOCAL || 75 mr->type == INTEL_MEMORY_STOLEN_LOCAL);
|
/linux-master/net/sunrpc/xprtrdma/ |
H A D | frwr_ops.c | 49 struct rpcrdma_mr *mr) 51 struct rpc_rdma_cid *cid = &mr->mr_cid; 54 cid->ci_completion_id = mr->mr_ibmr->res.id; 57 static void frwr_mr_unmap(struct rpcrdma_xprt *r_xprt, struct rpcrdma_mr *mr) argument 59 if (mr->mr_device) { 60 trace_xprtrdma_mr_unmap(mr); 61 ib_dma_unmap_sg(mr->mr_device, mr->mr_sg, mr->mr_nents, 62 mr 48 frwr_cid_init(struct rpcrdma_ep *ep, struct rpcrdma_mr *mr) argument 72 frwr_mr_release(struct rpcrdma_mr *mr) argument 85 frwr_mr_put(struct rpcrdma_mr *mr) argument 107 struct rpcrdma_mr *mr; local 121 frwr_mr_init(struct rpcrdma_xprt *r_xprt, struct rpcrdma_mr *mr) argument 284 frwr_map(struct rpcrdma_xprt *r_xprt, struct rpcrdma_mr_seg *seg, int nsegs, bool writing, __be32 xid, struct rpcrdma_mr *mr) argument 361 struct rpcrdma_mr *mr = container_of(cqe, struct rpcrdma_mr, mr_cqe); local 387 struct rpcrdma_mr *mr; local 430 struct rpcrdma_mr *mr; local 441 frwr_mr_done(struct ib_wc *wc, struct rpcrdma_mr *mr) argument 456 struct rpcrdma_mr *mr = container_of(cqe, struct rpcrdma_mr, mr_cqe); local 475 struct rpcrdma_mr *mr = container_of(cqe, struct rpcrdma_mr, mr_cqe); local 501 struct rpcrdma_mr *mr; local 572 struct rpcrdma_mr *mr = container_of(cqe, struct rpcrdma_mr, mr_cqe); local 606 struct rpcrdma_mr *mr; local 673 struct rpcrdma_mr *mr; local [all...] |
/linux-master/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_mr.c | 61 struct pvrdma_user_mr *mr; local 71 "unsupported dma mr access flags %#x\n", acc); 75 mr = kzalloc(sizeof(*mr), GFP_KERNEL); 76 if (!mr) 89 kfree(mr); 93 mr->mmr.mr_handle = resp->mr_handle; 94 mr->ibmr.lkey = resp->lkey; 95 mr->ibmr.rkey = resp->rkey; 97 return &mr 116 struct pvrdma_user_mr *mr = NULL; local 208 struct pvrdma_user_mr *mr; local 279 struct pvrdma_user_mr *mr = to_vmr(ibmr); local 304 struct pvrdma_user_mr *mr = to_vmr(ibmr); local 316 struct pvrdma_user_mr *mr = to_vmr(ibmr); local [all...] |
/linux-master/drivers/vdpa/mlx5/core/ |
H A D | mr.c | 35 static void populate_mtts(struct mlx5_vdpa_direct_mr *mr, __be64 *mtt) argument 38 int nsg = mr->nsg; 44 for_each_sg(mr->sg_head.sgl, sg, mr->nent, i) { 47 nsg--, dma_addr += BIT(mr->log_size), dma_len -= BIT(mr->log_size)) 52 static int create_direct_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_direct_mr *mr) argument 59 inlen = MLX5_ST_SZ_BYTES(create_mkey_in) + roundup(MLX5_ST_SZ_BYTES(mtt) * mr->nsg, 16); 66 MLX5_SET(mkc, mkc, lw, !!(mr->perm & VHOST_MAP_WO)); 67 MLX5_SET(mkc, mkc, lr, !!(mr 89 destroy_direct_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_direct_mr *mr) argument 94 map_start(struct vhost_iotlb_map *map, struct mlx5_vdpa_direct_mr *mr) argument 99 map_end(struct vhost_iotlb_map *map, struct mlx5_vdpa_direct_mr *mr) argument 104 maplen(struct vhost_iotlb_map *map, struct mlx5_vdpa_direct_mr *mr) argument 182 create_indirect_key(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) argument 223 map_direct_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_direct_mr *mr, struct vhost_iotlb *iotlb) argument 295 unmap_direct_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_direct_mr *mr) argument 304 add_direct_chain(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr, u64 start, u64 size, u8 perm, struct vhost_iotlb *iotlb) argument 360 create_user_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr, struct vhost_iotlb *iotlb) argument 424 create_dma_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) argument 452 destroy_dma_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) argument 486 destroy_user_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) argument 499 _mlx5_vdpa_destroy_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) argument 516 _mlx5_vdpa_put_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) argument 526 mlx5_vdpa_put_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) argument 534 _mlx5_vdpa_get_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) argument 543 mlx5_vdpa_get_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr) argument 567 struct mlx5_vdpa_mr *mr; local 592 _mlx5_vdpa_create_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_mr *mr, struct vhost_iotlb *iotlb) argument 635 struct mlx5_vdpa_mr *mr; local 679 struct mlx5_vdpa_mr *mr; local [all...] |
/linux-master/drivers/infiniband/hw/mlx4/ |
H A D | mr.c | 60 struct mlx4_ib_mr *mr; local 63 mr = kzalloc(sizeof(*mr), GFP_KERNEL); 64 if (!mr) 68 ~0ull, convert_access(acc), 0, 0, &mr->mmr); 72 err = mlx4_mr_enable(to_mdev(pd->device)->dev, &mr->mmr); 76 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; 77 mr 411 struct mlx4_ib_mr *mr; local 458 mlx4_ib_rereg_user_mr(struct ib_mr *mr, int flags, u64 start, u64 length, u64 virt_addr, int mr_access_flags, struct ib_pd *pd, struct ib_udata *udata) argument 549 mlx4_alloc_priv_pages(struct ib_device *device, struct mlx4_ib_mr *mr, int max_pages) argument 584 mlx4_free_priv_pages(struct mlx4_ib_mr *mr) argument 598 struct mlx4_ib_mr *mr = to_mmr(ibmr); local 648 struct mlx4_ib_mr *mr; local 690 struct mlx4_ib_mr *mr = to_mmr(ibmr); local 703 struct mlx4_ib_mr *mr = to_mmr(ibmr); local [all...] |
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | mr.c | 126 static int destroy_mkey(struct mlx5_ib_dev *dev, struct mlx5_ib_mr *mr) argument 128 WARN_ON(xa_load(&dev->odp_mkeys, mlx5_base_mkey(mr->mmkey.key))); 130 return mlx5_core_destroy_mkey(dev->mdev, mr->mmkey.key); 138 mlx5_ib_warn(dev, "async reg mr failed. status %d\n", status); 693 struct mlx5_ib_mr *mr; local 696 mr = kzalloc(sizeof(*mr), GFP_KERNEL); 697 if (!mr) 707 err = create_cache_mkey(ent, &mr->mmkey.key); 712 kfree(mr); 1043 struct mlx5_ib_mr *mr; local 1104 set_mr_fields(struct mlx5_ib_dev *dev, struct mlx5_ib_mr *mr, u64 length, int access_flags, u64 iova) argument 1135 struct mlx5_ib_mr *mr; local 1185 struct mlx5_ib_mr *mr; local 1273 struct mlx5_ib_mr *mr; local 1368 struct mlx5_ib_mr *mr = NULL; local 1413 struct mlx5_ib_mr *mr; local 1490 struct mlx5_ib_mr *mr = umem_dmabuf->private; local 1512 struct mlx5_ib_mr *mr = NULL; local 1579 can_use_umr_rereg_pas(struct mlx5_ib_mr *mr, struct ib_umem *new_umem, int new_access_flags, u64 iova, unsigned long *page_size) argument 1600 umr_rereg_pas(struct mlx5_ib_mr *mr, struct ib_pd *pd, int access_flags, int flags, struct ib_umem *new_umem, u64 iova, unsigned long page_size) argument 1653 struct mlx5_ib_mr *mr = to_mmr(ib_mr); local 1744 mlx5_alloc_priv_descs(struct ib_device *device, struct mlx5_ib_mr *mr, int ndescs, int desc_size) argument 1782 mlx5_free_priv_descs(struct mlx5_ib_mr *mr) argument 1796 cache_ent_find_and_store(struct mlx5_ib_dev *dev, struct mlx5_ib_mr *mr) argument 1840 struct mlx5_ib_mr *mr = to_mmr(ibmr); local 1929 _mlx5_alloc_mkey_descs(struct ib_pd *pd, struct mlx5_ib_mr *mr, int ndescs, int desc_size, int page_shift, int access_mode, u32 *in, int inlen) argument 1968 struct mlx5_ib_mr *mr; local 2005 mlx5_alloc_mem_reg_descs(struct ib_pd *pd, struct mlx5_ib_mr *mr, int ndescs, u32 *in, int inlen) argument 2013 mlx5_alloc_sg_gaps_descs(struct ib_pd *pd, struct mlx5_ib_mr *mr, int ndescs, u32 *in, int inlen) argument 2020 mlx5_alloc_integrity_descs(struct ib_pd *pd, struct mlx5_ib_mr *mr, int max_num_sg, int max_num_meta_sg, u32 *in, int inlen) argument 2105 struct mlx5_ib_mr *mr; local 2304 struct mlx5_ib_mr *mr = to_mmr(ibmr); local 2333 mlx5_ib_sg_to_klms(struct mlx5_ib_mr *mr, struct scatterlist *sgl, unsigned short sg_nents, unsigned int *sg_offset_p, struct scatterlist *meta_sgl, unsigned short meta_sg_nents, unsigned int *meta_sg_offset_p) argument 2394 struct mlx5_ib_mr *mr = to_mmr(ibmr); local 2408 struct mlx5_ib_mr *mr = to_mmr(ibmr); local 2427 struct mlx5_ib_mr *mr = to_mmr(ibmr); local 2492 struct mlx5_ib_mr *mr = to_mmr(ibmr); local 2525 struct mlx5_ib_mr *mr = to_mmr(ibmr); local 2584 struct mlx5_ib_mr *mr = to_mmr(ibmr); local [all...] |
/linux-master/drivers/sh/intc/ |
H A D | balancing.c | 44 struct intc_mask_reg *mr = desc->hw.mask_regs; local 48 for (i = 0; mr && enum_id && i < desc->hw.nr_mask_regs; i++) { 49 mr = desc->hw.mask_regs + i; 55 if (!mr->dist_reg) 58 for (j = 0; j < ARRAY_SIZE(mr->enum_ids); j++) { 59 if (mr->enum_ids[j] != enum_id) 64 reg_e = mr->dist_reg; 65 reg_d = mr->dist_reg; 67 fn += (mr->reg_width >> 3) - 1; 72 (mr [all...] |
/linux-master/drivers/scsi/ |
H A D | mesh.c | 305 volatile struct mesh_regs __iomem *mr = ms->mesh; local 311 ms, mr, md); 314 (mr->count_hi << 8) + mr->count_lo, mr->sequence, 315 (mr->bus_status1 << 8) + mr->bus_status0, mr->fifo_count, 316 mr->exception, mr 339 mesh_flush_io(volatile struct mesh_regs __iomem *mr) argument 351 volatile struct mesh_regs __iomem *mr = ms->mesh; local 398 volatile struct mesh_regs __iomem *mr = ms->mesh; local 635 volatile struct mesh_regs __iomem *mr = ms->mesh; local 676 volatile struct mesh_regs __iomem *mr = ms->mesh; local 820 volatile struct mesh_regs __iomem *mr = ms->mesh; local 852 volatile struct mesh_regs __iomem *mr = ms->mesh; local 985 volatile struct mesh_regs __iomem *mr = ms->mesh; local 1028 volatile struct mesh_regs __iomem *mr = ms->mesh; local 1124 volatile struct mesh_regs __iomem *mr = ms->mesh; local 1318 volatile struct mesh_regs __iomem *mr = ms->mesh; local 1363 volatile struct mesh_regs __iomem *mr = ms->mesh; local 1442 volatile struct mesh_regs __iomem *mr = ms->mesh; local 1657 volatile struct mesh_regs __iomem *mr = ms->mesh; local 1705 volatile struct mesh_regs __iomem *mr = ms->mesh; local 1816 volatile struct mesh_regs __iomem *mr; local [all...] |
/linux-master/net/xdp/ |
H A D | xdp_umem.h | 13 struct xdp_umem *xdp_umem_create(struct xdp_umem_reg *mr);
|
/linux-master/drivers/infiniband/hw/mana/ |
H A D | Makefile | 4 mana_ib-y := device.o main.o wq.o qp.o cq.o mr.o
|
H A D | mr.c | 28 static int mana_ib_gd_create_mr(struct mana_ib_dev *dev, struct mana_ib_mr *mr, argument 58 ibdev_dbg(&dev->ib_dev, "Failed to create mr %d, %u", err, 66 mr->ibmr.lkey = resp.lkey; 67 mr->ibmr.rkey = resp.rkey; 68 mr->mr_handle = resp.mr_handle; 105 struct mana_ib_mr *mr; local 118 mr = kzalloc(sizeof(*mr), GFP_KERNEL); 119 if (!mr) 122 mr 173 struct mana_ib_mr *mr = container_of(ibmr, struct mana_ib_mr, ibmr); local [all...] |
/linux-master/arch/powerpc/platforms/pseries/ |
H A D | hvCall.S | 64 mr r4,r3; \ 65 mr r3,r0; \ 77 mr r5,BUFREG; \ 160 mr r4,r5 161 mr r5,r6 162 mr r6,r7 163 mr r7,r8 164 mr r8,r9 165 mr r9,r10 187 mr r [all...] |
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_mr.c | 52 static int alloc_mr_key(struct hns_roce_dev *hr_dev, struct hns_roce_mr *mr) argument 59 /* Allocate a key for mr from mr_table */ 67 mr->key = hw_index_to_key(id); /* MR key */ 82 static void free_mr_key(struct hns_roce_dev *hr_dev, struct hns_roce_mr *mr) argument 84 unsigned long obj = key_to_hw_index(mr->key); 90 static int alloc_mr_pbl(struct hns_roce_dev *hr_dev, struct hns_roce_mr *mr, argument 94 bool is_fast = mr->type == MR_TYPE_FRMR; 98 mr->pbl_hop_num = is_fast ? 1 : hr_dev->caps.pbl_hop_num; 101 buf_attr.region[0].size = mr->size; 102 buf_attr.region[0].hopnum = mr 126 free_mr_pbl(struct hns_roce_dev *hr_dev, struct hns_roce_mr *mr) argument 131 hns_roce_mr_free(struct hns_roce_dev *hr_dev, struct hns_roce_mr *mr) argument 149 hns_roce_mr_enable(struct hns_roce_dev *hr_dev, struct hns_roce_mr *mr) argument 198 struct hns_roce_mr *mr; local 235 struct hns_roce_mr *mr; local 285 struct hns_roce_mr *mr = to_hr_mr(ibmr); local 362 struct hns_roce_mr *mr = to_hr_mr(ibmr); local 378 struct hns_roce_mr *mr; local 427 struct hns_roce_mr *mr = to_hr_mr(ibmr); local 442 struct hns_roce_mr *mr = to_hr_mr(ibmr); local [all...] |
/linux-master/drivers/rtc/ |
H A D | rtc-at91sam9.c | 133 u32 offset, alarm, mr; local 140 mr = rtt_readl(rtc, MR); 143 rtt_writel(rtc, MR, mr & ~(AT91_RTT_ALMIEN | AT91_RTT_RTTINCIEN)); 164 mr &= ~AT91_RTT_ALMIEN; 170 rtt_writel(rtc, MR, mr | AT91_RTT_RTTRST); 205 u32 mr; local 214 mr = rtt_readl(rtc, MR); 215 rtt_writel(rtc, MR, mr & ~AT91_RTT_ALMIEN); 226 rtt_writel(rtc, MR, mr | AT91_RTT_ALMIEN); 236 u32 mr local 252 u32 mr = rtt_readl(rtc, MR); local 261 u32 sr, mr; local 339 u32 mr; local 448 u32 mr = rtt_readl(rtc, MR); local 459 u32 mr = rtt_readl(rtc, MR); local 472 u32 mr = rtt_readl(rtc, MR); local 501 u32 mr; local [all...] |