/barrelfish-master/lib/devif/backends/net/mlx4/drivers/net/mlx4/ |
H A D | en_resources.c | 46 void mlx4_en_fill_qp_context(struct mlx4_en_priv *priv, int size, int stride, argument 57 context->rq_size_stride = ilog2(size) << 3 | (ilog2(stride) - 4); 59 context->sq_size_stride = ilog2(size) << 3 | (ilog2(stride) - 4);
|
H A D | en_rx.c | 63 + (ring->stride * index)); 72 * stride, remaining (unused) fragments must be padded with 75 possible_frags = (ring->stride - sizeof(struct mlx4_en_rx_desc)) / DS_SIZE; 178 // + (index * ring->stride)); 316 ring->stride = roundup_pow_of_two(sizeof(struct mlx4_en_rx_desc) + DS_SIZE); 317 ring->log_stride = ffs(ring->stride) - 1; 318 ring->buf_size = ring->size * ring->stride + TXBB_SIZE; 378 + (index * ring->stride)); 399 int stride = roundup_pow_of_two(sizeof(struct mlx4_en_rx_desc) + DS_SIZE); local 410 ring->stride [all...] |
H A D | en_tx.c | 79 struct mlx4_en_tx_ring **pring, u32 size, u16 stride, int node, 114 ring->stride = stride; 152 ring->buf_size = ALIGN(size * ring->stride, MLX4_EN_PAGE_SIZE); 257 mlx4_en_fill_qp_context(priv, ring->size, ring->stride, 1, 0, ring->qpn, 78 mlx4_en_create_tx_ring(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring **pring, u32 size, u16 stride, int node, int queue_idx) argument
|
H A D | mlx4_en.h | 263 u16 stride; member in struct:mlx4_en_tx_ring 298 /*actual number of entries depends on rx ring stride*/ 323 u16 stride; member in struct:mlx4_en_rx_ring 552 int stride; member in struct:mlx4_en_priv 783 struct mlx4_en_tx_ring **pring, u32 size, u16 stride, int node, int queue_idx); 800 u32 size, u16 stride); 814 void mlx4_en_fill_qp_context(struct mlx4_en_priv *priv, int size, int stride,
|
H A D | en_netdev.c | 1991 priv->prof->rx_ring_size, priv->stride); 2043 priv->stride); 2496 priv->stride = roundup_pow_of_two(sizeof(struct mlx4_en_rx_desc) + DS_SIZE);
|
/barrelfish-master/lib/barrelfish/vspace/ |
H A D | memobj_numa.c | 46 if (vregion->offset % mo_numa->stride) { 278 lvaddr_t map_size = mo_numa->stride; 316 map_offset += mo_numa->stride; 357 size_t stride) 361 assert((stride % BASE_PAGE_SIZE)==0); 380 mo_numa->stride = stride; 353 memobj_create_numa(struct memobj_numa *mo_numa, size_t size, memobj_flags_t flags, size_t node_count, size_t stride) argument
|
/barrelfish-master/include/barrelfish/ |
H A D | memobj.h | 141 size_t stride; ///< size of the regions to map member in struct:memobj_numa 188 memobj_flags_t flags, size_t node_count, size_t stride);
|
/barrelfish-master/lib/numa/ |
H A D | alloc.c | 340 size_t stride = pagesize; local 353 err = memobj_create_numa(memobj, size, 0, numa_num_configured_nodes(), stride);
|
/barrelfish-master/usr/pci/ |
H A D | pci.c | 731 uint16_t stride = pci_sr_iov_cap_stride_rd(sr_iov_cap); local 737 + stride * vf_number) 742 + stride * vf_number) 749 PCI_DEBUG("VF (bus=%d, device=%d, function=%d) offset is 0x%x, stride is 0x%x\n", 750 vf_addr->bus, vf_addr->device, vf_addr->function, offset, stride);
|