/barrelfish-2018-10-04/lib/tommath/ |
H A D | mess.sh | 2 if cvs log $1 >/dev/null 2>/dev/null; then exit 0; else echo "$1 shouldn't be here" ; exit 1; fi
|
/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/include/linux/ |
H A D | netdevice.h | 62 #define netif_running(dev) !!((dev)->if_drv_flags & IFF_DRV_RUNNING) 64 #define netif_oper_up(dev) !!((dev)->if_flags & IFF_UP) 65 #define netif_carrier_ok(dev) netif_running(dev) 68 netdev_priv(const struct net_device *dev) { 69 return (dev->if_softc); 152 static inline int dev_mc_delete(struct net_device *dev, void *addr, int alen, 164 return -if_delmulti(dev, (struc [all...] |
H A D | kdev_t.h | 33 #define MAJOR(dev) major((dev)) 34 #define MINOR(dev) minor((dev))
|
H A D | interrupt.h | 47 struct device *dev; member in struct:irq_ent 56 _irq_rid(struct device *dev, int irq) argument 58 if (irq == dev->irq) 60 return irq - dev->msix + 1; 73 _irq_ent(struct device *dev, int irq) argument 77 list_for_each_entry(irqe, &dev->irqents, links) 90 struct device *dev; local 94 dev = _pci_find_irq_dev(irq); 95 if (dev == NULL) 97 rid = _irq_rid(dev, ir 124 struct device *dev; local 141 struct device *dev; local [all...] |
/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/drivers/net/mlx4/ |
H A D | resource_tracker.c | 228 return priv->dev.caps.port_mask[port] == MLX4_PORT_TYPE_IB ? 0 : 1; 307 struct mlx4_priv *priv = mlx4_priv(&priv->dev); 315 res_alloc->allocated[(port - 1) * (priv->dev.num_vfs + 1) + slave] : 342 res_alloc->allocated[(port - 1) * (priv->dev.num_vfs + 1) + slave] += count; 360 struct mlx4_priv *priv = mlx4_priv(&priv->dev); 366 res_alloc->allocated[(port - 1) * (priv->dev.num_vfs + 1) + slave] -= count; 382 res_alloc->guaranteed[vf] = num_instances / (2 * (priv->dev.num_vfs + 1)); 384 if (vf == mlx4_master_func_num(&priv->dev)) { 388 res_alloc->res_free += priv->dev.caps.reserved_mtts; 389 res_alloc->guaranteed[vf] += priv->dev [all...] |
H A D | srq.c | 51 srq = radix_tree_lookup(&srq_table->tree, srqn & (priv->dev.caps.num_srqs - 1)); 70 return mlx4_cmd(dev, mailbox->dma, srq_num, 0, MLX4_CMD_SW2HW_SRQ, 76 return mlx4_cmd_box(dev, 0, mailbox ? mailbox->dma : 0, srq_num, 82 return mlx4_cmd(dev, limit_watermark, srq_num, 0, MLX4_CMD_ARM_SRQ, 88 return mlx4_cmd_box(dev, 0, mailbox->dma, srq_num, 0, MLX4_CMD_QUERY_SRQ, 93 struct mlx4_srq_table *srq_table = &mlx4_priv(&priv->dev)->srq_table; 100 err = mlx4_table_get(dev, &srq_table->table, *srqn); 104 err = mlx4_table_get(dev, &srq_table->cmpt_table, *srqn); 109 err_put: mlx4_table_put(dev, &srq_table->table, *srqn); 119 if (mlx4_is_mfunc(&priv->dev)) { [all...] |
/barrelfish-2018-10-04/usr/acpi/ |
H A D | lpc_ioapic_ioapic_impl.h | 18 static inline uint32_t lpc_ioapic_ioapic_read_32(lpc_ioapic_t *dev, argument 22 lpc_ioapic_ind_wr(dev, offset); 25 return lpc_ioapic_wdw_rd(dev); 28 static inline uint64_t lpc_ioapic_ioapic_read_64(lpc_ioapic_t *dev, argument 34 lpc_ioapic_ind_wr(dev, offset); 35 ret = lpc_ioapic_wdw_rd(dev); 38 lpc_ioapic_ind_wr(dev, offset + 1); 39 ret |= (uint64_t)lpc_ioapic_wdw_rd(dev) << 32; 44 static inline void lpc_ioapic_ioapic_write_32(lpc_ioapic_t *dev, size_t offset, argument 48 lpc_ioapic_ind_wr(dev, offse 54 lpc_ioapic_ioapic_write_64(lpc_ioapic_t *dev, size_t offset, uint64_t value) argument [all...] |
H A D | lpc_ioapic_spaces.h | 18 static inline uint32_t lpc_ioapic_ioapic_read_32(lpc_ioapic_t *dev, 20 static inline uint64_t lpc_ioapic_ioapic_read_64(lpc_ioapic_t *dev, 22 static inline void lpc_ioapic_ioapic_write_32(lpc_ioapic_t *dev, size_t offset, 24 static inline void lpc_ioapic_ioapic_write_64(lpc_ioapic_t *dev, size_t offset,
|
/barrelfish-2018-10-04/usr/drivers/usb/usb_manager/controller/ehci/ |
H A D | ehci_device.h | 19 #include <dev/omap/ehci_dev.h> 21 #include <dev/ehci_dev.h>
|
/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/drivers/infiniband/hw/mthca/ |
H A D | mthca_uar.c | 38 int mthca_uar_alloc(struct mthca_dev *dev, struct mthca_uar *uar) argument 40 uar->index = mthca_alloc(&dev->uar_table.alloc); 44 uar->pfn = (pci_resource_start(dev->pdev, 2) >> PAGE_SHIFT) + uar->index; 49 void mthca_uar_free(struct mthca_dev *dev, struct mthca_uar *uar) argument 51 mthca_free(&dev->uar_table.alloc, uar->index); 54 int mthca_init_uar_table(struct mthca_dev *dev) argument 58 ret = mthca_alloc_init(&dev->uar_table.alloc, 59 dev->limits.num_uars, 60 dev->limits.num_uars - 1, 61 dev 72 mthca_cleanup_uar_table(struct mthca_dev *dev) argument [all...] |
H A D | mthca_pd.c | 39 int mthca_pd_alloc(struct mthca_dev *dev, int privileged, struct mthca_pd *pd) argument 46 pd->pd_num = mthca_alloc(&dev->pd_table.alloc); 51 err = mthca_mr_alloc_notrans(dev, pd->pd_num, 56 mthca_free(&dev->pd_table.alloc, pd->pd_num); 62 void mthca_pd_free(struct mthca_dev *dev, struct mthca_pd *pd) argument 65 mthca_free_mr(dev, &pd->ntmr); 66 mthca_free(&dev->pd_table.alloc, pd->pd_num); 69 int mthca_init_pd_table(struct mthca_dev *dev) argument 71 return mthca_alloc_init(&dev->pd_table.alloc, 72 dev 77 mthca_cleanup_pd_table(struct mthca_dev *dev) argument [all...] |
H A D | mthca_catas.c | 60 struct mthca_dev *dev, *tmpdev; local 70 list_for_each_entry_safe(dev, tmpdev, &tlist, catas_err.list) { 71 struct pci_dev *pdev = dev->pdev; 72 ret = __mthca_restart_one(dev->pdev); 73 /* 'dev' now is not valid */ 86 static void handle_catas(struct mthca_dev *dev) argument 93 event.device = &dev->ib_dev; 96 dev->active = 0; 100 switch (swab32(readl(dev->catas_err.map)) >> 24) { 118 mthca_err(dev, "Catastrophi 134 struct mthca_dev *dev = (struct mthca_dev *) dev_ptr; local 147 mthca_start_catas_poll(struct mthca_dev *dev) argument 172 mthca_stop_catas_poll(struct mthca_dev *dev) argument [all...] |
H A D | mthca_eq.c | 165 static inline u64 async_mask(struct mthca_dev *dev) argument 167 return dev->mthca_flags & MTHCA_FLAG_SRQ ? 172 static inline void tavor_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) argument 184 dev->kar + MTHCA_EQ_DOORBELL, 185 MTHCA_GET_DOORBELL_LOCK(&dev->doorbell_lock)); 188 static inline void arbel_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) argument 193 dev->eq_regs.arbel.eq_set_ci_base + eq->eqn * 8); 198 static inline void set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) argument 200 if (mthca_is_memfree(dev)) 201 arbel_set_eq_ci(dev, e 206 tavor_eq_req_not(struct mthca_dev *dev, int eqn) argument 213 arbel_eq_req_not(struct mthca_dev *dev, u32 eqn_mask) argument 218 disarm_cq(struct mthca_dev *dev, int eqn, int cqn) argument 245 port_change(struct mthca_dev *dev, int port, int active) argument 259 mthca_eq_int(struct mthca_dev *dev, struct mthca_eq *eq) argument 392 struct mthca_dev *dev = dev_ptr; local 420 struct mthca_dev *dev = eq->dev; local 432 struct mthca_dev *dev = dev_ptr; local 454 struct mthca_dev *dev = eq->dev; local 464 mthca_create_eq(struct mthca_dev *dev, int nent, u8 intr, struct mthca_eq *eq) argument 594 mthca_free_eq(struct mthca_dev *dev, struct mthca_eq *eq) argument 637 mthca_free_irqs(struct mthca_dev *dev) argument 651 mthca_map_reg(struct mthca_dev *dev, unsigned long offset, unsigned long size, void __iomem **map) argument 664 mthca_map_eq_regs(struct mthca_dev *dev) argument 725 mthca_unmap_eq_regs(struct mthca_dev *dev) argument 737 mthca_map_eq_icm(struct mthca_dev *dev, u64 icm_virt) argument 771 mthca_unmap_eq_icm(struct mthca_dev *dev) argument 781 mthca_init_eq_table(struct mthca_dev *dev) argument 902 mthca_cleanup_eq_table(struct mthca_dev *dev) argument [all...] |
/barrelfish-2018-10-04/lib/dma/ |
H A D | dma_device.c | 24 * \param dev IOAT DMA device 30 errval_t dma_device_poll_channels(struct dma_device *dev) argument 32 if (dev->f.poll) { 33 return dev->f.poll(dev); 48 * \param dev DMA device 52 inline dma_dev_id_t dma_device_get_id(struct dma_device *dev) argument 54 return dev->id; 60 * \param dev DMA device 64 inline dma_dev_id_t dma_device_get_type(struct dma_device *dev) argument 76 dma_device_get_state(struct dma_device *dev) argument 89 dma_device_get_mmio_vbase(struct dma_device *dev) argument 102 dma_device_get_channel_count(struct dma_device *dev) argument 117 dma_device_get_channe_by_idx(struct dma_device *dev, uint8_t idx) argument 133 dma_device_get_channel(struct dma_device *dev) argument [all...] |
/barrelfish-2018-10-04/include/dma/ioat/ |
H A D | ioat_dma_device.h | 21 static inline struct ioat_dma_device *dma_device_to_ioat(struct dma_device *dev) argument 23 return (struct ioat_dma_device *)dev; 38 * \param dev returns a pointer to the device structure 45 struct ioat_dma_device **dev); 50 * \param dev IOAT DMA device to shutdown 55 errval_t ioat_dma_device_shutdown(struct ioat_dma_device *dev); 61 * \param dev returns a pointer to the device structure 66 errval_t ioat_dma_device_acquire(struct ioat_dma_device **dev); 72 * \param dev IOAT DMA device to be released 77 errval_t ioat_dma_device_release(struct ioat_dma_device *dev); [all...] |
/barrelfish-2018-10-04/usr/drivers/virtio/block/ |
H A D | request.h | 19 * \param dev vblock device to allocate the requests for 23 errval_t vblock_request_queue_init(struct vblock_device *dev); 28 * \param dev vblock device to free the requests 32 void vblock_request_queue_destroy(struct vblock_device *dev); 52 static inline struct vblock_req *vblock_request_alloc(struct vblock_device *dev) argument 54 return vblock_request_dequeue(&dev->free_queue); 73 static inline void vblock_request_free(struct vblock_device *dev, argument 76 vblock_request_enqueue(&dev->free_queue, req); 82 * \param dev the VirtIO block device 87 errval_t vblock_request_start(struct vblock_device *dev, [all...] |
/barrelfish-2018-10-04/include/dma/ |
H A D | dma_device.h | 38 * \param dev IOAT DMA device to shutdown 43 errval_t dma_device_shutdown(struct dma_device *dev); 49 * \param dev returns a pointer to the device structure 54 errval_t dma_device_acquire(struct dma_device **dev); 60 * \param dev IOAT DMA device to be released 65 errval_t dma_device_release(struct dma_device *dev); 70 * \param dev IOAT DMA device 76 errval_t dma_device_poll_channels(struct dma_device *dev); 87 * \param dev DMA device 91 dma_dev_id_t dma_device_get_id(struct dma_device *dev); [all...] |
/barrelfish-2018-10-04/lib/virtio/backends/ |
H A D | virtio_io.h | 19 errval_t virtio_device_pci_alloc(struct virtio_device **dev); 21 errval_t virtio_device_pci_free(struct virtio_device **dev);
|
H A D | virtio_pci.h | 19 errval_t virtio_device_pci_alloc(struct virtio_device **dev); 21 errval_t virtio_device_pci_free(struct virtio_device **dev);
|
/barrelfish-2018-10-04/usr/drivers/usb/usb_manager/controller/ohci/ |
H A D | ohci_device.h | 14 #include <dev/ohci_dev.h> 16 #include <dev/omap/ohci_dev.h>
|
/barrelfish-2018-10-04/lib/virtio/host/ |
H A D | channel.h | 14 errval_t virtio_host_flounder_init(struct virtio_device *dev);
|
/barrelfish-2018-10-04/usr/drivers/uhci/ |
H A D | uhci.h | 14 #include <dev/uhci_dev.h>
|
/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/include/ |
H A D | mlx4en.h | 5 void *mlx4_en_add(struct mlx4_dev *dev, struct mlx4_queue *queue);
|
/barrelfish-2018-10-04/include/dma/xeon_phi/ |
H A D | xeon_phi_dma_device.h | 51 static inline struct xeon_phi_dma_device *dma_device_to_xeon_phi(struct dma_device *dev) argument 53 return (struct xeon_phi_dma_device *) dev; 66 * \param dev returns a pointer to the device structure 72 struct xeon_phi_dma_device **dev); 77 * \param dev Xeon Phi DMA device to shutdown 82 errval_t xeon_phi_dma_device_shutdown(struct xeon_phi_dma_device *dev); 93 * \param dev Xeon Phi DMA device 98 errval_t xeon_phi_dma_device_intr_enable(struct xeon_phi_dma_device *dev, 106 * \param dev Xeon Phi DMA device 108 void xeon_phi_dma_device_intr_disable(struct xeon_phi_dma_device *dev); [all...] |
/barrelfish-2018-10-04/kernel/include/arch/x86_64/ |
H A D | cpuid_spaces.h | 21 static inline uint32_t cpuid_eax_read_32(cpuid_t *dev, size_t offset) argument 28 static inline uint32_t cpuid_ebx_read_32(cpuid_t *dev, size_t offset) argument 35 static inline uint32_t cpuid_ecx_read_32(cpuid_t *dev, size_t offset) argument 42 static inline uint32_t cpuid_edx_read_32(cpuid_t *dev, size_t offset) argument 49 static inline uint32_t cpuid_dcpa_read_32(cpuid_t *dev, size_t offset) argument 55 static inline uint32_t cpuid_dcpb_read_32(cpuid_t *dev, size_t offset) argument 60 static inline uint32_t cpuid_dcpc_read_32(cpuid_t *dev, size_t offset) argument 67 static inline void cpuid_eax_write_32(cpuid_t *dev, size_t offset, argument 72 static inline void cpuid_ebx_write_32(cpuid_t *dev, size_t offset, argument 77 static inline void cpuid_ecx_write_32(cpuid_t *dev, size_ argument 82 cpuid_edx_write_32(cpuid_t *dev, size_t offset, uint32_t value) argument 87 cpuid_dcpa_write_32(cpuid_t *dev, size_t offset, uint32_t value) argument 92 cpuid_dcpb_write_32(cpuid_t *dev, size_t offset, uint32_t value) argument 97 cpuid_dcpc_write_32(cpuid_t *dev, size_t offset, uint32_t value) argument [all...] |