Searched refs:dev (Results 1 - 25 of 400) sorted by relevance

1234567891011>>

/barrelfish-2018-10-04/lib/tommath/
H A Dmess.sh2 if cvs log $1 >/dev/null 2>/dev/null; then exit 0; else echo "$1 shouldn't be here" ; exit 1; fi
/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/include/linux/
H A Dnetdevice.h62 #define netif_running(dev) !!((dev)->if_drv_flags & IFF_DRV_RUNNING)
64 #define netif_oper_up(dev) !!((dev)->if_flags & IFF_UP)
65 #define netif_carrier_ok(dev) netif_running(dev)
68 netdev_priv(const struct net_device *dev) {
69 return (dev->if_softc);
152 static inline int dev_mc_delete(struct net_device *dev, void *addr, int alen,
164 return -if_delmulti(dev, (struc
[all...]
H A Dkdev_t.h33 #define MAJOR(dev) major((dev))
34 #define MINOR(dev) minor((dev))
H A Dinterrupt.h47 struct device *dev; member in struct:irq_ent
56 _irq_rid(struct device *dev, int irq) argument
58 if (irq == dev->irq)
60 return irq - dev->msix + 1;
73 _irq_ent(struct device *dev, int irq) argument
77 list_for_each_entry(irqe, &dev->irqents, links)
90 struct device *dev; local
94 dev = _pci_find_irq_dev(irq);
95 if (dev == NULL)
97 rid = _irq_rid(dev, ir
124 struct device *dev; local
141 struct device *dev; local
[all...]
/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/drivers/net/mlx4/
H A Dresource_tracker.c228 return priv->dev.caps.port_mask[port] == MLX4_PORT_TYPE_IB ? 0 : 1;
307 struct mlx4_priv *priv = mlx4_priv(&priv->dev);
315 res_alloc->allocated[(port - 1) * (priv->dev.num_vfs + 1) + slave] :
342 res_alloc->allocated[(port - 1) * (priv->dev.num_vfs + 1) + slave] += count;
360 struct mlx4_priv *priv = mlx4_priv(&priv->dev);
366 res_alloc->allocated[(port - 1) * (priv->dev.num_vfs + 1) + slave] -= count;
382 res_alloc->guaranteed[vf] = num_instances / (2 * (priv->dev.num_vfs + 1));
384 if (vf == mlx4_master_func_num(&priv->dev)) {
388 res_alloc->res_free += priv->dev.caps.reserved_mtts;
389 res_alloc->guaranteed[vf] += priv->dev
[all...]
H A Dsrq.c51 srq = radix_tree_lookup(&srq_table->tree, srqn & (priv->dev.caps.num_srqs - 1));
70 return mlx4_cmd(dev, mailbox->dma, srq_num, 0, MLX4_CMD_SW2HW_SRQ,
76 return mlx4_cmd_box(dev, 0, mailbox ? mailbox->dma : 0, srq_num,
82 return mlx4_cmd(dev, limit_watermark, srq_num, 0, MLX4_CMD_ARM_SRQ,
88 return mlx4_cmd_box(dev, 0, mailbox->dma, srq_num, 0, MLX4_CMD_QUERY_SRQ,
93 struct mlx4_srq_table *srq_table = &mlx4_priv(&priv->dev)->srq_table;
100 err = mlx4_table_get(dev, &srq_table->table, *srqn);
104 err = mlx4_table_get(dev, &srq_table->cmpt_table, *srqn);
109 err_put: mlx4_table_put(dev, &srq_table->table, *srqn);
119 if (mlx4_is_mfunc(&priv->dev)) {
[all...]
/barrelfish-2018-10-04/usr/acpi/
H A Dlpc_ioapic_ioapic_impl.h18 static inline uint32_t lpc_ioapic_ioapic_read_32(lpc_ioapic_t *dev, argument
22 lpc_ioapic_ind_wr(dev, offset);
25 return lpc_ioapic_wdw_rd(dev);
28 static inline uint64_t lpc_ioapic_ioapic_read_64(lpc_ioapic_t *dev, argument
34 lpc_ioapic_ind_wr(dev, offset);
35 ret = lpc_ioapic_wdw_rd(dev);
38 lpc_ioapic_ind_wr(dev, offset + 1);
39 ret |= (uint64_t)lpc_ioapic_wdw_rd(dev) << 32;
44 static inline void lpc_ioapic_ioapic_write_32(lpc_ioapic_t *dev, size_t offset, argument
48 lpc_ioapic_ind_wr(dev, offse
54 lpc_ioapic_ioapic_write_64(lpc_ioapic_t *dev, size_t offset, uint64_t value) argument
[all...]
H A Dlpc_ioapic_spaces.h18 static inline uint32_t lpc_ioapic_ioapic_read_32(lpc_ioapic_t *dev,
20 static inline uint64_t lpc_ioapic_ioapic_read_64(lpc_ioapic_t *dev,
22 static inline void lpc_ioapic_ioapic_write_32(lpc_ioapic_t *dev, size_t offset,
24 static inline void lpc_ioapic_ioapic_write_64(lpc_ioapic_t *dev, size_t offset,
/barrelfish-2018-10-04/usr/drivers/usb/usb_manager/controller/ehci/
H A Dehci_device.h19 #include <dev/omap/ehci_dev.h>
21 #include <dev/ehci_dev.h>
/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/drivers/infiniband/hw/mthca/
H A Dmthca_uar.c38 int mthca_uar_alloc(struct mthca_dev *dev, struct mthca_uar *uar) argument
40 uar->index = mthca_alloc(&dev->uar_table.alloc);
44 uar->pfn = (pci_resource_start(dev->pdev, 2) >> PAGE_SHIFT) + uar->index;
49 void mthca_uar_free(struct mthca_dev *dev, struct mthca_uar *uar) argument
51 mthca_free(&dev->uar_table.alloc, uar->index);
54 int mthca_init_uar_table(struct mthca_dev *dev) argument
58 ret = mthca_alloc_init(&dev->uar_table.alloc,
59 dev->limits.num_uars,
60 dev->limits.num_uars - 1,
61 dev
72 mthca_cleanup_uar_table(struct mthca_dev *dev) argument
[all...]
H A Dmthca_pd.c39 int mthca_pd_alloc(struct mthca_dev *dev, int privileged, struct mthca_pd *pd) argument
46 pd->pd_num = mthca_alloc(&dev->pd_table.alloc);
51 err = mthca_mr_alloc_notrans(dev, pd->pd_num,
56 mthca_free(&dev->pd_table.alloc, pd->pd_num);
62 void mthca_pd_free(struct mthca_dev *dev, struct mthca_pd *pd) argument
65 mthca_free_mr(dev, &pd->ntmr);
66 mthca_free(&dev->pd_table.alloc, pd->pd_num);
69 int mthca_init_pd_table(struct mthca_dev *dev) argument
71 return mthca_alloc_init(&dev->pd_table.alloc,
72 dev
77 mthca_cleanup_pd_table(struct mthca_dev *dev) argument
[all...]
H A Dmthca_catas.c60 struct mthca_dev *dev, *tmpdev; local
70 list_for_each_entry_safe(dev, tmpdev, &tlist, catas_err.list) {
71 struct pci_dev *pdev = dev->pdev;
72 ret = __mthca_restart_one(dev->pdev);
73 /* 'dev' now is not valid */
86 static void handle_catas(struct mthca_dev *dev) argument
93 event.device = &dev->ib_dev;
96 dev->active = 0;
100 switch (swab32(readl(dev->catas_err.map)) >> 24) {
118 mthca_err(dev, "Catastrophi
134 struct mthca_dev *dev = (struct mthca_dev *) dev_ptr; local
147 mthca_start_catas_poll(struct mthca_dev *dev) argument
172 mthca_stop_catas_poll(struct mthca_dev *dev) argument
[all...]
H A Dmthca_eq.c165 static inline u64 async_mask(struct mthca_dev *dev) argument
167 return dev->mthca_flags & MTHCA_FLAG_SRQ ?
172 static inline void tavor_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) argument
184 dev->kar + MTHCA_EQ_DOORBELL,
185 MTHCA_GET_DOORBELL_LOCK(&dev->doorbell_lock));
188 static inline void arbel_set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) argument
193 dev->eq_regs.arbel.eq_set_ci_base + eq->eqn * 8);
198 static inline void set_eq_ci(struct mthca_dev *dev, struct mthca_eq *eq, u32 ci) argument
200 if (mthca_is_memfree(dev))
201 arbel_set_eq_ci(dev, e
206 tavor_eq_req_not(struct mthca_dev *dev, int eqn) argument
213 arbel_eq_req_not(struct mthca_dev *dev, u32 eqn_mask) argument
218 disarm_cq(struct mthca_dev *dev, int eqn, int cqn) argument
245 port_change(struct mthca_dev *dev, int port, int active) argument
259 mthca_eq_int(struct mthca_dev *dev, struct mthca_eq *eq) argument
392 struct mthca_dev *dev = dev_ptr; local
420 struct mthca_dev *dev = eq->dev; local
432 struct mthca_dev *dev = dev_ptr; local
454 struct mthca_dev *dev = eq->dev; local
464 mthca_create_eq(struct mthca_dev *dev, int nent, u8 intr, struct mthca_eq *eq) argument
594 mthca_free_eq(struct mthca_dev *dev, struct mthca_eq *eq) argument
637 mthca_free_irqs(struct mthca_dev *dev) argument
651 mthca_map_reg(struct mthca_dev *dev, unsigned long offset, unsigned long size, void __iomem **map) argument
664 mthca_map_eq_regs(struct mthca_dev *dev) argument
725 mthca_unmap_eq_regs(struct mthca_dev *dev) argument
737 mthca_map_eq_icm(struct mthca_dev *dev, u64 icm_virt) argument
771 mthca_unmap_eq_icm(struct mthca_dev *dev) argument
781 mthca_init_eq_table(struct mthca_dev *dev) argument
902 mthca_cleanup_eq_table(struct mthca_dev *dev) argument
[all...]
/barrelfish-2018-10-04/lib/dma/
H A Ddma_device.c24 * \param dev IOAT DMA device
30 errval_t dma_device_poll_channels(struct dma_device *dev) argument
32 if (dev->f.poll) {
33 return dev->f.poll(dev);
48 * \param dev DMA device
52 inline dma_dev_id_t dma_device_get_id(struct dma_device *dev) argument
54 return dev->id;
60 * \param dev DMA device
64 inline dma_dev_id_t dma_device_get_type(struct dma_device *dev) argument
76 dma_device_get_state(struct dma_device *dev) argument
89 dma_device_get_mmio_vbase(struct dma_device *dev) argument
102 dma_device_get_channel_count(struct dma_device *dev) argument
117 dma_device_get_channe_by_idx(struct dma_device *dev, uint8_t idx) argument
133 dma_device_get_channel(struct dma_device *dev) argument
[all...]
/barrelfish-2018-10-04/include/dma/ioat/
H A Dioat_dma_device.h21 static inline struct ioat_dma_device *dma_device_to_ioat(struct dma_device *dev) argument
23 return (struct ioat_dma_device *)dev;
38 * \param dev returns a pointer to the device structure
45 struct ioat_dma_device **dev);
50 * \param dev IOAT DMA device to shutdown
55 errval_t ioat_dma_device_shutdown(struct ioat_dma_device *dev);
61 * \param dev returns a pointer to the device structure
66 errval_t ioat_dma_device_acquire(struct ioat_dma_device **dev);
72 * \param dev IOAT DMA device to be released
77 errval_t ioat_dma_device_release(struct ioat_dma_device *dev);
[all...]
/barrelfish-2018-10-04/usr/drivers/virtio/block/
H A Drequest.h19 * \param dev vblock device to allocate the requests for
23 errval_t vblock_request_queue_init(struct vblock_device *dev);
28 * \param dev vblock device to free the requests
32 void vblock_request_queue_destroy(struct vblock_device *dev);
52 static inline struct vblock_req *vblock_request_alloc(struct vblock_device *dev) argument
54 return vblock_request_dequeue(&dev->free_queue);
73 static inline void vblock_request_free(struct vblock_device *dev, argument
76 vblock_request_enqueue(&dev->free_queue, req);
82 * \param dev the VirtIO block device
87 errval_t vblock_request_start(struct vblock_device *dev,
[all...]
/barrelfish-2018-10-04/include/dma/
H A Ddma_device.h38 * \param dev IOAT DMA device to shutdown
43 errval_t dma_device_shutdown(struct dma_device *dev);
49 * \param dev returns a pointer to the device structure
54 errval_t dma_device_acquire(struct dma_device **dev);
60 * \param dev IOAT DMA device to be released
65 errval_t dma_device_release(struct dma_device *dev);
70 * \param dev IOAT DMA device
76 errval_t dma_device_poll_channels(struct dma_device *dev);
87 * \param dev DMA device
91 dma_dev_id_t dma_device_get_id(struct dma_device *dev);
[all...]
/barrelfish-2018-10-04/lib/virtio/backends/
H A Dvirtio_io.h19 errval_t virtio_device_pci_alloc(struct virtio_device **dev);
21 errval_t virtio_device_pci_free(struct virtio_device **dev);
H A Dvirtio_pci.h19 errval_t virtio_device_pci_alloc(struct virtio_device **dev);
21 errval_t virtio_device_pci_free(struct virtio_device **dev);
/barrelfish-2018-10-04/usr/drivers/usb/usb_manager/controller/ohci/
H A Dohci_device.h14 #include <dev/ohci_dev.h>
16 #include <dev/omap/ohci_dev.h>
/barrelfish-2018-10-04/lib/virtio/host/
H A Dchannel.h14 errval_t virtio_host_flounder_init(struct virtio_device *dev);
/barrelfish-2018-10-04/usr/drivers/uhci/
H A Duhci.h14 #include <dev/uhci_dev.h>
/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/include/
H A Dmlx4en.h5 void *mlx4_en_add(struct mlx4_dev *dev, struct mlx4_queue *queue);
/barrelfish-2018-10-04/include/dma/xeon_phi/
H A Dxeon_phi_dma_device.h51 static inline struct xeon_phi_dma_device *dma_device_to_xeon_phi(struct dma_device *dev) argument
53 return (struct xeon_phi_dma_device *) dev;
66 * \param dev returns a pointer to the device structure
72 struct xeon_phi_dma_device **dev);
77 * \param dev Xeon Phi DMA device to shutdown
82 errval_t xeon_phi_dma_device_shutdown(struct xeon_phi_dma_device *dev);
93 * \param dev Xeon Phi DMA device
98 errval_t xeon_phi_dma_device_intr_enable(struct xeon_phi_dma_device *dev,
106 * \param dev Xeon Phi DMA device
108 void xeon_phi_dma_device_intr_disable(struct xeon_phi_dma_device *dev);
[all...]
/barrelfish-2018-10-04/kernel/include/arch/x86_64/
H A Dcpuid_spaces.h21 static inline uint32_t cpuid_eax_read_32(cpuid_t *dev, size_t offset) argument
28 static inline uint32_t cpuid_ebx_read_32(cpuid_t *dev, size_t offset) argument
35 static inline uint32_t cpuid_ecx_read_32(cpuid_t *dev, size_t offset) argument
42 static inline uint32_t cpuid_edx_read_32(cpuid_t *dev, size_t offset) argument
49 static inline uint32_t cpuid_dcpa_read_32(cpuid_t *dev, size_t offset) argument
55 static inline uint32_t cpuid_dcpb_read_32(cpuid_t *dev, size_t offset) argument
60 static inline uint32_t cpuid_dcpc_read_32(cpuid_t *dev, size_t offset) argument
67 static inline void cpuid_eax_write_32(cpuid_t *dev, size_t offset, argument
72 static inline void cpuid_ebx_write_32(cpuid_t *dev, size_t offset, argument
77 static inline void cpuid_ecx_write_32(cpuid_t *dev, size_ argument
82 cpuid_edx_write_32(cpuid_t *dev, size_t offset, uint32_t value) argument
87 cpuid_dcpa_write_32(cpuid_t *dev, size_t offset, uint32_t value) argument
92 cpuid_dcpb_write_32(cpuid_t *dev, size_t offset, uint32_t value) argument
97 cpuid_dcpc_write_32(cpuid_t *dev, size_t offset, uint32_t value) argument
[all...]

Completed in 147 milliseconds

1234567891011>>