/linux-master/drivers/net/ethernet/marvell/octeon_ep/ |
H A D | octep_rx.c | 15 static void octep_oq_reset_indices(struct octep_oq *oq) argument 17 oq->host_read_idx = 0; 18 oq->host_refill_idx = 0; 19 oq->refill_count = 0; 20 oq->last_pkt_count = 0; 21 oq->pkts_pending = 0; 27 * @oq: Octeon Rx queue data structure. 32 static int octep_oq_fill_ring_buffers(struct octep_oq *oq) argument 34 struct octep_oq_desc_hw *desc_ring = oq->desc_ring; 38 for (i = 0; i < oq 79 octep_oq_refill(struct octep_device *oct, struct octep_oq *oq) argument 125 struct octep_oq *oq; local 198 octep_oq_free_ring_buffers(struct octep_oq *oq) argument 225 octep_free_oq(struct octep_oq *oq) argument 316 octep_oq_check_hw_for_pkts(struct octep_device *oct, struct octep_oq *oq) argument 352 __octep_oq_process_rx(struct octep_device *oct, struct octep_oq *oq, u16 pkts_to_process) argument 478 octep_oq_process_rx(struct octep_oq *oq, int budget) argument [all...] |
H A D | octep_config.h | 67 #define CFG_GET_OQ_NUM_DESC(cfg) ((cfg)->oq.num_descs) 68 #define CFG_GET_OQ_BUF_SIZE(cfg) ((cfg)->oq.buf_size) 69 #define CFG_GET_OQ_REFILL_THRESHOLD(cfg) ((cfg)->oq.refill_threshold) 70 #define CFG_GET_OQ_INTR_PKT(cfg) ((cfg)->oq.oq_intr_pkt) 71 #define CFG_GET_OQ_INTR_TIME(cfg) ((cfg)->oq.oq_intr_time) 72 #define CFG_GET_OQ_WMARK(cfg) ((cfg)->oq.wmark) 235 struct octep_oq_config oq; member in struct:octep_config
|
H A D | octep_cnxk_pf.c | 256 conf->oq.num_descs = OCTEP_OQ_MAX_DESCRIPTORS; 257 conf->oq.buf_size = OCTEP_OQ_BUF_SIZE; 258 conf->oq.refill_threshold = OCTEP_OQ_REFILL_THRESHOLD; 259 conf->oq.oq_intr_pkt = OCTEP_OQ_INTR_PKT_THRESHOLD; 260 conf->oq.oq_intr_time = OCTEP_OQ_INTR_TIME_THRESHOLD; 261 conf->oq.wmark = OCTEP_OQ_WMARK_MIN; 335 struct octep_oq *oq = oct->oq[oq_no]; local 360 oq->desc_ring_dma); 362 oq 641 struct octep_oq *oq = vector->oq; local [all...] |
H A D | octep_cn9k_pf.c | 237 conf->oq.num_descs = OCTEP_OQ_MAX_DESCRIPTORS; 238 conf->oq.buf_size = OCTEP_OQ_BUF_SIZE; 239 conf->oq.refill_threshold = OCTEP_OQ_REFILL_THRESHOLD; 240 conf->oq.oq_intr_pkt = OCTEP_OQ_INTR_PKT_THRESHOLD; 241 conf->oq.oq_intr_time = OCTEP_OQ_INTR_TIME_THRESHOLD; 315 struct octep_oq *oq = oct->oq[oq_no]; local 340 oq->desc_ring_dma); 342 oq->max_count); 346 oq_ctl |= (oq 620 struct octep_oq *oq = vector->oq; local [all...] |
H A D | octep_ethtool.c | 162 struct octep_oq *oq = oct->oq[q]; local 168 rx_packets += oq->stats.packets; 169 rx_bytes += oq->stats.bytes; 170 rx_alloc_errors += oq->stats.alloc_failures; 219 struct octep_oq *oq = oct->oq[q]; local 221 data[i++] = oq->stats.packets; 222 data[i++] = oq->stats.bytes; 223 data[i++] = oq [all...] |
H A D | octep_main.c | 66 ioq_vector->oq = oct->oq[i]; 561 * @oq: Octeon Rx queue data structure. 563 static void octep_enable_ioq_irq(struct octep_iq *iq, struct octep_oq *oq) argument 565 u32 pkts_pend = oq->pkts_pending; 573 if (oq->last_pkt_count - pkts_pend) { 574 writel(oq->last_pkt_count - pkts_pend, oq->pkts_sent_reg); 575 oq->last_pkt_count = pkts_pend; 580 writeq(1UL << OCTEP_OQ_INTR_RESEND_BIT, oq 1010 struct octep_oq *oq = oct->oq[q]; local [all...] |
H A D | octep_main.h | 151 struct octep_oq *oq; member in struct:octep_ioq_vector 263 struct octep_oq *oq[OCTEP_MAX_OQ]; member in struct:octep_device 404 int octep_oq_process_rx(struct octep_oq *oq, int budget);
|
/linux-master/drivers/net/ethernet/marvell/octeon_ep_vf/ |
H A D | octep_vf_rx.c | 15 static void octep_vf_oq_reset_indices(struct octep_vf_oq *oq) argument 17 oq->host_read_idx = 0; 18 oq->host_refill_idx = 0; 19 oq->refill_count = 0; 20 oq->last_pkt_count = 0; 21 oq->pkts_pending = 0; 27 * @oq: Octeon Rx queue data structure. 32 static int octep_vf_oq_fill_ring_buffers(struct octep_vf_oq *oq) argument 34 struct octep_vf_oq_desc_hw *desc_ring = oq->desc_ring; 38 for (i = 0; i < oq 79 octep_vf_oq_refill(struct octep_vf_device *oct, struct octep_vf_oq *oq) argument 125 struct octep_vf_oq *oq; local 199 octep_vf_oq_free_ring_buffers(struct octep_vf_oq *oq) argument 226 octep_vf_free_oq(struct octep_vf_oq *oq) argument 317 octep_vf_oq_check_hw_for_pkts(struct octep_vf_device *oct, struct octep_vf_oq *oq) argument 353 __octep_vf_oq_process_rx(struct octep_vf_device *oct, struct octep_vf_oq *oq, u16 pkts_to_process) argument 478 octep_vf_oq_process_rx(struct octep_vf_oq *oq, int budget) argument [all...] |
H A D | octep_vf_config.h | 63 #define CFG_GET_OQ_NUM_DESC(cfg) ((cfg)->oq.num_descs) 64 #define CFG_GET_OQ_BUF_SIZE(cfg) ((cfg)->oq.buf_size) 65 #define CFG_GET_OQ_REFILL_THRESHOLD(cfg) ((cfg)->oq.refill_threshold) 66 #define CFG_GET_OQ_INTR_PKT(cfg) ((cfg)->oq.oq_intr_pkt) 67 #define CFG_GET_OQ_INTR_TIME(cfg) ((cfg)->oq.oq_intr_time) 68 #define CFG_GET_OQ_WMARK(cfg) ((cfg)->oq.wmark) 152 struct octep_vf_oq_config oq; member in struct:octep_vf_config
|
H A D | octep_vf_cnxk.c | 153 conf->oq.num_descs = OCTEP_VF_OQ_MAX_DESCRIPTORS; 154 conf->oq.buf_size = OCTEP_VF_OQ_BUF_SIZE; 155 conf->oq.refill_threshold = OCTEP_VF_OQ_REFILL_THRESHOLD; 156 conf->oq.oq_intr_pkt = OCTEP_VF_OQ_INTR_PKT_THRESHOLD; 157 conf->oq.oq_intr_time = OCTEP_VF_OQ_INTR_TIME_THRESHOLD; 158 conf->oq.wmark = OCTEP_VF_OQ_WMARK_MIN; 204 struct octep_vf_oq *oq = oct->oq[oq_no]; local 230 octep_vf_write_csr64(oct, CNXK_VF_SDP_R_OUT_SLIST_BADDR(oq_no), oq->desc_ring_dma); 231 octep_vf_write_csr64(oct, CNXK_VF_SDP_R_OUT_SLIST_RSIZE(oq_no), oq 284 struct octep_vf_oq *oq; local [all...] |
H A D | octep_vf_cn9k.c | 151 conf->oq.num_descs = OCTEP_VF_OQ_MAX_DESCRIPTORS; 152 conf->oq.buf_size = OCTEP_VF_OQ_BUF_SIZE; 153 conf->oq.refill_threshold = OCTEP_VF_OQ_REFILL_THRESHOLD; 154 conf->oq.oq_intr_pkt = OCTEP_VF_OQ_INTR_PKT_THRESHOLD; 155 conf->oq.oq_intr_time = OCTEP_VF_OQ_INTR_TIME_THRESHOLD; 201 struct octep_vf_oq *oq = oct->oq[oq_no]; local 227 octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_SLIST_BADDR(oq_no), oq->desc_ring_dma); 228 octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_SLIST_RSIZE(oq_no), oq->max_count); 232 oq_ctl |= (oq 273 struct octep_vf_oq *oq; local [all...] |
H A D | octep_vf_ethtool.c | 126 struct octep_vf_oq *oq = oct->oq[q]; local 129 rx_alloc_errors += oq->stats.alloc_failures; 155 struct octep_vf_oq *oq = oct->oq[q]; local 157 data[i++] = oq->stats.packets; 158 data[i++] = oq->stats.bytes; 159 data[i++] = oq->stats.alloc_failures;
|
H A D | octep_vf_main.c | 65 ioq_vector->oq = oct->oq[i]; 294 * @oq: Octeon Rx queue data structure. 296 static void octep_vf_enable_ioq_irq(struct octep_vf_iq *iq, struct octep_vf_oq *oq) argument 298 u32 pkts_pend = oq->pkts_pending; 306 if (oq->last_pkt_count - pkts_pend) { 307 writel(oq->last_pkt_count - pkts_pend, oq->pkts_sent_reg); 308 oq->last_pkt_count = pkts_pend; 313 writeq(1UL << OCTEP_VF_OQ_INTR_RESEND_BIT, oq 791 struct octep_vf_oq *oq = oct->oq[q]; local [all...] |
H A D | octep_vf_main.h | 127 struct octep_vf_oq *oq; member in struct:octep_vf_ioq_vector 252 struct octep_vf_oq *oq[OCTEP_VF_MAX_OQ]; member in struct:octep_vf_device 329 int octep_vf_oq_process_rx(struct octep_vf_oq *oq, int budget);
|
/linux-master/drivers/net/ethernet/cavium/liquidio/ |
H A D | cn23xx_vf_regs.h | 164 #define CN23XX_VF_SLI_OQ_PKT_CONTROL(oq) \ 165 (CN23XX_VF_SLI_OQ_PKT_CONTROL_START + ((oq) * CN23XX_VF_OQ_OFFSET)) 167 #define CN23XX_VF_SLI_OQ_BASE_ADDR64(oq) \ 168 (CN23XX_VF_SLI_OQ_BASE_ADDR_START64 + ((oq) * CN23XX_VF_OQ_OFFSET)) 170 #define CN23XX_VF_SLI_OQ_SIZE(oq) \ 171 (CN23XX_VF_SLI_OQ_SIZE_START + ((oq) * CN23XX_VF_OQ_OFFSET)) 173 #define CN23XX_VF_SLI_OQ_BUFF_INFO_SIZE(oq) \ 174 (CN23XX_VF_SLI_OQ0_BUFF_INFO_SIZE + ((oq) * CN23XX_VF_OQ_OFFSET)) 176 #define CN23XX_VF_SLI_OQ_PKTS_SENT(oq) \ 177 (CN23XX_VF_SLI_OQ_PKT_SENT_START + ((oq) * CN23XX_VF_OQ_OFFSE [all...] |
H A D | cn23xx_pf_regs.h | 282 #define CN23XX_SLI_OQ_PKT_CONTROL(oq) \ 283 (CN23XX_SLI_OQ_PKT_CONTROL_START + ((oq) * CN23XX_OQ_OFFSET)) 285 #define CN23XX_SLI_OQ_BASE_ADDR64(oq) \ 286 (CN23XX_SLI_OQ_BASE_ADDR_START64 + ((oq) * CN23XX_OQ_OFFSET)) 288 #define CN23XX_SLI_OQ_SIZE(oq) \ 289 (CN23XX_SLI_OQ_SIZE_START + ((oq) * CN23XX_OQ_OFFSET)) 291 #define CN23XX_SLI_OQ_BUFF_INFO_SIZE(oq) \ 292 (CN23XX_SLI_OQ0_BUFF_INFO_SIZE + ((oq) * CN23XX_OQ_OFFSET)) 294 #define CN23XX_SLI_OQ_PKTS_SENT(oq) \ 295 (CN23XX_SLI_OQ_PKT_SENT_START + ((oq) * CN23XX_OQ_OFFSE [all...] |
H A D | octeon_config.h | 131 #define CFG_GET_OQ_MAX_Q(cfg) ((cfg)->oq.max_oqs) 132 #define CFG_GET_OQ_PKTS_PER_INTR(cfg) ((cfg)->oq.pkts_per_intr) 133 #define CFG_GET_OQ_REFILL_THRESHOLD(cfg) ((cfg)->oq.refill_threshold) 134 #define CFG_GET_OQ_INTR_PKT(cfg) ((cfg)->oq.oq_intr_pkt) 135 #define CFG_GET_OQ_INTR_TIME(cfg) ((cfg)->oq.oq_intr_time) 136 #define CFG_SET_OQ_INTR_PKT(cfg, val) (cfg)->oq.oq_intr_pkt = val 137 #define CFG_SET_OQ_INTR_TIME(cfg, val) (cfg)->oq.oq_intr_time = val 413 struct octeon_oq_config oq; member in struct:octeon_config
|
H A D | cn66xx_regs.h | 277 #define CN6XXX_SLI_OQ_BASE_ADDR64(oq) \ 278 (CN6XXX_SLI_OQ_BASE_ADDR_START64 + ((oq) * CN6XXX_OQ_OFFSET)) 280 #define CN6XXX_SLI_OQ_SIZE(oq) \ 281 (CN6XXX_SLI_OQ_SIZE_START + ((oq) * CN6XXX_OQ_OFFSET)) 283 #define CN6XXX_SLI_OQ_BUFF_INFO_SIZE(oq) \ 284 (CN6XXX_SLI_OQ0_BUFF_INFO_SIZE + ((oq) * CN6XXX_OQ_OFFSET)) 286 #define CN6XXX_SLI_OQ_PKTS_SENT(oq) \ 287 (CN6XXX_SLI_OQ_PKT_SENT_START + ((oq) * CN6XXX_OQ_OFFSET)) 289 #define CN6XXX_SLI_OQ_PKTS_CREDIT(oq) \ 290 (CN6XXX_SLI_OQ_PKT_CREDITS_START + ((oq) * CN6XXX_OQ_OFFSE [all...] |
H A D | cn66xx_device.c | 136 /* core clock per us / oq ticks will be fractional. TO avoid that 143 /* This gives the oq ticks (1024 core clock cycles) per millisecond */ 146 /* time_intr is in microseconds. The next 2 steps gives the oq ticks 343 mask |= oct->io_qmask.oq; 378 mask ^= oct->io_qmask.oq; 383 mask = (u32)oct->io_qmask.oq; 393 if (!(oct->io_qmask.oq & BIT_ULL(i))) 521 droq_mask &= oct->io_qmask.oq; 557 droq_time_mask &= oct->io_qmask.oq; 558 droq_cnt_mask &= oct->io_qmask.oq; [all...] |
H A D | octeon_device.c | 52 .oq = { 161 .oq = { 327 .oq = { 430 .oq = { 651 if (oct->io_qmask.oq & BIT_ULL(i)) 1296 (oct->io_qmask.oq & BIT_ULL(q_no)))
|
H A D | octeon_device.h | 189 u64 oq; member in struct:octeon_io_enable
|
H A D | octeon_droq.c | 213 oct->io_qmask.oq &= ~(1ULL << q_no); 308 oct->io_qmask.oq |= BIT_ULL(q_no);
|
H A D | cn23xx_vf_device.c | 38 /* This gives the oq ticks (1024 core clock cycles) per millisecond */ 41 /* time_intr is in microseconds. The next 2 steps gives the oq ticks 345 if (oct->io_qmask.oq & BIT_ULL(q_no)) {
|
H A D | lio_ethtool.c | 1037 "Sending iq/oq config failed status: %x\n", 1050 "iq/oq config failed: %x\n", retval); 1154 if (!(oct->io_qmask.oq & BIT_ULL(i))) 1683 if (!(oct_dev->io_qmask.oq & BIT_ULL(j))) 1882 if (!(oct_dev->io_qmask.oq & BIT_ULL(i))) 1930 if (!(oct_dev->io_qmask.oq & BIT_ULL(i)))
|
/linux-master/fs/xfs/ |
H A D | xfs_trans_dquot.c | 79 struct xfs_dqtrx *oq, *nq; local 97 oq = &oqa[i]; 100 if (oq->qt_blk_res && oq->qt_bcount_delta > 0) 101 blk_res_used = oq->qt_bcount_delta; 103 nq->qt_dquot = oq->qt_dquot; 110 nq->qt_blk_res = oq->qt_blk_res - blk_res_used; 111 oq->qt_blk_res = blk_res_used; 113 nq->qt_rtblk_res = oq->qt_rtblk_res - 114 oq [all...] |