Lines Matching refs:gwca

125 		iowrite32(BIT(priv->gwca.index), priv->addr + FWPBFC(i));
129 iowrite32(FWPC0_DEFAULT, priv->addr + FWPC0(priv->gwca.index));
130 iowrite32(FWPC1_DDE, priv->addr + FWPC1(priv->gwca.index));
131 iowrite32(0, priv->addr + FWPBFC(priv->gwca.index));
132 iowrite32(GENMASK(RSWITCH_NUM_PORTS - 1, 0), priv->addr + FWPBFC(priv->gwca.index));
141 if (!rswitch_agent_clock_is_enabled(priv->addr, priv->gwca.index))
142 rswitch_agent_clock_ctrl(priv->addr, priv->gwca.index, 1);
149 rswitch_agent_clock_ctrl(priv->addr, priv->gwca.index, 0);
170 u32 *mask = tx ? priv->gwca.tx_irq_bits : priv->gwca.rx_irq_bits;
294 struct rswitch_gwca_queue *gq = &priv->gwca.ts_queue;
341 priv->gwca.tx_irq_bits[i] |= bit;
343 priv->gwca.rx_irq_bits[i] |= bit;
394 linkfix = &priv->gwca.linkfix_table[gq->index];
419 struct rswitch_gwca_queue *gq = &priv->gwca.ts_queue;
495 linkfix = &priv->gwca.linkfix_table[gq->index];
508 unsigned int i, num_queues = priv->gwca.num_queues;
509 struct rswitch_gwca *gwca = &priv->gwca;
512 gwca->linkfix_table_size = sizeof(struct rswitch_desc) * num_queues;
513 gwca->linkfix_table = dma_alloc_coherent(dev, gwca->linkfix_table_size,
514 &gwca->linkfix_table_dma, GFP_KERNEL);
515 if (!gwca->linkfix_table)
518 gwca->linkfix_table[i].die_dt = DT_EOS;
525 struct rswitch_gwca *gwca = &priv->gwca;
527 if (gwca->linkfix_table)
528 dma_free_coherent(&priv->pdev->dev, gwca->linkfix_table_size,
529 gwca->linkfix_table, gwca->linkfix_table_dma);
530 gwca->linkfix_table = NULL;
535 struct rswitch_gwca_queue *gq = &priv->gwca.ts_queue;
550 INIT_LIST_HEAD(&priv->gwca.ts_info_list);
560 index = find_first_zero_bit(priv->gwca.used, priv->gwca.num_queues);
561 if (index >= priv->gwca.num_queues)
563 set_bit(index, priv->gwca.used);
564 gq = &priv->gwca.queues[index];
574 clear_bit(gq->index, priv->gwca.used);
667 iowrite32(lower_32_bits(priv->gwca.linkfix_table_dma), priv->addr + GWDCBAC1);
668 iowrite32(upper_32_bits(priv->gwca.linkfix_table_dma), priv->addr + GWDCBAC0);
669 iowrite32(lower_32_bits(priv->gwca.ts_queue.ring_dma), priv->addr + GWTDCAC10);
670 iowrite32(upper_32_bits(priv->gwca.ts_queue.ring_dma), priv->addr + GWTDCAC00);
943 for (i = 0; i < priv->gwca.num_queues; i++) {
944 gq = &priv->gwca.queues[i];
1003 struct rswitch_gwca_queue *gq = &priv->gwca.ts_queue;
1018 list_for_each_entry_safe(ts_info, ts_info2, &priv->gwca.ts_info_list, list) {
1571 list_for_each_entry_safe(ts_info, ts_info2, &rdev->priv->gwca.ts_info_list, list) {
1610 list_add_tail(&ts_info->list, &rdev->priv->gwca.ts_info_list);
1906 if (rdev->priv->gwca.speed < rdev->etha->speed)
1907 rdev->priv->gwca.speed = rdev->etha->speed;
2092 priv->gwca.index = AGENT_INDEX_GWCA;
2093 priv->gwca.num_queues = min(RSWITCH_NUM_PORTS * NUM_QUEUES_PER_NDEV,
2095 priv->gwca.queues = devm_kcalloc(&pdev->dev, priv->gwca.num_queues,
2096 sizeof(*priv->gwca.queues), GFP_KERNEL);
2097 if (!priv->gwca.queues)