/linux-master/arch/arm64/kvm/hyp/nvhe/ |
H A D | mem_protect.c | 574 const struct pkvm_mem_transition tx; member in struct:pkvm_mem_share 579 const struct pkvm_mem_transition tx; member in struct:pkvm_mem_donation 639 const struct pkvm_mem_transition *tx) 641 u64 size = tx->nr_pages * PAGE_SIZE; 642 u64 addr = tx->initiator.addr; 644 *completer_addr = tx->initiator.host.completer_addr; 649 const struct pkvm_mem_transition *tx) 651 u64 size = tx->nr_pages * PAGE_SIZE; 652 u64 addr = tx->initiator.addr; 654 *completer_addr = tx 638 host_request_owned_transition(u64 *completer_addr, const struct pkvm_mem_transition *tx) argument 648 host_request_unshare(u64 *completer_addr, const struct pkvm_mem_transition *tx) argument 658 host_initiate_share(u64 *completer_addr, const struct pkvm_mem_transition *tx) argument 668 host_initiate_unshare(u64 *completer_addr, const struct pkvm_mem_transition *tx) argument 678 host_initiate_donation(u64 *completer_addr, const struct pkvm_mem_transition *tx) argument 688 __host_ack_skip_pgtable_check(const struct pkvm_mem_transition *tx) argument 694 __host_ack_transition(u64 addr, const struct pkvm_mem_transition *tx, enum pkvm_page_state state) argument 705 host_ack_donation(u64 addr, const struct pkvm_mem_transition *tx) argument 710 host_complete_donation(u64 addr, const struct pkvm_mem_transition *tx) argument 738 hyp_request_donation(u64 *completer_addr, const struct pkvm_mem_transition *tx) argument 748 hyp_initiate_donation(u64 *completer_addr, const struct pkvm_mem_transition *tx) argument 759 __hyp_ack_skip_pgtable_check(const struct pkvm_mem_transition *tx) argument 765 hyp_ack_share(u64 addr, const struct pkvm_mem_transition *tx, enum kvm_pgtable_prot perms) argument 779 hyp_ack_unshare(u64 addr, const struct pkvm_mem_transition *tx) argument 793 hyp_ack_donation(u64 addr, const struct pkvm_mem_transition *tx) argument 803 hyp_complete_share(u64 addr, const struct pkvm_mem_transition *tx, enum kvm_pgtable_prot perms) argument 813 hyp_complete_unshare(u64 addr, const struct pkvm_mem_transition *tx) argument 821 hyp_complete_donation(u64 addr, const struct pkvm_mem_transition *tx) argument 832 const struct pkvm_mem_transition *tx = &share->tx; local 867 const struct pkvm_mem_transition *tx = &share->tx; local 922 const struct pkvm_mem_transition *tx = &share->tx; local 954 const struct pkvm_mem_transition *tx = &share->tx; local 1006 const struct pkvm_mem_transition *tx = &donation->tx; local 1040 const struct pkvm_mem_transition *tx = &donation->tx; local [all...] |
/linux-master/include/uapi/linux/ |
H A D | atm_eni.h | 15 int tx,rx; /* values are in percent and must be > 100 */ member in struct:eni_multipliers
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | sdma.h | 190 * The tx request, once initialized, is manipulated with calls to 199 * in the tx. Memory locations added with sdma_txadd_page() 201 * to the tx and nmapped as part of the progress processing in the 205 * tx. An example of a use case would be a pre-allocated 212 * a tx to the ring after the appropriate number of 219 * long as the tx isn't in flight. 440 struct sdma_txreq *tx, 448 * @tx: tx request to initialize 479 * being submitted. The callback will be provided this tx, 498 sdma_txinit_ahg( struct sdma_txreq *tx, u16 flags, u16 tlen, u8 ahg_entry, u8 num_ahg, u32 *ahg, u8 ahg_hlen, void (*cb)(struct sdma_txreq *, int)) argument 566 sdma_txinit( struct sdma_txreq *tx, u16 flags, u16 tlen, void (*cb)(struct sdma_txreq *, int)) argument 594 make_tx_sdma_desc( struct sdma_txreq *tx, int type, dma_addr_t addr, size_t len, void *pinning_ctx, void (*ctx_get)(void *), void (*ctx_put)(void *)) argument 632 sdma_txclean(struct hfi1_devdata *dd, struct sdma_txreq *tx) argument 639 _sdma_close_tx(struct hfi1_devdata *dd, struct sdma_txreq *tx) argument 651 _sdma_txadd_daddr( struct hfi1_devdata *dd, int type, struct sdma_txreq *tx, dma_addr_t addr, u16 len, void *pinning_ctx, void (*ctx_get)(void *), void (*ctx_put)(void *)) argument 708 sdma_txadd_page( struct hfi1_devdata *dd, struct sdma_txreq *tx, struct page *page, unsigned long offset, u16 len, void *pinning_ctx, void (*ctx_get)(void *), void (*ctx_put)(void *)) argument 760 sdma_txadd_daddr( struct hfi1_devdata *dd, struct sdma_txreq *tx, dma_addr_t addr, u16 len) argument 795 sdma_txadd_kvaddr( struct hfi1_devdata *dd, struct sdma_txreq *tx, void *kvaddr, u16 len) argument 880 sdma_progress(struct sdma_engine *sde, unsigned seq, struct sdma_txreq *tx) argument [all...] |
H A D | verbs_txreq.c | 13 void hfi1_put_txreq(struct verbs_txreq *tx) argument 21 qp = tx->qp; 24 if (tx->mr) 25 rvt_put_mr(tx->mr); 27 sdma_txclean(dd_from_dev(dev), &tx->txreq); 30 kmem_cache_free(dev->verbs_txreq_cache, tx); 55 struct verbs_txreq *tx = NULL; variable in typeref:struct:verbs_txreq 61 tx = kmem_cache_alloc(dev->verbs_txreq_cache, VERBS_TXREQ_GFP); 62 if (tx) 77 return tx; [all...] |
H A D | ipoib_tx.c | 101 * The tx queue len can be adjusted upward while the interface is 103 * The tx queue len can be large enough to overflow the txreq_ring. 114 static void hfi1_ipoib_free_tx(struct ipoib_txreq *tx, int budget) argument 116 struct hfi1_ipoib_dev_priv *priv = tx->txq->priv; 118 if (likely(!tx->sdma_status)) { 119 dev_sw_netstats_tx_add(priv->netdev, 1, tx->skb->len); 124 __func__, tx->sdma_status, 125 le64_to_cpu(tx->sdma_hdr->pbc), tx->txq->q_idx, 126 tx 138 struct ipoib_txreq *tx; local 162 struct ipoib_txreq *tx = hfi1_txreq_from_idx(tx_ring, head); local 190 struct ipoib_txreq *tx = container_of(txreq, struct ipoib_txreq, txreq); local 199 hfi1_ipoib_build_ulp_payload(struct ipoib_txreq *tx, struct ipoib_txparms *txp) argument 230 hfi1_ipoib_build_tx_desc(struct ipoib_txreq *tx, struct ipoib_txparms *txp) argument 256 hfi1_ipoib_build_ib_tx_headers(struct ipoib_txreq *tx, struct ipoib_txparms *txp) argument 354 struct ipoib_txreq *tx; local 439 hfi1_ipoib_submit_tx(struct hfi1_ipoib_txq *txq, struct ipoib_txreq *tx) argument 462 struct ipoib_txreq *tx; local 513 struct ipoib_txreq *tx; local 706 struct ipoib_txreq *tx; local 776 struct ipoib_txreq *tx = local [all...] |
/linux-master/crypto/async_tx/ |
H A D | async_raid6_recov.c | 36 struct dma_async_tx_descriptor *tx; local 54 tx = dma->device_prep_dma_pq(chan, pq, unmap->addr, 2, coef, 56 if (tx) { 57 dma_set_unmap(tx, unmap); 58 async_tx_submit(chan, tx, submit); 60 return tx; 104 struct dma_async_tx_descriptor *tx; local 122 tx = dma->device_prep_dma_pq(chan, dma_dest, unmap->addr, 125 if (tx) { 126 dma_set_unmap(tx, unma 157 struct dma_async_tx_descriptor *tx = NULL; local 207 struct dma_async_tx_descriptor *tx = NULL; local 298 struct dma_async_tx_descriptor *tx = NULL; local 475 struct dma_async_tx_descriptor *tx = NULL; local [all...] |
H A D | async_memcpy.c | 39 struct dma_async_tx_descriptor *tx = NULL; local 61 tx = device->device_prep_dma_memcpy(chan, unmap->addr[1], 66 if (tx) { 69 dma_set_unmap(tx, unmap); 70 async_tx_submit(chan, tx, submit); 91 return tx;
|
/linux-master/drivers/i2c/busses/ |
H A D | i2c-dln2.c | 58 } tx; local 60 tx.port = dln2->port; 67 return dln2_transfer_tx(dln2->pdev, cmd, &tx, sizeof(tx)); 81 } __packed *tx = dln2->buf; local 84 BUILD_BUG_ON(sizeof(*tx) > DLN2_I2C_BUF_SIZE); 86 tx->port = dln2->port; 87 tx->addr = addr; 88 tx->mem_addr_len = 0; 89 tx 111 } __packed tx; local [all...] |
/linux-master/drivers/dma/ |
H A D | virt-dma.c | 14 static struct virt_dma_desc *to_virt_desc(struct dma_async_tx_descriptor *tx) argument 16 return container_of(tx, struct virt_dma_desc, tx); 19 dma_cookie_t vchan_tx_submit(struct dma_async_tx_descriptor *tx) argument 21 struct virt_dma_chan *vc = to_virt_chan(tx->chan); 22 struct virt_dma_desc *vd = to_virt_desc(tx); 27 cookie = dma_cookie_assign(tx); 41 * @tx: the transfer 49 int vchan_tx_desc_free(struct dma_async_tx_descriptor *tx) argument 51 struct virt_dma_chan *vc = to_virt_chan(tx [all...] |
H A D | virt-dma.h | 16 struct dma_async_tx_descriptor tx; member in struct:virt_dma_desc 61 dma_async_tx_descriptor_init(&vd->tx, &vc->chan); 62 vd->tx.flags = tx_flags; 63 vd->tx.tx_submit = vchan_tx_submit; 64 vd->tx.desc_free = vchan_tx_desc_free; 73 return &vd->tx; 98 struct virt_dma_chan *vc = to_virt_chan(vd->tx.chan); 103 cookie = vd->tx.cookie; 104 dma_cookie_complete(&vd->tx); 118 struct virt_dma_chan *vc = to_virt_chan(vd->tx [all...] |
/linux-master/drivers/iio/pressure/ |
H A D | mpl115_spi.c | 19 u8 tx[4]; member in struct:mpl115_spi_buf 42 .tx_buf = buf->tx, 48 buf->tx[0] = MPL115_SPI_READ(address); 49 buf->tx[2] = MPL115_SPI_READ(address + 1); 63 .tx_buf = buf->tx, 67 buf->tx[0] = MPL115_SPI_WRITE(address); 68 buf->tx[1] = value;
|
/linux-master/drivers/net/ethernet/engleder/ |
H A D | tsnep_main.c | 277 static void tsnep_tx_ring_cleanup(struct tsnep_tx *tx) argument 279 struct device *dmadev = tx->adapter->dmadev; 282 memset(tx->entry, 0, sizeof(tx->entry)); 285 if (tx->page[i]) { 286 dma_free_coherent(dmadev, PAGE_SIZE, tx->page[i], 287 tx->page_dma[i]); 288 tx->page[i] = NULL; 289 tx->page_dma[i] = 0; 294 static int tsnep_tx_ring_create(struct tsnep_tx *tx) argument 333 tsnep_tx_init(struct tsnep_tx *tx) argument 346 tsnep_tx_enable(struct tsnep_tx *tx) argument 357 tsnep_tx_disable(struct tsnep_tx *tx, struct napi_struct *napi) argument 380 tsnep_tx_activate(struct tsnep_tx *tx, int index, int length, bool last) argument 442 tsnep_tx_desc_available(struct tsnep_tx *tx) argument 483 tsnep_tx_map(struct sk_buff *skb, struct tsnep_tx *tx, int count) argument 530 tsnep_tx_unmap(struct tsnep_tx *tx, int index, int count) argument 559 tsnep_xmit_frame_ring(struct sk_buff *skb, struct tsnep_tx *tx) argument 618 tsnep_xdp_tx_map(struct xdp_frame *xdpf, struct tsnep_tx *tx, struct skb_shared_info *shinfo, int count, u32 type) argument 674 tsnep_xdp_xmit_frame_ring(struct xdp_frame *xdpf, struct tsnep_tx *tx, u32 type) argument 717 tsnep_xdp_xmit_flush(struct tsnep_tx *tx) argument 722 tsnep_xdp_xmit_back(struct tsnep_adapter *adapter, struct xdp_buff *xdp, struct netdev_queue *tx_nq, struct tsnep_tx *tx, bool zc) argument 753 tsnep_xdp_tx_map_zc(struct xdp_desc *xdpd, struct tsnep_tx *tx) argument 772 tsnep_xdp_xmit_frame_ring_zc(struct xdp_desc *xdpd, struct tsnep_tx *tx) argument 783 tsnep_xdp_xmit_zc(struct tsnep_tx *tx) argument 811 tsnep_tx_poll(struct tsnep_tx *tx, int napi_budget) argument 904 tsnep_tx_pending(struct tsnep_tx *tx) argument 926 tsnep_tx_open(struct tsnep_tx *tx) argument 939 tsnep_tx_close(struct tsnep_tx *tx) argument 1279 tsnep_xdp_run_prog(struct tsnep_rx *rx, struct bpf_prog *prog, struct xdp_buff *xdp, int *status, struct netdev_queue *tx_nq, struct tsnep_tx *tx) argument 1323 tsnep_xdp_run_prog_zc(struct tsnep_rx *rx, struct bpf_prog *prog, struct xdp_buff *xdp, int *status, struct netdev_queue *tx_nq, struct tsnep_tx *tx) argument 1361 tsnep_finalize_xdp(struct tsnep_adapter *adapter, int status, struct netdev_queue *tx_nq, struct tsnep_tx *tx) argument 1436 struct tsnep_tx *tx; local 1539 struct tsnep_tx *tx; local 1912 struct tsnep_tx *tx = queue->tx; local 2308 struct tsnep_tx *tx; local [all...] |
/linux-master/drivers/iio/gyro/ |
H A D | adxrs450.c | 69 * @buf_lock: mutex to protect tx and rx 70 * @tx: transmit buffer 76 __be32 tx __aligned(IIO_DMA_MINALIGN); 93 u32 tx; local 97 .tx_buf = &st->tx, 99 .len = sizeof(st->tx), 109 tx = ADXRS450_READ_DATA | (reg_address << 17); 111 if (!(hweight32(tx) & 1)) 112 tx |= ADXRS450_P; 114 st->tx 141 u32 tx; local 208 u32 tx; local [all...] |
/linux-master/drivers/staging/gdm724x/ |
H A D | gdm_usb.c | 149 static struct usb_tx_sdu *get_tx_sdu_struct(struct tx_cxt *tx, int *no_spc) argument 153 if (list_empty(&tx->free_list)) 156 t_sdu = list_entry(tx->free_list.next, struct usb_tx_sdu, list); 159 tx->avail_count--; 161 *no_spc = list_empty(&tx->free_list) ? 1 : 0; 166 static void put_tx_struct(struct tx_cxt *tx, struct usb_tx_sdu *t_sdu) argument 168 list_add_tail(&t_sdu->list, &tx->free_list); 169 tx->avail_count++; 251 struct tx_cxt *tx = &udev->tx; local 304 struct tx_cxt *tx = &udev->tx; local 546 struct tx_cxt *tx = t->tx; local 594 struct tx_cxt *tx = &udev->tx; local 643 struct tx_cxt *tx = &udev->tx; local 710 struct tx_cxt *tx = &udev->tx; local 766 struct tx_cxt *tx = &udev->tx; local 943 struct tx_cxt *tx; local [all...] |
/linux-master/net/ipv4/ |
H A D | tcp_rate.c | 65 TCP_SKB_CB(skb)->tx.first_tx_mstamp = tp->first_tx_mstamp; 66 TCP_SKB_CB(skb)->tx.delivered_mstamp = tp->delivered_mstamp; 67 TCP_SKB_CB(skb)->tx.delivered = tp->delivered; 68 TCP_SKB_CB(skb)->tx.delivered_ce = tp->delivered_ce; 69 TCP_SKB_CB(skb)->tx.is_app_limited = tp->app_limited ? 1 : 0; 87 if (!scb->tx.delivered_mstamp) 94 rs->prior_delivered_ce = scb->tx.delivered_ce; 95 rs->prior_delivered = scb->tx.delivered; 96 rs->prior_mstamp = scb->tx.delivered_mstamp; 97 rs->is_app_limited = scb->tx [all...] |
/linux-master/drivers/peci/ |
H A D | request.c | 155 if (WARN_ON(req->tx.len == 0)) 169 req->tx.buf[1] |= PECI_RETRY_BIT; 211 req->tx.len = tx_len; 237 req->tx.buf[0] = PECI_GET_DIB_CMD; 258 req->tx.buf[0] = PECI_GET_TEMP_CMD; 280 req->tx.buf[0] = PECI_RDPKGCFG_CMD; 281 req->tx.buf[1] = 0; 282 req->tx.buf[2] = index; 283 put_unaligned_le16(param, &req->tx.buf[3]); 313 req->tx [all...] |
/linux-master/drivers/media/dvb-frontends/cxd2880/ |
H A D | cxd2880_spi_device.c | 19 struct spi_transfer tx; local 27 memset(&tx, 0, sizeof(tx)); 28 tx.tx_buf = data; 29 tx.len = size; 32 spi_message_add_tail(&tx, &msg);
|
/linux-master/drivers/net/ethernet/microchip/lan966x/ |
H A D | lan966x_fdma.c | 233 static void lan966x_fdma_tx_add_dcb(struct lan966x_tx *tx, argument 240 static int lan966x_fdma_tx_alloc(struct lan966x_tx *tx) argument 242 struct lan966x *lan966x = tx->lan966x; 248 tx->dcbs_buf = kcalloc(FDMA_DCB_MAX, sizeof(struct lan966x_tx_dcb_buf), 250 if (!tx->dcbs_buf) 256 tx->dcbs = dma_alloc_coherent(lan966x->dev, size, &tx->dma, GFP_KERNEL); 257 if (!tx->dcbs) 262 dcb = &tx->dcbs[i]; 270 lan966x_fdma_tx_add_dcb(tx, dc 280 lan966x_fdma_tx_free(struct lan966x_tx *tx) argument 292 lan966x_fdma_tx_activate(struct lan966x_tx *tx) argument 330 lan966x_fdma_tx_disable(struct lan966x_tx *tx) argument 352 lan966x_fdma_tx_reload(struct lan966x_tx *tx) argument 393 struct lan966x_tx *tx = &lan966x->tx; local 650 lan966x_fdma_get_next_dcb(struct lan966x_tx *tx) argument 664 lan966x_fdma_tx_setup_dcb(struct lan966x_tx *tx, int next_to_use, int len, dma_addr_t dma_addr) argument 683 lan966x_fdma_tx_start(struct lan966x_tx *tx, int next_to_use) argument 709 struct lan966x_tx *tx = &lan966x->tx; local 805 struct lan966x_tx *tx = &lan966x->tx; local [all...] |
/linux-master/sound/soc/mediatek/common/ |
H A D | mtk-btcvsd.c | 130 spinlock_t tx_lock; /* spinlock for bt tx stream control */ 135 struct mtk_btcvsd_snd_stream *tx; member in struct:mtk_btcvsd_snd 209 dev_dbg(bt->dev, "%s(), stream %d, state %d, tx->state %d, rx->state %d, irq_disabled %d\n", 212 bt->tx->state, bt->rx->state, bt->irq_disabled); 216 if (bt->tx->state == BT_SCO_STATE_IDLE && 234 memset(bt->tx, 0, sizeof(*bt->tx)); 237 bt->tx->packet_size = BTCVSD_TX_PACKET_SIZE; 238 bt->tx->buf_size = BTCVSD_TX_BUF_SIZE; 239 bt->tx [all...] |
/linux-master/drivers/misc/echo/ |
H A D | oslec.h | 65 * @tx: The transmitted audio sample. 70 int16_t oslec_update(struct oslec_state *ec, int16_t tx, int16_t rx); 73 * oslec_hpf_tx: Process to high pass filter the tx signal. 75 * @tx: The transmitted auio sample. 79 int16_t oslec_hpf_tx(struct oslec_state *ec, int16_t tx);
|
/linux-master/drivers/net/phy/ |
H A D | linkmode.c | 49 * @tx: boolean from ethtool struct ethtool_pauseparam tx_pause member 53 * capabilities of provided in @tx and @rx. 56 * tx rx Pause AsymDir 62 * Note: this translation from ethtool tx/rx notation to the advertisement 65 * For tx=0 rx=1, meaning transmit is unsupported, receive is supported: 72 * For tx=1 rx=1, meaning we have the capability to transmit and receive 77 * 1 0 0 1 Disabled - but since we do support tx and rx, 81 * rx=1 tx=0 gives Pause+AsymDir advertisement, but we may end up 82 * resolving to tx+rx pause or only rx pause depending on 84 * rx=0 tx 89 linkmode_set_pause(unsigned long *advertisement, bool tx, bool rx) argument [all...] |
/linux-master/sound/soc/fsl/ |
H A D | fsl_sai.h | 81 #define FSL_SAI_xCSR(tx, ofs) (tx ? FSL_SAI_TCSR(ofs) : FSL_SAI_RCSR(ofs)) 82 #define FSL_SAI_xCR1(tx, ofs) (tx ? FSL_SAI_TCR1(ofs) : FSL_SAI_RCR1(ofs)) 83 #define FSL_SAI_xCR2(tx, ofs) (tx ? FSL_SAI_TCR2(ofs) : FSL_SAI_RCR2(ofs)) 84 #define FSL_SAI_xCR3(tx, ofs) (tx ? FSL_SAI_TCR3(ofs) : FSL_SAI_RCR3(ofs)) 85 #define FSL_SAI_xCR4(tx, ofs) (tx [all...] |
/linux-master/tools/spi/ |
H A D | spidev_test.c | 118 static void transfer(int fd, uint8_t const *tx, uint8_t const *rx, size_t len) argument 123 .tx_buf = (unsigned long)tx, 155 hex_dump(tx, len, 32, "TX"); 205 " -v --verbose Verbose (show tx buffer)\n"); 336 uint8_t *tx; local 339 tx = malloc(size); 340 if (!tx) 341 pabort("can't allocate tx buffer"); 347 size = unescape((char *)tx, str, size); 348 transfer(fd, tx, r 358 uint8_t *tx; local 405 uint8_t *tx; local [all...] |
/linux-master/drivers/net/ethernet/microchip/sparx5/ |
H A D | sparx5_fdma.c | 97 static void sparx5_fdma_tx_add_dcb(struct sparx5_tx *tx, argument 159 static void sparx5_fdma_tx_activate(struct sparx5 *sparx5, struct sparx5_tx *tx) argument 162 spx5_wr(((u64)tx->dma) & GENMASK(31, 0), sparx5, 163 FDMA_DCB_LLP(tx->channel_id)); 164 spx5_wr(((u64)tx->dma) >> 32, sparx5, FDMA_DCB_LLP1(tx->channel_id)); 170 sparx5, FDMA_CH_CFG(tx->channel_id)); 177 spx5_wr(BIT(tx->channel_id), sparx5, FDMA_CH_ACTIVATE); 180 static void sparx5_fdma_tx_deactivate(struct sparx5 *sparx5, struct sparx5_tx *tx) argument 183 spx5_rmw(0, BIT(tx 193 sparx5_fdma_tx_reload(struct sparx5 *sparx5, struct sparx5_tx *tx) argument 295 sparx5_fdma_next_dcb(struct sparx5_tx *tx, struct sparx5_tx_dcb_hw *dcb) argument 312 struct sparx5_tx *tx = &sparx5->tx; local 393 struct sparx5_tx *tx = &sparx5->tx; local 456 sparx5_fdma_tx_init(struct sparx5 *sparx5, struct sparx5_tx *tx, int channel) argument [all...] |
/linux-master/drivers/net/can/spi/mcp251xfd/ |
H A D | mcp251xfd-ram.c | 71 num_tx = config->tx.def[fd_mode]; 72 num_tx = can_ram_rounddown_pow_of_two(config, &config->tx, 0, num_tx); 75 ram_free -= config->tx.size[fd_mode] * num_tx; 85 ram_free -= config->tx.size[fd_mode] * config->tx.min; 90 num_tx = ram_free / config->tx.size[fd_mode]; 93 layout->max_tx = can_ram_rounddown_pow_of_two(config, &config->tx, 0, num_tx); 121 num_tx = ram_free / config->tx.size[fd_mode]; 123 num_tx = can_ram_rounddown_pow_of_two(config, &config->tx, 0, num_tx); 135 (u32)config->tx [all...] |