/linux-master/drivers/net/netdevsim/ |
H A D | netdevsim.h | 104 struct nsim_rq *rq; member in struct:netdevsim
|
/linux-master/drivers/net/ethernet/xircom/ |
H A D | xirc2ps_cs.c | 298 static int do_ioctl(struct net_device *dev, struct ifreq *rq, int cmd); 1418 do_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) argument 1422 struct mii_ioctl_data *data = if_mii(rq); 1425 dev->name, rq->ifr_ifrn.ifrn_name, cmd,
|
/linux-master/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_main.c | 241 struct hinic_rq *rq = hinic_hwdev_get_rq(nic_dev->hwdev, i); local 243 err = hinic_init_rxq(&nic_dev->rxqs[i], rq, netdev); 264 hinic_rq_debug_rem(nic_dev->rxqs[j].rq); 287 hinic_rq_debug_rem(nic_dev->rxqs[i].rq);
|
/linux-master/drivers/ptp/ |
H A D | ptp_dte.c | 211 struct ptp_clock_request *rq, int on) 210 ptp_dte_enable(struct ptp_clock_info *ptp, struct ptp_clock_request *rq, int on) argument
|
/linux-master/include/scsi/ |
H A D | scsi_host.h | 315 bool (* dma_need_drain)(struct request *rq);
|
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc.h | 533 void intel_guc_context_ban(struct intel_context *ce, struct i915_request *rq);
|
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_main.c | 1039 spin_lock_irqsave(&hr_qp->rq.lock, flags_qp); 1040 if ((!hr_qp->ibqp.srq) && (hr_qp->rq.tail != hr_qp->rq.head)) 1042 spin_unlock_irqrestore(&hr_qp->rq.lock, flags_qp);
|
/linux-master/drivers/net/ethernet/chelsio/cxgb3/ |
H A D | sge.c | 1949 * @rq: the response queue that received the packet 1957 static inline int rx_offload(struct t3cdev *tdev, struct sge_rspq *rq, argument 1965 if (rq->polling) { 1970 rq->offload_bundles++; 1973 offload_enqueue(rq, skb); 2075 * @rq: the response queue that received the packet 2084 static void rx_eth(struct adapter *adap, struct sge_rspq *rq, argument 2088 struct sge_qset *qs = rspq_to_qset(rq); 2106 if (rq->polling) { 2701 struct sge_rspq *rq) 2700 process_responses_gts(struct adapter *adap, struct sge_rspq *rq) argument [all...] |
/linux-master/drivers/net/phy/ |
H A D | bcm-phy-ptp.c | 689 struct ptp_clock_request *rq, int on) 696 switch (rq->type) { 699 err = bcm_ptp_perout_locked(priv, &rq->perout, on); 688 bcm_ptp_enable(struct ptp_clock_info *info, struct ptp_clock_request *rq, int on) argument
|
/linux-master/drivers/net/ethernet/ti/ |
H A D | cpts.c | 300 struct ptp_clock_request *rq, int on) 304 switch (rq->type) { 306 return cpts_extts_enable(cpts, rq->extts.index, on); 299 cpts_ptp_enable(struct ptp_clock_info *ptp, struct ptp_clock_request *rq, int on) argument
|
/linux-master/drivers/net/slip/ |
H A D | slip.c | 112 static int sl_siocdevprivate(struct net_device *dev, struct ifreq *rq, void __user *data, int cmd); 1186 static int sl_siocdevprivate(struct net_device *dev, struct ifreq *rq, argument 1190 unsigned long *p = (unsigned long *)&rq->ifr_ifru;
|
/linux-master/drivers/isdn/mISDN/ |
H A D | tei.c | 788 struct channel_req rq; local 797 rq.protocol = ISDN_P_NT_E1; 799 rq.protocol = ISDN_P_NT_S0; 836 rq.adr.dev = mgr->ch.st->dev->id; 837 id = mgr->ch.st->own.ctrl(&mgr->ch.st->own, OPEN_CHANNEL, &rq);
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
H A D | dr_send.c | 268 dr_qp->rq.pc = 0; 269 dr_qp->rq.cc = 0; 270 dr_qp->rq.wqe_cnt = 256; 277 MLX5_SET(qpc, temp_qpc, log_rq_size, ilog2(dr_qp->rq.wqe_cnt)); 317 MLX5_SET(qpc, qpc, log_rq_size, ilog2(dr_qp->rq.wqe_cnt));
|
/linux-master/drivers/net/ethernet/cavium/thunder/ |
H A D | nicvf_main.c | 530 struct rcv_queue *rq, struct sk_buff **skb) 551 &rq->xdp_rxq); 773 struct snd_queue *sq, struct rcv_queue *rq) 799 if (nicvf_xdp_rx(snic, nic->xdp_prog, cqe_rx, sq, rq, &skb)) 857 struct rcv_queue *rq = &qs->rq[cq_idx]; local 888 nicvf_rcv_pkt_handler(netdev, napi, cq_desc, sq, rq); 528 nicvf_xdp_rx(struct nicvf *nic, struct bpf_prog *prog, struct cqe_rx_t *cqe_rx, struct snd_queue *sq, struct rcv_queue *rq, struct sk_buff **skb) argument 770 nicvf_rcv_pkt_handler(struct net_device *netdev, struct napi_struct *napi, struct cqe_rx_t *cqe_rx, struct snd_queue *sq, struct rcv_queue *rq) argument
|
/linux-master/drivers/block/ |
H A D | nbd.c | 888 struct request *rq; local 914 rq = blk_mq_rq_from_pdu(cmd); 915 if (likely(!blk_should_fake_timeout(rq->q))) { 923 blk_mq_complete_request(rq); 1095 struct nbd_cmd *cmd = blk_mq_rq_to_pdu(bd->rq); 1783 static int nbd_init_request(struct blk_mq_tag_set *set, struct request *rq, argument 1786 struct nbd_cmd *cmd = blk_mq_rq_to_pdu(rq);
|
/linux-master/drivers/scsi/mpi3mr/ |
H A D | mpi3mr_os.c | 443 * @rq: Block request 450 static bool mpi3mr_print_scmd(struct request *rq, void *data) argument 453 struct scsi_cmnd *scmd = blk_mq_rq_to_pdu(rq); 472 * @rq: Block request 481 static bool mpi3mr_flush_scmd(struct request *rq, void *data) argument 484 struct scsi_cmnd *scmd = blk_mq_rq_to_pdu(rq); 509 * @rq: Block request 520 static bool mpi3mr_count_dev_pending(struct request *rq, void *data) argument 524 struct scsi_cmnd *scmd = blk_mq_rq_to_pdu(rq); 541 * @rq 552 mpi3mr_count_tgt_pending(struct request *rq, void *data) argument 4729 struct request *rq = scsi_cmd_to_rq(scmd); local [all...] |
/linux-master/drivers/scsi/elx/libefc_sli/ |
H A D | sli4.c | 371 struct sli4_rqst_rq_create_v1 *rq; local 376 rq = sli_config_cmd_init(sli4, buf, SLI4_CFG_PYLD_LENGTH(rq_create_v1), 378 if (!rq) 381 sli_cmd_fill_hdr(&rq->hdr, SLI4_OPC_RQ_CREATE, SLI4_SUBSYSTEM_FC, 384 rq->dim_dfd_dnb |= SLI4_RQ_CREATE_V1_DNB; 388 rq->num_pages = cpu_to_le16(num_pages); 399 rq->rqe_count = cpu_to_le16(qmem->size / SLI4_RQE_SIZE); 401 rq->rqe_size_byte |= SLI4_RQE_SIZE_8; 403 rq->page_size = SLI4_RQ_PAGE_SIZE_4096; 412 rq [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_context.c | 1357 struct i915_request *rq; local 1366 * rq->link is only SLAB_TYPESAFE_BY_RCU, we need to hold a reference 1371 list_for_each_entry_reverse(rq, &ce->timeline->requests, link) { 1375 if (!i915_request_get_rcu(rq)) 1380 if (likely(rcu_access_pointer(rq->timeline) == ce->timeline)) 1381 found = i915_request_active_engine(rq, &engine); 1383 i915_request_put(rq);
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_stats.c | 467 mlx5e_stats_grp_sw_update_stats_rq_stats(s, &priv->ptp_stats.rq); 496 struct mlx5e_rq_stats *rq_stats = c->rq.stats; 497 struct page_pool *pool = c->rq.page_pool; 538 mlx5e_stats_grp_sw_update_stats_rq_stats(s, &channel_stats->rq); 2423 MLX5E_READ_CTR64_CPU(&priv->ptp_stats.rq, 2493 &priv->channel_stats[i]->rq,
|
/linux-master/drivers/net/ethernet/8390/ |
H A D | axnet_cs.c | 81 static int axnet_ioctl(struct net_device *dev, struct ifreq *rq, int cmd); 608 static int axnet_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) argument 611 struct mii_ioctl_data *data = if_mii(rq);
|
/linux-master/drivers/tty/ipwireless/ |
H A D | hardware.c | 1734 struct ipw_rx_packet *rp, *rq; local 1750 list_for_each_entry_safe(rp, rq, &hw->rx_queue, queue) { 1755 list_for_each_entry_safe(rp, rq, &hw->rx_pool, queue) {
|
/linux-master/drivers/net/ethernet/sgi/ |
H A D | ioc3-eth.c | 104 static int ioc3_ioctl(struct net_device *dev, struct ifreq *rq, int cmd); 1222 static int ioc3_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) argument 1228 rc = generic_mii_ioctl(&ip->mii, if_mii(rq), cmd, NULL);
|
/linux-master/drivers/memstick/core/ |
H A D | mspro_block.c | 768 blk_mq_start_request(bd->rq); 772 msb->block_req = bd->rq; 773 blk_mq_start_request(bd->rq);
|
/linux-master/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | sge.c | 4850 void free_rspq_fl(struct adapter *adap, struct sge_rspq *rq, argument 4856 adap->sge.ingr_map[rq->cntxt_id - adap->sge.ingr_start] = NULL; 4858 rq->cntxt_id, fl_id, 0xffff); 4859 dma_free_coherent(adap->pdev_dev, (rq->size + 1) * rq->iqe_len, 4860 rq->desc, rq->phys_addr); 4861 netif_napi_del(&rq->napi); 4862 rq->netdev = NULL; 4863 rq [all...] |
/linux-master/drivers/infiniband/hw/mlx5/ |
H A D | mlx5_ib.h | 444 struct mlx5_ib_wq *rq; member in struct:mlx5_ib_rq 464 struct mlx5_ib_rq rq; member in struct:mlx5_ib_raw_packet_qp 506 struct mlx5_ib_wq rq; member in struct:mlx5_ib_qp
|