/linux-master/drivers/scsi/ |
H A D | scsi_debugfs.h | 2 struct request; 5 void scsi_show_rq(struct seq_file *m, struct request *rq);
|
/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | mock_request.c | 33 struct i915_request *request; local 36 request = intel_context_create_request(ce); 37 if (IS_ERR(request)) 40 request->mock.delay = delay; 41 return request; 44 bool mock_cancel_request(struct i915_request *request) argument 47 container_of(request->engine, typeof(*engine), base); 51 was_queued = !list_empty(&request->mock.link); 52 list_del_init(&request->mock.link); 56 i915_request_unsubmit(request); [all...] |
H A D | mock_request.h | 35 bool mock_cancel_request(struct i915_request *request);
|
H A D | i915_perf_selftests.h | 19 selftest(request, i915_request_perf_selftests)
|
/linux-master/drivers/s390/char/ |
H A D | tape_std.c | 38 struct tape_request * request = from_timer(request, t, timer); local 39 struct tape_device * device = request->device; 46 rc = tape_cancel_io(device, request); 56 struct tape_request *request; local 58 request = tape_alloc_request(2, 11); 59 if (IS_ERR(request)) 60 return PTR_ERR(request); 62 request->op = TO_ASSIGN; 63 tape_ccw_cc(request 95 struct tape_request *request; local 126 struct tape_request *request; local 156 struct tape_request *request; local 268 struct tape_request *request; local 292 struct tape_request *request; local 324 struct tape_request *request; local 355 struct tape_request *request; local 380 struct tape_request *request; local 404 struct tape_request *request; local 435 struct tape_request *request; local 465 struct tape_request *request; local 488 struct tape_request *request; local 509 struct tape_request *request; local 560 struct tape_request *request; local 583 struct tape_request *request; local 617 struct tape_request *request; local 644 struct tape_request *request; local 667 tape_std_read_backward(struct tape_device *device, struct tape_request *request) argument 688 struct tape_request *request; local [all...] |
H A D | tape_34xx.c | 56 static void __tape_34xx_medium_sense(struct tape_request *request) argument 58 struct tape_device *device = request->device; 61 if (request->rc == 0) { 62 sense = request->cpdata; 81 request->rc); 82 tape_free_request(request); 87 struct tape_request *request; local 90 request = tape_alloc_request(1, 32); 91 if (IS_ERR(request)) { 93 return PTR_ERR(request); 105 struct tape_request *request; local 175 tape_34xx_done(struct tape_request *request) argument 195 tape_34xx_erp_failed(struct tape_request *request, int rc) argument 203 tape_34xx_erp_succeeded(struct tape_request *request) argument 211 tape_34xx_erp_retry(struct tape_request *request) argument 241 tape_34xx_erp_read_opposite(struct tape_device *device, struct tape_request *request) argument 262 tape_34xx_erp_bug(struct tape_device *device, struct tape_request *request, struct irb *irb, int no) argument 278 tape_34xx_erp_overrun(struct tape_device *device, struct tape_request *request, struct irb *irb) argument 293 tape_34xx_erp_sequence(struct tape_device *device, struct tape_request *request, struct irb *irb) argument 317 tape_34xx_unit_check(struct tape_device *device, struct tape_request *request, struct irb *irb) argument 832 tape_34xx_irq(struct tape_device *device, struct tape_request *request, struct irb *irb) argument 1089 struct tape_request *request; local [all...] |
H A D | tape_3590.c | 204 struct tape_request *request; local 213 request = tape_alloc_request(2, sizeof(*order)); 214 if (IS_ERR(request)) { 215 rc = PTR_ERR(request); 218 order = request->cpdata; 222 request->op = TO_KEKL_QUERY; 223 tape_ccw_cc(request->cpaddr, PERF_SUBSYS_FUNC, sizeof(*order), order); 224 tape_ccw_end(request->cpaddr + 1, READ_SS_DATA, sizeof(*int_kekls), 226 rc = tape_do_io(device, request); 233 tape_free_request(request); 277 struct tape_request *request; local 329 struct tape_request *request; local 357 struct tape_request *request; local 367 struct tape_request *request; local 379 struct tape_request *request; local 405 struct tape_request *request; local 415 struct tape_request *request; local 495 struct tape_request *request; local 507 struct tape_request *request; local 538 struct tape_request *request; local 557 tape_3590_read_opposite(struct tape_device *device, struct tape_request *request) argument 595 struct tape_request *request; local 706 tape_3590_done(struct tape_device *device, struct tape_request *request) argument 764 tape_3590_erp_succeeded(struct tape_device *device, struct tape_request *request) argument 775 tape_3590_erp_failed(struct tape_device *device, struct tape_request *request, struct irb *irb, int rc) argument 788 tape_3590_erp_retry(struct tape_device *device, struct tape_request *request, struct irb *irb) argument 823 tape_3590_erp_basic(struct tape_device *device, struct tape_request *request, struct irb *irb, int rc) argument 849 tape_3590_erp_read_buf_log(struct tape_device *device, struct tape_request *request, struct irb *irb) argument 863 tape_3590_erp_swap(struct tape_device *device, struct tape_request *request, struct irb *irb) argument 881 tape_3590_erp_long_busy(struct tape_device *device, struct tape_request *request, struct irb *irb) argument 892 tape_3590_erp_special_interrupt(struct tape_device *device, struct tape_request *request, struct irb *irb) argument 902 tape_3590_erp_read_alternate(struct tape_device *device, struct tape_request *request, struct irb *irb) argument 932 tape_3590_erp_read_opposite(struct tape_device *device, struct tape_request *request, struct irb *irb) argument 1301 tape_3590_crypt_error(struct tape_device *device, struct tape_request *request, struct irb *irb) argument 1330 tape_3590_unit_check(struct tape_device *device, struct tape_request *request, struct irb *irb) argument 1451 tape_3590_irq(struct tape_device *device, struct tape_request *request, struct irb *irb) argument 1499 struct tape_request *request; local [all...] |
H A D | tape_core.c | 292 __tape_cancel_io(struct tape_device *device, struct tape_request *request) argument 298 if (request->callback == NULL) 303 rc = ccw_device_clear(device->cdev, (long) request); 307 request->status = TAPE_REQUEST_DONE; 310 request->status = TAPE_REQUEST_CANCEL; 434 * request. We may prevent this by returning an error. 592 struct tape_request * request; local 596 request = list_entry(l, struct tape_request, list); 597 if (request->status == TAPE_REQUEST_IN_IO) 598 request 676 struct tape_request *request; local 717 tape_free_request(struct tape_request * request) argument 729 __tape_start_io(struct tape_device *device, struct tape_request *request) argument 758 struct tape_request *request; local 825 struct tape_request *request; local 837 __tape_end_request( struct tape_device * device, struct tape_request * request, int rc) argument 864 tape_dump_sense_dbf(struct tape_device *device, struct tape_request *request, struct irb *irb) argument 890 __tape_start_request(struct tape_device *device, struct tape_request *request) argument 936 tape_do_io_async(struct tape_device *device, struct tape_request *request) argument 955 __tape_wake_up(struct tape_request *request, void *data) argument 962 tape_do_io(struct tape_device *device, struct tape_request *request) argument 987 __tape_wake_up_interruptible(struct tape_request *request, void *data) argument 994 tape_do_io_interruptible(struct tape_device *device, struct tape_request *request) argument 1037 tape_cancel_io(struct tape_device *device, struct tape_request *request) argument 1054 struct tape_request *request; local [all...] |
/linux-master/block/ |
H A D | elevator.h | 33 bool (*allow_merge)(struct request_queue *, struct request *, struct bio *); 35 int (*request_merge)(struct request_queue *q, struct request **, struct bio *); 36 void (*request_merged)(struct request_queue *, struct request *, enum elv_merge); 37 void (*requests_merged)(struct request_queue *, struct request *, struct request *); 39 void (*prepare_request)(struct request *); 40 void (*finish_request)(struct request *); 43 struct request *(*dispatch_request)(struct blk_mq_hw_ctx *); 45 void (*completed_request)(struct request *, u64); 46 void (*requeue_request)(struct request *); [all...] |
H A D | blk-crypto-internal.h | 31 bool bio_crypt_rq_ctx_compatible(struct request *rq, struct bio *bio); 36 static inline bool bio_crypt_ctx_back_mergeable(struct request *req, 43 static inline bool bio_crypt_ctx_front_mergeable(struct request *req, 50 static inline bool bio_crypt_ctx_merge_rq(struct request *req, 51 struct request *next) 57 static inline void blk_crypto_rq_set_defaults(struct request *rq) 63 static inline bool blk_crypto_rq_is_encrypted(struct request *rq) 68 static inline bool blk_crypto_rq_has_keyslot(struct request *rq) 96 static inline bool bio_crypt_rq_ctx_compatible(struct request *rq, 102 static inline bool bio_crypt_ctx_front_mergeable(struct request *re [all...] |
/linux-master/drivers/staging/rtl8723bs/include/ |
H A D | rtw_wifi_regd.h | 13 struct regulatory_request *request)); 14 void rtw_reg_notifier(struct wiphy *wiphy, struct regulatory_request *request);
|
/linux-master/drivers/s390/cio/ |
H A D | chsc_sch.h | 11 struct chsc_request *request; member in struct:chsc_private
|
/linux-master/drivers/mmc/core/ |
H A D | block.h | 6 struct request; 12 enum mmc_issued mmc_blk_mq_issue_rq(struct mmc_queue *mq, struct request *req); 13 void mmc_blk_mq_complete(struct request *req);
|
/linux-master/include/linux/ |
H A D | blk-mq.h | 24 typedef enum rq_end_io_ret (rq_end_io_fn)(struct request *, blk_status_t); 27 * request flags */ 32 /* request for flush sequence */ 40 /* use an I/O scheduler for this request */ 48 /* runtime pm request */ 57 /* The per-zone write lock is held for this request */ 79 struct request { struct 101 struct request *rq_next; 106 /* Time that the first bio started allocating this request. */ 109 /* Time that this request wa [all...] |
H A D | bsg-lib.h | 15 struct request; 21 typedef enum blk_eh_timer_return (bsg_timeout_fn)(struct request *); 36 /* Transport/driver specific request/reply structs */ 37 void *request; member in struct:bsg_job 51 /* DMA payloads for the request/response */ 59 struct request *bidi_rq;
|
/linux-master/sound/soc/intel/catpt/ |
H A D | messages.c | 17 struct catpt_ipc_msg request = {{0}}, reply; local 20 request.header = msg.val; 24 ret = catpt_dsp_send_msg(cdev, request, &reply); 58 struct catpt_ipc_msg request, reply; local 91 request.header = msg.val; 92 request.size = size; 93 request.data = payload; 97 ret = catpt_dsp_send_msg(cdev, request, &reply); 109 struct catpt_ipc_msg request; local 112 request 128 struct catpt_ipc_msg request; local 146 struct catpt_ipc_msg request, reply; local 166 struct catpt_ipc_msg request = {{0}}, reply; local 183 struct catpt_ipc_msg request = {{0}}; local 200 struct catpt_ipc_msg request = {{0}}; local 217 struct catpt_ipc_msg request = {{0}}; local 244 struct catpt_ipc_msg request; local 276 struct catpt_ipc_msg request; local 300 struct catpt_ipc_msg request; local [all...] |
/linux-master/drivers/platform/chrome/wilco_ec/ |
H A D | keyboard_leds.c | 54 /* Send a request, get a response, and check that the response is good. */ 56 struct wilco_keyboard_leds_msg *request, 64 msg.request_data = request; 65 msg.request_size = sizeof(*request); 81 struct wilco_keyboard_leds_msg request; local 85 memset(&request, 0, sizeof(request)); 86 request.command = WILCO_EC_COMMAND_KBBL; 87 request.subcmd = WILCO_KBBL_SUBCMD_SET_STATE; 88 request 55 send_kbbl_msg(struct wilco_ec_device *ec, struct wilco_keyboard_leds_msg *request, struct wilco_keyboard_leds_msg *response) argument 107 struct wilco_keyboard_leds_msg request; local 136 struct wilco_keyboard_leds_msg request; local [all...] |
/linux-master/drivers/greybus/ |
H A D | svc.c | 176 struct gb_svc_pwrmon_sample_get_request request; local 180 request.rail_id = rail_id; 181 request.measurement_type = measurement_type; 184 &request, sizeof(request), 213 struct gb_svc_pwrmon_intf_sample_get_request request; local 217 request.intf_id = intf_id; 218 request.measurement_type = measurement_type; 222 &request, sizeof(request), 260 struct gb_svc_intf_device_id_request request; local 271 struct gb_svc_intf_eject_request request; local 294 struct gb_svc_intf_vsys_request request; local 317 struct gb_svc_intf_refclk_request request; local 340 struct gb_svc_intf_unipro_request request; local 363 struct gb_svc_intf_activate_request request; local 389 struct gb_svc_intf_resume_request request; local 418 struct gb_svc_dme_peer_get_request request; local 452 struct gb_svc_dme_peer_set_request request; local 486 struct gb_svc_conn_create_request request; local 502 struct gb_svc_conn_destroy_request request; local 523 struct gb_svc_route_create_request request; local 537 struct gb_svc_route_destroy_request request; local 559 struct gb_svc_intf_set_pwrm_request request; local 601 struct gb_svc_intf_set_pwrm_request request; local 645 struct gb_svc_version_request *request; local 956 struct gb_svc_module_inserted_request *request; local 1004 struct gb_svc_module_removed_request *request; local 1032 struct gb_svc_intf_oops_request *request; local 1063 struct gb_svc_intf_mailbox_event_request *request; local 1147 struct gb_message *request = op->request; local 1165 struct gb_svc_module_inserted_request *request; local 1184 struct gb_svc_module_removed_request *request; local 1203 struct gb_svc_intf_oops_request *request; local 1217 struct gb_svc_intf_mailbox_event_request *request; local [all...] |
/linux-master/net/ethtool/ |
H A D | eeprom.c | 29 static int fallback_set_params(struct eeprom_req_info *request, argument 33 u32 offset = request->offset; 34 u32 length = request->length; 36 if (request->page) 37 offset = request->page * ETH_MODULE_EEPROM_PAGE_LEN + offset; 40 request->i2c_address == 0x51) 53 static int eeprom_fallback(struct eeprom_req_info *request, argument 67 err = fallback_set_params(request, &modinfo, &eeprom); 108 struct eeprom_req_info *request = MODULE_EEPROM_REQINFO(req_base); local 113 page_data.offset = request 149 struct eeprom_req_info *request = MODULE_EEPROM_REQINFO(req_info); local 194 const struct eeprom_req_info *request = MODULE_EEPROM_REQINFO(req_base); local [all...] |
/linux-master/drivers/staging/greybus/ |
H A D | pwm.c | 38 * The request returns the highest allowed PWM id parameter. So add one 47 struct gb_pwm_activate_request request; local 51 request.which = which; 59 &request, sizeof(request), NULL, 0); 69 struct gb_pwm_deactivate_request request; local 73 request.which = which; 81 &request, sizeof(request), NULL, 0); 92 struct gb_pwm_config_request request; local 117 struct gb_pwm_polarity_request request; local 140 struct gb_pwm_enable_request request; local 162 struct gb_pwm_disable_request request; local [all...] |
/linux-master/include/trace/events/ |
H A D | nbd.h | 12 TP_PROTO(struct request *req, u64 handle), 17 __field(struct request *, req) 27 "nbd transport event: request %p, handle 0x%016llx", 35 TP_PROTO(struct request *req, u64 handle), 42 TP_PROTO(struct request *req, u64 handle), 49 TP_PROTO(struct request *req, u64 handle), 56 TP_PROTO(struct request *req, u64 handle), 64 struct request *rq), 71 __field(struct request *, request) [all...] |
/linux-master/tools/testing/kunit/ |
H A D | kunit.py | 78 request: KunitConfigRequest) -> KunitResult: 82 success = linux.build_reconfig(request.build_dir, request.make_options) 88 request: KunitBuildRequest) -> KunitResult: 92 success = linux.build_kernel(request.jobs, 93 request.build_dir, 94 request.make_options) 100 request: KunitBuildRequest) -> KunitResult: 101 config_result = config_tests(linux, request) 105 return build_tests(linux, request) [all...] |
/linux-master/drivers/media/usb/as102/ |
H A D | as102_fw.h | 17 unsigned char request[2]; member in union:as10x_fw_pkt_t::__anon144
|
/linux-master/drivers/media/rc/img-ir/ |
H A D | img-ir-rc5.c | 12 struct img_ir_scancode_req *request) 32 request->protocol = RC_PROTO_RC5; 33 request->scancode = addr << 8 | cmd; 34 request->toggle = tgl; 11 img_ir_rc5_scancode(int len, u64 raw, u64 enabled_protocols, struct img_ir_scancode_req *request) argument
|
/linux-master/drivers/md/dm-vdo/indexer/ |
H A D | funnel-requestqueue.c | 19 * immediately to each new request. The wait time between batches is dynamically adjusted up or 23 * awoken when a new request is enqueued. The enqueue operation updates "newest" in the funnel 52 /* Function to process a request */ 90 * Determine if there is a next request to process, and return it if there is. Also return flags 92 * the thread did sleep before returning a new request. 97 struct uds_request *request = poll_queues(queue); local 99 if (request != NULL) { 100 *request_ptr = request; 116 unsigned long timeout, struct uds_request **request, 121 (dequeue_request(queue, request, waite 115 wait_for_request(struct uds_request_queue *queue, bool dormant, unsigned long timeout, struct uds_request **request, bool *waited) argument 134 struct uds_request *request = NULL; local 240 uds_request_queue_enqueue(struct uds_request_queue *queue, struct uds_request *request) argument [all...] |