Lines Matching refs:tcmu_dev

100 struct tcmu_dev;
106 struct tcmu_dev *udev;
111 struct tcmu_dev {
174 #define TCMU_DEV(_se_dev) container_of(_se_dev, struct tcmu_dev, se_dev)
178 struct tcmu_dev *tcmu_dev;
212 * mutex_lock(&tcmu_dev->cmdr_lock);
213 * mutex_unlock(&tcmu_dev->cmdr_lock);
305 struct tcmu_dev *udev = nl_cmd->udev;
374 struct tcmu_dev *udev = NULL;
499 struct tcmu_dev *udev = tcmu_cmd->tcmu_dev;
506 static inline int tcmu_get_empty_block(struct tcmu_dev *udev,
554 static int tcmu_get_empty_blocks(struct tcmu_dev *udev,
582 uint32_t blk_size = cmd->tcmu_dev->data_blk_size;
596 static int new_block_to_iov(struct tcmu_dev *udev, struct tcmu_cmd *cmd,
622 static void tcmu_setup_iovs(struct tcmu_dev *udev, struct tcmu_cmd *cmd,
636 struct tcmu_dev *udev = TCMU_DEV(se_dev);
645 tcmu_cmd->tcmu_dev = udev;
702 static inline void tcmu_copy_data(struct tcmu_dev *udev,
774 static void scatter_data_area(struct tcmu_dev *udev, struct tcmu_cmd *tcmu_cmd,
783 static void gather_data_area(struct tcmu_dev *udev, struct tcmu_cmd *tcmu_cmd,
820 static bool is_ring_space_avail(struct tcmu_dev *udev, size_t cmd_size)
854 static int tcmu_alloc_data_space(struct tcmu_dev *udev, struct tcmu_cmd *cmd,
926 tcmu_cmd->tcmu_dev->name, tmo / MSEC_PER_SEC);
931 struct tcmu_dev *udev = tcmu_cmd->tcmu_dev;
955 static uint32_t ring_insert_padding(struct tcmu_dev *udev, size_t cmd_size)
986 struct tcmu_dev *udev = TCMU_DEV(se_dev);
994 struct tcmu_dev *udev = TCMU_DEV(se_dev);
1014 struct tcmu_dev *udev = tcmu_cmd->tcmu_dev;
1150 * @udev: related tcmu_dev
1158 queue_tmr_ring(struct tcmu_dev *udev, struct tcmu_tmr *tmr)
1206 struct tcmu_dev *udev = TCMU_DEV(se_dev);
1263 struct tcmu_dev *udev = TCMU_DEV(se_dev);
1324 struct tcmu_dev *udev = cmd->tcmu_dev;
1411 static int tcmu_run_tmr_queue(struct tcmu_dev *udev)
1443 static bool tcmu_handle_completions(struct tcmu_dev *udev)
1532 cmd->cmd_id, cmd->tcmu_dev->name);
1545 cmd, cmd->tcmu_dev->name);
1555 static void tcmu_device_timedout(struct tcmu_dev *udev)
1567 struct tcmu_dev *udev = from_timer(udev, t, cmd_timer);
1575 struct tcmu_dev *udev = from_timer(udev, t, qfull_timer);
1603 struct tcmu_dev *udev;
1605 udev = kzalloc(sizeof(struct tcmu_dev), GFP_KERNEL);
1645 struct tcmu_dev *udev = TCMU_DEV(dev);
1662 static u32 tcmu_blocks_release(struct tcmu_dev *udev, unsigned long first,
1702 static void tcmu_remove_all_queued_tmr(struct tcmu_dev *udev)
1714 struct tcmu_dev *udev = container_of(kref, struct tcmu_dev, kref);
1750 static void run_qfull_queue(struct tcmu_dev *udev, bool fail)
1812 struct tcmu_dev *udev = container_of(info, struct tcmu_dev, uio_info);
1828 struct tcmu_dev *udev = vma->vm_private_data;
1839 static struct page *tcmu_try_get_data_page(struct tcmu_dev *udev, uint32_t dpi)
1865 struct tcmu_dev *udev = vma->vm_private_data;
1874 struct tcmu_dev *udev = vma->vm_private_data;
1884 struct tcmu_dev *udev = vmf->vma->vm_private_data;
1929 struct tcmu_dev *udev = container_of(info, struct tcmu_dev, uio_info);
1947 struct tcmu_dev *udev = container_of(info, struct tcmu_dev, uio_info);
1962 struct tcmu_dev *udev = container_of(info, struct tcmu_dev, uio_info);
2003 static int tcmu_init_genl_cmd_reply(struct tcmu_dev *udev, int cmd)
2041 static void tcmu_destroy_genl_cmd_reply(struct tcmu_dev *udev)
2059 static int tcmu_wait_genl_cmd_reply(struct tcmu_dev *udev)
2081 static int tcmu_netlink_event_init(struct tcmu_dev *udev,
2118 static int tcmu_netlink_event_send(struct tcmu_dev *udev,
2145 static int tcmu_send_dev_add_event(struct tcmu_dev *udev)
2159 static int tcmu_send_dev_remove_event(struct tcmu_dev *udev)
2173 static int tcmu_update_uio_info(struct tcmu_dev *udev)
2199 struct tcmu_dev *udev = TCMU_DEV(dev);
2314 struct tcmu_dev *udev = TCMU_DEV(dev);
2322 struct tcmu_dev *udev = TCMU_DEV(dev);
2339 static void tcmu_unblock_dev(struct tcmu_dev *udev)
2346 static void tcmu_block_dev(struct tcmu_dev *udev)
2362 static void tcmu_reset_ring(struct tcmu_dev *udev, u8 err_level)
2465 static int tcmu_set_max_blocks_param(struct tcmu_dev *udev, substring_t *arg)
2506 static int tcmu_set_data_pages_per_blk(struct tcmu_dev *udev, substring_t *arg)
2539 static int tcmu_set_cmd_ring_size(struct tcmu_dev *udev, substring_t *arg)
2577 struct tcmu_dev *udev = TCMU_DEV(dev);
2645 struct tcmu_dev *udev = TCMU_DEV(dev);
2661 struct tcmu_dev *udev = TCMU_DEV(dev);
2677 struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
2687 struct tcmu_dev *udev = container_of(da->da_dev,
2688 struct tcmu_dev, se_dev);
2710 struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
2722 struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
2746 struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
2757 struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
2767 struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
2778 struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
2783 static int tcmu_send_dev_config_event(struct tcmu_dev *udev,
2809 struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
2840 struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
2845 static int tcmu_send_dev_size_event(struct tcmu_dev *udev, u64 size)
2870 struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
2896 struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
2906 struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
2928 static int tcmu_send_emulate_write_cache(struct tcmu_dev *udev, u8 val)
2952 struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
2978 struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
2989 struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
3012 struct tcmu_dev *udev = TCMU_DEV(se_dev);
3026 struct tcmu_dev *udev = TCMU_DEV(se_dev);
3058 struct tcmu_dev *udev = TCMU_DEV(se_dev);
3087 struct tcmu_dev *udev = TCMU_DEV(se_dev);
3188 struct tcmu_dev *udev;
3265 struct tcmu_dev *udev, *tmp_dev;