1/* SPDX-License-Identifier: GPL-2.0 */
2/*
3 * Copyright (c) 2018-2020, The Linux Foundation. All rights reserved.
4 *
5 */
6#ifndef _MHI_H_
7#define _MHI_H_
8
9#include <linux/device.h>
10#include <linux/dma-direction.h>
11#include <linux/mutex.h>
12#include <linux/skbuff.h>
13#include <linux/slab.h>
14#include <linux/spinlock.h>
15#include <linux/wait.h>
16#include <linux/workqueue.h>
17
18#define MHI_MAX_OEM_PK_HASH_SEGMENTS 16
19
20struct mhi_chan;
21struct mhi_event;
22struct mhi_ctxt;
23struct mhi_cmd;
24struct mhi_buf_info;
25
26/**
27 * enum mhi_callback - MHI callback
28 * @MHI_CB_IDLE: MHI entered idle state
29 * @MHI_CB_PENDING_DATA: New data available for client to process
30 * @MHI_CB_LPM_ENTER: MHI host entered low power mode
31 * @MHI_CB_LPM_EXIT: MHI host about to exit low power mode
32 * @MHI_CB_EE_RDDM: MHI device entered RDDM exec env
33 * @MHI_CB_EE_MISSION_MODE: MHI device entered Mission Mode exec env
34 * @MHI_CB_SYS_ERROR: MHI device entered error state (may recover)
35 * @MHI_CB_FATAL_ERROR: MHI device entered fatal error state
36 * @MHI_CB_BW_REQ: Received a bandwidth switch request from device
37 */
38enum mhi_callback {
39	MHI_CB_IDLE,
40	MHI_CB_PENDING_DATA,
41	MHI_CB_LPM_ENTER,
42	MHI_CB_LPM_EXIT,
43	MHI_CB_EE_RDDM,
44	MHI_CB_EE_MISSION_MODE,
45	MHI_CB_SYS_ERROR,
46	MHI_CB_FATAL_ERROR,
47	MHI_CB_BW_REQ,
48};
49
50/**
51 * enum mhi_flags - Transfer flags
52 * @MHI_EOB: End of buffer for bulk transfer
53 * @MHI_EOT: End of transfer
54 * @MHI_CHAIN: Linked transfer
55 */
56enum mhi_flags {
57	MHI_EOB = BIT(0),
58	MHI_EOT = BIT(1),
59	MHI_CHAIN = BIT(2),
60};
61
62/**
63 * enum mhi_device_type - Device types
64 * @MHI_DEVICE_XFER: Handles data transfer
65 * @MHI_DEVICE_CONTROLLER: Control device
66 */
67enum mhi_device_type {
68	MHI_DEVICE_XFER,
69	MHI_DEVICE_CONTROLLER,
70};
71
72/**
73 * enum mhi_ch_type - Channel types
74 * @MHI_CH_TYPE_INVALID: Invalid channel type
75 * @MHI_CH_TYPE_OUTBOUND: Outbound channel to the device
76 * @MHI_CH_TYPE_INBOUND: Inbound channel from the device
77 * @MHI_CH_TYPE_INBOUND_COALESCED: Coalesced channel for the device to combine
78 *				   multiple packets and send them as a single
79 *				   large packet to reduce CPU consumption
80 */
81enum mhi_ch_type {
82	MHI_CH_TYPE_INVALID = 0,
83	MHI_CH_TYPE_OUTBOUND = DMA_TO_DEVICE,
84	MHI_CH_TYPE_INBOUND = DMA_FROM_DEVICE,
85	MHI_CH_TYPE_INBOUND_COALESCED = 3,
86};
87
88/**
89 * struct image_info - Firmware and RDDM table
90 * @mhi_buf: Buffer for firmware and RDDM table
91 * @entries: # of entries in table
92 */
93struct image_info {
94	struct mhi_buf *mhi_buf;
95	/* private: from internal.h */
96	struct bhi_vec_entry *bhi_vec;
97	/* public: */
98	u32 entries;
99};
100
101/**
102 * struct mhi_link_info - BW requirement
103 * target_link_speed - Link speed as defined by TLS bits in LinkControl reg
104 * target_link_width - Link width as defined by NLW bits in LinkStatus reg
105 */
106struct mhi_link_info {
107	unsigned int target_link_speed;
108	unsigned int target_link_width;
109};
110
111/**
112 * enum mhi_ee_type - Execution environment types
113 * @MHI_EE_PBL: Primary Bootloader
114 * @MHI_EE_SBL: Secondary Bootloader
115 * @MHI_EE_AMSS: Modem, aka the primary runtime EE
116 * @MHI_EE_RDDM: Ram dump download mode
117 * @MHI_EE_WFW: WLAN firmware mode
118 * @MHI_EE_PTHRU: Passthrough
119 * @MHI_EE_EDL: Embedded downloader
120 * @MHI_EE_FP: Flash Programmer Environment
121 */
122enum mhi_ee_type {
123	MHI_EE_PBL,
124	MHI_EE_SBL,
125	MHI_EE_AMSS,
126	MHI_EE_RDDM,
127	MHI_EE_WFW,
128	MHI_EE_PTHRU,
129	MHI_EE_EDL,
130	MHI_EE_FP,
131	MHI_EE_MAX_SUPPORTED = MHI_EE_FP,
132	MHI_EE_DISABLE_TRANSITION, /* local EE, not related to mhi spec */
133	MHI_EE_NOT_SUPPORTED,
134	MHI_EE_MAX,
135};
136
137/**
138 * enum mhi_state - MHI states
139 * @MHI_STATE_RESET: Reset state
140 * @MHI_STATE_READY: Ready state
141 * @MHI_STATE_M0: M0 state
142 * @MHI_STATE_M1: M1 state
143 * @MHI_STATE_M2: M2 state
144 * @MHI_STATE_M3: M3 state
145 * @MHI_STATE_M3_FAST: M3 Fast state
146 * @MHI_STATE_BHI: BHI state
147 * @MHI_STATE_SYS_ERR: System Error state
148 */
149enum mhi_state {
150	MHI_STATE_RESET = 0x0,
151	MHI_STATE_READY = 0x1,
152	MHI_STATE_M0 = 0x2,
153	MHI_STATE_M1 = 0x3,
154	MHI_STATE_M2 = 0x4,
155	MHI_STATE_M3 = 0x5,
156	MHI_STATE_M3_FAST = 0x6,
157	MHI_STATE_BHI = 0x7,
158	MHI_STATE_SYS_ERR = 0xFF,
159	MHI_STATE_MAX,
160};
161
162/**
163 * enum mhi_ch_ee_mask - Execution environment mask for channel
164 * @MHI_CH_EE_PBL: Allow channel to be used in PBL EE
165 * @MHI_CH_EE_SBL: Allow channel to be used in SBL EE
166 * @MHI_CH_EE_AMSS: Allow channel to be used in AMSS EE
167 * @MHI_CH_EE_RDDM: Allow channel to be used in RDDM EE
168 * @MHI_CH_EE_PTHRU: Allow channel to be used in PTHRU EE
169 * @MHI_CH_EE_WFW: Allow channel to be used in WFW EE
170 * @MHI_CH_EE_EDL: Allow channel to be used in EDL EE
171 */
172enum mhi_ch_ee_mask {
173	MHI_CH_EE_PBL = BIT(MHI_EE_PBL),
174	MHI_CH_EE_SBL = BIT(MHI_EE_SBL),
175	MHI_CH_EE_AMSS = BIT(MHI_EE_AMSS),
176	MHI_CH_EE_RDDM = BIT(MHI_EE_RDDM),
177	MHI_CH_EE_PTHRU = BIT(MHI_EE_PTHRU),
178	MHI_CH_EE_WFW = BIT(MHI_EE_WFW),
179	MHI_CH_EE_EDL = BIT(MHI_EE_EDL),
180};
181
182/**
183 * enum mhi_er_data_type - Event ring data types
184 * @MHI_ER_DATA: Only client data over this ring
185 * @MHI_ER_CTRL: MHI control data and client data
186 */
187enum mhi_er_data_type {
188	MHI_ER_DATA,
189	MHI_ER_CTRL,
190};
191
192/**
193 * enum mhi_db_brst_mode - Doorbell mode
194 * @MHI_DB_BRST_DISABLE: Burst mode disable
195 * @MHI_DB_BRST_ENABLE: Burst mode enable
196 */
197enum mhi_db_brst_mode {
198	MHI_DB_BRST_DISABLE = 0x2,
199	MHI_DB_BRST_ENABLE = 0x3,
200};
201
202/**
203 * struct mhi_channel_config - Channel configuration structure for controller
204 * @name: The name of this channel
205 * @num: The number assigned to this channel
206 * @num_elements: The number of elements that can be queued to this channel
207 * @local_elements: The local ring length of the channel
208 * @event_ring: The event ring index that services this channel
209 * @dir: Direction that data may flow on this channel
210 * @type: Channel type
211 * @ee_mask: Execution Environment mask for this channel
212 * @pollcfg: Polling configuration for burst mode.  0 is default.  milliseconds
213	     for UL channels, multiple of 8 ring elements for DL channels
214 * @doorbell: Doorbell mode
215 * @lpm_notify: The channel master requires low power mode notifications
216 * @offload_channel: The client manages the channel completely
217 * @doorbell_mode_switch: Channel switches to doorbell mode on M0 transition
218 * @auto_queue: Framework will automatically queue buffers for DL traffic
219 * @wake-capable: Channel capable of waking up the system
220 */
221struct mhi_channel_config {
222	char *name;
223	u32 num;
224	u32 num_elements;
225	u32 local_elements;
226	u32 event_ring;
227	enum dma_data_direction dir;
228	enum mhi_ch_type type;
229	u32 ee_mask;
230	u32 pollcfg;
231	enum mhi_db_brst_mode doorbell;
232	bool lpm_notify;
233	bool offload_channel;
234	bool doorbell_mode_switch;
235	bool auto_queue;
236	bool wake_capable;
237};
238
239/**
240 * struct mhi_event_config - Event ring configuration structure for controller
241 * @num_elements: The number of elements that can be queued to this ring
242 * @irq_moderation_ms: Delay irq for additional events to be aggregated
243 * @irq: IRQ associated with this ring
244 * @channel: Dedicated channel number. U32_MAX indicates a non-dedicated ring
245 * @priority: Priority of this ring. Use 1 for now
246 * @mode: Doorbell mode
247 * @data_type: Type of data this ring will process
248 * @hardware_event: This ring is associated with hardware channels
249 * @client_managed: This ring is client managed
250 * @offload_channel: This ring is associated with an offloaded channel
251 */
252struct mhi_event_config {
253	u32 num_elements;
254	u32 irq_moderation_ms;
255	u32 irq;
256	u32 channel;
257	u32 priority;
258	enum mhi_db_brst_mode mode;
259	enum mhi_er_data_type data_type;
260	bool hardware_event;
261	bool client_managed;
262	bool offload_channel;
263};
264
265/**
266 * struct mhi_controller_config - Root MHI controller configuration
267 * @max_channels: Maximum number of channels supported
268 * @timeout_ms: Timeout value for operations. 0 means use default
269 * @ready_timeout_ms: Timeout value for waiting device to be ready (optional)
270 * @buf_len: Size of automatically allocated buffers. 0 means use default
271 * @num_channels: Number of channels defined in @ch_cfg
272 * @ch_cfg: Array of defined channels
273 * @num_events: Number of event rings defined in @event_cfg
274 * @event_cfg: Array of defined event rings
275 * @use_bounce_buf: Use a bounce buffer pool due to limited DDR access
276 * @m2_no_db: Host is not allowed to ring DB in M2 state
277 */
278struct mhi_controller_config {
279	u32 max_channels;
280	u32 timeout_ms;
281	u32 ready_timeout_ms;
282	u32 buf_len;
283	u32 num_channels;
284	const struct mhi_channel_config *ch_cfg;
285	u32 num_events;
286	struct mhi_event_config *event_cfg;
287	bool use_bounce_buf;
288	bool m2_no_db;
289};
290
291/**
292 * struct mhi_controller - Master MHI controller structure
293 * @cntrl_dev: Pointer to the struct device of physical bus acting as the MHI
294 *            controller (required)
295 * @mhi_dev: MHI device instance for the controller
296 * @debugfs_dentry: MHI controller debugfs directory
297 * @regs: Base address of MHI MMIO register space (required)
298 * @bhi: Points to base of MHI BHI register space
299 * @bhie: Points to base of MHI BHIe register space
300 * @wake_db: MHI WAKE doorbell register address
301 * @iova_start: IOMMU starting address for data (required)
302 * @iova_stop: IOMMU stop address for data (required)
303 * @fw_image: Firmware image name for normal booting (optional)
304 * @fw_data: Firmware image data content for normal booting, used only
305 *           if fw_image is NULL and fbc_download is true (optional)
306 * @fw_sz: Firmware image data size for normal booting, used only if fw_image
307 *         is NULL and fbc_download is true (optional)
308 * @edl_image: Firmware image name for emergency download mode (optional)
309 * @rddm_size: RAM dump size that host should allocate for debugging purpose
310 * @sbl_size: SBL image size downloaded through BHIe (optional)
311 * @seg_len: BHIe vector size (optional)
312 * @reg_len: Length of the MHI MMIO region (required)
313 * @fbc_image: Points to firmware image buffer
314 * @rddm_image: Points to RAM dump buffer
315 * @mhi_chan: Points to the channel configuration table
316 * @lpm_chans: List of channels that require LPM notifications
317 * @irq: base irq # to request (required)
318 * @max_chan: Maximum number of channels the controller supports
319 * @total_ev_rings: Total # of event rings allocated
320 * @hw_ev_rings: Number of hardware event rings
321 * @sw_ev_rings: Number of software event rings
322 * @nr_irqs: Number of IRQ allocated by bus master (required)
323 * @serial_number: MHI controller serial number obtained from BHI
324 * @mhi_event: MHI event ring configurations table
325 * @mhi_cmd: MHI command ring configurations table
326 * @mhi_ctxt: MHI device context, shared memory between host and device
327 * @pm_mutex: Mutex for suspend/resume operation
328 * @pm_lock: Lock for protecting MHI power management state
329 * @timeout_ms: Timeout in ms for state transitions
330 * @ready_timeout_ms: Timeout in ms for waiting device to be ready (optional)
331 * @pm_state: MHI power management state
332 * @db_access: DB access states
333 * @ee: MHI device execution environment
334 * @dev_state: MHI device state
335 * @dev_wake: Device wakeup count
336 * @pending_pkts: Pending packets for the controller
337 * @M0, M2, M3: Counters to track number of device MHI state changes
338 * @transition_list: List of MHI state transitions
339 * @transition_lock: Lock for protecting MHI state transition list
340 * @wlock: Lock for protecting device wakeup
341 * @mhi_link_info: Device bandwidth info
342 * @st_worker: State transition worker
343 * @hiprio_wq: High priority workqueue for MHI work such as state transitions
344 * @state_event: State change event
345 * @status_cb: CB function to notify power states of the device (required)
346 * @wake_get: CB function to assert device wake (optional)
347 * @wake_put: CB function to de-assert device wake (optional)
348 * @wake_toggle: CB function to assert and de-assert device wake (optional)
349 * @runtime_get: CB function to controller runtime resume (required)
350 * @runtime_put: CB function to decrement pm usage (required)
351 * @map_single: CB function to create TRE buffer
352 * @unmap_single: CB function to destroy TRE buffer
353 * @read_reg: Read a MHI register via the physical link (required)
354 * @write_reg: Write a MHI register via the physical link (required)
355 * @reset: Controller specific reset function (optional)
356 * @buffer_len: Bounce buffer length
357 * @index: Index of the MHI controller instance
358 * @bounce_buf: Use of bounce buffer
359 * @fbc_download: MHI host needs to do complete image transfer (optional)
360 * @wake_set: Device wakeup set flag
361 * @irq_flags: irq flags passed to request_irq (optional)
362 * @mru: the default MRU for the MHI device
363 *
364 * Fields marked as (required) need to be populated by the controller driver
365 * before calling mhi_register_controller(). For the fields marked as (optional)
366 * they can be populated depending on the usecase.
367 */
368struct mhi_controller {
369	struct device *cntrl_dev;
370	struct mhi_device *mhi_dev;
371	struct dentry *debugfs_dentry;
372	void __iomem *regs;
373	void __iomem *bhi;
374	void __iomem *bhie;
375	void __iomem *wake_db;
376
377	dma_addr_t iova_start;
378	dma_addr_t iova_stop;
379	const char *fw_image;
380	const u8 *fw_data;
381	size_t fw_sz;
382	const char *edl_image;
383	size_t rddm_size;
384	size_t sbl_size;
385	size_t seg_len;
386	size_t reg_len;
387	struct image_info *fbc_image;
388	struct image_info *rddm_image;
389	struct mhi_chan *mhi_chan;
390	struct list_head lpm_chans;
391	int *irq;
392	u32 max_chan;
393	u32 total_ev_rings;
394	u32 hw_ev_rings;
395	u32 sw_ev_rings;
396	u32 nr_irqs;
397	u32 serial_number;
398
399	struct mhi_event *mhi_event;
400	struct mhi_cmd *mhi_cmd;
401	struct mhi_ctxt *mhi_ctxt;
402
403	struct mutex pm_mutex;
404	rwlock_t pm_lock;
405	u32 timeout_ms;
406	u32 ready_timeout_ms;
407	u32 pm_state;
408	u32 db_access;
409	enum mhi_ee_type ee;
410	enum mhi_state dev_state;
411	atomic_t dev_wake;
412	atomic_t pending_pkts;
413	u32 M0, M2, M3;
414	struct list_head transition_list;
415	spinlock_t transition_lock;
416	spinlock_t wlock;
417	struct mhi_link_info mhi_link_info;
418	struct work_struct st_worker;
419	struct workqueue_struct *hiprio_wq;
420	wait_queue_head_t state_event;
421
422	void (*status_cb)(struct mhi_controller *mhi_cntrl,
423			  enum mhi_callback cb);
424	void (*wake_get)(struct mhi_controller *mhi_cntrl, bool override);
425	void (*wake_put)(struct mhi_controller *mhi_cntrl, bool override);
426	void (*wake_toggle)(struct mhi_controller *mhi_cntrl);
427	int (*runtime_get)(struct mhi_controller *mhi_cntrl);
428	void (*runtime_put)(struct mhi_controller *mhi_cntrl);
429	int (*map_single)(struct mhi_controller *mhi_cntrl,
430			  struct mhi_buf_info *buf);
431	void (*unmap_single)(struct mhi_controller *mhi_cntrl,
432			     struct mhi_buf_info *buf);
433	int (*read_reg)(struct mhi_controller *mhi_cntrl, void __iomem *addr,
434			u32 *out);
435	void (*write_reg)(struct mhi_controller *mhi_cntrl, void __iomem *addr,
436			  u32 val);
437	void (*reset)(struct mhi_controller *mhi_cntrl);
438
439	size_t buffer_len;
440	int index;
441	bool bounce_buf;
442	bool fbc_download;
443	bool wake_set;
444	unsigned long irq_flags;
445	u32 mru;
446};
447
448/**
449 * struct mhi_device - Structure representing an MHI device which binds
450 *                     to channels or is associated with controllers
451 * @id: Pointer to MHI device ID struct
452 * @name: Name of the associated MHI device
453 * @mhi_cntrl: Controller the device belongs to
454 * @ul_chan: UL channel for the device
455 * @dl_chan: DL channel for the device
456 * @dev: Driver model device node for the MHI device
457 * @dev_type: MHI device type
458 * @ul_chan_id: MHI channel id for UL transfer
459 * @dl_chan_id: MHI channel id for DL transfer
460 * @dev_wake: Device wakeup counter
461 */
462struct mhi_device {
463	const struct mhi_device_id *id;
464	const char *name;
465	struct mhi_controller *mhi_cntrl;
466	struct mhi_chan *ul_chan;
467	struct mhi_chan *dl_chan;
468	struct device dev;
469	enum mhi_device_type dev_type;
470	int ul_chan_id;
471	int dl_chan_id;
472	u32 dev_wake;
473};
474
475/**
476 * struct mhi_result - Completed buffer information
477 * @buf_addr: Address of data buffer
478 * @bytes_xferd: # of bytes transferred
479 * @dir: Channel direction
480 * @transaction_status: Status of last transaction
481 */
482struct mhi_result {
483	void *buf_addr;
484	size_t bytes_xferd;
485	enum dma_data_direction dir;
486	int transaction_status;
487};
488
489/**
490 * struct mhi_buf - MHI Buffer description
491 * @buf: Virtual address of the buffer
492 * @name: Buffer label. For offload channel, configurations name must be:
493 *        ECA - Event context array data
494 *        CCA - Channel context array data
495 * @dma_addr: IOMMU address of the buffer
496 * @len: # of bytes
497 */
498struct mhi_buf {
499	void *buf;
500	const char *name;
501	dma_addr_t dma_addr;
502	size_t len;
503};
504
505/**
506 * struct mhi_driver - Structure representing a MHI client driver
507 * @probe: CB function for client driver probe function
508 * @remove: CB function for client driver remove function
509 * @ul_xfer_cb: CB function for UL data transfer
510 * @dl_xfer_cb: CB function for DL data transfer
511 * @status_cb: CB functions for asynchronous status
512 * @driver: Device driver model driver
513 */
514struct mhi_driver {
515	const struct mhi_device_id *id_table;
516	int (*probe)(struct mhi_device *mhi_dev,
517		     const struct mhi_device_id *id);
518	void (*remove)(struct mhi_device *mhi_dev);
519	void (*ul_xfer_cb)(struct mhi_device *mhi_dev,
520			   struct mhi_result *result);
521	void (*dl_xfer_cb)(struct mhi_device *mhi_dev,
522			   struct mhi_result *result);
523	void (*status_cb)(struct mhi_device *mhi_dev, enum mhi_callback mhi_cb);
524	struct device_driver driver;
525};
526
527#define to_mhi_driver(drv) container_of(drv, struct mhi_driver, driver)
528#define to_mhi_device(dev) container_of(dev, struct mhi_device, dev)
529
530/**
531 * mhi_alloc_controller - Allocate the MHI Controller structure
532 * Allocate the mhi_controller structure using zero initialized memory
533 */
534struct mhi_controller *mhi_alloc_controller(void);
535
536/**
537 * mhi_free_controller - Free the MHI Controller structure
538 * Free the mhi_controller structure which was previously allocated
539 */
540void mhi_free_controller(struct mhi_controller *mhi_cntrl);
541
542/**
543 * mhi_register_controller - Register MHI controller
544 * @mhi_cntrl: MHI controller to register
545 * @config: Configuration to use for the controller
546 */
547int mhi_register_controller(struct mhi_controller *mhi_cntrl,
548			const struct mhi_controller_config *config);
549
550/**
551 * mhi_unregister_controller - Unregister MHI controller
552 * @mhi_cntrl: MHI controller to unregister
553 */
554void mhi_unregister_controller(struct mhi_controller *mhi_cntrl);
555
556/*
557 * module_mhi_driver() - Helper macro for drivers that don't do
558 * anything special other than using default mhi_driver_register() and
559 * mhi_driver_unregister().  This eliminates a lot of boilerplate.
560 * Each module may only use this macro once.
561 */
562#define module_mhi_driver(mhi_drv) \
563	module_driver(mhi_drv, mhi_driver_register, \
564		      mhi_driver_unregister)
565
566/*
567 * Macro to avoid include chaining to get THIS_MODULE
568 */
569#define mhi_driver_register(mhi_drv) \
570	__mhi_driver_register(mhi_drv, THIS_MODULE)
571
572/**
573 * __mhi_driver_register - Register driver with MHI framework
574 * @mhi_drv: Driver associated with the device
575 * @owner: The module owner
576 */
577int __mhi_driver_register(struct mhi_driver *mhi_drv, struct module *owner);
578
579/**
580 * mhi_driver_unregister - Unregister a driver for mhi_devices
581 * @mhi_drv: Driver associated with the device
582 */
583void mhi_driver_unregister(struct mhi_driver *mhi_drv);
584
585/**
586 * mhi_set_mhi_state - Set MHI device state
587 * @mhi_cntrl: MHI controller
588 * @state: State to set
589 */
590void mhi_set_mhi_state(struct mhi_controller *mhi_cntrl,
591		       enum mhi_state state);
592
593/**
594 * mhi_notify - Notify the MHI client driver about client device status
595 * @mhi_dev: MHI device instance
596 * @cb_reason: MHI callback reason
597 */
598void mhi_notify(struct mhi_device *mhi_dev, enum mhi_callback cb_reason);
599
600/**
601 * mhi_get_free_desc_count - Get transfer ring length
602 * Get # of TD available to queue buffers
603 * @mhi_dev: Device associated with the channels
604 * @dir: Direction of the channel
605 */
606int mhi_get_free_desc_count(struct mhi_device *mhi_dev,
607				enum dma_data_direction dir);
608
609/**
610 * mhi_prepare_for_power_up - Do pre-initialization before power up.
611 *                            This is optional, call this before power up if
612 *                            the controller does not want bus framework to
613 *                            automatically free any allocated memory during
614 *                            shutdown process.
615 * @mhi_cntrl: MHI controller
616 */
617int mhi_prepare_for_power_up(struct mhi_controller *mhi_cntrl);
618
619/**
620 * mhi_async_power_up - Start MHI power up sequence
621 * @mhi_cntrl: MHI controller
622 */
623int mhi_async_power_up(struct mhi_controller *mhi_cntrl);
624
625/**
626 * mhi_sync_power_up - Start MHI power up sequence and wait till the device
627 *                     enters valid EE state
628 * @mhi_cntrl: MHI controller
629 */
630int mhi_sync_power_up(struct mhi_controller *mhi_cntrl);
631
632/**
633 * mhi_power_down - Start MHI power down sequence
634 * @mhi_cntrl: MHI controller
635 * @graceful: Link is still accessible, so do a graceful shutdown process
636 */
637void mhi_power_down(struct mhi_controller *mhi_cntrl, bool graceful);
638
639/**
640 * mhi_unprepare_after_power_down - Free any allocated memory after power down
641 * @mhi_cntrl: MHI controller
642 */
643void mhi_unprepare_after_power_down(struct mhi_controller *mhi_cntrl);
644
645/**
646 * mhi_pm_suspend - Move MHI into a suspended state
647 * @mhi_cntrl: MHI controller
648 */
649int mhi_pm_suspend(struct mhi_controller *mhi_cntrl);
650
651/**
652 * mhi_pm_resume - Resume MHI from suspended state
653 * @mhi_cntrl: MHI controller
654 */
655int mhi_pm_resume(struct mhi_controller *mhi_cntrl);
656
657/**
658 * mhi_pm_resume_force - Force resume MHI from suspended state
659 * @mhi_cntrl: MHI controller
660 *
661 * Resume the device irrespective of its MHI state. As per the MHI spec, devices
662 * has to be in M3 state during resume. But some devices seem to be in a
663 * different MHI state other than M3 but they continue working fine if allowed.
664 * This API is intented to be used for such devices.
665 *
666 * Return: 0 if the resume succeeds, a negative error code otherwise
667 */
668int mhi_pm_resume_force(struct mhi_controller *mhi_cntrl);
669
670/**
671 * mhi_download_rddm_image - Download ramdump image from device for
672 *                           debugging purpose.
673 * @mhi_cntrl: MHI controller
674 * @in_panic: Download rddm image during kernel panic
675 */
676int mhi_download_rddm_image(struct mhi_controller *mhi_cntrl, bool in_panic);
677
678/**
679 * mhi_force_rddm_mode - Force device into rddm mode
680 * @mhi_cntrl: MHI controller
681 */
682int mhi_force_rddm_mode(struct mhi_controller *mhi_cntrl);
683
684/**
685 * mhi_get_exec_env - Get BHI execution environment of the device
686 * @mhi_cntrl: MHI controller
687 */
688enum mhi_ee_type mhi_get_exec_env(struct mhi_controller *mhi_cntrl);
689
690/**
691 * mhi_get_mhi_state - Get MHI state of the device
692 * @mhi_cntrl: MHI controller
693 */
694enum mhi_state mhi_get_mhi_state(struct mhi_controller *mhi_cntrl);
695
696/**
697 * mhi_soc_reset - Trigger a device reset. This can be used as a last resort
698 *		   to reset and recover a device.
699 * @mhi_cntrl: MHI controller
700 */
701void mhi_soc_reset(struct mhi_controller *mhi_cntrl);
702
703/**
704 * mhi_device_get - Disable device low power mode
705 * @mhi_dev: Device associated with the channel
706 */
707void mhi_device_get(struct mhi_device *mhi_dev);
708
709/**
710 * mhi_device_get_sync - Disable device low power mode. Synchronously
711 *                       take the controller out of suspended state
712 * @mhi_dev: Device associated with the channel
713 */
714int mhi_device_get_sync(struct mhi_device *mhi_dev);
715
716/**
717 * mhi_device_put - Re-enable device low power mode
718 * @mhi_dev: Device associated with the channel
719 */
720void mhi_device_put(struct mhi_device *mhi_dev);
721
722/**
723 * mhi_prepare_for_transfer - Setup UL and DL channels for data transfer.
724 * @mhi_dev: Device associated with the channels
725 *
726 * Allocate and initialize the channel context and also issue the START channel
727 * command to both channels. Channels can be started only if both host and
728 * device execution environments match and channels are in a DISABLED state.
729 */
730int mhi_prepare_for_transfer(struct mhi_device *mhi_dev);
731
732/**
733 * mhi_prepare_for_transfer_autoqueue - Setup UL and DL channels with auto queue
734 *                                      buffers for DL traffic
735 * @mhi_dev: Device associated with the channels
736 *
737 * Allocate and initialize the channel context and also issue the START channel
738 * command to both channels. Channels can be started only if both host and
739 * device execution environments match and channels are in a DISABLED state.
740 * The MHI core will automatically allocate and queue buffers for the DL traffic.
741 */
742int mhi_prepare_for_transfer_autoqueue(struct mhi_device *mhi_dev);
743
744/**
745 * mhi_unprepare_from_transfer - Reset UL and DL channels for data transfer.
746 *                               Issue the RESET channel command and let the
747 *                               device clean-up the context so no incoming
748 *                               transfers are seen on the host. Free memory
749 *                               associated with the context on host. If device
750 *                               is unresponsive, only perform a host side
751 *                               clean-up. Channels can be reset only if both
752 *                               host and device execution environments match
753 *                               and channels are in an ENABLED, STOPPED or
754 *                               SUSPENDED state.
755 * @mhi_dev: Device associated with the channels
756 */
757void mhi_unprepare_from_transfer(struct mhi_device *mhi_dev);
758
759/**
760 * mhi_queue_dma - Send or receive DMA mapped buffers from client device
761 *                 over MHI channel
762 * @mhi_dev: Device associated with the channels
763 * @dir: DMA direction for the channel
764 * @mhi_buf: Buffer for holding the DMA mapped data
765 * @len: Buffer length
766 * @mflags: MHI transfer flags used for the transfer
767 */
768int mhi_queue_dma(struct mhi_device *mhi_dev, enum dma_data_direction dir,
769		  struct mhi_buf *mhi_buf, size_t len, enum mhi_flags mflags);
770
771/**
772 * mhi_queue_buf - Send or receive raw buffers from client device over MHI
773 *                 channel
774 * @mhi_dev: Device associated with the channels
775 * @dir: DMA direction for the channel
776 * @buf: Buffer for holding the data
777 * @len: Buffer length
778 * @mflags: MHI transfer flags used for the transfer
779 */
780int mhi_queue_buf(struct mhi_device *mhi_dev, enum dma_data_direction dir,
781		  void *buf, size_t len, enum mhi_flags mflags);
782
783/**
784 * mhi_queue_skb - Send or receive SKBs from client device over MHI channel
785 * @mhi_dev: Device associated with the channels
786 * @dir: DMA direction for the channel
787 * @skb: Buffer for holding SKBs
788 * @len: Buffer length
789 * @mflags: MHI transfer flags used for the transfer
790 */
791int mhi_queue_skb(struct mhi_device *mhi_dev, enum dma_data_direction dir,
792		  struct sk_buff *skb, size_t len, enum mhi_flags mflags);
793
794/**
795 * mhi_queue_is_full - Determine whether queueing new elements is possible
796 * @mhi_dev: Device associated with the channels
797 * @dir: DMA direction for the channel
798 */
799bool mhi_queue_is_full(struct mhi_device *mhi_dev, enum dma_data_direction dir);
800
801#endif /* _MHI_H_ */
802