1/* SPDX-License-Identifier: GPL-2.0 */
2#ifndef S390_IO_SCH_H
3#define S390_IO_SCH_H
4
5#include <linux/types.h>
6#include <asm/schid.h>
7#include <asm/ccwdev.h>
8#include <asm/irq.h>
9#include "css.h"
10#include "orb.h"
11
12struct io_subchannel_dma_area {
13	struct ccw1 sense_ccw;	/* static ccw for sense command */
14};
15
16struct io_subchannel_private {
17	union orb orb;		/* operation request block */
18	struct ccw_device *cdev;/* pointer to the child ccw device */
19	struct {
20		unsigned int suspend:1;	/* allow suspend */
21		unsigned int prefetch:1;/* deny prefetch */
22		unsigned int inter:1;	/* suppress intermediate interrupts */
23	} __packed options;
24	struct io_subchannel_dma_area *dma_area;
25	dma_addr_t dma_area_dma;
26} __aligned(8);
27
28#define to_io_private(n) ((struct io_subchannel_private *) \
29			  dev_get_drvdata(&(n)->dev))
30#define set_io_private(n, p) (dev_set_drvdata(&(n)->dev, p))
31
32static inline struct ccw_device *sch_get_cdev(struct subchannel *sch)
33{
34	struct io_subchannel_private *priv = to_io_private(sch);
35	return priv ? priv->cdev : NULL;
36}
37
38static inline void sch_set_cdev(struct subchannel *sch,
39				struct ccw_device *cdev)
40{
41	struct io_subchannel_private *priv = to_io_private(sch);
42	if (priv)
43		priv->cdev = cdev;
44}
45
46#define MAX_CIWS 8
47
48/*
49 * Possible status values for a CCW request's I/O.
50 */
51enum io_status {
52	IO_DONE,
53	IO_RUNNING,
54	IO_STATUS_ERROR,
55	IO_PATH_ERROR,
56	IO_REJECTED,
57	IO_KILLED
58};
59
60/**
61 * ccw_request - Internal CCW request.
62 * @cp: channel program to start
63 * @timeout: maximum allowable time in jiffies between start I/O and interrupt
64 * @maxretries: number of retries per I/O operation and path
65 * @lpm: mask of paths to use
66 * @check: optional callback that determines if results are final
67 * @filter: optional callback to adjust request status based on IRB data
68 * @callback: final callback
69 * @data: user-defined pointer passed to all callbacks
70 * @singlepath: if set, use only one path from @lpm per start I/O
71 * @cancel: non-zero if request was cancelled
72 * @done: non-zero if request was finished
73 * @mask: current path mask
74 * @retries: current number of retries
75 * @drc: delayed return code
76 */
77struct ccw_request {
78	struct ccw1 *cp;
79	unsigned long timeout;
80	u16 maxretries;
81	u8 lpm;
82	int (*check)(struct ccw_device *, void *);
83	enum io_status (*filter)(struct ccw_device *, void *, struct irb *,
84				 enum io_status);
85	void (*callback)(struct ccw_device *, void *, int);
86	void *data;
87	unsigned int singlepath:1;
88	/* These fields are used internally. */
89	unsigned int cancel:1;
90	unsigned int done:1;
91	u16 mask;
92	u16 retries;
93	int drc;
94} __attribute__((packed));
95
96/*
97 * sense-id response buffer layout
98 */
99struct senseid {
100	/* common part */
101	u8  reserved;	/* always 0x'FF' */
102	u16 cu_type;	/* control unit type */
103	u8  cu_model;	/* control unit model */
104	u16 dev_type;	/* device type */
105	u8  dev_model;	/* device model */
106	u8  unused;	/* padding byte */
107	/* extended part */
108	struct ciw ciw[MAX_CIWS];	/* variable # of CIWs */
109}  __attribute__ ((packed, aligned(4)));
110
111enum cdev_todo {
112	CDEV_TODO_NOTHING,
113	CDEV_TODO_ENABLE_CMF,
114	CDEV_TODO_REBIND,
115	CDEV_TODO_REGISTER,
116	CDEV_TODO_UNREG,
117	CDEV_TODO_UNREG_EVAL,
118};
119
120#define FAKE_CMD_IRB	1
121#define FAKE_TM_IRB	2
122
123struct ccw_device_dma_area {
124	struct senseid senseid;	/* SenseID info */
125	struct ccw1 iccws[2];	/* ccws for SNID/SID/SPGID commands */
126	struct irb irb;		/* device status */
127	struct pgid pgid[8];	/* path group IDs per chpid*/
128};
129
130struct ccw_device_private {
131	struct ccw_device *cdev;
132	struct subchannel *sch;
133	int state;		/* device state */
134	atomic_t onoff;
135	struct ccw_dev_id dev_id;	/* device id */
136	struct ccw_request req;		/* internal I/O request */
137	int iretry;
138	u8 pgid_valid_mask;	/* mask of valid PGIDs */
139	u8 pgid_todo_mask;	/* mask of PGIDs to be adjusted */
140	u8 pgid_reset_mask;	/* mask of PGIDs which were reset */
141	u8 path_noirq_mask;	/* mask of paths for which no irq was
142				   received */
143	u8 path_notoper_mask;	/* mask of paths which were found
144				   not operable */
145	u8 path_gone_mask;	/* mask of paths, that became unavailable */
146	u8 path_new_mask;	/* mask of paths, that became available */
147	u8 path_broken_mask;	/* mask of paths, which were found to be
148				   unusable */
149	struct {
150		unsigned int fast:1;	/* post with "channel end" */
151		unsigned int repall:1;	/* report every interrupt status */
152		unsigned int pgroup:1;	/* do path grouping */
153		unsigned int force:1;	/* allow forced online */
154		unsigned int mpath:1;	/* do multipathing */
155	} __attribute__ ((packed)) options;
156	struct {
157		unsigned int esid:1;	    /* Ext. SenseID supported by HW */
158		unsigned int dosense:1;	    /* delayed SENSE required */
159		unsigned int doverify:1;    /* delayed path verification */
160		unsigned int donotify:1;    /* call notify function */
161		unsigned int recog_done:1;  /* dev. recog. complete */
162		unsigned int fake_irb:2;    /* deliver faked irb */
163		unsigned int pgroup:1;	    /* pathgroup is set up */
164		unsigned int mpath:1;	    /* multipathing is set up */
165		unsigned int pgid_unknown:1;/* unknown pgid state */
166		unsigned int initialized:1; /* set if initial reference held */
167	} __attribute__((packed)) flags;
168	unsigned long intparm;	/* user interruption parameter */
169	struct qdio_irq *qdio_data;
170	int async_kill_io_rc;
171	struct work_struct todo_work;
172	enum cdev_todo todo;
173	wait_queue_head_t wait_q;
174	struct timer_list timer;
175	void *cmb;			/* measurement information */
176	struct list_head cmb_list;	/* list of measured devices */
177	u64 cmb_start_time;		/* clock value of cmb reset */
178	void *cmb_wait;			/* deferred cmb enable/disable */
179	struct gen_pool *dma_pool;
180	struct ccw_device_dma_area *dma_area;
181	enum interruption_class int_class;
182};
183
184#endif
185