WIP FPC-III support
[linux/fpc-iii.git] / drivers / s390 / cio / io_sch.h
blob85a11c1836e5f1b931f72fc87d153616fcc77d1d
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef S390_IO_SCH_H
3 #define S390_IO_SCH_H
5 #include <linux/types.h>
6 #include <asm/schid.h>
7 #include <asm/ccwdev.h>
8 #include <asm/irq.h>
9 #include "css.h"
10 #include "orb.h"
12 struct io_subchannel_dma_area {
13 struct ccw1 sense_ccw; /* static ccw for sense command */
16 struct io_subchannel_private {
17 union orb orb; /* operation request block */
18 struct ccw_device *cdev;/* pointer to the child ccw device */
19 struct {
20 unsigned int suspend:1; /* allow suspend */
21 unsigned int prefetch:1;/* deny prefetch */
22 unsigned int inter:1; /* suppress intermediate interrupts */
23 } __packed options;
24 struct io_subchannel_dma_area *dma_area;
25 dma_addr_t dma_area_dma;
26 } __aligned(8);
28 #define to_io_private(n) ((struct io_subchannel_private *) \
29 dev_get_drvdata(&(n)->dev))
30 #define set_io_private(n, p) (dev_set_drvdata(&(n)->dev, p))
32 static inline struct ccw_device *sch_get_cdev(struct subchannel *sch)
34 struct io_subchannel_private *priv = to_io_private(sch);
35 return priv ? priv->cdev : NULL;
38 static inline void sch_set_cdev(struct subchannel *sch,
39 struct ccw_device *cdev)
41 struct io_subchannel_private *priv = to_io_private(sch);
42 if (priv)
43 priv->cdev = cdev;
46 #define MAX_CIWS 8
49 * Possible status values for a CCW request's I/O.
51 enum io_status {
52 IO_DONE,
53 IO_RUNNING,
54 IO_STATUS_ERROR,
55 IO_PATH_ERROR,
56 IO_REJECTED,
57 IO_KILLED
60 /**
61 * ccw_request - Internal CCW request.
62 * @cp: channel program to start
63 * @timeout: maximum allowable time in jiffies between start I/O and interrupt
64 * @maxretries: number of retries per I/O operation and path
65 * @lpm: mask of paths to use
66 * @check: optional callback that determines if results are final
67 * @filter: optional callback to adjust request status based on IRB data
68 * @callback: final callback
69 * @data: user-defined pointer passed to all callbacks
70 * @singlepath: if set, use only one path from @lpm per start I/O
71 * @cancel: non-zero if request was cancelled
72 * @done: non-zero if request was finished
73 * @mask: current path mask
74 * @retries: current number of retries
75 * @drc: delayed return code
77 struct ccw_request {
78 struct ccw1 *cp;
79 unsigned long timeout;
80 u16 maxretries;
81 u8 lpm;
82 int (*check)(struct ccw_device *, void *);
83 enum io_status (*filter)(struct ccw_device *, void *, struct irb *,
84 enum io_status);
85 void (*callback)(struct ccw_device *, void *, int);
86 void *data;
87 unsigned int singlepath:1;
88 /* These fields are used internally. */
89 unsigned int cancel:1;
90 unsigned int done:1;
91 u16 mask;
92 u16 retries;
93 int drc;
94 } __attribute__((packed));
97 * sense-id response buffer layout
99 struct senseid {
100 /* common part */
101 u8 reserved; /* always 0x'FF' */
102 u16 cu_type; /* control unit type */
103 u8 cu_model; /* control unit model */
104 u16 dev_type; /* device type */
105 u8 dev_model; /* device model */
106 u8 unused; /* padding byte */
107 /* extended part */
108 struct ciw ciw[MAX_CIWS]; /* variable # of CIWs */
109 } __attribute__ ((packed, aligned(4)));
111 enum cdev_todo {
112 CDEV_TODO_NOTHING,
113 CDEV_TODO_ENABLE_CMF,
114 CDEV_TODO_REBIND,
115 CDEV_TODO_REGISTER,
116 CDEV_TODO_UNREG,
117 CDEV_TODO_UNREG_EVAL,
120 #define FAKE_CMD_IRB 1
121 #define FAKE_TM_IRB 2
123 struct ccw_device_dma_area {
124 struct senseid senseid; /* SenseID info */
125 struct ccw1 iccws[2]; /* ccws for SNID/SID/SPGID commands */
126 struct irb irb; /* device status */
127 struct pgid pgid[8]; /* path group IDs per chpid*/
130 struct ccw_device_private {
131 struct ccw_device *cdev;
132 struct subchannel *sch;
133 int state; /* device state */
134 atomic_t onoff;
135 struct ccw_dev_id dev_id; /* device id */
136 struct ccw_request req; /* internal I/O request */
137 int iretry;
138 u8 pgid_valid_mask; /* mask of valid PGIDs */
139 u8 pgid_todo_mask; /* mask of PGIDs to be adjusted */
140 u8 pgid_reset_mask; /* mask of PGIDs which were reset */
141 u8 path_noirq_mask; /* mask of paths for which no irq was
142 received */
143 u8 path_notoper_mask; /* mask of paths which were found
144 not operable */
145 u8 path_gone_mask; /* mask of paths, that became unavailable */
146 u8 path_new_mask; /* mask of paths, that became available */
147 u8 path_broken_mask; /* mask of paths, which were found to be
148 unusable */
149 struct {
150 unsigned int fast:1; /* post with "channel end" */
151 unsigned int repall:1; /* report every interrupt status */
152 unsigned int pgroup:1; /* do path grouping */
153 unsigned int force:1; /* allow forced online */
154 unsigned int mpath:1; /* do multipathing */
155 } __attribute__ ((packed)) options;
156 struct {
157 unsigned int esid:1; /* Ext. SenseID supported by HW */
158 unsigned int dosense:1; /* delayed SENSE required */
159 unsigned int doverify:1; /* delayed path verification */
160 unsigned int donotify:1; /* call notify function */
161 unsigned int recog_done:1; /* dev. recog. complete */
162 unsigned int fake_irb:2; /* deliver faked irb */
163 unsigned int pgroup:1; /* pathgroup is set up */
164 unsigned int mpath:1; /* multipathing is set up */
165 unsigned int pgid_unknown:1;/* unknown pgid state */
166 unsigned int initialized:1; /* set if initial reference held */
167 } __attribute__((packed)) flags;
168 unsigned long intparm; /* user interruption parameter */
169 struct qdio_irq *qdio_data;
170 int async_kill_io_rc;
171 struct work_struct todo_work;
172 enum cdev_todo todo;
173 wait_queue_head_t wait_q;
174 struct timer_list timer;
175 void *cmb; /* measurement information */
176 struct list_head cmb_list; /* list of measured devices */
177 u64 cmb_start_time; /* clock value of cmb reset */
178 void *cmb_wait; /* deferred cmb enable/disable */
179 struct gen_pool *dma_pool;
180 struct ccw_device_dma_area *dma_area;
181 enum interruption_class int_class;
184 #endif