Lines Matching refs:irb

29 ccw_device_msg_control_check(struct ccw_device *cdev, struct irb *irb)
34 if (!scsw_is_valid_cstat(&irb->scsw) ||
35 !(scsw_cstat(&irb->scsw) & (SCHN_STAT_CHN_DATA_CHK |
44 scsw_dstat(&irb->scsw), scsw_cstat(&irb->scsw));
47 CIO_HEX_EVENT(0, irb, sizeof(struct irb));
73 * Copy valid bits from the extended control word to device irb.
76 ccw_device_accumulate_ecw(struct ccw_device *cdev, struct irb *irb)
83 cdev->private->dma_area->irb.scsw.cmd.ectl = 0;
84 if ((irb->scsw.cmd.stctl & SCSW_STCTL_ALERT_STATUS) &&
85 !(irb->scsw.cmd.stctl & SCSW_STCTL_INTER_STATUS))
86 cdev->private->dma_area->irb.scsw.cmd.ectl = irb->scsw.cmd.ectl;
88 if (!cdev->private->dma_area->irb.scsw.cmd.ectl)
91 memcpy(&cdev->private->dma_area->irb.ecw, irb->ecw, sizeof(irb->ecw));
98 ccw_device_accumulate_esw_valid(struct irb *irb)
100 if (!irb->scsw.cmd.eswf &&
101 (irb->scsw.cmd.stctl == SCSW_STCTL_STATUS_PEND))
103 if (irb->scsw.cmd.stctl ==
105 !(irb->scsw.cmd.actl & SCSW_ACTL_SUSPENDED))
111 * Copy valid bits from the extended status word to device irb.
114 ccw_device_accumulate_esw(struct ccw_device *cdev, struct irb *irb)
116 struct irb *cdev_irb;
119 if (!ccw_device_accumulate_esw_valid(irb))
122 cdev_irb = &cdev->private->dma_area->irb;
125 cdev_irb->esw.esw1.lpum = irb->esw.esw1.lpum;
128 if (irb->scsw.cmd.eswf) {
130 sublog = &irb->esw.esw0.sublog;
137 if (irb->scsw.cmd.cstat & (SCHN_STAT_CHN_DATA_CHK |
158 if (irb->scsw.cmd.cstat & SCHN_STAT_INTF_CTRL_CHK)
159 cdev_irb->esw.esw0.erw.cpt = irb->esw.esw0.erw.cpt;
161 cdev_irb->esw.esw0.erw.fsavf = irb->esw.esw0.erw.fsavf;
164 memcpy(cdev_irb->esw.esw0.faddr, irb->esw.esw0.faddr,
165 sizeof (irb->esw.esw0.faddr));
167 cdev_irb->esw.esw0.erw.fsaf = irb->esw.esw0.erw.fsaf;
170 cdev_irb->esw.esw0.erw.scavf = irb->esw.esw0.erw.scavf;
171 if (irb->esw.esw0.erw.scavf)
173 cdev_irb->esw.esw0.saddr = irb->esw.esw0.saddr;
179 cdev_irb->esw.esw0.erw.auth = irb->esw.esw0.erw.auth;
181 cdev_irb->esw.esw0.erw.pvrf = irb->esw.esw0.erw.pvrf;
182 if (irb->esw.esw0.erw.pvrf)
185 cdev_irb->esw.esw0.erw.cons = irb->esw.esw0.erw.cons;
186 if (irb->esw.esw0.erw.cons)
187 cdev_irb->esw.esw0.erw.scnt = irb->esw.esw0.erw.scnt;
191 * Accumulate status from irb to devstat.
194 ccw_device_accumulate_irb(struct ccw_device *cdev, struct irb *irb)
196 struct irb *cdev_irb;
203 if (!(scsw_stctl(&irb->scsw) & SCSW_STCTL_STATUS_PEND))
207 ccw_device_msg_control_check(cdev, irb);
210 if (scsw_is_valid_pno(&irb->scsw) && scsw_pno(&irb->scsw))
212 /* No irb accumulation for transport mode irbs. */
213 if (scsw_is_tm(&irb->scsw)) {
214 memcpy(&cdev->private->dma_area->irb, irb, sizeof(struct irb));
220 if (!scsw_is_solicited(&irb->scsw))
223 cdev_irb = &cdev->private->dma_area->irb;
230 if (irb->scsw.cmd.fctl & SCSW_FCTL_CLEAR_FUNC)
231 memset(&cdev->private->dma_area->irb, 0, sizeof(struct irb));
234 if (irb->scsw.cmd.fctl & SCSW_FCTL_START_FUNC) {
236 cdev_irb->scsw.cmd.key = irb->scsw.cmd.key;
238 cdev_irb->scsw.cmd.sctl = irb->scsw.cmd.sctl;
240 cdev_irb->scsw.cmd.cc |= irb->scsw.cmd.cc;
242 cdev_irb->scsw.cmd.fmt = irb->scsw.cmd.fmt;
244 cdev_irb->scsw.cmd.pfch = irb->scsw.cmd.pfch;
246 cdev_irb->scsw.cmd.isic = irb->scsw.cmd.isic;
248 cdev_irb->scsw.cmd.alcc = irb->scsw.cmd.alcc;
250 cdev_irb->scsw.cmd.ssi = irb->scsw.cmd.ssi;
254 ccw_device_accumulate_ecw(cdev, irb);
257 cdev_irb->scsw.cmd.fctl |= irb->scsw.cmd.fctl;
259 cdev_irb->scsw.cmd.actl = irb->scsw.cmd.actl;
261 cdev_irb->scsw.cmd.stctl |= irb->scsw.cmd.stctl;
266 if ((irb->scsw.cmd.stctl & SCSW_STCTL_PRIM_STATUS) ||
267 ((irb->scsw.cmd.stctl ==
269 (irb->scsw.cmd.actl & SCSW_ACTL_DEVACT) &&
270 (irb->scsw.cmd.actl & SCSW_ACTL_SCHACT)) ||
271 (irb->scsw.cmd.actl & SCSW_ACTL_SUSPENDED))
272 cdev_irb->scsw.cmd.cpa = irb->scsw.cmd.cpa;
276 if (irb->scsw.cmd.stctl &
279 cdev_irb->scsw.cmd.dstat |= irb->scsw.cmd.dstat;
281 cdev_irb->scsw.cmd.cstat |= irb->scsw.cmd.cstat;
283 if ((irb->scsw.cmd.stctl & SCSW_STCTL_PRIM_STATUS) &&
284 (irb->scsw.cmd.cstat & ~(SCHN_STAT_PCI | SCHN_STAT_INCORR_LEN))
286 cdev_irb->scsw.cmd.count = irb->scsw.cmd.count;
289 ccw_device_accumulate_esw(cdev, irb);
312 ccw_device_do_sense(struct ccw_device *cdev, struct irb *irb)
321 if (scsw_actl(&irb->scsw) & (SCSW_ACTL_DEVACT | SCSW_ACTL_SCHACT))
335 sense_ccw->cda = virt_to_dma32(cdev->private->dma_area->irb.ecw);
349 ccw_device_accumulate_basic_sense(struct ccw_device *cdev, struct irb *irb)
356 if (!(scsw_stctl(&irb->scsw) & SCSW_STCTL_STATUS_PEND))
360 ccw_device_msg_control_check(cdev, irb);
363 if (scsw_is_valid_pno(&irb->scsw) && scsw_pno(&irb->scsw))
366 if (!(irb->scsw.cmd.dstat & DEV_STAT_UNIT_CHECK) &&
367 (irb->scsw.cmd.dstat & DEV_STAT_CHN_END)) {
368 cdev->private->dma_area->irb.esw.esw0.erw.cons = 1;
372 if (ccw_device_accumulate_esw_valid(irb) &&
373 irb->esw.esw0.erw.pvrf)
382 ccw_device_accumulate_and_sense(struct ccw_device *cdev, struct irb *irb)
384 ccw_device_accumulate_irb(cdev, irb);
385 if ((irb->scsw.cmd.actl & (SCSW_ACTL_DEVACT | SCSW_ACTL_SCHACT)) != 0)
389 !(irb->scsw.cmd.dstat & DEV_STAT_UNIT_CHECK)) {
390 cdev->private->dma_area->irb.esw.esw0.erw.cons = 1;
395 ccw_device_do_sense(cdev, irb);