Lines Matching refs:irb

28 ccw_device_msg_control_check(struct ccw_device *cdev, struct irb *irb)
33 if (!scsw_is_valid_cstat(&irb->scsw) ||
34 !(scsw_cstat(&irb->scsw) & (SCHN_STAT_CHN_DATA_CHK |
43 scsw_dstat(&irb->scsw), scsw_cstat(&irb->scsw));
46 CIO_HEX_EVENT(0, irb, sizeof(struct irb));
72 * Copy valid bits from the extended control word to device irb.
75 ccw_device_accumulate_ecw(struct ccw_device *cdev, struct irb *irb)
82 cdev->private->dma_area->irb.scsw.cmd.ectl = 0;
83 if ((irb->scsw.cmd.stctl & SCSW_STCTL_ALERT_STATUS) &&
84 !(irb->scsw.cmd.stctl & SCSW_STCTL_INTER_STATUS))
85 cdev->private->dma_area->irb.scsw.cmd.ectl = irb->scsw.cmd.ectl;
87 if (!cdev->private->dma_area->irb.scsw.cmd.ectl)
90 memcpy(&cdev->private->dma_area->irb.ecw, irb->ecw, sizeof(irb->ecw));
97 ccw_device_accumulate_esw_valid(struct irb *irb)
99 if (!irb->scsw.cmd.eswf &&
100 (irb->scsw.cmd.stctl == SCSW_STCTL_STATUS_PEND))
102 if (irb->scsw.cmd.stctl ==
104 !(irb->scsw.cmd.actl & SCSW_ACTL_SUSPENDED))
110 * Copy valid bits from the extended status word to device irb.
113 ccw_device_accumulate_esw(struct ccw_device *cdev, struct irb *irb)
115 struct irb *cdev_irb;
118 if (!ccw_device_accumulate_esw_valid(irb))
121 cdev_irb = &cdev->private->dma_area->irb;
124 cdev_irb->esw.esw1.lpum = irb->esw.esw1.lpum;
127 if (irb->scsw.cmd.eswf) {
129 sublog = &irb->esw.esw0.sublog;
136 if (irb->scsw.cmd.cstat & (SCHN_STAT_CHN_DATA_CHK |
157 if (irb->scsw.cmd.cstat & SCHN_STAT_INTF_CTRL_CHK)
158 cdev_irb->esw.esw0.erw.cpt = irb->esw.esw0.erw.cpt;
160 cdev_irb->esw.esw0.erw.fsavf = irb->esw.esw0.erw.fsavf;
163 memcpy(cdev_irb->esw.esw0.faddr, irb->esw.esw0.faddr,
164 sizeof (irb->esw.esw0.faddr));
166 cdev_irb->esw.esw0.erw.fsaf = irb->esw.esw0.erw.fsaf;
169 cdev_irb->esw.esw0.erw.scavf = irb->esw.esw0.erw.scavf;
170 if (irb->esw.esw0.erw.scavf)
172 cdev_irb->esw.esw0.saddr = irb->esw.esw0.saddr;
178 cdev_irb->esw.esw0.erw.auth = irb->esw.esw0.erw.auth;
180 cdev_irb->esw.esw0.erw.pvrf = irb->esw.esw0.erw.pvrf;
181 if (irb->esw.esw0.erw.pvrf)
184 cdev_irb->esw.esw0.erw.cons = irb->esw.esw0.erw.cons;
185 if (irb->esw.esw0.erw.cons)
186 cdev_irb->esw.esw0.erw.scnt = irb->esw.esw0.erw.scnt;
190 * Accumulate status from irb to devstat.
193 ccw_device_accumulate_irb(struct ccw_device *cdev, struct irb *irb)
195 struct irb *cdev_irb;
202 if (!(scsw_stctl(&irb->scsw) & SCSW_STCTL_STATUS_PEND))
206 ccw_device_msg_control_check(cdev, irb);
209 if (scsw_is_valid_pno(&irb->scsw) && scsw_pno(&irb->scsw))
211 /* No irb accumulation for transport mode irbs. */
212 if (scsw_is_tm(&irb->scsw)) {
213 memcpy(&cdev->private->dma_area->irb, irb, sizeof(struct irb));
219 if (!scsw_is_solicited(&irb->scsw))
222 cdev_irb = &cdev->private->dma_area->irb;
229 if (irb->scsw.cmd.fctl & SCSW_FCTL_CLEAR_FUNC)
230 memset(&cdev->private->dma_area->irb, 0, sizeof(struct irb));
233 if (irb->scsw.cmd.fctl & SCSW_FCTL_START_FUNC) {
235 cdev_irb->scsw.cmd.key = irb->scsw.cmd.key;
237 cdev_irb->scsw.cmd.sctl = irb->scsw.cmd.sctl;
239 cdev_irb->scsw.cmd.cc |= irb->scsw.cmd.cc;
241 cdev_irb->scsw.cmd.fmt = irb->scsw.cmd.fmt;
243 cdev_irb->scsw.cmd.pfch = irb->scsw.cmd.pfch;
245 cdev_irb->scsw.cmd.isic = irb->scsw.cmd.isic;
247 cdev_irb->scsw.cmd.alcc = irb->scsw.cmd.alcc;
249 cdev_irb->scsw.cmd.ssi = irb->scsw.cmd.ssi;
253 ccw_device_accumulate_ecw(cdev, irb);
256 cdev_irb->scsw.cmd.fctl |= irb->scsw.cmd.fctl;
258 cdev_irb->scsw.cmd.actl = irb->scsw.cmd.actl;
260 cdev_irb->scsw.cmd.stctl |= irb->scsw.cmd.stctl;
265 if ((irb->scsw.cmd.stctl & SCSW_STCTL_PRIM_STATUS) ||
266 ((irb->scsw.cmd.stctl ==
268 (irb->scsw.cmd.actl & SCSW_ACTL_DEVACT) &&
269 (irb->scsw.cmd.actl & SCSW_ACTL_SCHACT)) ||
270 (irb->scsw.cmd.actl & SCSW_ACTL_SUSPENDED))
271 cdev_irb->scsw.cmd.cpa = irb->scsw.cmd.cpa;
275 if (irb->scsw.cmd.stctl &
278 cdev_irb->scsw.cmd.dstat |= irb->scsw.cmd.dstat;
280 cdev_irb->scsw.cmd.cstat |= irb->scsw.cmd.cstat;
282 if ((irb->scsw.cmd.stctl & SCSW_STCTL_PRIM_STATUS) &&
283 (irb->scsw.cmd.cstat & ~(SCHN_STAT_PCI | SCHN_STAT_INCORR_LEN))
285 cdev_irb->scsw.cmd.count = irb->scsw.cmd.count;
288 ccw_device_accumulate_esw(cdev, irb);
311 ccw_device_do_sense(struct ccw_device *cdev, struct irb *irb)
320 if (scsw_actl(&irb->scsw) & (SCSW_ACTL_DEVACT | SCSW_ACTL_SCHACT))
334 sense_ccw->cda = (__u32) __pa(cdev->private->dma_area->irb.ecw);
348 ccw_device_accumulate_basic_sense(struct ccw_device *cdev, struct irb *irb)
355 if (!(scsw_stctl(&irb->scsw) & SCSW_STCTL_STATUS_PEND))
359 ccw_device_msg_control_check(cdev, irb);
362 if (scsw_is_valid_pno(&irb->scsw) && scsw_pno(&irb->scsw))
365 if (!(irb->scsw.cmd.dstat & DEV_STAT_UNIT_CHECK) &&
366 (irb->scsw.cmd.dstat & DEV_STAT_CHN_END)) {
367 cdev->private->dma_area->irb.esw.esw0.erw.cons = 1;
371 if (ccw_device_accumulate_esw_valid(irb) &&
372 irb->esw.esw0.erw.pvrf)
381 ccw_device_accumulate_and_sense(struct ccw_device *cdev, struct irb *irb)
383 ccw_device_accumulate_irb(cdev, irb);
384 if ((irb->scsw.cmd.actl & (SCSW_ACTL_DEVACT | SCSW_ACTL_SCHACT)) != 0)
388 !(irb->scsw.cmd.dstat & DEV_STAT_UNIT_CHECK)) {
389 cdev->private->dma_area->irb.esw.esw0.erw.cons = 1;
394 ccw_device_do_sense(cdev, irb);