Lines Matching refs:sch

59 static int eadm_subchannel_start(struct subchannel *sch, struct aob *aob)
61 union orb *orb = &get_eadm_private(sch)->orb;
66 orb->eadm.intparm = (u32)(addr_t)sch;
70 EADM_LOG_HEX(6, &sch->schid, sizeof(sch->schid));
72 cc = ssch(sch->schid, orb);
75 sch->schib.scsw.eadm.actl |= SCSW_ACTL_START_PEND;
86 static int eadm_subchannel_clear(struct subchannel *sch)
90 cc = csch(sch->schid);
94 sch->schib.scsw.eadm.actl |= SCSW_ACTL_CLEAR_PEND;
101 struct subchannel *sch = private->sch;
103 spin_lock_irq(sch->lock);
105 EADM_LOG_HEX(1, &sch->schid, sizeof(sch->schid));
106 if (eadm_subchannel_clear(sch))
108 spin_unlock_irq(sch->lock);
111 static void eadm_subchannel_set_timeout(struct subchannel *sch, int expires)
113 struct eadm_private *private = get_eadm_private(sch);
127 static void eadm_subchannel_irq(struct subchannel *sch)
129 struct eadm_private *private = get_eadm_private(sch);
130 struct eadm_scsw *scsw = &sch->schib.scsw.eadm;
146 eadm_subchannel_set_timeout(sch, 0);
152 css_sched_sch_todo(sch, SCH_TODO_EVAL);
165 struct subchannel *sch;
170 sch = private->sch;
171 spin_lock(sch->lock);
175 spin_unlock(sch->lock);
178 return sch;
180 spin_unlock(sch->lock);
190 struct subchannel *sch;
194 sch = eadm_get_idle_sch();
195 if (!sch)
198 spin_lock_irqsave(sch->lock, flags);
199 eadm_subchannel_set_timeout(sch, EADM_TIMEOUT);
200 ret = eadm_subchannel_start(sch, aob);
205 eadm_subchannel_set_timeout(sch, 0);
206 private = get_eadm_private(sch);
208 css_sched_sch_todo(sch, SCH_TODO_EVAL);
211 spin_unlock_irqrestore(sch->lock, flags);
217 static int eadm_subchannel_probe(struct subchannel *sch)
229 spin_lock_irq(sch->lock);
230 set_eadm_private(sch, private);
232 private->sch = sch;
233 sch->isc = EADM_SCH_ISC;
234 ret = cio_enable_subchannel(sch, (u32)(unsigned long)sch);
236 set_eadm_private(sch, NULL);
237 spin_unlock_irq(sch->lock);
241 spin_unlock_irq(sch->lock);
247 if (dev_get_uevent_suppress(&sch->dev)) {
248 dev_set_uevent_suppress(&sch->dev, 0);
249 kobject_uevent(&sch->dev.kobj, KOBJ_ADD);
255 static void eadm_quiesce(struct subchannel *sch)
257 struct eadm_private *private = get_eadm_private(sch);
261 spin_lock_irq(sch->lock);
265 if (eadm_subchannel_clear(sch))
269 spin_unlock_irq(sch->lock);
273 spin_lock_irq(sch->lock);
277 eadm_subchannel_set_timeout(sch, 0);
279 ret = cio_disable_subchannel(sch);
282 spin_unlock_irq(sch->lock);
285 static int eadm_subchannel_remove(struct subchannel *sch)
287 struct eadm_private *private = get_eadm_private(sch);
293 eadm_quiesce(sch);
295 spin_lock_irq(sch->lock);
296 set_eadm_private(sch, NULL);
297 spin_unlock_irq(sch->lock);
304 static void eadm_subchannel_shutdown(struct subchannel *sch)
306 eadm_quiesce(sch);
309 static int eadm_subchannel_freeze(struct subchannel *sch)
311 return cio_disable_subchannel(sch);
314 static int eadm_subchannel_restore(struct subchannel *sch)
316 return cio_enable_subchannel(sch, (u32)(unsigned long)sch);
321 * @sch: subchannel
329 static int eadm_subchannel_sch_event(struct subchannel *sch, int process)
334 spin_lock_irqsave(sch->lock, flags);
335 if (!device_is_registered(&sch->dev))
338 if (work_pending(&sch->todo_work))
341 if (cio_update_schib(sch)) {
342 css_sched_sch_todo(sch, SCH_TODO_UNREG);
345 private = get_eadm_private(sch);
350 spin_unlock_irqrestore(sch->lock, flags);