Lines Matching refs:vbi_mode
193 struct cx231xx_dmaqueue *vidq = &dev->vbi_mode.vidq;
196 spin_lock_irqsave(&dev->vbi_mode.slock, flags);
198 spin_unlock_irqrestore(&dev->vbi_mode.slock, flags);
204 struct cx231xx_dmaqueue *vidq = &dev->vbi_mode.vidq;
208 spin_lock_irqsave(&dev->vbi_mode.slock, flags);
209 dev->vbi_mode.bulk_ctl.buf = NULL;
214 spin_unlock_irqrestore(&dev->vbi_mode.slock, flags);
220 struct cx231xx_dmaqueue *vidq = &dev->vbi_mode.vidq;
226 dev->vbi_mode.alt_max_pkt_size[0],
262 struct cx231xx *dev = container_of(vmode, struct cx231xx, vbi_mode);
280 spin_lock_irqsave(&dev->vbi_mode.slock, flags);
281 dev->vbi_mode.bulk_ctl.bulk_copy(dev, urb);
282 spin_unlock_irqrestore(&dev->vbi_mode.slock, flags);
304 dev->vbi_mode.bulk_ctl.nfields = -1;
305 for (i = 0; i < dev->vbi_mode.bulk_ctl.num_bufs; i++) {
306 urb = dev->vbi_mode.bulk_ctl.urb[i];
313 if (dev->vbi_mode.bulk_ctl.transfer_buffer[i]) {
315 kfree(dev->vbi_mode.bulk_ctl.
317 dev->vbi_mode.bulk_ctl.transfer_buffer[i] =
321 dev->vbi_mode.bulk_ctl.urb[i] = NULL;
323 dev->vbi_mode.bulk_ctl.transfer_buffer[i] = NULL;
326 kfree(dev->vbi_mode.bulk_ctl.urb);
327 kfree(dev->vbi_mode.bulk_ctl.transfer_buffer);
329 dev->vbi_mode.bulk_ctl.urb = NULL;
330 dev->vbi_mode.bulk_ctl.transfer_buffer = NULL;
331 dev->vbi_mode.bulk_ctl.num_bufs = 0;
345 struct cx231xx_dmaqueue *dma_q = &dev->vbi_mode.vidq;
359 dev->vbi_mode.end_point_addr));
361 dev->vbi_mode.bulk_ctl.bulk_copy = bulk_copy;
362 dev->vbi_mode.bulk_ctl.num_bufs = num_bufs;
374 dev->vbi_mode.bulk_ctl.urb = kcalloc(num_bufs, sizeof(void *),
376 if (!dev->vbi_mode.bulk_ctl.urb) {
382 dev->vbi_mode.bulk_ctl.transfer_buffer =
384 if (!dev->vbi_mode.bulk_ctl.transfer_buffer) {
387 kfree(dev->vbi_mode.bulk_ctl.urb);
391 dev->vbi_mode.bulk_ctl.max_pkt_size = max_pkt_size;
392 dev->vbi_mode.bulk_ctl.buf = NULL;
394 sb_size = max_packets * dev->vbi_mode.bulk_ctl.max_pkt_size;
397 for (i = 0; i < dev->vbi_mode.bulk_ctl.num_bufs; i++) {
404 dev->vbi_mode.bulk_ctl.urb[i] = urb;
407 dev->vbi_mode.bulk_ctl.transfer_buffer[i] =
409 if (!dev->vbi_mode.bulk_ctl.transfer_buffer[i]) {
417 pipe = usb_rcvbulkpipe(dev->udev, dev->vbi_mode.end_point_addr);
419 dev->vbi_mode.bulk_ctl.transfer_buffer[i],
426 for (i = 0; i < dev->vbi_mode.bulk_ctl.num_bufs; i++) {
427 rc = usb_submit_urb(dev->vbi_mode.bulk_ctl.urb[i], GFP_ATOMIC);
486 dev->vbi_mode.bulk_ctl.buf = NULL;
508 buf = dev->vbi_mode.bulk_ctl.buf;
568 struct cx231xx *dev = container_of(vmode, struct cx231xx, vbi_mode);
573 dev->vbi_mode.bulk_ctl.buf = NULL;
585 dev->vbi_mode.bulk_ctl.buf = *buf;
595 buf = dev->vbi_mode.bulk_ctl.buf;
619 buf = dev->vbi_mode.bulk_ctl.buf;