Lines Matching refs:vbi_mode
105 buf = dev->vbi_mode.bulk_ctl.buf; in cx231xx_isoc_vbi_copy()
211 spin_lock_irqsave(&dev->vbi_mode.slock, flags); in free_buffer()
212 if (dev->vbi_mode.bulk_ctl.buf == buf) in free_buffer()
213 dev->vbi_mode.bulk_ctl.buf = NULL; in free_buffer()
214 spin_unlock_irqrestore(&dev->vbi_mode.slock, flags); in free_buffer()
249 if (!dev->vbi_mode.bulk_ctl.num_bufs) in vbi_buffer_prepare()
255 dev->vbi_mode.alt_max_pkt_size[0], in vbi_buffer_prepare()
276 struct cx231xx_dmaqueue *vidq = &dev->vbi_mode.vidq; in vbi_buffer_queue()
312 struct cx231xx *dev = container_of(vmode, struct cx231xx, vbi_mode); in cx231xx_irq_vbi_callback()
330 spin_lock(&dev->vbi_mode.slock); in cx231xx_irq_vbi_callback()
331 rc = dev->vbi_mode.bulk_ctl.bulk_copy(dev, urb); in cx231xx_irq_vbi_callback()
332 spin_unlock(&dev->vbi_mode.slock); in cx231xx_irq_vbi_callback()
354 dev->vbi_mode.bulk_ctl.nfields = -1; in cx231xx_uninit_vbi_isoc()
355 for (i = 0; i < dev->vbi_mode.bulk_ctl.num_bufs; i++) { in cx231xx_uninit_vbi_isoc()
356 urb = dev->vbi_mode.bulk_ctl.urb[i]; in cx231xx_uninit_vbi_isoc()
363 if (dev->vbi_mode.bulk_ctl.transfer_buffer[i]) { in cx231xx_uninit_vbi_isoc()
365 kfree(dev->vbi_mode.bulk_ctl. in cx231xx_uninit_vbi_isoc()
367 dev->vbi_mode.bulk_ctl.transfer_buffer[i] = in cx231xx_uninit_vbi_isoc()
371 dev->vbi_mode.bulk_ctl.urb[i] = NULL; in cx231xx_uninit_vbi_isoc()
373 dev->vbi_mode.bulk_ctl.transfer_buffer[i] = NULL; in cx231xx_uninit_vbi_isoc()
376 kfree(dev->vbi_mode.bulk_ctl.urb); in cx231xx_uninit_vbi_isoc()
377 kfree(dev->vbi_mode.bulk_ctl.transfer_buffer); in cx231xx_uninit_vbi_isoc()
379 dev->vbi_mode.bulk_ctl.urb = NULL; in cx231xx_uninit_vbi_isoc()
380 dev->vbi_mode.bulk_ctl.transfer_buffer = NULL; in cx231xx_uninit_vbi_isoc()
381 dev->vbi_mode.bulk_ctl.num_bufs = 0; in cx231xx_uninit_vbi_isoc()
395 struct cx231xx_dmaqueue *dma_q = &dev->vbi_mode.vidq; in cx231xx_init_vbi_isoc()
409 dev->vbi_mode.end_point_addr)); in cx231xx_init_vbi_isoc()
411 dev->vbi_mode.bulk_ctl.bulk_copy = bulk_copy; in cx231xx_init_vbi_isoc()
412 dev->vbi_mode.bulk_ctl.num_bufs = num_bufs; in cx231xx_init_vbi_isoc()
424 dev->vbi_mode.bulk_ctl.urb = kzalloc(sizeof(void *) * num_bufs, in cx231xx_init_vbi_isoc()
426 if (!dev->vbi_mode.bulk_ctl.urb) { in cx231xx_init_vbi_isoc()
431 dev->vbi_mode.bulk_ctl.transfer_buffer = in cx231xx_init_vbi_isoc()
433 if (!dev->vbi_mode.bulk_ctl.transfer_buffer) { in cx231xx_init_vbi_isoc()
435 kfree(dev->vbi_mode.bulk_ctl.urb); in cx231xx_init_vbi_isoc()
439 dev->vbi_mode.bulk_ctl.max_pkt_size = max_pkt_size; in cx231xx_init_vbi_isoc()
440 dev->vbi_mode.bulk_ctl.buf = NULL; in cx231xx_init_vbi_isoc()
442 sb_size = max_packets * dev->vbi_mode.bulk_ctl.max_pkt_size; in cx231xx_init_vbi_isoc()
445 for (i = 0; i < dev->vbi_mode.bulk_ctl.num_bufs; i++) { in cx231xx_init_vbi_isoc()
454 dev->vbi_mode.bulk_ctl.urb[i] = urb; in cx231xx_init_vbi_isoc()
457 dev->vbi_mode.bulk_ctl.transfer_buffer[i] = in cx231xx_init_vbi_isoc()
459 if (!dev->vbi_mode.bulk_ctl.transfer_buffer[i]) { in cx231xx_init_vbi_isoc()
468 pipe = usb_rcvbulkpipe(dev->udev, dev->vbi_mode.end_point_addr); in cx231xx_init_vbi_isoc()
470 dev->vbi_mode.bulk_ctl.transfer_buffer[i], in cx231xx_init_vbi_isoc()
477 for (i = 0; i < dev->vbi_mode.bulk_ctl.num_bufs; i++) { in cx231xx_init_vbi_isoc()
478 rc = usb_submit_urb(dev->vbi_mode.bulk_ctl.urb[i], GFP_ATOMIC); in cx231xx_init_vbi_isoc()
539 dev->vbi_mode.bulk_ctl.buf = NULL; in vbi_buffer_filled()
561 buf = dev->vbi_mode.bulk_ctl.buf; in cx231xx_copy_vbi_line()
621 struct cx231xx *dev = container_of(vmode, struct cx231xx, vbi_mode); in get_next_vbi_buf()
626 dev->vbi_mode.bulk_ctl.buf = NULL; in get_next_vbi_buf()
638 dev->vbi_mode.bulk_ctl.buf = *buf; in get_next_vbi_buf()
648 buf = dev->vbi_mode.bulk_ctl.buf; in cx231xx_reset_vbi_buffer()
672 buf = dev->vbi_mode.bulk_ctl.buf; in cx231xx_do_vbi_copy()