Lines Matching refs:xport
33 struct efct_xport *xport;
39 xport = efct->xport;
41 io = efct_io_pool_io_alloc(efct->xport->io_pool);
44 atomic_add_return(1, &xport->io_alloc_failed_count);
95 efct_io_pool_io_free(efct->xport->io_pool, io);
391 struct efct_xport *xport = efct->xport;
397 spin_lock_irqsave(&xport->io_pending_lock, flags);
399 if (!list_empty(&xport->io_pending_list)) {
400 io = list_first_entry(&xport->io_pending_list, struct efct_io,
406 spin_unlock_irqrestore(&xport->io_pending_lock, flags);
419 list_add(&xport->io_pending_list, &io->io_pending_link);
427 spin_unlock_irqrestore(&xport->io_pending_lock, flags);
436 atomic_sub_return(1, &xport->io_pending_count);
460 struct efct_xport *xport = efct->xport;
467 if (atomic_add_return(1, &xport->io_pending_recursing)) {
469 atomic_sub_return(1, &xport->io_pending_recursing);
477 atomic_sub_return(1, &xport->io_pending_recursing);
488 spin_lock_irqsave(&xport->io_pending_lock, flags);
490 list_for_each_entry(io, &xport->io_pending_list, io_pending_link) {
497 atomic_sub_return(1, &xport->io_pending_count);
502 spin_unlock_irqrestore(&xport->io_pending_lock, flags);
513 atomic_sub_return(1, &xport->io_pending_recursing);
521 struct efct_xport *xport = efct->xport;
538 spin_lock_irqsave(&xport->io_pending_lock, flags);
539 if (!list_empty(&xport->io_pending_list)) {
547 list_add(&xport->io_pending_list, &io->io_pending_link);
551 &xport->io_pending_list);
553 spin_unlock_irqrestore(&xport->io_pending_lock, flags);
554 atomic_add_return(1, &xport->io_pending_count);
555 atomic_add_return(1, &xport->io_total_pending);
561 spin_unlock_irqrestore(&xport->io_pending_lock, flags);
570 spin_lock_irqsave(&xport->io_pending_lock, flags);
572 list_add_tail(&io->io_pending_link, &xport->io_pending_list);
573 spin_unlock_irqrestore(&xport->io_pending_lock, flags);
575 atomic_add_return(1, &xport->io_total_pending);
576 atomic_add_return(1, &xport->io_pending_count);
588 struct efct_xport *xport = efct->xport;
599 spin_lock_irqsave(&xport->io_pending_lock, flags);
600 if (!list_empty(&xport->io_pending_list)) {
602 list_add_tail(&io->io_pending_link, &xport->io_pending_list);
603 spin_unlock_irqrestore(&xport->io_pending_lock, flags);
604 atomic_add_return(1, &xport->io_pending_count);
605 atomic_add_return(1, &xport->io_total_pending);
611 spin_unlock_irqrestore(&xport->io_pending_lock, flags);
696 if (efct->xport) {
697 struct efct_xport *xport = efct->xport;
700 xport->fcp_stats.input_requests++;
701 xport->fcp_stats.input_bytes += xwire_len;
703 xport->fcp_stats.output_requests++;
704 xport->fcp_stats.output_bytes += xwire_len;
1090 efct_io_pool_io_free(efct->xport->io_pool, io);
1100 struct efct_xport *xport;
1105 xport = efct->xport;
1119 abort_io = efct_io_pool_io_alloc(efct->xport->io_pool);
1121 atomic_add_return(1, &xport->io_alloc_failed_count);