Lines Matching refs:path
61 struct pqi_queue_group *queue_group, enum pqi_io_path path,
2325 /* Process via normal I/O path. */
2643 static inline void pqi_take_device_offline(struct scsi_device *sdev, char *path)
2656 path, ctrl_info->scsi_host->host_no, device->bus,
3892 struct pqi_queue_group *queue_group, enum pqi_io_path path,
3906 spin_lock_irqsave(&queue_group->submit_lock[path], flags);
3911 &queue_group->request_list[path]);
3914 iq_pi = queue_group->iq_pi_copy[path];
3917 &queue_group->request_list[path], request_list_entry) {
3927 iq_ci = readl(queue_group->iq_ci[path]);
3936 next_element = queue_group->iq_element_array[path] +
3948 memcpy(queue_group->iq_element_array[path],
3959 if (iq_pi != queue_group->iq_pi_copy[path]) {
3960 queue_group->iq_pi_copy[path] = iq_pi;
3965 writel(iq_pi, queue_group->iq_pi[path]);
3968 spin_unlock_irqrestore(&queue_group->submit_lock[path], flags);
4346 * RAID path.
4381 * Advanced I/O (AIO) path.
4415 * Designate the 2nd IQ as the AIO path. By default, all IQs are
4416 * assumed to be for RAID path I/O unless we change the queue's
5405 unsigned int path;
5409 for (path = 0; path < 2; path++) {
5412 &queue_group->submit_lock[path], flags);
5414 list_empty(&queue_group->request_list[path]);
5416 &queue_group->submit_lock[path], flags);
5433 unsigned int path;
5445 for (path = 0; path < 2; path++) {
5446 iq_pi = queue_group->iq_pi_copy[path];
5449 iq_ci = readl(queue_group->iq_ci[path]);
5467 unsigned int path;
5478 for (path = 0; path < 2; path++) {
5480 &queue_group->submit_lock[path], flags);
5483 &queue_group->request_list[path],
5501 &queue_group->submit_lock[path], flags);
5509 unsigned int path;
5519 for (path = 0; path < 2; path++) {
5520 spin_lock_irqsave(&queue_group->submit_lock[path],
5524 &queue_group->request_list[path],
5539 &queue_group->submit_lock[path], flags);