Lines Matching defs:queue_group

75 	struct pqi_queue_group *queue_group, enum pqi_io_path path,
82 unsigned int cdb_length, struct pqi_queue_group *queue_group,
85 struct scsi_cmnd *scmd, struct pqi_queue_group *queue_group,
89 struct scsi_cmnd *scmd, struct pqi_queue_group *queue_group,
2966 struct pqi_queue_group *queue_group)
3047 return pqi_aio_submit_r1_write_io(ctrl_info, scmd, queue_group,
3051 return pqi_aio_submit_r56_write_io(ctrl_info, scmd, queue_group,
3057 rmd.cdb, rmd.cdb_length, queue_group,
3396 static int pqi_process_io_intr(struct pqi_ctrl_info *ctrl_info, struct pqi_queue_group *queue_group)
3406 oq_ci = queue_group->oq_ci_copy;
3409 oq_pi = readl(queue_group->oq_pi);
3421 response = queue_group->oq_element_array +
3488 queue_group->oq_ci_copy = oq_ci;
3489 writel(oq_ci, queue_group->oq_ci);
3515 struct pqi_queue_group *queue_group;
3517 queue_group = &ctrl_info->queue_groups[PQI_DEFAULT_QUEUE_GROUP];
3518 put_unaligned_le16(queue_group->oq_id, &iu->header.response_queue_id);
3521 spin_lock_irqsave(&queue_group->submit_lock[RAID_PATH], flags);
3523 iq_pi = queue_group->iq_pi_copy[RAID_PATH];
3524 iq_ci = readl(queue_group->iq_ci[RAID_PATH]);
3531 &queue_group->submit_lock[RAID_PATH], flags);
3537 next_element = queue_group->iq_element_array[RAID_PATH] +
3543 queue_group->iq_pi_copy[RAID_PATH] = iq_pi;
3549 writel(iq_pi, queue_group->iq_pi[RAID_PATH]);
3551 spin_unlock_irqrestore(&queue_group->submit_lock[RAID_PATH], flags);
4013 struct pqi_queue_group *queue_group;
4017 queue_group = data;
4018 ctrl_info = queue_group->ctrl_info;
4023 num_io_responses_handled = pqi_process_io_intr(ctrl_info, queue_group);
4038 pqi_start_io(ctrl_info, queue_group, RAID_PATH, NULL);
4039 pqi_start_io(ctrl_info, queue_group, AIO_PATH, NULL);
4122 struct pqi_queue_group *queue_group;
4178 queue_group = &ctrl_info->queue_groups[i];
4179 queue_group->iq_element_array[RAID_PATH] = element_array;
4180 queue_group->iq_element_array_bus_addr[RAID_PATH] =
4186 queue_group->iq_element_array[AIO_PATH] = element_array;
4187 queue_group->iq_element_array_bus_addr[AIO_PATH] =
4196 queue_group = &ctrl_info->queue_groups[i];
4197 queue_group->oq_element_array = element_array;
4198 queue_group->oq_element_array_bus_addr =
4217 queue_group = &ctrl_info->queue_groups[i];
4218 queue_group->iq_ci[RAID_PATH] = next_queue_index;
4219 queue_group->iq_ci_bus_addr[RAID_PATH] =
4226 queue_group->iq_ci[AIO_PATH] = next_queue_index;
4227 queue_group->iq_ci_bus_addr[AIO_PATH] =
4234 queue_group->oq_pi = next_queue_index;
4235 queue_group->oq_pi_bus_addr =
4466 struct pqi_queue_group *queue_group, enum pqi_io_path path,
4480 spin_lock_irqsave(&queue_group->submit_lock[path], flags);
4483 io_request->queue_group = queue_group;
4485 &queue_group->request_list[path]);
4488 iq_pi = queue_group->iq_pi_copy[path];
4491 &queue_group->request_list[path], request_list_entry) {
4501 iq_ci = readl(queue_group->iq_ci[path]);
4507 put_unaligned_le16(queue_group->oq_id,
4510 next_element = queue_group->iq_element_array[path] +
4522 memcpy(queue_group->iq_element_array[path],
4533 if (iq_pi != queue_group->iq_pi_copy[path]) {
4534 queue_group->iq_pi_copy[path] = iq_pi;
4539 writel(iq_pi, queue_group->iq_pi[path]);
4542 spin_unlock_irqrestore(&queue_group->submit_lock[path], flags);
4877 struct pqi_queue_group *queue_group;
4881 queue_group = &ctrl_info->queue_groups[group_number];
4892 put_unaligned_le16(queue_group->iq_id[RAID_PATH],
4895 (u64)queue_group->iq_element_array_bus_addr[RAID_PATH],
4897 put_unaligned_le64((u64)queue_group->iq_ci_bus_addr[RAID_PATH],
4913 queue_group->iq_pi[RAID_PATH] = ctrl_info->iomem_base +
4927 put_unaligned_le16(queue_group->iq_id[AIO_PATH],
4929 put_unaligned_le64((u64)queue_group->
4932 put_unaligned_le64((u64)queue_group->iq_ci_bus_addr[AIO_PATH],
4948 queue_group->iq_pi[AIO_PATH] = ctrl_info->iomem_base +
4963 put_unaligned_le16(queue_group->iq_id[AIO_PATH],
4984 put_unaligned_le16(queue_group->oq_id,
4986 put_unaligned_le64((u64)queue_group->oq_element_array_bus_addr,
4988 put_unaligned_le64((u64)queue_group->oq_pi_bus_addr,
4995 put_unaligned_le16(queue_group->int_msg_num,
5006 queue_group->oq_ci = ctrl_info->iomem_base +
5531 struct pqi_queue_group *queue_group, bool io_high_prio)
5608 pqi_start_io(ctrl_info, queue_group, RAID_PATH, io_request);
5615 struct pqi_queue_group *queue_group)
5621 return pqi_raid_submit_io(ctrl_info, device, scmd, queue_group, io_high_prio);
5668 struct pqi_queue_group *queue_group)
5675 scmd->cmnd, scmd->cmd_len, queue_group, NULL,
5681 unsigned int cdb_length, struct pqi_queue_group *queue_group,
5750 pqi_start_io(ctrl_info, queue_group, AIO_PATH, io_request);
5756 struct scsi_cmnd *scmd, struct pqi_queue_group *queue_group,
5811 pqi_start_io(ctrl_info, queue_group, AIO_PATH, io_request);
5817 struct scsi_cmnd *scmd, struct pqi_queue_group *queue_group,
5878 pqi_start_io(ctrl_info, queue_group, AIO_PATH, io_request);
6009 struct pqi_queue_group *queue_group;
6047 queue_group = &ctrl_info->queue_groups[hw_queue];
6054 rc = pqi_raid_bypass_submit_scsi_cmd(ctrl_info, device, scmd, queue_group);
6061 rc = pqi_raid_submit_scsi_cmd(ctrl_info, device, scmd, queue_group);
6064 rc = pqi_aio_submit_scsi_cmd(ctrl_info, device, scmd, queue_group);
6066 rc = pqi_raid_submit_scsi_cmd(ctrl_info, device, scmd, queue_group);
6084 struct pqi_queue_group *queue_group;
6090 queue_group = &ctrl_info->queue_groups[i];
6092 spin_lock_irqsave(&queue_group->submit_lock[path], flags);
6093 list_for_each_entry(io_request, &queue_group->request_list[path], request_list_entry)
6095 spin_unlock_irqrestore(&queue_group->submit_lock[path], flags);
6107 struct pqi_queue_group *queue_group;
6114 queue_group = &ctrl_info->queue_groups[i];
6116 iq_pi = queue_group->iq_pi_copy[path];
6117 iq_ci = readl(queue_group->iq_ci[path]);
6171 struct pqi_queue_group *queue_group;
6179 queue_group = &ctrl_info->queue_groups[i];
6183 &queue_group->submit_lock[path], flags);
6186 &queue_group->request_list[path],
6208 &queue_group->submit_lock[path], flags);