Lines Matching defs:queue_group
61 struct pqi_queue_group *queue_group, enum pqi_io_path path,
68 unsigned int cdb_length, struct pqi_queue_group *queue_group,
2246 struct pqi_queue_group *queue_group)
2567 cdb, cdb_length, queue_group, encryption_info_ptr, true);
2890 static int pqi_process_io_intr(struct pqi_ctrl_info *ctrl_info, struct pqi_queue_group *queue_group)
2900 oq_ci = queue_group->oq_ci_copy;
2903 oq_pi = readl(queue_group->oq_pi);
2915 response = queue_group->oq_element_array +
2984 queue_group->oq_ci_copy = oq_ci;
2985 writel(oq_ci, queue_group->oq_ci);
3011 struct pqi_queue_group *queue_group;
3013 queue_group = &ctrl_info->queue_groups[PQI_DEFAULT_QUEUE_GROUP];
3014 put_unaligned_le16(queue_group->oq_id, &iu->header.response_queue_id);
3017 spin_lock_irqsave(&queue_group->submit_lock[RAID_PATH], flags);
3019 iq_pi = queue_group->iq_pi_copy[RAID_PATH];
3020 iq_ci = readl(queue_group->iq_ci[RAID_PATH]);
3027 &queue_group->submit_lock[RAID_PATH], flags);
3033 next_element = queue_group->iq_element_array[RAID_PATH] +
3039 queue_group->iq_pi_copy[RAID_PATH] = iq_pi;
3045 writel(iq_pi, queue_group->iq_pi[RAID_PATH]);
3047 spin_unlock_irqrestore(&queue_group->submit_lock[RAID_PATH], flags);
3445 struct pqi_queue_group *queue_group;
3449 queue_group = data;
3450 ctrl_info = queue_group->ctrl_info;
3455 num_io_responses_handled = pqi_process_io_intr(ctrl_info, queue_group);
3470 pqi_start_io(ctrl_info, queue_group, RAID_PATH, NULL);
3471 pqi_start_io(ctrl_info, queue_group, AIO_PATH, NULL);
3550 struct pqi_queue_group *queue_group;
3606 queue_group = &ctrl_info->queue_groups[i];
3607 queue_group->iq_element_array[RAID_PATH] = element_array;
3608 queue_group->iq_element_array_bus_addr[RAID_PATH] =
3614 queue_group->iq_element_array[AIO_PATH] = element_array;
3615 queue_group->iq_element_array_bus_addr[AIO_PATH] =
3624 queue_group = &ctrl_info->queue_groups[i];
3625 queue_group->oq_element_array = element_array;
3626 queue_group->oq_element_array_bus_addr =
3645 queue_group = &ctrl_info->queue_groups[i];
3646 queue_group->iq_ci[RAID_PATH] = next_queue_index;
3647 queue_group->iq_ci_bus_addr[RAID_PATH] =
3654 queue_group->iq_ci[AIO_PATH] = next_queue_index;
3655 queue_group->iq_ci_bus_addr[AIO_PATH] =
3662 queue_group->oq_pi = next_queue_index;
3663 queue_group->oq_pi_bus_addr =
3892 struct pqi_queue_group *queue_group, enum pqi_io_path path,
3906 spin_lock_irqsave(&queue_group->submit_lock[path], flags);
3909 io_request->queue_group = queue_group;
3911 &queue_group->request_list[path]);
3914 iq_pi = queue_group->iq_pi_copy[path];
3917 &queue_group->request_list[path], request_list_entry) {
3927 iq_ci = readl(queue_group->iq_ci[path]);
3933 put_unaligned_le16(queue_group->oq_id,
3936 next_element = queue_group->iq_element_array[path] +
3948 memcpy(queue_group->iq_element_array[path],
3959 if (iq_pi != queue_group->iq_pi_copy[path]) {
3960 queue_group->iq_pi_copy[path] = iq_pi;
3965 writel(iq_pi, queue_group->iq_pi[path]);
3968 spin_unlock_irqrestore(&queue_group->submit_lock[path], flags);
4338 struct pqi_queue_group *queue_group;
4342 queue_group = &ctrl_info->queue_groups[group_number];
4353 put_unaligned_le16(queue_group->iq_id[RAID_PATH],
4356 (u64)queue_group->iq_element_array_bus_addr[RAID_PATH],
4358 put_unaligned_le64((u64)queue_group->iq_ci_bus_addr[RAID_PATH],
4374 queue_group->iq_pi[RAID_PATH] = ctrl_info->iomem_base +
4388 put_unaligned_le16(queue_group->iq_id[AIO_PATH],
4390 put_unaligned_le64((u64)queue_group->
4393 put_unaligned_le64((u64)queue_group->iq_ci_bus_addr[AIO_PATH],
4409 queue_group->iq_pi[AIO_PATH] = ctrl_info->iomem_base +
4424 put_unaligned_le16(queue_group->iq_id[AIO_PATH],
4445 put_unaligned_le16(queue_group->oq_id,
4447 put_unaligned_le64((u64)queue_group->oq_element_array_bus_addr,
4449 put_unaligned_le64((u64)queue_group->oq_pi_bus_addr,
4456 put_unaligned_le16(queue_group->int_msg_num,
4467 queue_group->oq_ci = ctrl_info->iomem_base +
4943 struct pqi_queue_group *queue_group)
5025 pqi_start_io(ctrl_info, queue_group, RAID_PATH, io_request);
5032 struct pqi_queue_group *queue_group)
5039 device, scmd, queue_group);
5119 struct pqi_queue_group *queue_group;
5131 queue_group = io_request->queue_group;
5136 device, scmd, queue_group);
5219 struct pqi_queue_group *queue_group)
5222 scmd->cmnd, scmd->cmd_len, queue_group, NULL, false);
5227 unsigned int cdb_length, struct pqi_queue_group *queue_group,
5290 pqi_start_io(ctrl_info, queue_group, AIO_PATH, io_request);
5337 struct pqi_queue_group *queue_group;
5372 queue_group = &ctrl_info->queue_groups[hw_queue];
5379 scmd, queue_group);
5386 rc = pqi_raid_submit_scsi_cmd(ctrl_info, device, scmd, queue_group);
5389 rc = pqi_aio_submit_scsi_cmd(ctrl_info, device, scmd, queue_group);
5391 rc = pqi_raid_submit_scsi_cmd(ctrl_info, device, scmd, queue_group);
5403 struct pqi_queue_group *queue_group)
5412 &queue_group->submit_lock[path], flags);
5414 list_empty(&queue_group->request_list[path]);
5416 &queue_group->submit_lock[path], flags);
5434 struct pqi_queue_group *queue_group;
5439 queue_group = &ctrl_info->queue_groups[i];
5441 rc = pqi_wait_until_queued_io_drained(ctrl_info, queue_group);
5446 iq_pi = queue_group->iq_pi_copy[path];
5449 iq_ci = readl(queue_group->iq_ci[path]);
5468 struct pqi_queue_group *queue_group;
5476 queue_group = &ctrl_info->queue_groups[i];
5480 &queue_group->submit_lock[path], flags);
5483 &queue_group->request_list[path],
5501 &queue_group->submit_lock[path], flags);
5510 struct pqi_queue_group *queue_group;
5517 queue_group = &ctrl_info->queue_groups[i];
5520 spin_lock_irqsave(&queue_group->submit_lock[path],
5524 &queue_group->request_list[path],
5539 &queue_group->submit_lock[path], flags);