Lines Matching refs:raid

573 	/* create sense buffer for the raid 1/10 fp */
1495 struct MR_LD_RAID *raid;
1531 raid = MR_LdRaidGet(i, map);
1533 ld_sync->seqNum = raid->seqNum;
2389 struct MR_LD_RAID *raid;
2398 raid = MR_LdRaidGet(ld, local_map_ptr);
2399 if (raid->capability.ldPiMode == MR_PROT_INFO_TYPE_CONTROLLER) {
2582 /* if we found a stream, update the raid
2635 * affinity (cpu of the controller) and raid_flags in the raid context
2640 * @raid: LD raid map
2649 struct MR_LD_RAID *raid, bool fp_possible,
2658 if ((raid->cpuAffinity.pdRead.cpu0) &&
2659 (raid->cpuAffinity.pdRead.cpu1))
2661 else if (raid->cpuAffinity.pdRead.cpu1)
2664 if ((raid->cpuAffinity.pdWrite.cpu0) &&
2665 (raid->cpuAffinity.pdWrite.cpu1))
2667 else if (raid->cpuAffinity.pdWrite.cpu1)
2670 if ((raid->level <= 1) &&
2671 (raid->capability.fp_cache_bypass_capable)) {
2681 if ((raid->cpuAffinity.ldRead.cpu0) &&
2682 (raid->cpuAffinity.ldRead.cpu1))
2684 else if (raid->cpuAffinity.ldRead.cpu1)
2687 if ((raid->cpuAffinity.ldWrite.cpu0) &&
2688 (raid->cpuAffinity.ldWrite.cpu1))
2690 else if (raid->cpuAffinity.ldWrite.cpu1)
2694 ((raid->level == 5) || (raid->level == 6)) &&
2695 (raid->writeMode == MR_RL_WRITE_THROUGH_MODE) &&
2708 if ((fusion->pcie_bw_limitation) && (raid->level == 1) && (!is_read) &&
2740 struct MR_LD_RAID *raid = NULL;
2827 raid = MR_LdRaidGet(ld, local_map_ptr);
2829 if (!raid || (!fusion->fast_path_io)) {
2841 /* FP for Optimal raid level 1.
2881 /* If raid is NULL, set CPU affinity to default CPU0 */
2882 if (raid)
2884 raid, fp_possible, io_info.isRead,
2993 struct MR_LD_RAID *raid;
3015 raid = MR_LdRaidGet(ld, local_map_ptr);
3016 if (!(raid->capability.fpNonRWCapable))
3034 pRAID_Context->config_seq_num = raid->seqNum;
3038 cpu_to_le16(raid->fpIoTimeoutForLd);
3054 memcpy(io_request->LUN, raid->LUN, 8);
3316 * It prepares the raid 1 second IO
3414 * if it is raid 1/10 fp write capable.