Lines Matching refs:sd_idx

15  * @sd_idx: pointer to return index of the segment descriptor in question
23 u32 idx, u32 cnt, u32 *sd_idx,
31 *sd_idx = (u32)(fpm_addr / IRDMA_HMC_DIRECT_BP_SIZE);
101 * @sd_idx: segment descriptor index
104 static inline void irdma_invalidate_pf_hmc_pd(struct irdma_sc_dev *dev, u32 sd_idx,
107 u32 val = FIELD_PREP(IRDMA_PFHMC_PDINV_PMSDIDX, sd_idx) |
119 * @sd_idx: sd index
123 int irdma_hmc_sd_one(struct irdma_sc_dev *dev, u8 hmc_fn_id, u64 pa, u32 sd_idx,
131 irdma_set_sd_entry(pa, sd_idx, type, sdinfo.entry);
133 irdma_clr_sd_entry(sd_idx, type, sdinfo.entry);
224 u32 sd_idx, sd_lmt;
244 info->start_idx, info->count, &sd_idx,
246 if (sd_idx >= info->hmc_info->sd_table.sd_cnt ||
255 for (j = sd_idx; j < sd_lmt; j++) {
296 while (j && (j > sd_idx)) {
330 u32 i, sd_idx;
341 sd_idx = info->hmc_info->sd_indexes[i];
342 sd_entry = &info->hmc_info->sd_table.sd_entry[sd_idx];
374 u32 sd_idx, sd_lmt;
401 sd_idx = j / IRDMA_HMC_PD_CNT_IN_SD;
403 if (!info->hmc_info->sd_table.sd_entry[sd_idx].valid)
406 if (info->hmc_info->sd_table.sd_entry[sd_idx].entry_type !=
411 pd_table = &info->hmc_info->sd_table.sd_entry[sd_idx].u.pd_table;
424 info->start_idx, info->count, &sd_idx,
426 if (sd_idx >= info->hmc_info->sd_table.sd_cnt ||
428 ibdev_dbg(to_ibdev(dev), "HMC: invalid sd_idx\n");
432 for (i = sd_idx; i < sd_lmt; i++) {
551 u32 sd_idx, rel_pd_idx;
558 sd_idx = (pd_index / IRDMA_HMC_PD_CNT_IN_SD);
559 if (hmc_info->sd_table.sd_entry[sd_idx].entry_type !=
564 pd_table = &hmc_info->sd_table.sd_entry[sd_idx].u.pd_table;
589 pd_entry->sd_index = sd_idx;
592 irdma_invalidate_pf_hmc_pd(dev, sd_idx, rel_pd_idx);
620 u32 sd_idx, rel_pd_idx;
624 sd_idx = idx / IRDMA_HMC_PD_CNT_IN_SD;
626 if (sd_idx >= hmc_info->sd_table.sd_cnt)
629 sd_entry = &hmc_info->sd_table.sd_entry[sd_idx];
633 pd_table = &hmc_info->sd_table.sd_entry[sd_idx].u.pd_table;
643 irdma_invalidate_pf_hmc_pd(dev, sd_idx, idx);