Lines Matching refs:vf
23 struct ifcvf_hw *vf = arg;
25 if (vf->config_cb.callback)
26 return vf->config_cb.callback(vf->config_cb.private);
49 struct ifcvf_hw *vf = &adapter->vf;
54 devm_free_irq(&pdev->dev, vf->vring[i].irq, &vf->vring[i]);
55 vf->vring[i].irq = -EINVAL;
58 devm_free_irq(&pdev->dev, vf->config_irq, vf);
65 struct ifcvf_hw *vf = &adapter->vf;
75 snprintf(vf->config_msix_name, 256, "ifcvf[%s]-config\n",
78 vf->config_irq = pci_irq_vector(pdev, vector);
79 ret = devm_request_irq(&pdev->dev, vf->config_irq,
81 vf->config_msix_name, vf);
88 snprintf(vf->vring[i].msix_name, 256, "ifcvf[%s]-%d\n",
94 vf->vring[i].msix_name,
95 &vf->vring[i]);
104 vf->vring[i].irq = irq;
112 struct ifcvf_hw *vf = ifcvf_private_to_vf(private);
116 vf->nr_vring = IFCVF_MAX_QUEUE_PAIRS * 2;
117 ret = ifcvf_start_hw(vf);
119 status = ifcvf_get_status(vf);
121 ifcvf_set_status(vf, status);
129 struct ifcvf_hw *vf = ifcvf_private_to_vf(private);
133 vf->vring[i].cb.callback = NULL;
135 ifcvf_stop_hw(vf);
142 struct ifcvf_hw *vf = ifcvf_private_to_vf(adapter);
146 vf->vring[i].last_avail_idx = 0;
147 vf->vring[i].desc = 0;
148 vf->vring[i].avail = 0;
149 vf->vring[i].used = 0;
150 vf->vring[i].ready = 0;
151 vf->vring[i].cb.callback = NULL;
152 vf->vring[i].cb.private = NULL;
155 ifcvf_reset(vf);
167 return &adapter->vf;
172 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
175 features = ifcvf_get_features(vf) & IFCVF_SUPPORTED_FEATURES;
182 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
184 vf->req_features = features;
191 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
193 return ifcvf_get_status(vf);
199 struct ifcvf_hw *vf;
203 vf = vdpa_to_vf(vdpa_dev);
205 status_old = ifcvf_get_status(vf);
225 status = ifcvf_get_status(vf);
227 ifcvf_set_status(vf, status);
237 ifcvf_set_status(vf, status);
248 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
250 state->avail_index = ifcvf_get_vq_state(vf, qid);
257 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
259 return ifcvf_set_vq_state(vf, qid, state->avail_index);
265 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
267 vf->vring[qid].cb = *cb;
273 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
275 vf->vring[qid].ready = ready;
280 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
282 return vf->vring[qid].ready;
288 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
290 vf->vring[qid].size = num;
297 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
299 vf->vring[qid].desc = desc_area;
300 vf->vring[qid].avail = driver_area;
301 vf->vring[qid].used = device_area;
308 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
310 ifcvf_notify_queue(vf, qid);
315 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
317 return ioread8(&vf->common_cfg->config_generation);
339 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
342 ifcvf_read_net_config(vf, offset, buf, len);
349 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
352 ifcvf_write_net_config(vf, offset, buf, len);
358 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
360 vf->config_cb.callback = cb->callback;
361 vf->config_cb.private = cb->private;
367 struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev);
369 return vf->vring[qid].irq;
404 struct ifcvf_hw *vf;
451 vf = &adapter->vf;
452 vf->base = pcim_iomap_table(pdev);
457 ret = ifcvf_init_hw(vf, pdev);
464 vf->vring[i].irq = -EINVAL;