Lines Matching refs:ret

470 	int ret;
481 ret = intel_gvt_scan_and_shadow_ringbuffer(workload);
482 if (ret)
483 return ret;
487 ret = intel_gvt_scan_and_shadow_wa_ctx(&workload->wa_ctx);
488 if (ret)
497 return ret;
507 int ret;
534 ret = PTR_ERR(bb->vma);
543 ret = i915_vma_move_to_active(bb->vma,
546 if (ret)
556 return ret;
640 int ret = 0;
642 ret = intel_vgpu_pin_mm(workload->shadow_mm);
643 if (ret) {
645 return ret;
658 ret = intel_vgpu_pin_mm(m);
659 if (ret) {
670 if (ret)
673 return ret;
693 int ret = 0;
695 ret = intel_vgpu_shadow_mm_pin(workload);
696 if (ret) {
698 return ret;
705 ret = intel_vgpu_sync_oos_pages(workload->vgpu);
706 if (ret) {
711 ret = intel_vgpu_flush_post_shadow(workload->vgpu);
712 if (ret) {
717 ret = copy_workload_to_ring_buffer(workload);
718 if (ret) {
723 ret = prepare_shadow_batch_buffer(workload);
724 if (ret) {
729 ret = prepare_shadow_wa_ctx(&workload->wa_ctx);
730 if (ret) {
736 ret = workload->prepare(workload);
737 if (ret)
748 return ret;
755 int ret;
762 ret = intel_gvt_workload_req_alloc(workload);
763 if (ret)
766 ret = intel_gvt_scan_and_shadow_workload(workload);
767 if (ret)
770 ret = populate_shadow_context(workload);
771 if (ret) {
776 ret = prepare_workload(workload);
778 if (ret) {
793 if (ret)
794 workload->status = ret;
796 return ret;
1111 int ret;
1152 ret = dispatch_workload(workload);
1154 if (ret) {
1175 if (ret && (vgpu_is_vm_unhealthy(ret)))
1216 int ret;
1229 ret = PTR_ERR(scheduler->thread[i]);
1243 return ret;
1342 int ret;
1358 ret = PTR_ERR(ce);
1386 ret = -ENOMEM;
1407 return ret;
1432 int ret;
1452 ret = ops[interface]->init(vgpu, engine_mask);
1453 if (ret)
1454 return ret;
1594 int ret;
1700 ret = prepare_mm(workload);
1701 if (ret) {
1703 return ERR_PTR(ret);
1713 ret = intel_gvt_scan_and_shadow_workload(workload);
1716 if (ret) {
1717 if (vgpu_is_vm_unhealthy(ret))
1720 return ERR_PTR(ret);
1723 ret = intel_context_pin(s->shadow[engine->id]);
1724 if (ret) {
1726 return ERR_PTR(ret);