Lines Matching defs:ring

607 static int safexcel_handle_req_result(struct safexcel_crypto_priv *priv, int ring,
627 rdesc = safexcel_ring_next_rptr(priv, &priv->ring[ring].rdr);
641 safexcel_complete(priv, ring);
673 static int safexcel_send_req(struct crypto_async_request *base, int ring,
781 first_cdesc = safexcel_add_cdesc(priv, ring,
786 /* No space left in the command descriptor ring */
802 cdesc = safexcel_add_cdesc(priv, ring, !n_cdesc,
807 /* No space left in the command descriptor ring */
849 rdesc = safexcel_add_rdesc(priv, ring, first, last,
855 rdesc = safexcel_add_rdesc(priv, ring, first, last,
860 /* No space left in the result descriptor ring */
878 rdesc = safexcel_add_rdesc(priv, ring, true, true, 0, 0);
880 /* No space left in the result descriptor ring */
888 safexcel_rdr_req_set(priv, ring, first_rdesc, base);
896 safexcel_ring_rollback_wptr(priv, &priv->ring[ring].rdr);
899 safexcel_ring_rollback_wptr(priv, &priv->ring[ring].cdr);
918 int ring,
933 rdesc = safexcel_ring_next_rptr(priv, &priv->ring[ring].rdr);
947 safexcel_complete(priv, ring);
958 ring = safexcel_select_ring(priv);
959 ctx->base.ring = ring;
961 spin_lock_bh(&priv->ring[ring].queue_lock);
962 enq_ret = crypto_enqueue_request(&priv->ring[ring].queue, base);
963 spin_unlock_bh(&priv->ring[ring].queue_lock);
968 queue_work(priv->ring[ring].workqueue,
969 &priv->ring[ring].work_data.work);
977 int ring,
987 err = safexcel_handle_inv_result(priv, ring, async, sreq,
990 err = safexcel_handle_req_result(priv, ring, async, req->src,
999 int ring,
1010 err = safexcel_handle_inv_result(priv, ring, async, sreq,
1013 err = safexcel_handle_req_result(priv, ring, async, req->src,
1023 int ring, int *commands, int *results)
1029 ret = safexcel_invalidate_cache(base, priv, ctx->base.ctxr_dma, ring);
1039 static int safexcel_skcipher_send(struct crypto_async_request *async, int ring,
1051 ret = safexcel_cipher_send_inv(async, ring, commands, results);
1062 ret = safexcel_send_req(async, ring, sreq, req->src,
1071 static int safexcel_aead_send(struct crypto_async_request *async, int ring,
1084 ret = safexcel_cipher_send_inv(async, ring, commands, results);
1086 ret = safexcel_send_req(async, ring, sreq, req->src, req->dst,
1101 int ring = ctx->base.ring;
1108 spin_lock_bh(&priv->ring[ring].queue_lock);
1109 crypto_enqueue_request(&priv->ring[ring].queue, base);
1110 spin_unlock_bh(&priv->ring[ring].queue_lock);
1112 queue_work(priv->ring[ring].workqueue,
1113 &priv->ring[ring].work_data.work);
1163 int ret, ring;
1174 ctx->base.ring = safexcel_select_ring(priv);
1182 ring = ctx->base.ring;
1184 spin_lock_bh(&priv->ring[ring].queue_lock);
1185 ret = crypto_enqueue_request(&priv->ring[ring].queue, base);
1186 spin_unlock_bh(&priv->ring[ring].queue_lock);
1188 queue_work(priv->ring[ring].workqueue,
1189 &priv->ring[ring].work_data.work);