Lines Matching defs:pdb
62 struct rsa_pub_pdb *pdb = &edesc->pdb.pub;
64 dma_unmap_single(dev, pdb->n_dma, key->n_sz, DMA_TO_DEVICE);
65 dma_unmap_single(dev, pdb->e_dma, key->e_sz, DMA_TO_DEVICE);
74 struct rsa_priv_f1_pdb *pdb = &edesc->pdb.priv_f1;
76 dma_unmap_single(dev, pdb->n_dma, key->n_sz, DMA_TO_DEVICE);
77 dma_unmap_single(dev, pdb->d_dma, key->d_sz, DMA_TO_DEVICE);
86 struct rsa_priv_f2_pdb *pdb = &edesc->pdb.priv_f2;
90 dma_unmap_single(dev, pdb->d_dma, key->d_sz, DMA_TO_DEVICE);
91 dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE);
92 dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE);
93 dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_BIDIRECTIONAL);
94 dma_unmap_single(dev, pdb->tmp2_dma, q_sz, DMA_BIDIRECTIONAL);
103 struct rsa_priv_f3_pdb *pdb = &edesc->pdb.priv_f3;
107 dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE);
108 dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE);
109 dma_unmap_single(dev, pdb->dp_dma, p_sz, DMA_TO_DEVICE);
110 dma_unmap_single(dev, pdb->dq_dma, q_sz, DMA_TO_DEVICE);
111 dma_unmap_single(dev, pdb->c_dma, p_sz, DMA_TO_DEVICE);
112 dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_BIDIRECTIONAL);
113 dma_unmap_single(dev, pdb->tmp2_dma, q_sz, DMA_BIDIRECTIONAL);
403 struct rsa_pub_pdb *pdb = &edesc->pdb.pub;
406 pdb->n_dma = dma_map_single(dev, key->n, key->n_sz, DMA_TO_DEVICE);
407 if (dma_mapping_error(dev, pdb->n_dma)) {
412 pdb->e_dma = dma_map_single(dev, key->e, key->e_sz, DMA_TO_DEVICE);
413 if (dma_mapping_error(dev, pdb->e_dma)) {
415 dma_unmap_single(dev, pdb->n_dma, key->n_sz, DMA_TO_DEVICE);
420 pdb->sgf |= RSA_PDB_SGF_F;
421 pdb->f_dma = edesc->sec4_sg_dma;
424 pdb->f_dma = sg_dma_address(req_ctx->fixup_src);
428 pdb->sgf |= RSA_PDB_SGF_G;
429 pdb->g_dma = edesc->sec4_sg_dma +
432 pdb->g_dma = sg_dma_address(req->dst);
435 pdb->sgf |= (key->e_sz << RSA_PDB_E_SHIFT) | key->n_sz;
436 pdb->f_len = req_ctx->fixup_src_len;
448 struct rsa_priv_f1_pdb *pdb = &edesc->pdb.priv_f1;
451 pdb->n_dma = dma_map_single(dev, key->n, key->n_sz, DMA_TO_DEVICE);
452 if (dma_mapping_error(dev, pdb->n_dma)) {
457 pdb->d_dma = dma_map_single(dev, key->d, key->d_sz, DMA_TO_DEVICE);
458 if (dma_mapping_error(dev, pdb->d_dma)) {
460 dma_unmap_single(dev, pdb->n_dma, key->n_sz, DMA_TO_DEVICE);
465 pdb->sgf |= RSA_PRIV_PDB_SGF_G;
466 pdb->g_dma = edesc->sec4_sg_dma;
472 pdb->g_dma = sg_dma_address(req_ctx->fixup_src);
476 pdb->sgf |= RSA_PRIV_PDB_SGF_F;
477 pdb->f_dma = edesc->sec4_sg_dma +
480 pdb->f_dma = sg_dma_address(req->dst);
483 pdb->sgf |= (key->d_sz << RSA_PDB_D_SHIFT) | key->n_sz;
495 struct rsa_priv_f2_pdb *pdb = &edesc->pdb.priv_f2;
500 pdb->d_dma = dma_map_single(dev, key->d, key->d_sz, DMA_TO_DEVICE);
501 if (dma_mapping_error(dev, pdb->d_dma)) {
506 pdb->p_dma = dma_map_single(dev, key->p, p_sz, DMA_TO_DEVICE);
507 if (dma_mapping_error(dev, pdb->p_dma)) {
512 pdb->q_dma = dma_map_single(dev, key->q, q_sz, DMA_TO_DEVICE);
513 if (dma_mapping_error(dev, pdb->q_dma)) {
518 pdb->tmp1_dma = dma_map_single(dev, key->tmp1, p_sz, DMA_BIDIRECTIONAL);
519 if (dma_mapping_error(dev, pdb->tmp1_dma)) {
524 pdb->tmp2_dma = dma_map_single(dev, key->tmp2, q_sz, DMA_BIDIRECTIONAL);
525 if (dma_mapping_error(dev, pdb->tmp2_dma)) {
531 pdb->sgf |= RSA_PRIV_PDB_SGF_G;
532 pdb->g_dma = edesc->sec4_sg_dma;
537 pdb->g_dma = sg_dma_address(req_ctx->fixup_src);
541 pdb->sgf |= RSA_PRIV_PDB_SGF_F;
542 pdb->f_dma = edesc->sec4_sg_dma +
545 pdb->f_dma = sg_dma_address(req->dst);
548 pdb->sgf |= (key->d_sz << RSA_PDB_D_SHIFT) | key->n_sz;
549 pdb->p_q_len = (q_sz << RSA_PDB_Q_SHIFT) | p_sz;
554 dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_BIDIRECTIONAL);
556 dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE);
558 dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE);
560 dma_unmap_single(dev, pdb->d_dma, key->d_sz, DMA_TO_DEVICE);
572 struct rsa_priv_f3_pdb *pdb = &edesc->pdb.priv_f3;
577 pdb->p_dma = dma_map_single(dev, key->p, p_sz, DMA_TO_DEVICE);
578 if (dma_mapping_error(dev, pdb->p_dma)) {
583 pdb->q_dma = dma_map_single(dev, key->q, q_sz, DMA_TO_DEVICE);
584 if (dma_mapping_error(dev, pdb->q_dma)) {
589 pdb->dp_dma = dma_map_single(dev, key->dp, p_sz, DMA_TO_DEVICE);
590 if (dma_mapping_error(dev, pdb->dp_dma)) {
595 pdb->dq_dma = dma_map_single(dev, key->dq, q_sz, DMA_TO_DEVICE);
596 if (dma_mapping_error(dev, pdb->dq_dma)) {
601 pdb->c_dma = dma_map_single(dev, key->qinv, p_sz, DMA_TO_DEVICE);
602 if (dma_mapping_error(dev, pdb->c_dma)) {
607 pdb->tmp1_dma = dma_map_single(dev, key->tmp1, p_sz, DMA_BIDIRECTIONAL);
608 if (dma_mapping_error(dev, pdb->tmp1_dma)) {
613 pdb->tmp2_dma = dma_map_single(dev, key->tmp2, q_sz, DMA_BIDIRECTIONAL);
614 if (dma_mapping_error(dev, pdb->tmp2_dma)) {
620 pdb->sgf |= RSA_PRIV_PDB_SGF_G;
621 pdb->g_dma = edesc->sec4_sg_dma;
626 pdb->g_dma = sg_dma_address(req_ctx->fixup_src);
630 pdb->sgf |= RSA_PRIV_PDB_SGF_F;
631 pdb->f_dma = edesc->sec4_sg_dma +
634 pdb->f_dma = sg_dma_address(req->dst);
637 pdb->sgf |= key->n_sz;
638 pdb->p_q_len = (q_sz << RSA_PDB_Q_SHIFT) | p_sz;
643 dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_BIDIRECTIONAL);
645 dma_unmap_single(dev, pdb->c_dma, p_sz, DMA_TO_DEVICE);
647 dma_unmap_single(dev, pdb->dq_dma, q_sz, DMA_TO_DEVICE);
649 dma_unmap_single(dev, pdb->dp_dma, p_sz, DMA_TO_DEVICE);
651 dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE);
653 dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE);
734 init_rsa_pub_desc(edesc->hw_desc, &edesc->pdb.pub);
763 init_rsa_priv_f1_desc(edesc->hw_desc, &edesc->pdb.priv_f1);
792 init_rsa_priv_f2_desc(edesc->hw_desc, &edesc->pdb.priv_f2);
821 init_rsa_priv_f3_desc(edesc->hw_desc, &edesc->pdb.priv_f3);