Lines Matching defs:pdb

68 	struct rsa_pub_pdb *pdb = &edesc->pdb.pub;
70 dma_unmap_single(dev, pdb->n_dma, key->n_sz, DMA_TO_DEVICE);
71 dma_unmap_single(dev, pdb->e_dma, key->e_sz, DMA_TO_DEVICE);
80 struct rsa_priv_f1_pdb *pdb = &edesc->pdb.priv_f1;
82 dma_unmap_single(dev, pdb->n_dma, key->n_sz, DMA_TO_DEVICE);
83 dma_unmap_single(dev, pdb->d_dma, key->d_sz, DMA_TO_DEVICE);
92 struct rsa_priv_f2_pdb *pdb = &edesc->pdb.priv_f2;
96 dma_unmap_single(dev, pdb->d_dma, key->d_sz, DMA_TO_DEVICE);
97 dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE);
98 dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE);
99 dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_BIDIRECTIONAL);
100 dma_unmap_single(dev, pdb->tmp2_dma, q_sz, DMA_BIDIRECTIONAL);
109 struct rsa_priv_f3_pdb *pdb = &edesc->pdb.priv_f3;
113 dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE);
114 dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE);
115 dma_unmap_single(dev, pdb->dp_dma, p_sz, DMA_TO_DEVICE);
116 dma_unmap_single(dev, pdb->dq_dma, q_sz, DMA_TO_DEVICE);
117 dma_unmap_single(dev, pdb->c_dma, p_sz, DMA_TO_DEVICE);
118 dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_BIDIRECTIONAL);
119 dma_unmap_single(dev, pdb->tmp2_dma, q_sz, DMA_BIDIRECTIONAL);
412 struct rsa_pub_pdb *pdb = &edesc->pdb.pub;
415 pdb->n_dma = dma_map_single(dev, key->n, key->n_sz, DMA_TO_DEVICE);
416 if (dma_mapping_error(dev, pdb->n_dma)) {
421 pdb->e_dma = dma_map_single(dev, key->e, key->e_sz, DMA_TO_DEVICE);
422 if (dma_mapping_error(dev, pdb->e_dma)) {
424 dma_unmap_single(dev, pdb->n_dma, key->n_sz, DMA_TO_DEVICE);
429 pdb->sgf |= RSA_PDB_SGF_F;
430 pdb->f_dma = edesc->sec4_sg_dma;
433 pdb->f_dma = sg_dma_address(req_ctx->fixup_src);
437 pdb->sgf |= RSA_PDB_SGF_G;
438 pdb->g_dma = edesc->sec4_sg_dma +
441 pdb->g_dma = sg_dma_address(req->dst);
444 pdb->sgf |= (key->e_sz << RSA_PDB_E_SHIFT) | key->n_sz;
445 pdb->f_len = req_ctx->fixup_src_len;
457 struct rsa_priv_f1_pdb *pdb = &edesc->pdb.priv_f1;
460 pdb->n_dma = dma_map_single(dev, key->n, key->n_sz, DMA_TO_DEVICE);
461 if (dma_mapping_error(dev, pdb->n_dma)) {
466 pdb->d_dma = dma_map_single(dev, key->d, key->d_sz, DMA_TO_DEVICE);
467 if (dma_mapping_error(dev, pdb->d_dma)) {
469 dma_unmap_single(dev, pdb->n_dma, key->n_sz, DMA_TO_DEVICE);
474 pdb->sgf |= RSA_PRIV_PDB_SGF_G;
475 pdb->g_dma = edesc->sec4_sg_dma;
481 pdb->g_dma = sg_dma_address(req_ctx->fixup_src);
485 pdb->sgf |= RSA_PRIV_PDB_SGF_F;
486 pdb->f_dma = edesc->sec4_sg_dma +
489 pdb->f_dma = sg_dma_address(req->dst);
492 pdb->sgf |= (key->d_sz << RSA_PDB_D_SHIFT) | key->n_sz;
504 struct rsa_priv_f2_pdb *pdb = &edesc->pdb.priv_f2;
509 pdb->d_dma = dma_map_single(dev, key->d, key->d_sz, DMA_TO_DEVICE);
510 if (dma_mapping_error(dev, pdb->d_dma)) {
515 pdb->p_dma = dma_map_single(dev, key->p, p_sz, DMA_TO_DEVICE);
516 if (dma_mapping_error(dev, pdb->p_dma)) {
521 pdb->q_dma = dma_map_single(dev, key->q, q_sz, DMA_TO_DEVICE);
522 if (dma_mapping_error(dev, pdb->q_dma)) {
527 pdb->tmp1_dma = dma_map_single(dev, key->tmp1, p_sz, DMA_BIDIRECTIONAL);
528 if (dma_mapping_error(dev, pdb->tmp1_dma)) {
533 pdb->tmp2_dma = dma_map_single(dev, key->tmp2, q_sz, DMA_BIDIRECTIONAL);
534 if (dma_mapping_error(dev, pdb->tmp2_dma)) {
540 pdb->sgf |= RSA_PRIV_PDB_SGF_G;
541 pdb->g_dma = edesc->sec4_sg_dma;
546 pdb->g_dma = sg_dma_address(req_ctx->fixup_src);
550 pdb->sgf |= RSA_PRIV_PDB_SGF_F;
551 pdb->f_dma = edesc->sec4_sg_dma +
554 pdb->f_dma = sg_dma_address(req->dst);
557 pdb->sgf |= (key->d_sz << RSA_PDB_D_SHIFT) | key->n_sz;
558 pdb->p_q_len = (q_sz << RSA_PDB_Q_SHIFT) | p_sz;
563 dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_BIDIRECTIONAL);
565 dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE);
567 dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE);
569 dma_unmap_single(dev, pdb->d_dma, key->d_sz, DMA_TO_DEVICE);
581 struct rsa_priv_f3_pdb *pdb = &edesc->pdb.priv_f3;
586 pdb->p_dma = dma_map_single(dev, key->p, p_sz, DMA_TO_DEVICE);
587 if (dma_mapping_error(dev, pdb->p_dma)) {
592 pdb->q_dma = dma_map_single(dev, key->q, q_sz, DMA_TO_DEVICE);
593 if (dma_mapping_error(dev, pdb->q_dma)) {
598 pdb->dp_dma = dma_map_single(dev, key->dp, p_sz, DMA_TO_DEVICE);
599 if (dma_mapping_error(dev, pdb->dp_dma)) {
604 pdb->dq_dma = dma_map_single(dev, key->dq, q_sz, DMA_TO_DEVICE);
605 if (dma_mapping_error(dev, pdb->dq_dma)) {
610 pdb->c_dma = dma_map_single(dev, key->qinv, p_sz, DMA_TO_DEVICE);
611 if (dma_mapping_error(dev, pdb->c_dma)) {
616 pdb->tmp1_dma = dma_map_single(dev, key->tmp1, p_sz, DMA_BIDIRECTIONAL);
617 if (dma_mapping_error(dev, pdb->tmp1_dma)) {
622 pdb->tmp2_dma = dma_map_single(dev, key->tmp2, q_sz, DMA_BIDIRECTIONAL);
623 if (dma_mapping_error(dev, pdb->tmp2_dma)) {
629 pdb->sgf |= RSA_PRIV_PDB_SGF_G;
630 pdb->g_dma = edesc->sec4_sg_dma;
635 pdb->g_dma = sg_dma_address(req_ctx->fixup_src);
639 pdb->sgf |= RSA_PRIV_PDB_SGF_F;
640 pdb->f_dma = edesc->sec4_sg_dma +
643 pdb->f_dma = sg_dma_address(req->dst);
646 pdb->sgf |= key->n_sz;
647 pdb->p_q_len = (q_sz << RSA_PDB_Q_SHIFT) | p_sz;
652 dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_BIDIRECTIONAL);
654 dma_unmap_single(dev, pdb->c_dma, p_sz, DMA_TO_DEVICE);
656 dma_unmap_single(dev, pdb->dq_dma, q_sz, DMA_TO_DEVICE);
658 dma_unmap_single(dev, pdb->dp_dma, p_sz, DMA_TO_DEVICE);
660 dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE);
662 dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE);
743 init_rsa_pub_desc(edesc->hw_desc, &edesc->pdb.pub);
772 init_rsa_priv_f1_desc(edesc->hw_desc, &edesc->pdb.priv_f1);
801 init_rsa_priv_f2_desc(edesc->hw_desc, &edesc->pdb.priv_f2);
830 init_rsa_priv_f3_desc(edesc->hw_desc, &edesc->pdb.priv_f3);