Lines Matching defs:sctx

75 	struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm);
77 sctx->fallback.cip->base.crt_flags &= ~CRYPTO_TFM_REQ_MASK;
78 sctx->fallback.cip->base.crt_flags |= (tfm->crt_flags &
81 return crypto_cipher_setkey(sctx->fallback.cip, in_key, key_len);
87 struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm);
96 sctx->fc = (fc && cpacf_test_func(&km_functions, fc)) ? fc : 0;
97 if (!sctx->fc)
100 sctx->key_len = key_len;
101 memcpy(sctx->key, in_key, key_len);
107 struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm);
109 if (unlikely(!sctx->fc)) {
110 crypto_cipher_encrypt_one(sctx->fallback.cip, out, in);
113 cpacf_km(sctx->fc, &sctx->key, out, in, AES_BLOCK_SIZE);
118 struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm);
120 if (unlikely(!sctx->fc)) {
121 crypto_cipher_decrypt_one(sctx->fallback.cip, out, in);
124 cpacf_km(sctx->fc | CPACF_DECRYPT,
125 &sctx->key, out, in, AES_BLOCK_SIZE);
131 struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm);
133 sctx->fallback.cip = crypto_alloc_cipher(name, 0,
136 if (IS_ERR(sctx->fallback.cip)) {
139 return PTR_ERR(sctx->fallback.cip);
147 struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm);
149 crypto_free_cipher(sctx->fallback.cip);
150 sctx->fallback.cip = NULL;
178 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
180 crypto_skcipher_clear_flags(sctx->fallback.skcipher,
182 crypto_skcipher_set_flags(sctx->fallback.skcipher,
185 return crypto_skcipher_setkey(sctx->fallback.skcipher, key, len);
188 static int fallback_skcipher_crypt(struct s390_aes_ctx *sctx,
195 skcipher_request_set_tfm(subreq, sctx->fallback.skcipher);
204 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
213 sctx->fc = (fc && cpacf_test_func(&km_functions, fc)) ? fc : 0;
214 if (!sctx->fc)
217 sctx->key_len = key_len;
218 memcpy(sctx->key, in_key, key_len);
225 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
230 if (unlikely(!sctx->fc))
231 return fallback_skcipher_crypt(sctx, req, modifier);
237 cpacf_km(sctx->fc | modifier, sctx->key,
257 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
259 sctx->fallback.skcipher = crypto_alloc_skcipher(name, 0,
262 if (IS_ERR(sctx->fallback.skcipher)) {
265 return PTR_ERR(sctx->fallback.skcipher);
269 crypto_skcipher_reqsize(sctx->fallback.skcipher));
275 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
277 crypto_free_skcipher(sctx->fallback.skcipher);
300 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
309 sctx->fc = (fc && cpacf_test_func(&kmc_functions, fc)) ? fc : 0;
310 if (!sctx->fc)
313 sctx->key_len = key_len;
314 memcpy(sctx->key, in_key, key_len);
321 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
330 if (unlikely(!sctx->fc))
331 return fallback_skcipher_crypt(sctx, req, modifier);
337 memcpy(param.key, sctx->key, sctx->key_len);
341 cpacf_kmc(sctx->fc | modifier, &param,
532 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
541 sctx->fc = (fc && cpacf_test_func(&kmctr_functions, fc)) ? fc : 0;
542 if (!sctx->fc)
545 sctx->key_len = key_len;
546 memcpy(sctx->key, in_key, key_len);
568 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
574 if (unlikely(!sctx->fc))
575 return fallback_skcipher_crypt(sctx, req, 0);
586 cpacf_kmctr(sctx->fc, sctx->key, walk.dst.virt.addr,
602 cpacf_kmctr(sctx->fc, sctx->key, buf, buf,