Lines Matching defs:sctx

74 	struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm);
76 sctx->fallback.cip->base.crt_flags &= ~CRYPTO_TFM_REQ_MASK;
77 sctx->fallback.cip->base.crt_flags |= (tfm->crt_flags &
80 return crypto_cipher_setkey(sctx->fallback.cip, in_key, key_len);
86 struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm);
95 sctx->fc = (fc && cpacf_test_func(&km_functions, fc)) ? fc : 0;
96 if (!sctx->fc)
99 sctx->key_len = key_len;
100 memcpy(sctx->key, in_key, key_len);
106 struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm);
108 if (unlikely(!sctx->fc)) {
109 crypto_cipher_encrypt_one(sctx->fallback.cip, out, in);
112 cpacf_km(sctx->fc, &sctx->key, out, in, AES_BLOCK_SIZE);
117 struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm);
119 if (unlikely(!sctx->fc)) {
120 crypto_cipher_decrypt_one(sctx->fallback.cip, out, in);
123 cpacf_km(sctx->fc | CPACF_DECRYPT,
124 &sctx->key, out, in, AES_BLOCK_SIZE);
130 struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm);
132 sctx->fallback.cip = crypto_alloc_cipher(name, 0,
135 if (IS_ERR(sctx->fallback.cip)) {
138 return PTR_ERR(sctx->fallback.cip);
146 struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm);
148 crypto_free_cipher(sctx->fallback.cip);
149 sctx->fallback.cip = NULL;
177 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
179 crypto_skcipher_clear_flags(sctx->fallback.skcipher,
181 crypto_skcipher_set_flags(sctx->fallback.skcipher,
184 return crypto_skcipher_setkey(sctx->fallback.skcipher, key, len);
187 static int fallback_skcipher_crypt(struct s390_aes_ctx *sctx,
194 skcipher_request_set_tfm(subreq, sctx->fallback.skcipher);
203 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
212 sctx->fc = (fc && cpacf_test_func(&km_functions, fc)) ? fc : 0;
213 if (!sctx->fc)
216 sctx->key_len = key_len;
217 memcpy(sctx->key, in_key, key_len);
224 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
229 if (unlikely(!sctx->fc))
230 return fallback_skcipher_crypt(sctx, req, modifier);
236 cpacf_km(sctx->fc | modifier, sctx->key,
256 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
258 sctx->fallback.skcipher = crypto_alloc_skcipher(name, 0,
261 if (IS_ERR(sctx->fallback.skcipher)) {
264 return PTR_ERR(sctx->fallback.skcipher);
268 crypto_skcipher_reqsize(sctx->fallback.skcipher));
274 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
276 crypto_free_skcipher(sctx->fallback.skcipher);
299 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
308 sctx->fc = (fc && cpacf_test_func(&kmc_functions, fc)) ? fc : 0;
309 if (!sctx->fc)
312 sctx->key_len = key_len;
313 memcpy(sctx->key, in_key, key_len);
320 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
329 if (unlikely(!sctx->fc))
330 return fallback_skcipher_crypt(sctx, req, modifier);
336 memcpy(param.key, sctx->key, sctx->key_len);
340 cpacf_kmc(sctx->fc | modifier, &param,
535 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
544 sctx->fc = (fc && cpacf_test_func(&kmctr_functions, fc)) ? fc : 0;
545 if (!sctx->fc)
548 sctx->key_len = key_len;
549 memcpy(sctx->key, in_key, key_len);
571 struct s390_aes_ctx *sctx = crypto_skcipher_ctx(tfm);
577 if (unlikely(!sctx->fc))
578 return fallback_skcipher_crypt(sctx, req, 0);
589 cpacf_kmctr(sctx->fc, sctx->key, walk.dst.virt.addr,
605 cpacf_kmctr(sctx->fc, sctx->key, buf, buf,