 5d26a105b5
			
		
	
	
	5d26a105b5
	
	
	
		
			
			This prefixes all crypto module loading with "crypto-" so we never run the risk of exposing module auto-loading to userspace via a crypto API, as demonstrated by Mathias Krause: https://lkml.org/lkml/2013/3/4/70 Signed-off-by: Kees Cook <keescook@chromium.org> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
		
			
				
	
	
		
			626 lines
		
	
	
	
		
			16 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			626 lines
		
	
	
	
		
			16 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Cryptographic API.
 | |
|  *
 | |
|  * s390 implementation of the DES Cipher Algorithm.
 | |
|  *
 | |
|  * Copyright IBM Corp. 2003, 2011
 | |
|  * Author(s): Thomas Spatzier
 | |
|  *	      Jan Glauber (jan.glauber@de.ibm.com)
 | |
|  *
 | |
|  * This program is free software; you can redistribute it and/or modify
 | |
|  * it under the terms of the GNU General Public License as published by
 | |
|  * the Free Software Foundation; either version 2 of the License, or
 | |
|  * (at your option) any later version.
 | |
|  *
 | |
|  */
 | |
| 
 | |
| #include <linux/init.h>
 | |
| #include <linux/module.h>
 | |
| #include <linux/crypto.h>
 | |
| #include <crypto/algapi.h>
 | |
| #include <crypto/des.h>
 | |
| 
 | |
| #include "crypt_s390.h"
 | |
| 
 | |
| #define DES3_KEY_SIZE	(3 * DES_KEY_SIZE)
 | |
| 
 | |
| static u8 *ctrblk;
 | |
| static DEFINE_SPINLOCK(ctrblk_lock);
 | |
| 
 | |
| struct s390_des_ctx {
 | |
| 	u8 iv[DES_BLOCK_SIZE];
 | |
| 	u8 key[DES3_KEY_SIZE];
 | |
| };
 | |
| 
 | |
| static int des_setkey(struct crypto_tfm *tfm, const u8 *key,
 | |
| 		      unsigned int key_len)
 | |
| {
 | |
| 	struct s390_des_ctx *ctx = crypto_tfm_ctx(tfm);
 | |
| 	u32 *flags = &tfm->crt_flags;
 | |
| 	u32 tmp[DES_EXPKEY_WORDS];
 | |
| 
 | |
| 	/* check for weak keys */
 | |
| 	if (!des_ekey(tmp, key) && (*flags & CRYPTO_TFM_REQ_WEAK_KEY)) {
 | |
| 		*flags |= CRYPTO_TFM_RES_WEAK_KEY;
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	memcpy(ctx->key, key, key_len);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void des_encrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in)
 | |
| {
 | |
| 	struct s390_des_ctx *ctx = crypto_tfm_ctx(tfm);
 | |
| 
 | |
| 	crypt_s390_km(KM_DEA_ENCRYPT, ctx->key, out, in, DES_BLOCK_SIZE);
 | |
| }
 | |
| 
 | |
| static void des_decrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in)
 | |
| {
 | |
| 	struct s390_des_ctx *ctx = crypto_tfm_ctx(tfm);
 | |
| 
 | |
| 	crypt_s390_km(KM_DEA_DECRYPT, ctx->key, out, in, DES_BLOCK_SIZE);
 | |
| }
 | |
| 
 | |
| static struct crypto_alg des_alg = {
 | |
| 	.cra_name		=	"des",
 | |
| 	.cra_driver_name	=	"des-s390",
 | |
| 	.cra_priority		=	CRYPT_S390_PRIORITY,
 | |
| 	.cra_flags		=	CRYPTO_ALG_TYPE_CIPHER,
 | |
| 	.cra_blocksize		=	DES_BLOCK_SIZE,
 | |
| 	.cra_ctxsize		=	sizeof(struct s390_des_ctx),
 | |
| 	.cra_module		=	THIS_MODULE,
 | |
| 	.cra_u			=	{
 | |
| 		.cipher = {
 | |
| 			.cia_min_keysize	=	DES_KEY_SIZE,
 | |
| 			.cia_max_keysize	=	DES_KEY_SIZE,
 | |
| 			.cia_setkey		=	des_setkey,
 | |
| 			.cia_encrypt		=	des_encrypt,
 | |
| 			.cia_decrypt		=	des_decrypt,
 | |
| 		}
 | |
| 	}
 | |
| };
 | |
| 
 | |
| static int ecb_desall_crypt(struct blkcipher_desc *desc, long func,
 | |
| 			    u8 *key, struct blkcipher_walk *walk)
 | |
| {
 | |
| 	int ret = blkcipher_walk_virt(desc, walk);
 | |
| 	unsigned int nbytes;
 | |
| 
 | |
| 	while ((nbytes = walk->nbytes)) {
 | |
| 		/* only use complete blocks */
 | |
| 		unsigned int n = nbytes & ~(DES_BLOCK_SIZE - 1);
 | |
| 		u8 *out = walk->dst.virt.addr;
 | |
| 		u8 *in = walk->src.virt.addr;
 | |
| 
 | |
| 		ret = crypt_s390_km(func, key, out, in, n);
 | |
| 		if (ret < 0 || ret != n)
 | |
| 			return -EIO;
 | |
| 
 | |
| 		nbytes &= DES_BLOCK_SIZE - 1;
 | |
| 		ret = blkcipher_walk_done(desc, walk, nbytes);
 | |
| 	}
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int cbc_desall_crypt(struct blkcipher_desc *desc, long func,
 | |
| 			    struct blkcipher_walk *walk)
 | |
| {
 | |
| 	struct s390_des_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
 | |
| 	int ret = blkcipher_walk_virt(desc, walk);
 | |
| 	unsigned int nbytes = walk->nbytes;
 | |
| 	struct {
 | |
| 		u8 iv[DES_BLOCK_SIZE];
 | |
| 		u8 key[DES3_KEY_SIZE];
 | |
| 	} param;
 | |
| 
 | |
| 	if (!nbytes)
 | |
| 		goto out;
 | |
| 
 | |
| 	memcpy(param.iv, walk->iv, DES_BLOCK_SIZE);
 | |
| 	memcpy(param.key, ctx->key, DES3_KEY_SIZE);
 | |
| 	do {
 | |
| 		/* only use complete blocks */
 | |
| 		unsigned int n = nbytes & ~(DES_BLOCK_SIZE - 1);
 | |
| 		u8 *out = walk->dst.virt.addr;
 | |
| 		u8 *in = walk->src.virt.addr;
 | |
| 
 | |
| 		ret = crypt_s390_kmc(func, ¶m, out, in, n);
 | |
| 		if (ret < 0 || ret != n)
 | |
| 			return -EIO;
 | |
| 
 | |
| 		nbytes &= DES_BLOCK_SIZE - 1;
 | |
| 		ret = blkcipher_walk_done(desc, walk, nbytes);
 | |
| 	} while ((nbytes = walk->nbytes));
 | |
| 	memcpy(walk->iv, param.iv, DES_BLOCK_SIZE);
 | |
| 
 | |
| out:
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int ecb_des_encrypt(struct blkcipher_desc *desc,
 | |
| 			   struct scatterlist *dst, struct scatterlist *src,
 | |
| 			   unsigned int nbytes)
 | |
| {
 | |
| 	struct s390_des_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
 | |
| 	struct blkcipher_walk walk;
 | |
| 
 | |
| 	blkcipher_walk_init(&walk, dst, src, nbytes);
 | |
| 	return ecb_desall_crypt(desc, KM_DEA_ENCRYPT, ctx->key, &walk);
 | |
| }
 | |
| 
 | |
| static int ecb_des_decrypt(struct blkcipher_desc *desc,
 | |
| 			   struct scatterlist *dst, struct scatterlist *src,
 | |
| 			   unsigned int nbytes)
 | |
| {
 | |
| 	struct s390_des_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
 | |
| 	struct blkcipher_walk walk;
 | |
| 
 | |
| 	blkcipher_walk_init(&walk, dst, src, nbytes);
 | |
| 	return ecb_desall_crypt(desc, KM_DEA_DECRYPT, ctx->key, &walk);
 | |
| }
 | |
| 
 | |
| static struct crypto_alg ecb_des_alg = {
 | |
| 	.cra_name		=	"ecb(des)",
 | |
| 	.cra_driver_name	=	"ecb-des-s390",
 | |
| 	.cra_priority		=	CRYPT_S390_COMPOSITE_PRIORITY,
 | |
| 	.cra_flags		=	CRYPTO_ALG_TYPE_BLKCIPHER,
 | |
| 	.cra_blocksize		=	DES_BLOCK_SIZE,
 | |
| 	.cra_ctxsize		=	sizeof(struct s390_des_ctx),
 | |
| 	.cra_type		=	&crypto_blkcipher_type,
 | |
| 	.cra_module		=	THIS_MODULE,
 | |
| 	.cra_u			=	{
 | |
| 		.blkcipher = {
 | |
| 			.min_keysize		=	DES_KEY_SIZE,
 | |
| 			.max_keysize		=	DES_KEY_SIZE,
 | |
| 			.setkey			=	des_setkey,
 | |
| 			.encrypt		=	ecb_des_encrypt,
 | |
| 			.decrypt		=	ecb_des_decrypt,
 | |
| 		}
 | |
| 	}
 | |
| };
 | |
| 
 | |
| static int cbc_des_encrypt(struct blkcipher_desc *desc,
 | |
| 			   struct scatterlist *dst, struct scatterlist *src,
 | |
| 			   unsigned int nbytes)
 | |
| {
 | |
| 	struct blkcipher_walk walk;
 | |
| 
 | |
| 	blkcipher_walk_init(&walk, dst, src, nbytes);
 | |
| 	return cbc_desall_crypt(desc, KMC_DEA_ENCRYPT, &walk);
 | |
| }
 | |
| 
 | |
| static int cbc_des_decrypt(struct blkcipher_desc *desc,
 | |
| 			   struct scatterlist *dst, struct scatterlist *src,
 | |
| 			   unsigned int nbytes)
 | |
| {
 | |
| 	struct blkcipher_walk walk;
 | |
| 
 | |
| 	blkcipher_walk_init(&walk, dst, src, nbytes);
 | |
| 	return cbc_desall_crypt(desc, KMC_DEA_DECRYPT, &walk);
 | |
| }
 | |
| 
 | |
| static struct crypto_alg cbc_des_alg = {
 | |
| 	.cra_name		=	"cbc(des)",
 | |
| 	.cra_driver_name	=	"cbc-des-s390",
 | |
| 	.cra_priority		=	CRYPT_S390_COMPOSITE_PRIORITY,
 | |
| 	.cra_flags		=	CRYPTO_ALG_TYPE_BLKCIPHER,
 | |
| 	.cra_blocksize		=	DES_BLOCK_SIZE,
 | |
| 	.cra_ctxsize		=	sizeof(struct s390_des_ctx),
 | |
| 	.cra_type		=	&crypto_blkcipher_type,
 | |
| 	.cra_module		=	THIS_MODULE,
 | |
| 	.cra_u			=	{
 | |
| 		.blkcipher = {
 | |
| 			.min_keysize		=	DES_KEY_SIZE,
 | |
| 			.max_keysize		=	DES_KEY_SIZE,
 | |
| 			.ivsize			=	DES_BLOCK_SIZE,
 | |
| 			.setkey			=	des_setkey,
 | |
| 			.encrypt		=	cbc_des_encrypt,
 | |
| 			.decrypt		=	cbc_des_decrypt,
 | |
| 		}
 | |
| 	}
 | |
| };
 | |
| 
 | |
| /*
 | |
|  * RFC2451:
 | |
|  *
 | |
|  *   For DES-EDE3, there is no known need to reject weak or
 | |
|  *   complementation keys.  Any weakness is obviated by the use of
 | |
|  *   multiple keys.
 | |
|  *
 | |
|  *   However, if the first two or last two independent 64-bit keys are
 | |
|  *   equal (k1 == k2 or k2 == k3), then the DES3 operation is simply the
 | |
|  *   same as DES.  Implementers MUST reject keys that exhibit this
 | |
|  *   property.
 | |
|  *
 | |
|  */
 | |
| static int des3_setkey(struct crypto_tfm *tfm, const u8 *key,
 | |
| 		       unsigned int key_len)
 | |
| {
 | |
| 	struct s390_des_ctx *ctx = crypto_tfm_ctx(tfm);
 | |
| 	u32 *flags = &tfm->crt_flags;
 | |
| 
 | |
| 	if (!(crypto_memneq(key, &key[DES_KEY_SIZE], DES_KEY_SIZE) &&
 | |
| 	    crypto_memneq(&key[DES_KEY_SIZE], &key[DES_KEY_SIZE * 2],
 | |
| 			  DES_KEY_SIZE)) &&
 | |
| 	    (*flags & CRYPTO_TFM_REQ_WEAK_KEY)) {
 | |
| 		*flags |= CRYPTO_TFM_RES_WEAK_KEY;
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 	memcpy(ctx->key, key, key_len);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void des3_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
 | |
| {
 | |
| 	struct s390_des_ctx *ctx = crypto_tfm_ctx(tfm);
 | |
| 
 | |
| 	crypt_s390_km(KM_TDEA_192_ENCRYPT, ctx->key, dst, src, DES_BLOCK_SIZE);
 | |
| }
 | |
| 
 | |
| static void des3_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
 | |
| {
 | |
| 	struct s390_des_ctx *ctx = crypto_tfm_ctx(tfm);
 | |
| 
 | |
| 	crypt_s390_km(KM_TDEA_192_DECRYPT, ctx->key, dst, src, DES_BLOCK_SIZE);
 | |
| }
 | |
| 
 | |
| static struct crypto_alg des3_alg = {
 | |
| 	.cra_name		=	"des3_ede",
 | |
| 	.cra_driver_name	=	"des3_ede-s390",
 | |
| 	.cra_priority		=	CRYPT_S390_PRIORITY,
 | |
| 	.cra_flags		=	CRYPTO_ALG_TYPE_CIPHER,
 | |
| 	.cra_blocksize		=	DES_BLOCK_SIZE,
 | |
| 	.cra_ctxsize		=	sizeof(struct s390_des_ctx),
 | |
| 	.cra_module		=	THIS_MODULE,
 | |
| 	.cra_u			=	{
 | |
| 		.cipher = {
 | |
| 			.cia_min_keysize	=	DES3_KEY_SIZE,
 | |
| 			.cia_max_keysize	=	DES3_KEY_SIZE,
 | |
| 			.cia_setkey		=	des3_setkey,
 | |
| 			.cia_encrypt		=	des3_encrypt,
 | |
| 			.cia_decrypt		=	des3_decrypt,
 | |
| 		}
 | |
| 	}
 | |
| };
 | |
| 
 | |
| static int ecb_des3_encrypt(struct blkcipher_desc *desc,
 | |
| 			    struct scatterlist *dst, struct scatterlist *src,
 | |
| 			    unsigned int nbytes)
 | |
| {
 | |
| 	struct s390_des_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
 | |
| 	struct blkcipher_walk walk;
 | |
| 
 | |
| 	blkcipher_walk_init(&walk, dst, src, nbytes);
 | |
| 	return ecb_desall_crypt(desc, KM_TDEA_192_ENCRYPT, ctx->key, &walk);
 | |
| }
 | |
| 
 | |
| static int ecb_des3_decrypt(struct blkcipher_desc *desc,
 | |
| 			    struct scatterlist *dst, struct scatterlist *src,
 | |
| 			    unsigned int nbytes)
 | |
| {
 | |
| 	struct s390_des_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
 | |
| 	struct blkcipher_walk walk;
 | |
| 
 | |
| 	blkcipher_walk_init(&walk, dst, src, nbytes);
 | |
| 	return ecb_desall_crypt(desc, KM_TDEA_192_DECRYPT, ctx->key, &walk);
 | |
| }
 | |
| 
 | |
| static struct crypto_alg ecb_des3_alg = {
 | |
| 	.cra_name		=	"ecb(des3_ede)",
 | |
| 	.cra_driver_name	=	"ecb-des3_ede-s390",
 | |
| 	.cra_priority		=	CRYPT_S390_COMPOSITE_PRIORITY,
 | |
| 	.cra_flags		=	CRYPTO_ALG_TYPE_BLKCIPHER,
 | |
| 	.cra_blocksize		=	DES_BLOCK_SIZE,
 | |
| 	.cra_ctxsize		=	sizeof(struct s390_des_ctx),
 | |
| 	.cra_type		=	&crypto_blkcipher_type,
 | |
| 	.cra_module		=	THIS_MODULE,
 | |
| 	.cra_u			=	{
 | |
| 		.blkcipher = {
 | |
| 			.min_keysize		=	DES3_KEY_SIZE,
 | |
| 			.max_keysize		=	DES3_KEY_SIZE,
 | |
| 			.setkey			=	des3_setkey,
 | |
| 			.encrypt		=	ecb_des3_encrypt,
 | |
| 			.decrypt		=	ecb_des3_decrypt,
 | |
| 		}
 | |
| 	}
 | |
| };
 | |
| 
 | |
| static int cbc_des3_encrypt(struct blkcipher_desc *desc,
 | |
| 			    struct scatterlist *dst, struct scatterlist *src,
 | |
| 			    unsigned int nbytes)
 | |
| {
 | |
| 	struct blkcipher_walk walk;
 | |
| 
 | |
| 	blkcipher_walk_init(&walk, dst, src, nbytes);
 | |
| 	return cbc_desall_crypt(desc, KMC_TDEA_192_ENCRYPT, &walk);
 | |
| }
 | |
| 
 | |
| static int cbc_des3_decrypt(struct blkcipher_desc *desc,
 | |
| 			    struct scatterlist *dst, struct scatterlist *src,
 | |
| 			    unsigned int nbytes)
 | |
| {
 | |
| 	struct blkcipher_walk walk;
 | |
| 
 | |
| 	blkcipher_walk_init(&walk, dst, src, nbytes);
 | |
| 	return cbc_desall_crypt(desc, KMC_TDEA_192_DECRYPT, &walk);
 | |
| }
 | |
| 
 | |
| static struct crypto_alg cbc_des3_alg = {
 | |
| 	.cra_name		=	"cbc(des3_ede)",
 | |
| 	.cra_driver_name	=	"cbc-des3_ede-s390",
 | |
| 	.cra_priority		=	CRYPT_S390_COMPOSITE_PRIORITY,
 | |
| 	.cra_flags		=	CRYPTO_ALG_TYPE_BLKCIPHER,
 | |
| 	.cra_blocksize		=	DES_BLOCK_SIZE,
 | |
| 	.cra_ctxsize		=	sizeof(struct s390_des_ctx),
 | |
| 	.cra_type		=	&crypto_blkcipher_type,
 | |
| 	.cra_module		=	THIS_MODULE,
 | |
| 	.cra_u			=	{
 | |
| 		.blkcipher = {
 | |
| 			.min_keysize		=	DES3_KEY_SIZE,
 | |
| 			.max_keysize		=	DES3_KEY_SIZE,
 | |
| 			.ivsize			=	DES_BLOCK_SIZE,
 | |
| 			.setkey			=	des3_setkey,
 | |
| 			.encrypt		=	cbc_des3_encrypt,
 | |
| 			.decrypt		=	cbc_des3_decrypt,
 | |
| 		}
 | |
| 	}
 | |
| };
 | |
| 
 | |
| static unsigned int __ctrblk_init(u8 *ctrptr, unsigned int nbytes)
 | |
| {
 | |
| 	unsigned int i, n;
 | |
| 
 | |
| 	/* align to block size, max. PAGE_SIZE */
 | |
| 	n = (nbytes > PAGE_SIZE) ? PAGE_SIZE : nbytes & ~(DES_BLOCK_SIZE - 1);
 | |
| 	for (i = DES_BLOCK_SIZE; i < n; i += DES_BLOCK_SIZE) {
 | |
| 		memcpy(ctrptr + i, ctrptr + i - DES_BLOCK_SIZE, DES_BLOCK_SIZE);
 | |
| 		crypto_inc(ctrptr + i, DES_BLOCK_SIZE);
 | |
| 	}
 | |
| 	return n;
 | |
| }
 | |
| 
 | |
| static int ctr_desall_crypt(struct blkcipher_desc *desc, long func,
 | |
| 			    struct s390_des_ctx *ctx,
 | |
| 			    struct blkcipher_walk *walk)
 | |
| {
 | |
| 	int ret = blkcipher_walk_virt_block(desc, walk, DES_BLOCK_SIZE);
 | |
| 	unsigned int n, nbytes;
 | |
| 	u8 buf[DES_BLOCK_SIZE], ctrbuf[DES_BLOCK_SIZE];
 | |
| 	u8 *out, *in, *ctrptr = ctrbuf;
 | |
| 
 | |
| 	if (!walk->nbytes)
 | |
| 		return ret;
 | |
| 
 | |
| 	if (spin_trylock(&ctrblk_lock))
 | |
| 		ctrptr = ctrblk;
 | |
| 
 | |
| 	memcpy(ctrptr, walk->iv, DES_BLOCK_SIZE);
 | |
| 	while ((nbytes = walk->nbytes) >= DES_BLOCK_SIZE) {
 | |
| 		out = walk->dst.virt.addr;
 | |
| 		in = walk->src.virt.addr;
 | |
| 		while (nbytes >= DES_BLOCK_SIZE) {
 | |
| 			if (ctrptr == ctrblk)
 | |
| 				n = __ctrblk_init(ctrptr, nbytes);
 | |
| 			else
 | |
| 				n = DES_BLOCK_SIZE;
 | |
| 			ret = crypt_s390_kmctr(func, ctx->key, out, in,
 | |
| 					       n, ctrptr);
 | |
| 			if (ret < 0 || ret != n) {
 | |
| 				if (ctrptr == ctrblk)
 | |
| 					spin_unlock(&ctrblk_lock);
 | |
| 				return -EIO;
 | |
| 			}
 | |
| 			if (n > DES_BLOCK_SIZE)
 | |
| 				memcpy(ctrptr, ctrptr + n - DES_BLOCK_SIZE,
 | |
| 				       DES_BLOCK_SIZE);
 | |
| 			crypto_inc(ctrptr, DES_BLOCK_SIZE);
 | |
| 			out += n;
 | |
| 			in += n;
 | |
| 			nbytes -= n;
 | |
| 		}
 | |
| 		ret = blkcipher_walk_done(desc, walk, nbytes);
 | |
| 	}
 | |
| 	if (ctrptr == ctrblk) {
 | |
| 		if (nbytes)
 | |
| 			memcpy(ctrbuf, ctrptr, DES_BLOCK_SIZE);
 | |
| 		else
 | |
| 			memcpy(walk->iv, ctrptr, DES_BLOCK_SIZE);
 | |
| 		spin_unlock(&ctrblk_lock);
 | |
| 	} else {
 | |
| 		if (!nbytes)
 | |
| 			memcpy(walk->iv, ctrptr, DES_BLOCK_SIZE);
 | |
| 	}
 | |
| 	/* final block may be < DES_BLOCK_SIZE, copy only nbytes */
 | |
| 	if (nbytes) {
 | |
| 		out = walk->dst.virt.addr;
 | |
| 		in = walk->src.virt.addr;
 | |
| 		ret = crypt_s390_kmctr(func, ctx->key, buf, in,
 | |
| 				       DES_BLOCK_SIZE, ctrbuf);
 | |
| 		if (ret < 0 || ret != DES_BLOCK_SIZE)
 | |
| 			return -EIO;
 | |
| 		memcpy(out, buf, nbytes);
 | |
| 		crypto_inc(ctrbuf, DES_BLOCK_SIZE);
 | |
| 		ret = blkcipher_walk_done(desc, walk, 0);
 | |
| 		memcpy(walk->iv, ctrbuf, DES_BLOCK_SIZE);
 | |
| 	}
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int ctr_des_encrypt(struct blkcipher_desc *desc,
 | |
| 			   struct scatterlist *dst, struct scatterlist *src,
 | |
| 			   unsigned int nbytes)
 | |
| {
 | |
| 	struct s390_des_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
 | |
| 	struct blkcipher_walk walk;
 | |
| 
 | |
| 	blkcipher_walk_init(&walk, dst, src, nbytes);
 | |
| 	return ctr_desall_crypt(desc, KMCTR_DEA_ENCRYPT, ctx, &walk);
 | |
| }
 | |
| 
 | |
| static int ctr_des_decrypt(struct blkcipher_desc *desc,
 | |
| 			   struct scatterlist *dst, struct scatterlist *src,
 | |
| 			   unsigned int nbytes)
 | |
| {
 | |
| 	struct s390_des_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
 | |
| 	struct blkcipher_walk walk;
 | |
| 
 | |
| 	blkcipher_walk_init(&walk, dst, src, nbytes);
 | |
| 	return ctr_desall_crypt(desc, KMCTR_DEA_DECRYPT, ctx, &walk);
 | |
| }
 | |
| 
 | |
| static struct crypto_alg ctr_des_alg = {
 | |
| 	.cra_name		=	"ctr(des)",
 | |
| 	.cra_driver_name	=	"ctr-des-s390",
 | |
| 	.cra_priority		=	CRYPT_S390_COMPOSITE_PRIORITY,
 | |
| 	.cra_flags		=	CRYPTO_ALG_TYPE_BLKCIPHER,
 | |
| 	.cra_blocksize		=	1,
 | |
| 	.cra_ctxsize		=	sizeof(struct s390_des_ctx),
 | |
| 	.cra_type		=	&crypto_blkcipher_type,
 | |
| 	.cra_module		=	THIS_MODULE,
 | |
| 	.cra_u			=	{
 | |
| 		.blkcipher = {
 | |
| 			.min_keysize		=	DES_KEY_SIZE,
 | |
| 			.max_keysize		=	DES_KEY_SIZE,
 | |
| 			.ivsize			=	DES_BLOCK_SIZE,
 | |
| 			.setkey			=	des_setkey,
 | |
| 			.encrypt		=	ctr_des_encrypt,
 | |
| 			.decrypt		=	ctr_des_decrypt,
 | |
| 		}
 | |
| 	}
 | |
| };
 | |
| 
 | |
| static int ctr_des3_encrypt(struct blkcipher_desc *desc,
 | |
| 			    struct scatterlist *dst, struct scatterlist *src,
 | |
| 			    unsigned int nbytes)
 | |
| {
 | |
| 	struct s390_des_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
 | |
| 	struct blkcipher_walk walk;
 | |
| 
 | |
| 	blkcipher_walk_init(&walk, dst, src, nbytes);
 | |
| 	return ctr_desall_crypt(desc, KMCTR_TDEA_192_ENCRYPT, ctx, &walk);
 | |
| }
 | |
| 
 | |
| static int ctr_des3_decrypt(struct blkcipher_desc *desc,
 | |
| 			    struct scatterlist *dst, struct scatterlist *src,
 | |
| 			    unsigned int nbytes)
 | |
| {
 | |
| 	struct s390_des_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
 | |
| 	struct blkcipher_walk walk;
 | |
| 
 | |
| 	blkcipher_walk_init(&walk, dst, src, nbytes);
 | |
| 	return ctr_desall_crypt(desc, KMCTR_TDEA_192_DECRYPT, ctx, &walk);
 | |
| }
 | |
| 
 | |
| static struct crypto_alg ctr_des3_alg = {
 | |
| 	.cra_name		=	"ctr(des3_ede)",
 | |
| 	.cra_driver_name	=	"ctr-des3_ede-s390",
 | |
| 	.cra_priority		=	CRYPT_S390_COMPOSITE_PRIORITY,
 | |
| 	.cra_flags		=	CRYPTO_ALG_TYPE_BLKCIPHER,
 | |
| 	.cra_blocksize		=	1,
 | |
| 	.cra_ctxsize		=	sizeof(struct s390_des_ctx),
 | |
| 	.cra_type		=	&crypto_blkcipher_type,
 | |
| 	.cra_module		=	THIS_MODULE,
 | |
| 	.cra_u			=	{
 | |
| 		.blkcipher = {
 | |
| 			.min_keysize		=	DES3_KEY_SIZE,
 | |
| 			.max_keysize		=	DES3_KEY_SIZE,
 | |
| 			.ivsize			=	DES_BLOCK_SIZE,
 | |
| 			.setkey			=	des3_setkey,
 | |
| 			.encrypt		=	ctr_des3_encrypt,
 | |
| 			.decrypt		=	ctr_des3_decrypt,
 | |
| 		}
 | |
| 	}
 | |
| };
 | |
| 
 | |
| static int __init des_s390_init(void)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	if (!crypt_s390_func_available(KM_DEA_ENCRYPT, CRYPT_S390_MSA) ||
 | |
| 	    !crypt_s390_func_available(KM_TDEA_192_ENCRYPT, CRYPT_S390_MSA))
 | |
| 		return -EOPNOTSUPP;
 | |
| 
 | |
| 	ret = crypto_register_alg(&des_alg);
 | |
| 	if (ret)
 | |
| 		goto des_err;
 | |
| 	ret = crypto_register_alg(&ecb_des_alg);
 | |
| 	if (ret)
 | |
| 		goto ecb_des_err;
 | |
| 	ret = crypto_register_alg(&cbc_des_alg);
 | |
| 	if (ret)
 | |
| 		goto cbc_des_err;
 | |
| 	ret = crypto_register_alg(&des3_alg);
 | |
| 	if (ret)
 | |
| 		goto des3_err;
 | |
| 	ret = crypto_register_alg(&ecb_des3_alg);
 | |
| 	if (ret)
 | |
| 		goto ecb_des3_err;
 | |
| 	ret = crypto_register_alg(&cbc_des3_alg);
 | |
| 	if (ret)
 | |
| 		goto cbc_des3_err;
 | |
| 
 | |
| 	if (crypt_s390_func_available(KMCTR_DEA_ENCRYPT,
 | |
| 			CRYPT_S390_MSA | CRYPT_S390_MSA4) &&
 | |
| 	    crypt_s390_func_available(KMCTR_TDEA_192_ENCRYPT,
 | |
| 			CRYPT_S390_MSA | CRYPT_S390_MSA4)) {
 | |
| 		ret = crypto_register_alg(&ctr_des_alg);
 | |
| 		if (ret)
 | |
| 			goto ctr_des_err;
 | |
| 		ret = crypto_register_alg(&ctr_des3_alg);
 | |
| 		if (ret)
 | |
| 			goto ctr_des3_err;
 | |
| 		ctrblk = (u8 *) __get_free_page(GFP_KERNEL);
 | |
| 		if (!ctrblk) {
 | |
| 			ret = -ENOMEM;
 | |
| 			goto ctr_mem_err;
 | |
| 		}
 | |
| 	}
 | |
| out:
 | |
| 	return ret;
 | |
| 
 | |
| ctr_mem_err:
 | |
| 	crypto_unregister_alg(&ctr_des3_alg);
 | |
| ctr_des3_err:
 | |
| 	crypto_unregister_alg(&ctr_des_alg);
 | |
| ctr_des_err:
 | |
| 	crypto_unregister_alg(&cbc_des3_alg);
 | |
| cbc_des3_err:
 | |
| 	crypto_unregister_alg(&ecb_des3_alg);
 | |
| ecb_des3_err:
 | |
| 	crypto_unregister_alg(&des3_alg);
 | |
| des3_err:
 | |
| 	crypto_unregister_alg(&cbc_des_alg);
 | |
| cbc_des_err:
 | |
| 	crypto_unregister_alg(&ecb_des_alg);
 | |
| ecb_des_err:
 | |
| 	crypto_unregister_alg(&des_alg);
 | |
| des_err:
 | |
| 	goto out;
 | |
| }
 | |
| 
 | |
| static void __exit des_s390_exit(void)
 | |
| {
 | |
| 	if (ctrblk) {
 | |
| 		crypto_unregister_alg(&ctr_des_alg);
 | |
| 		crypto_unregister_alg(&ctr_des3_alg);
 | |
| 		free_page((unsigned long) ctrblk);
 | |
| 	}
 | |
| 	crypto_unregister_alg(&cbc_des3_alg);
 | |
| 	crypto_unregister_alg(&ecb_des3_alg);
 | |
| 	crypto_unregister_alg(&des3_alg);
 | |
| 	crypto_unregister_alg(&cbc_des_alg);
 | |
| 	crypto_unregister_alg(&ecb_des_alg);
 | |
| 	crypto_unregister_alg(&des_alg);
 | |
| }
 | |
| 
 | |
| module_init(des_s390_init);
 | |
| module_exit(des_s390_exit);
 | |
| 
 | |
| MODULE_ALIAS_CRYPTO("des");
 | |
| MODULE_ALIAS_CRYPTO("des3_ede");
 | |
| 
 | |
| MODULE_LICENSE("GPL");
 | |
| MODULE_DESCRIPTION("DES & Triple DES EDE Cipher Algorithms");
 |