mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-02 08:34:20 +08:00
91b05a7e7d
Currently, NETLINK_CRYPTO works only in the init network namespace. It doesn't make much sense to cut it out of the other network namespaces, so do the minor plumbing work necessary to make it work in any network namespace. Code inspired by net/core/sock_diag.c. Tested using kcapi-dgst from libkcapi [1]: Before: # unshare -n kcapi-dgst -c sha256 </dev/null | wc -c libkcapi - Error: Netlink error: sendmsg failed libkcapi - Error: Netlink error: sendmsg failed libkcapi - Error: NETLINK_CRYPTO: cannot obtain cipher information for hmac(sha512) (is required crypto_user.c patch missing? see documentation) 0 After: # unshare -n kcapi-dgst -c sha256 </dev/null | wc -c 32 [1] https://github.com/smuellerDD/libkcapi Signed-off-by: Ondrej Mosnacek <omosnace@redhat.com> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
338 lines
9.7 KiB
C
338 lines
9.7 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Crypto user configuration API.
|
|
*
|
|
* Copyright (C) 2017-2018 Corentin Labbe <clabbe@baylibre.com>
|
|
*
|
|
*/
|
|
|
|
#include <linux/crypto.h>
|
|
#include <linux/cryptouser.h>
|
|
#include <linux/sched.h>
|
|
#include <net/netlink.h>
|
|
#include <net/sock.h>
|
|
#include <crypto/internal/skcipher.h>
|
|
#include <crypto/internal/rng.h>
|
|
#include <crypto/akcipher.h>
|
|
#include <crypto/kpp.h>
|
|
#include <crypto/internal/cryptouser.h>
|
|
|
|
#include "internal.h"
|
|
|
|
#define null_terminated(x) (strnlen(x, sizeof(x)) < sizeof(x))
|
|
|
|
struct crypto_dump_info {
|
|
struct sk_buff *in_skb;
|
|
struct sk_buff *out_skb;
|
|
u32 nlmsg_seq;
|
|
u16 nlmsg_flags;
|
|
};
|
|
|
|
static int crypto_report_aead(struct sk_buff *skb, struct crypto_alg *alg)
|
|
{
|
|
struct crypto_stat_aead raead;
|
|
|
|
memset(&raead, 0, sizeof(raead));
|
|
|
|
strscpy(raead.type, "aead", sizeof(raead.type));
|
|
|
|
raead.stat_encrypt_cnt = atomic64_read(&alg->stats.aead.encrypt_cnt);
|
|
raead.stat_encrypt_tlen = atomic64_read(&alg->stats.aead.encrypt_tlen);
|
|
raead.stat_decrypt_cnt = atomic64_read(&alg->stats.aead.decrypt_cnt);
|
|
raead.stat_decrypt_tlen = atomic64_read(&alg->stats.aead.decrypt_tlen);
|
|
raead.stat_err_cnt = atomic64_read(&alg->stats.aead.err_cnt);
|
|
|
|
return nla_put(skb, CRYPTOCFGA_STAT_AEAD, sizeof(raead), &raead);
|
|
}
|
|
|
|
static int crypto_report_cipher(struct sk_buff *skb, struct crypto_alg *alg)
|
|
{
|
|
struct crypto_stat_cipher rcipher;
|
|
|
|
memset(&rcipher, 0, sizeof(rcipher));
|
|
|
|
strscpy(rcipher.type, "cipher", sizeof(rcipher.type));
|
|
|
|
rcipher.stat_encrypt_cnt = atomic64_read(&alg->stats.cipher.encrypt_cnt);
|
|
rcipher.stat_encrypt_tlen = atomic64_read(&alg->stats.cipher.encrypt_tlen);
|
|
rcipher.stat_decrypt_cnt = atomic64_read(&alg->stats.cipher.decrypt_cnt);
|
|
rcipher.stat_decrypt_tlen = atomic64_read(&alg->stats.cipher.decrypt_tlen);
|
|
rcipher.stat_err_cnt = atomic64_read(&alg->stats.cipher.err_cnt);
|
|
|
|
return nla_put(skb, CRYPTOCFGA_STAT_CIPHER, sizeof(rcipher), &rcipher);
|
|
}
|
|
|
|
static int crypto_report_comp(struct sk_buff *skb, struct crypto_alg *alg)
|
|
{
|
|
struct crypto_stat_compress rcomp;
|
|
|
|
memset(&rcomp, 0, sizeof(rcomp));
|
|
|
|
strscpy(rcomp.type, "compression", sizeof(rcomp.type));
|
|
rcomp.stat_compress_cnt = atomic64_read(&alg->stats.compress.compress_cnt);
|
|
rcomp.stat_compress_tlen = atomic64_read(&alg->stats.compress.compress_tlen);
|
|
rcomp.stat_decompress_cnt = atomic64_read(&alg->stats.compress.decompress_cnt);
|
|
rcomp.stat_decompress_tlen = atomic64_read(&alg->stats.compress.decompress_tlen);
|
|
rcomp.stat_err_cnt = atomic64_read(&alg->stats.compress.err_cnt);
|
|
|
|
return nla_put(skb, CRYPTOCFGA_STAT_COMPRESS, sizeof(rcomp), &rcomp);
|
|
}
|
|
|
|
static int crypto_report_acomp(struct sk_buff *skb, struct crypto_alg *alg)
|
|
{
|
|
struct crypto_stat_compress racomp;
|
|
|
|
memset(&racomp, 0, sizeof(racomp));
|
|
|
|
strscpy(racomp.type, "acomp", sizeof(racomp.type));
|
|
racomp.stat_compress_cnt = atomic64_read(&alg->stats.compress.compress_cnt);
|
|
racomp.stat_compress_tlen = atomic64_read(&alg->stats.compress.compress_tlen);
|
|
racomp.stat_decompress_cnt = atomic64_read(&alg->stats.compress.decompress_cnt);
|
|
racomp.stat_decompress_tlen = atomic64_read(&alg->stats.compress.decompress_tlen);
|
|
racomp.stat_err_cnt = atomic64_read(&alg->stats.compress.err_cnt);
|
|
|
|
return nla_put(skb, CRYPTOCFGA_STAT_ACOMP, sizeof(racomp), &racomp);
|
|
}
|
|
|
|
static int crypto_report_akcipher(struct sk_buff *skb, struct crypto_alg *alg)
|
|
{
|
|
struct crypto_stat_akcipher rakcipher;
|
|
|
|
memset(&rakcipher, 0, sizeof(rakcipher));
|
|
|
|
strscpy(rakcipher.type, "akcipher", sizeof(rakcipher.type));
|
|
rakcipher.stat_encrypt_cnt = atomic64_read(&alg->stats.akcipher.encrypt_cnt);
|
|
rakcipher.stat_encrypt_tlen = atomic64_read(&alg->stats.akcipher.encrypt_tlen);
|
|
rakcipher.stat_decrypt_cnt = atomic64_read(&alg->stats.akcipher.decrypt_cnt);
|
|
rakcipher.stat_decrypt_tlen = atomic64_read(&alg->stats.akcipher.decrypt_tlen);
|
|
rakcipher.stat_sign_cnt = atomic64_read(&alg->stats.akcipher.sign_cnt);
|
|
rakcipher.stat_verify_cnt = atomic64_read(&alg->stats.akcipher.verify_cnt);
|
|
rakcipher.stat_err_cnt = atomic64_read(&alg->stats.akcipher.err_cnt);
|
|
|
|
return nla_put(skb, CRYPTOCFGA_STAT_AKCIPHER,
|
|
sizeof(rakcipher), &rakcipher);
|
|
}
|
|
|
|
static int crypto_report_kpp(struct sk_buff *skb, struct crypto_alg *alg)
|
|
{
|
|
struct crypto_stat_kpp rkpp;
|
|
|
|
memset(&rkpp, 0, sizeof(rkpp));
|
|
|
|
strscpy(rkpp.type, "kpp", sizeof(rkpp.type));
|
|
|
|
rkpp.stat_setsecret_cnt = atomic64_read(&alg->stats.kpp.setsecret_cnt);
|
|
rkpp.stat_generate_public_key_cnt = atomic64_read(&alg->stats.kpp.generate_public_key_cnt);
|
|
rkpp.stat_compute_shared_secret_cnt = atomic64_read(&alg->stats.kpp.compute_shared_secret_cnt);
|
|
rkpp.stat_err_cnt = atomic64_read(&alg->stats.kpp.err_cnt);
|
|
|
|
return nla_put(skb, CRYPTOCFGA_STAT_KPP, sizeof(rkpp), &rkpp);
|
|
}
|
|
|
|
static int crypto_report_ahash(struct sk_buff *skb, struct crypto_alg *alg)
|
|
{
|
|
struct crypto_stat_hash rhash;
|
|
|
|
memset(&rhash, 0, sizeof(rhash));
|
|
|
|
strscpy(rhash.type, "ahash", sizeof(rhash.type));
|
|
|
|
rhash.stat_hash_cnt = atomic64_read(&alg->stats.hash.hash_cnt);
|
|
rhash.stat_hash_tlen = atomic64_read(&alg->stats.hash.hash_tlen);
|
|
rhash.stat_err_cnt = atomic64_read(&alg->stats.hash.err_cnt);
|
|
|
|
return nla_put(skb, CRYPTOCFGA_STAT_HASH, sizeof(rhash), &rhash);
|
|
}
|
|
|
|
static int crypto_report_shash(struct sk_buff *skb, struct crypto_alg *alg)
|
|
{
|
|
struct crypto_stat_hash rhash;
|
|
|
|
memset(&rhash, 0, sizeof(rhash));
|
|
|
|
strscpy(rhash.type, "shash", sizeof(rhash.type));
|
|
|
|
rhash.stat_hash_cnt = atomic64_read(&alg->stats.hash.hash_cnt);
|
|
rhash.stat_hash_tlen = atomic64_read(&alg->stats.hash.hash_tlen);
|
|
rhash.stat_err_cnt = atomic64_read(&alg->stats.hash.err_cnt);
|
|
|
|
return nla_put(skb, CRYPTOCFGA_STAT_HASH, sizeof(rhash), &rhash);
|
|
}
|
|
|
|
static int crypto_report_rng(struct sk_buff *skb, struct crypto_alg *alg)
|
|
{
|
|
struct crypto_stat_rng rrng;
|
|
|
|
memset(&rrng, 0, sizeof(rrng));
|
|
|
|
strscpy(rrng.type, "rng", sizeof(rrng.type));
|
|
|
|
rrng.stat_generate_cnt = atomic64_read(&alg->stats.rng.generate_cnt);
|
|
rrng.stat_generate_tlen = atomic64_read(&alg->stats.rng.generate_tlen);
|
|
rrng.stat_seed_cnt = atomic64_read(&alg->stats.rng.seed_cnt);
|
|
rrng.stat_err_cnt = atomic64_read(&alg->stats.rng.err_cnt);
|
|
|
|
return nla_put(skb, CRYPTOCFGA_STAT_RNG, sizeof(rrng), &rrng);
|
|
}
|
|
|
|
static int crypto_reportstat_one(struct crypto_alg *alg,
|
|
struct crypto_user_alg *ualg,
|
|
struct sk_buff *skb)
|
|
{
|
|
memset(ualg, 0, sizeof(*ualg));
|
|
|
|
strscpy(ualg->cru_name, alg->cra_name, sizeof(ualg->cru_name));
|
|
strscpy(ualg->cru_driver_name, alg->cra_driver_name,
|
|
sizeof(ualg->cru_driver_name));
|
|
strscpy(ualg->cru_module_name, module_name(alg->cra_module),
|
|
sizeof(ualg->cru_module_name));
|
|
|
|
ualg->cru_type = 0;
|
|
ualg->cru_mask = 0;
|
|
ualg->cru_flags = alg->cra_flags;
|
|
ualg->cru_refcnt = refcount_read(&alg->cra_refcnt);
|
|
|
|
if (nla_put_u32(skb, CRYPTOCFGA_PRIORITY_VAL, alg->cra_priority))
|
|
goto nla_put_failure;
|
|
if (alg->cra_flags & CRYPTO_ALG_LARVAL) {
|
|
struct crypto_stat_larval rl;
|
|
|
|
memset(&rl, 0, sizeof(rl));
|
|
strscpy(rl.type, "larval", sizeof(rl.type));
|
|
if (nla_put(skb, CRYPTOCFGA_STAT_LARVAL, sizeof(rl), &rl))
|
|
goto nla_put_failure;
|
|
goto out;
|
|
}
|
|
|
|
switch (alg->cra_flags & (CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_LARVAL)) {
|
|
case CRYPTO_ALG_TYPE_AEAD:
|
|
if (crypto_report_aead(skb, alg))
|
|
goto nla_put_failure;
|
|
break;
|
|
case CRYPTO_ALG_TYPE_SKCIPHER:
|
|
if (crypto_report_cipher(skb, alg))
|
|
goto nla_put_failure;
|
|
break;
|
|
case CRYPTO_ALG_TYPE_BLKCIPHER:
|
|
if (crypto_report_cipher(skb, alg))
|
|
goto nla_put_failure;
|
|
break;
|
|
case CRYPTO_ALG_TYPE_CIPHER:
|
|
if (crypto_report_cipher(skb, alg))
|
|
goto nla_put_failure;
|
|
break;
|
|
case CRYPTO_ALG_TYPE_COMPRESS:
|
|
if (crypto_report_comp(skb, alg))
|
|
goto nla_put_failure;
|
|
break;
|
|
case CRYPTO_ALG_TYPE_ACOMPRESS:
|
|
if (crypto_report_acomp(skb, alg))
|
|
goto nla_put_failure;
|
|
break;
|
|
case CRYPTO_ALG_TYPE_SCOMPRESS:
|
|
if (crypto_report_acomp(skb, alg))
|
|
goto nla_put_failure;
|
|
break;
|
|
case CRYPTO_ALG_TYPE_AKCIPHER:
|
|
if (crypto_report_akcipher(skb, alg))
|
|
goto nla_put_failure;
|
|
break;
|
|
case CRYPTO_ALG_TYPE_KPP:
|
|
if (crypto_report_kpp(skb, alg))
|
|
goto nla_put_failure;
|
|
break;
|
|
case CRYPTO_ALG_TYPE_AHASH:
|
|
if (crypto_report_ahash(skb, alg))
|
|
goto nla_put_failure;
|
|
break;
|
|
case CRYPTO_ALG_TYPE_HASH:
|
|
if (crypto_report_shash(skb, alg))
|
|
goto nla_put_failure;
|
|
break;
|
|
case CRYPTO_ALG_TYPE_RNG:
|
|
if (crypto_report_rng(skb, alg))
|
|
goto nla_put_failure;
|
|
break;
|
|
default:
|
|
pr_err("ERROR: Unhandled alg %d in %s\n",
|
|
alg->cra_flags & (CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_LARVAL),
|
|
__func__);
|
|
}
|
|
|
|
out:
|
|
return 0;
|
|
|
|
nla_put_failure:
|
|
return -EMSGSIZE;
|
|
}
|
|
|
|
static int crypto_reportstat_alg(struct crypto_alg *alg,
|
|
struct crypto_dump_info *info)
|
|
{
|
|
struct sk_buff *in_skb = info->in_skb;
|
|
struct sk_buff *skb = info->out_skb;
|
|
struct nlmsghdr *nlh;
|
|
struct crypto_user_alg *ualg;
|
|
int err = 0;
|
|
|
|
nlh = nlmsg_put(skb, NETLINK_CB(in_skb).portid, info->nlmsg_seq,
|
|
CRYPTO_MSG_GETSTAT, sizeof(*ualg), info->nlmsg_flags);
|
|
if (!nlh) {
|
|
err = -EMSGSIZE;
|
|
goto out;
|
|
}
|
|
|
|
ualg = nlmsg_data(nlh);
|
|
|
|
err = crypto_reportstat_one(alg, ualg, skb);
|
|
if (err) {
|
|
nlmsg_cancel(skb, nlh);
|
|
goto out;
|
|
}
|
|
|
|
nlmsg_end(skb, nlh);
|
|
|
|
out:
|
|
return err;
|
|
}
|
|
|
|
int crypto_reportstat(struct sk_buff *in_skb, struct nlmsghdr *in_nlh,
|
|
struct nlattr **attrs)
|
|
{
|
|
struct net *net = sock_net(in_skb->sk);
|
|
struct crypto_user_alg *p = nlmsg_data(in_nlh);
|
|
struct crypto_alg *alg;
|
|
struct sk_buff *skb;
|
|
struct crypto_dump_info info;
|
|
int err;
|
|
|
|
if (!null_terminated(p->cru_name) || !null_terminated(p->cru_driver_name))
|
|
return -EINVAL;
|
|
|
|
alg = crypto_alg_match(p, 0);
|
|
if (!alg)
|
|
return -ENOENT;
|
|
|
|
err = -ENOMEM;
|
|
skb = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_ATOMIC);
|
|
if (!skb)
|
|
goto drop_alg;
|
|
|
|
info.in_skb = in_skb;
|
|
info.out_skb = skb;
|
|
info.nlmsg_seq = in_nlh->nlmsg_seq;
|
|
info.nlmsg_flags = 0;
|
|
|
|
err = crypto_reportstat_alg(alg, &info);
|
|
|
|
drop_alg:
|
|
crypto_mod_put(alg);
|
|
|
|
if (err)
|
|
return err;
|
|
|
|
return nlmsg_unicast(net->crypto_nlsk, skb, NETLINK_CB(in_skb).portid);
|
|
}
|
|
|
|
MODULE_LICENSE("GPL");
|