linux/crypto/acompress.c
<<
>>
Prefs
   1// SPDX-License-Identifier: GPL-2.0-or-later
   2/*
   3 * Asynchronous Compression operations
   4 *
   5 * Copyright (c) 2016, Intel Corporation
   6 * Authors: Weigang Li <weigang.li@intel.com>
   7 *          Giovanni Cabiddu <giovanni.cabiddu@intel.com>
   8 */
   9
  10#include <crypto/internal/acompress.h>
  11#include <linux/cryptouser.h>
  12#include <linux/errno.h>
  13#include <linux/kernel.h>
  14#include <linux/module.h>
  15#include <linux/seq_file.h>
  16#include <linux/slab.h>
  17#include <linux/string.h>
  18#include <net/netlink.h>
  19
  20#include "compress.h"
  21
  22struct crypto_scomp;
  23
  24static const struct crypto_type crypto_acomp_type;
  25
  26static inline struct acomp_alg *__crypto_acomp_alg(struct crypto_alg *alg)
  27{
  28        return container_of(alg, struct acomp_alg, calg.base);
  29}
  30
  31static inline struct acomp_alg *crypto_acomp_alg(struct crypto_acomp *tfm)
  32{
  33        return __crypto_acomp_alg(crypto_acomp_tfm(tfm)->__crt_alg);
  34}
  35
  36static int __maybe_unused crypto_acomp_report(
  37        struct sk_buff *skb, struct crypto_alg *alg)
  38{
  39        struct crypto_report_acomp racomp;
  40
  41        memset(&racomp, 0, sizeof(racomp));
  42
  43        strscpy(racomp.type, "acomp", sizeof(racomp.type));
  44
  45        return nla_put(skb, CRYPTOCFGA_REPORT_ACOMP, sizeof(racomp), &racomp);
  46}
  47
  48static void crypto_acomp_show(struct seq_file *m, struct crypto_alg *alg)
  49        __maybe_unused;
  50
  51static void crypto_acomp_show(struct seq_file *m, struct crypto_alg *alg)
  52{
  53        seq_puts(m, "type         : acomp\n");
  54}
  55
  56static void crypto_acomp_exit_tfm(struct crypto_tfm *tfm)
  57{
  58        struct crypto_acomp *acomp = __crypto_acomp_tfm(tfm);
  59        struct acomp_alg *alg = crypto_acomp_alg(acomp);
  60
  61        alg->exit(acomp);
  62}
  63
  64static int crypto_acomp_init_tfm(struct crypto_tfm *tfm)
  65{
  66        struct crypto_acomp *acomp = __crypto_acomp_tfm(tfm);
  67        struct acomp_alg *alg = crypto_acomp_alg(acomp);
  68
  69        if (tfm->__crt_alg->cra_type != &crypto_acomp_type)
  70                return crypto_init_scomp_ops_async(tfm);
  71
  72        acomp->compress = alg->compress;
  73        acomp->decompress = alg->decompress;
  74        acomp->dst_free = alg->dst_free;
  75        acomp->reqsize = alg->reqsize;
  76
  77        if (alg->exit)
  78                acomp->base.exit = crypto_acomp_exit_tfm;
  79
  80        if (alg->init)
  81                return alg->init(acomp);
  82
  83        return 0;
  84}
  85
  86static unsigned int crypto_acomp_extsize(struct crypto_alg *alg)
  87{
  88        int extsize = crypto_alg_extsize(alg);
  89
  90        if (alg->cra_type != &crypto_acomp_type)
  91                extsize += sizeof(struct crypto_scomp *);
  92
  93        return extsize;
  94}
  95
  96static inline int __crypto_acomp_report_stat(struct sk_buff *skb,
  97                                             struct crypto_alg *alg)
  98{
  99        struct comp_alg_common *calg = __crypto_comp_alg_common(alg);
 100        struct crypto_istat_compress *istat = comp_get_stat(calg);
 101        struct crypto_stat_compress racomp;
 102
 103        memset(&racomp, 0, sizeof(racomp));
 104
 105        strscpy(racomp.type, "acomp", sizeof(racomp.type));
 106        racomp.stat_compress_cnt = atomic64_read(&istat->compress_cnt);
 107        racomp.stat_compress_tlen = atomic64_read(&istat->compress_tlen);
 108        racomp.stat_decompress_cnt =  atomic64_read(&istat->decompress_cnt);
 109        racomp.stat_decompress_tlen = atomic64_read(&istat->decompress_tlen);
 110        racomp.stat_err_cnt = atomic64_read(&istat->err_cnt);
 111
 112        return nla_put(skb, CRYPTOCFGA_STAT_ACOMP, sizeof(racomp), &racomp);
 113}
 114
 115#ifdef CONFIG_CRYPTO_STATS
 116int crypto_acomp_report_stat(struct sk_buff *skb, struct crypto_alg *alg)
 117{
 118        return __crypto_acomp_report_stat(skb, alg);
 119}
 120#endif
 121
 122static const struct crypto_type crypto_acomp_type = {
 123        .extsize = crypto_acomp_extsize,
 124        .init_tfm = crypto_acomp_init_tfm,
 125#ifdef CONFIG_PROC_FS
 126        .show = crypto_acomp_show,
 127#endif
 128#if IS_ENABLED(CONFIG_CRYPTO_USER)
 129        .report = crypto_acomp_report,
 130#endif
 131#ifdef CONFIG_CRYPTO_STATS
 132        .report_stat = crypto_acomp_report_stat,
 133#endif
 134        .maskclear = ~CRYPTO_ALG_TYPE_MASK,
 135        .maskset = CRYPTO_ALG_TYPE_ACOMPRESS_MASK,
 136        .type = CRYPTO_ALG_TYPE_ACOMPRESS,
 137        .tfmsize = offsetof(struct crypto_acomp, base),
 138};
 139
 140struct crypto_acomp *crypto_alloc_acomp(const char *alg_name, u32 type,
 141                                        u32 mask)
 142{
 143        return crypto_alloc_tfm(alg_name, &crypto_acomp_type, type, mask);
 144}
 145EXPORT_SYMBOL_GPL(crypto_alloc_acomp);
 146
 147struct crypto_acomp *crypto_alloc_acomp_node(const char *alg_name, u32 type,
 148                                        u32 mask, int node)
 149{
 150        return crypto_alloc_tfm_node(alg_name, &crypto_acomp_type, type, mask,
 151                                node);
 152}
 153EXPORT_SYMBOL_GPL(crypto_alloc_acomp_node);
 154
 155struct acomp_req *acomp_request_alloc(struct crypto_acomp *acomp)
 156{
 157        struct crypto_tfm *tfm = crypto_acomp_tfm(acomp);
 158        struct acomp_req *req;
 159
 160        req = __acomp_request_alloc(acomp);
 161        if (req && (tfm->__crt_alg->cra_type != &crypto_acomp_type))
 162                return crypto_acomp_scomp_alloc_ctx(req);
 163
 164        return req;
 165}
 166EXPORT_SYMBOL_GPL(acomp_request_alloc);
 167
 168void acomp_request_free(struct acomp_req *req)
 169{
 170        struct crypto_acomp *acomp = crypto_acomp_reqtfm(req);
 171        struct crypto_tfm *tfm = crypto_acomp_tfm(acomp);
 172
 173        if (tfm->__crt_alg->cra_type != &crypto_acomp_type)
 174                crypto_acomp_scomp_free_ctx(req);
 175
 176        if (req->flags & CRYPTO_ACOMP_ALLOC_OUTPUT) {
 177                acomp->dst_free(req->dst);
 178                req->dst = NULL;
 179        }
 180
 181        __acomp_request_free(req);
 182}
 183EXPORT_SYMBOL_GPL(acomp_request_free);
 184
 185void comp_prepare_alg(struct comp_alg_common *alg)
 186{
 187        struct crypto_istat_compress *istat = comp_get_stat(alg);
 188        struct crypto_alg *base = &alg->base;
 189
 190        base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK;
 191
 192        if (IS_ENABLED(CONFIG_CRYPTO_STATS))
 193                memset(istat, 0, sizeof(*istat));
 194}
 195
 196int crypto_register_acomp(struct acomp_alg *alg)
 197{
 198        struct crypto_alg *base = &alg->calg.base;
 199
 200        comp_prepare_alg(&alg->calg);
 201
 202        base->cra_type = &crypto_acomp_type;
 203        base->cra_flags |= CRYPTO_ALG_TYPE_ACOMPRESS;
 204
 205        return crypto_register_alg(base);
 206}
 207EXPORT_SYMBOL_GPL(crypto_register_acomp);
 208
 209void crypto_unregister_acomp(struct acomp_alg *alg)
 210{
 211        crypto_unregister_alg(&alg->base);
 212}
 213EXPORT_SYMBOL_GPL(crypto_unregister_acomp);
 214
 215int crypto_register_acomps(struct acomp_alg *algs, int count)
 216{
 217        int i, ret;
 218
 219        for (i = 0; i < count; i++) {
 220                ret = crypto_register_acomp(&algs[i]);
 221                if (ret)
 222                        goto err;
 223        }
 224
 225        return 0;
 226
 227err:
 228        for (--i; i >= 0; --i)
 229                crypto_unregister_acomp(&algs[i]);
 230
 231        return ret;
 232}
 233EXPORT_SYMBOL_GPL(crypto_register_acomps);
 234
 235void crypto_unregister_acomps(struct acomp_alg *algs, int count)
 236{
 237        int i;
 238
 239        for (i = count - 1; i >= 0; --i)
 240                crypto_unregister_acomp(&algs[i]);
 241}
 242EXPORT_SYMBOL_GPL(crypto_unregister_acomps);
 243
 244MODULE_LICENSE("GPL");
 245MODULE_DESCRIPTION("Asynchronous compression type");
 246