Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at master 246 lines 6.8 kB view raw
1/* SPDX-License-Identifier: GPL-2.0-or-later */ 2/* 3 * Asynchronous Compression operations 4 * 5 * Copyright (c) 2016, Intel Corporation 6 * Authors: Weigang Li <weigang.li@intel.com> 7 * Giovanni Cabiddu <giovanni.cabiddu@intel.com> 8 */ 9#ifndef _CRYPTO_ACOMP_INT_H 10#define _CRYPTO_ACOMP_INT_H 11 12#include <crypto/acompress.h> 13#include <crypto/algapi.h> 14#include <crypto/scatterwalk.h> 15#include <linux/compiler_types.h> 16#include <linux/cpumask_types.h> 17#include <linux/spinlock.h> 18#include <linux/workqueue_types.h> 19 20#define ACOMP_FBREQ_ON_STACK(name, req) \ 21 char __##name##_req[sizeof(struct acomp_req) + \ 22 MAX_SYNC_COMP_REQSIZE] CRYPTO_MINALIGN_ATTR; \ 23 struct acomp_req *name = acomp_fbreq_on_stack_init( \ 24 __##name##_req, (req)) 25 26/** 27 * struct acomp_alg - asynchronous compression algorithm 28 * 29 * @compress: Function performs a compress operation 30 * @decompress: Function performs a de-compress operation 31 * @init: Initialize the cryptographic transformation object. 32 * This function is used to initialize the cryptographic 33 * transformation object. This function is called only once at 34 * the instantiation time, right after the transformation context 35 * was allocated. In case the cryptographic hardware has some 36 * special requirements which need to be handled by software, this 37 * function shall check for the precise requirement of the 38 * transformation and put any software fallbacks in place. 39 * @exit: Deinitialize the cryptographic transformation object. This is a 40 * counterpart to @init, used to remove various changes set in 41 * @init. 42 * 43 * @base: Common crypto API algorithm data structure 44 * @calg: Cmonn algorithm data structure shared with scomp 45 * @COMP_ALG_COMMON: see struct comp_alg_common 46 */ 47struct acomp_alg { 48 int (*compress)(struct acomp_req *req); 49 int (*decompress)(struct acomp_req *req); 50 int (*init)(struct crypto_acomp *tfm); 51 void (*exit)(struct crypto_acomp *tfm); 52 53 union { 54 struct COMP_ALG_COMMON; 55 struct comp_alg_common calg; 56 }; 57}; 58 59struct crypto_acomp_stream { 60 spinlock_t lock; 61 void *ctx; 62}; 63 64struct crypto_acomp_streams { 65 /* These must come first because of struct scomp_alg. */ 66 void *(*alloc_ctx)(void); 67 void (*free_ctx)(void *); 68 69 struct crypto_acomp_stream __percpu *streams; 70 struct work_struct stream_work; 71 cpumask_t stream_want; 72}; 73 74struct acomp_walk { 75 union { 76 /* Virtual address of the source. */ 77 struct { 78 struct { 79 const void *const addr; 80 } virt; 81 } src; 82 83 /* Private field for the API, do not use. */ 84 struct scatter_walk in; 85 }; 86 87 union { 88 /* Virtual address of the destination. */ 89 struct { 90 struct { 91 void *const addr; 92 } virt; 93 } dst; 94 95 /* Private field for the API, do not use. */ 96 struct scatter_walk out; 97 }; 98 99 unsigned int slen; 100 unsigned int dlen; 101 102 int flags; 103}; 104 105/* 106 * Transform internal helpers. 107 */ 108static inline void *acomp_request_ctx(struct acomp_req *req) 109{ 110 return req->__ctx; 111} 112 113static inline void *acomp_tfm_ctx(struct crypto_acomp *tfm) 114{ 115 return tfm->base.__crt_ctx; 116} 117 118static inline void acomp_request_complete(struct acomp_req *req, 119 int err) 120{ 121 crypto_request_complete(&req->base, err); 122} 123 124/** 125 * crypto_register_acomp() -- Register asynchronous compression algorithm 126 * 127 * Function registers an implementation of an asynchronous 128 * compression algorithm 129 * 130 * @alg: algorithm definition 131 * 132 * Return: zero on success; error code in case of error 133 */ 134int crypto_register_acomp(struct acomp_alg *alg); 135 136/** 137 * crypto_unregister_acomp() -- Unregister asynchronous compression algorithm 138 * 139 * Function unregisters an implementation of an asynchronous 140 * compression algorithm 141 * 142 * @alg: algorithm definition 143 */ 144void crypto_unregister_acomp(struct acomp_alg *alg); 145 146int crypto_register_acomps(struct acomp_alg *algs, int count); 147void crypto_unregister_acomps(struct acomp_alg *algs, int count); 148 149static inline bool acomp_request_issg(struct acomp_req *req) 150{ 151 return !(req->base.flags & (CRYPTO_ACOMP_REQ_SRC_VIRT | 152 CRYPTO_ACOMP_REQ_DST_VIRT)); 153} 154 155static inline bool acomp_request_src_isvirt(struct acomp_req *req) 156{ 157 return req->base.flags & CRYPTO_ACOMP_REQ_SRC_VIRT; 158} 159 160static inline bool acomp_request_dst_isvirt(struct acomp_req *req) 161{ 162 return req->base.flags & CRYPTO_ACOMP_REQ_DST_VIRT; 163} 164 165static inline bool acomp_request_isvirt(struct acomp_req *req) 166{ 167 return req->base.flags & (CRYPTO_ACOMP_REQ_SRC_VIRT | 168 CRYPTO_ACOMP_REQ_DST_VIRT); 169} 170 171static inline bool acomp_request_src_isnondma(struct acomp_req *req) 172{ 173 return req->base.flags & CRYPTO_ACOMP_REQ_SRC_NONDMA; 174} 175 176static inline bool acomp_request_dst_isnondma(struct acomp_req *req) 177{ 178 return req->base.flags & CRYPTO_ACOMP_REQ_DST_NONDMA; 179} 180 181static inline bool acomp_request_isnondma(struct acomp_req *req) 182{ 183 return req->base.flags & (CRYPTO_ACOMP_REQ_SRC_NONDMA | 184 CRYPTO_ACOMP_REQ_DST_NONDMA); 185} 186 187static inline bool crypto_acomp_req_virt(struct crypto_acomp *tfm) 188{ 189 return crypto_tfm_req_virt(&tfm->base); 190} 191 192void crypto_acomp_free_streams(struct crypto_acomp_streams *s); 193int crypto_acomp_alloc_streams(struct crypto_acomp_streams *s); 194 195#define crypto_acomp_lock_stream_bh(...) __acquire_ret(_crypto_acomp_lock_stream_bh(__VA_ARGS__), &__ret->lock); 196struct crypto_acomp_stream *_crypto_acomp_lock_stream_bh( 197 struct crypto_acomp_streams *s) __acquires_ret; 198 199static inline void crypto_acomp_unlock_stream_bh( 200 struct crypto_acomp_stream *stream) __releases(&stream->lock) 201{ 202 spin_unlock_bh(&stream->lock); 203} 204 205void acomp_walk_done_src(struct acomp_walk *walk, int used); 206void acomp_walk_done_dst(struct acomp_walk *walk, int used); 207int acomp_walk_next_src(struct acomp_walk *walk); 208int acomp_walk_next_dst(struct acomp_walk *walk); 209int acomp_walk_virt(struct acomp_walk *__restrict walk, 210 struct acomp_req *__restrict req, bool atomic); 211 212static inline bool acomp_walk_more_src(const struct acomp_walk *walk, int cur) 213{ 214 return walk->slen != cur; 215} 216 217static inline u32 acomp_request_flags(struct acomp_req *req) 218{ 219 return crypto_request_flags(&req->base) & ~CRYPTO_ACOMP_REQ_PRIVATE; 220} 221 222static inline struct crypto_acomp *crypto_acomp_fb(struct crypto_acomp *tfm) 223{ 224 return __crypto_acomp_tfm(crypto_acomp_tfm(tfm)->fb); 225} 226 227static inline struct acomp_req *acomp_fbreq_on_stack_init( 228 char *buf, struct acomp_req *old) 229{ 230 struct crypto_acomp *tfm = crypto_acomp_reqtfm(old); 231 struct acomp_req *req = (void *)buf; 232 233 crypto_stack_request_init(&req->base, 234 crypto_acomp_tfm(crypto_acomp_fb(tfm))); 235 acomp_request_set_callback(req, acomp_request_flags(old), NULL, NULL); 236 req->base.flags &= ~CRYPTO_ACOMP_REQ_PRIVATE; 237 req->base.flags |= old->base.flags & CRYPTO_ACOMP_REQ_PRIVATE; 238 req->src = old->src; 239 req->dst = old->dst; 240 req->slen = old->slen; 241 req->dlen = old->dlen; 242 243 return req; 244} 245 246#endif