/linux-4.1.27/drivers/crypto/nx/ |
H A D | nx-sha256.c | 79 u32 max_sg_len; nx_sha256_update() local 99 max_sg_len = min_t(u64, nx_ctx->ap->sglen, nx_sha256_update() 100 nx_driver.of.max_sg_len/sizeof(struct nx_sg)); nx_sha256_update() 101 max_sg_len = min_t(u64, max_sg_len, nx_sha256_update() 106 &data_len, max_sg_len); nx_sha256_update() 123 max_sg_len); nx_sha256_update() 136 * In ideal case, we could allow NX_PAGE_SIZE * max_sg_len, nx_sha256_update() 140 (max_sg_len - 1 - used_sgs) * NX_PAGE_SIZE); nx_sha256_update() 145 &data_len, max_sg_len); nx_sha256_update() 196 u32 max_sg_len; nx_sha256_final() local 202 max_sg_len = min_t(u64, nx_ctx->ap->sglen, nx_sha256_final() 203 nx_driver.of.max_sg_len/sizeof(struct nx_sg)); nx_sha256_final() 204 max_sg_len = min_t(u64, max_sg_len, nx_sha256_final() 224 &len, max_sg_len); nx_sha256_final() 232 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, max_sg_len); nx_sha256_final()
|
H A D | nx-sha512.c | 79 u32 max_sg_len; nx_sha512_update() local 99 max_sg_len = min_t(u64, nx_ctx->ap->sglen, nx_sha512_update() 100 nx_driver.of.max_sg_len/sizeof(struct nx_sg)); nx_sha512_update() 101 max_sg_len = min_t(u64, max_sg_len, nx_sha512_update() 106 &data_len, max_sg_len); nx_sha512_update() 122 &data_len, max_sg_len); nx_sha512_update() 135 * In ideal case, we could allow NX_PAGE_SIZE * max_sg_len, nx_sha512_update() 139 (max_sg_len - 1 - used_sgs) * NX_PAGE_SIZE); nx_sha512_update() 144 &data_len, max_sg_len); nx_sha512_update() 198 u32 max_sg_len; nx_sha512_final() local 206 max_sg_len = min_t(u64, nx_ctx->ap->sglen, nx_sha512_final() 207 nx_driver.of.max_sg_len/sizeof(struct nx_sg)); nx_sha512_final() 208 max_sg_len = min_t(u64, max_sg_len, nx_sha512_final() 233 max_sg_len); nx_sha512_final() 242 max_sg_len); nx_sha512_final()
|
H A D | nx-aes-gcm.c | 134 unsigned int max_sg_len; nx_gca() local 146 max_sg_len = min_t(u64, nx_driver.of.max_sg_len/sizeof(struct nx_sg), nx_gca() 148 max_sg_len = min_t(u64, max_sg_len, nx_gca() 159 NX_PAGE_SIZE * (max_sg_len - 1)); nx_gca() 161 nx_sg = nx_walk_and_build(nx_ctx->in_sg, max_sg_len, nx_gca() 201 unsigned int max_sg_len; gmac() local 209 max_sg_len = min_t(u64, nx_driver.of.max_sg_len/sizeof(struct nx_sg), gmac() 211 max_sg_len = min_t(u64, max_sg_len, gmac() 225 NX_PAGE_SIZE * (max_sg_len - 1)); gmac() 227 nx_sg = nx_walk_and_build(nx_ctx->in_sg, max_sg_len, gmac()
|
H A D | nx-aes-xcbc.c | 189 unsigned int max_sg_len; nx_xcbc_update() local 210 max_sg_len = min_t(u64, nx_driver.of.max_sg_len/sizeof(struct nx_sg), nx_xcbc_update() 212 max_sg_len = min_t(u64, max_sg_len, nx_xcbc_update() 247 max_sg_len); nx_xcbc_update() 258 max_sg_len); nx_xcbc_update()
|
H A D | nx.c | 283 unsigned int max_sg_len; nx_build_sg_lists() local 285 max_sg_len = min_t(u64, nx_ctx->ap->sglen, nx_build_sg_lists() 286 nx_driver.of.max_sg_len/sizeof(struct nx_sg)); nx_build_sg_lists() 287 max_sg_len = min_t(u64, max_sg_len, nx_build_sg_lists() 295 nx_outsg = nx_walk_and_build(nx_outsg, max_sg_len, dst, nx_build_sg_lists() 297 nx_insg = nx_walk_and_build(nx_insg, max_sg_len, src, nx_build_sg_lists() 356 if (p->length != sizeof(props->max_sg_len)) { nx_of_update_sglen() 361 p->length, sizeof(props->max_sg_len)); nx_of_update_sglen() 365 props->max_sg_len = *(u32 *)p->value; nx_of_update_sglen()
|
H A D | nx-aes-ccm.c | 184 unsigned int max_sg_len; generate_pat() local 291 max_sg_len = min_t(u64, nx_ctx->ap->sglen, generate_pat() 292 nx_driver.of.max_sg_len/sizeof(struct nx_sg)); generate_pat() 293 max_sg_len = min_t(u64, max_sg_len, generate_pat()
|
H A D | nx.h | 58 u32 max_sg_len; member in struct:nx_of
|
H A D | nx-842.c | 96 unsigned int max_sg_len; member in struct:nx842_devdata 713 devdata->max_sg_len = 0; nx842_OF_set_defaults() 783 devdata->max_sg_len = (unsigned int)min(*maxsglen, nx842_OF_upd_maxsglen() 968 dev_info(old_devdata->dev, "%s: max_sg_len new:%u old:%u\n", nx842_OF_upd() 969 __func__, new_devdata->max_sg_len, nx842_OF_upd() 970 old_devdata->max_sg_len); nx842_OF_upd()
|