/linux-6.6.21/arch/powerpc/crypto/ |
D | aes-gcm-p10-glue.c | 87 static void set_aad(struct gcm_ctx *gctx, struct Hash_ctx *hash, in set_aad() argument 93 gctx->aadLen = alen; in set_aad() 104 memset(gctx->aad_hash, 0, 16); in set_aad() 105 gcm_ghash_p10(gctx->aad_hash, hash->Htable+32, nXi, 16); in set_aad() 107 memcpy(gctx->aad_hash, nXi, 16); in set_aad() 110 memcpy(hash->Htable, gctx->aad_hash, 16); in set_aad() 113 static void gcmp10_init(struct gcm_ctx *gctx, u8 *iv, unsigned char *rdkey, in gcmp10_init() argument 124 gctx->Plen = 0; in gcmp10_init() 129 aes_p10_encrypt(iv, gctx->ivtag, rdkey); in gcmp10_init() 133 memcpy(gctx->iv, iv, 16); in gcmp10_init() [all …]
|
/linux-6.6.21/drivers/gpu/drm/radeon/ |
D | atom.c | 183 struct atom_context *gctx = ctx->ctx; in atom_get_src_int() local 192 idx += gctx->reg_block; in atom_get_src_int() 193 switch (gctx->io_mode) { in atom_get_src_int() 195 val = gctx->card->reg_read(gctx->card, idx); in atom_get_src_int() 204 if (!(gctx->io_mode & 0x80)) { in atom_get_src_int() 208 if (!gctx->iio[gctx->io_mode & 0x7F]) { in atom_get_src_int() 210 gctx->io_mode & 0x7F); in atom_get_src_int() 214 atom_iio_execute(gctx, in atom_get_src_int() 215 gctx->iio[gctx->io_mode & 0x7F], in atom_get_src_int() 235 val = gctx->divmul[0]; in atom_get_src_int() [all …]
|
/linux-6.6.21/drivers/gpu/drm/amd/amdgpu/ |
D | atom.c | 185 struct atom_context *gctx = ctx->ctx; in atom_get_src_int() local 194 idx += gctx->reg_block; in atom_get_src_int() 195 switch (gctx->io_mode) { in atom_get_src_int() 197 val = gctx->card->reg_read(gctx->card, idx); in atom_get_src_int() 206 if (!(gctx->io_mode & 0x80)) { in atom_get_src_int() 210 if (!gctx->iio[gctx->io_mode & 0x7F]) { in atom_get_src_int() 212 gctx->io_mode & 0x7F); in atom_get_src_int() 216 atom_iio_execute(gctx, in atom_get_src_int() 217 gctx->iio[gctx->io_mode & 0x7F], in atom_get_src_int() 237 val = gctx->divmul[0]; in atom_get_src_int() [all …]
|
/linux-6.6.21/net/sunrpc/auth_gss/ |
D | gss_krb5_mech.c | 514 static u32 gss_krb5_get_mic(struct gss_ctx *gctx, struct xdr_buf *text, in gss_krb5_get_mic() argument 517 struct krb5_ctx *kctx = gctx->internal_ctx_id; in gss_krb5_get_mic() 535 static u32 gss_krb5_verify_mic(struct gss_ctx *gctx, in gss_krb5_verify_mic() argument 539 struct krb5_ctx *kctx = gctx->internal_ctx_id; in gss_krb5_verify_mic() 556 static u32 gss_krb5_wrap(struct gss_ctx *gctx, int offset, in gss_krb5_wrap() argument 559 struct krb5_ctx *kctx = gctx->internal_ctx_id; in gss_krb5_wrap() 578 static u32 gss_krb5_unwrap(struct gss_ctx *gctx, int offset, in gss_krb5_unwrap() argument 581 struct krb5_ctx *kctx = gctx->internal_ctx_id; in gss_krb5_unwrap() 584 &gctx->slack, &gctx->align); in gss_krb5_unwrap()
|
/linux-6.6.21/crypto/ |
D | gcm.c | 227 struct crypto_gcm_ghash_ctx *gctx = &pctx->ghash_ctx; in gcm_hash_len() local 231 lengths.b = cpu_to_be64(gctx->cryptlen * 8); in gcm_hash_len() 244 struct crypto_gcm_ghash_ctx *gctx = &pctx->ghash_ctx; in gcm_hash_len_continue() local 246 return gctx->complete(req, flags); in gcm_hash_len_continue() 288 struct crypto_gcm_ghash_ctx *gctx = &pctx->ghash_ctx; in gcm_hash_crypt_continue() local 291 remain = gcm_remain(gctx->cryptlen); in gcm_hash_crypt_continue() 318 struct crypto_gcm_ghash_ctx *gctx = &pctx->ghash_ctx; in gcm_hash_assoc_remain_continue() local 320 if (gctx->cryptlen) in gcm_hash_assoc_remain_continue() 322 gctx->src, gctx->cryptlen, flags) ?: in gcm_hash_assoc_remain_continue() 425 struct crypto_gcm_ghash_ctx *gctx = &pctx->ghash_ctx; in gcm_encrypt_continue() local [all …]
|
/linux-6.6.21/drivers/accel/ivpu/ |
D | ivpu_gem.c | 581 bo = ivpu_bo_alloc(vdev, &vdev->gctx, size, flags, &internal_ops, range, 0); in ivpu_bo_alloc_internal() 735 mutex_lock(&vdev->gctx.lock); in ivpu_bo_list() 736 list_for_each_entry(bo, &vdev->gctx.bo_list, ctx_node) in ivpu_bo_list() 738 mutex_unlock(&vdev->gctx.lock); in ivpu_bo_list()
|
D | ivpu_drv.h | 112 struct ivpu_mmu_context gctx; member
|
D | ivpu_mmu_context.c | 459 return ivpu_mmu_context_init(vdev, &vdev->gctx, IVPU_GLOBAL_CONTEXT_MMU_SSID); in ivpu_mmu_global_context_init() 464 return ivpu_mmu_context_fini(vdev, &vdev->gctx); in ivpu_mmu_global_context_fini()
|
D | ivpu_mmu.c | 680 ret = ivpu_mmu_cd_add(vdev, 0, vdev->gctx.pgtable.pgd_dma); in ivpu_mmu_cd_add_gbl()
|
/linux-6.6.21/drivers/crypto/chelsio/ |
D | chcr_algo.c | 114 static inline struct chcr_gcm_ctx *GCM_CTX(struct chcr_aead_ctx *gctx) in GCM_CTX() argument 116 return gctx->ctx->gcm; in GCM_CTX() 119 static inline struct chcr_authenc_ctx *AUTHENC_CTX(struct chcr_aead_ctx *gctx) in AUTHENC_CTX() argument 121 return gctx->ctx->authenc; in AUTHENC_CTX() 3492 struct chcr_gcm_ctx *gctx = GCM_CTX(aeadctx); in chcr_gcm_setkey() local 3538 memset(gctx->ghash_h, 0, AEAD_H_SIZE); in chcr_gcm_setkey() 3539 aes_encrypt(&aes, gctx->ghash_h, gctx->ghash_h); in chcr_gcm_setkey()
|