/linux-6.6.21/net/netfilter/ |
D | nft_set_pipapo_avx2.c | 45 #define NFT_PIPAPO_AVX2_BUCKET_LOAD4(reg, lt, group, v, bsize) \ argument 48 (v)) * (bsize)]) 49 #define NFT_PIPAPO_AVX2_BUCKET_LOAD8(reg, lt, group, v, bsize) \ argument 52 (v)) * (bsize)]) 218 int i, ret = -1, m256_size = f->bsize / NFT_PIPAPO_LONGS_PER_M256, b; in nft_pipapo_avx2_lookup_4b_2() 220 unsigned long *lt = f->lt, bsize = f->bsize; in nft_pipapo_avx2_lookup_4b_2() local 227 NFT_PIPAPO_AVX2_BUCKET_LOAD4(0, lt, 0, pg[0], bsize); in nft_pipapo_avx2_lookup_4b_2() 228 NFT_PIPAPO_AVX2_BUCKET_LOAD4(1, lt, 1, pg[1], bsize); in nft_pipapo_avx2_lookup_4b_2() 231 NFT_PIPAPO_AVX2_BUCKET_LOAD4(0, lt, 0, pg[0], bsize); in nft_pipapo_avx2_lookup_4b_2() 233 NFT_PIPAPO_AVX2_BUCKET_LOAD4(1, lt, 1, pg[1], bsize); in nft_pipapo_avx2_lookup_4b_2() [all …]
|
D | nft_set_pipapo.h | 124 size_t bsize; member 207 __bitmap_and(dst, dst, lt + v * f->bsize, in pipapo_and_field_buckets_4bit() 208 f->bsize * BITS_PER_LONG); in pipapo_and_field_buckets_4bit() 209 lt += f->bsize * NFT_PIPAPO_BUCKETS(4); in pipapo_and_field_buckets_4bit() 212 __bitmap_and(dst, dst, lt + v * f->bsize, in pipapo_and_field_buckets_4bit() 213 f->bsize * BITS_PER_LONG); in pipapo_and_field_buckets_4bit() 214 lt += f->bsize * NFT_PIPAPO_BUCKETS(4); in pipapo_and_field_buckets_4bit() 232 __bitmap_and(dst, dst, lt + *data * f->bsize, in pipapo_and_field_buckets_8bit() 233 f->bsize * BITS_PER_LONG); in pipapo_and_field_buckets_8bit() 234 lt += f->bsize * NFT_PIPAPO_BUCKETS(8); in pipapo_and_field_buckets_8bit()
|
D | nft_set_pipapo.c | 460 b = pipapo_refill(res_map, f->bsize, f->rules, fill_map, f->mt, in nft_pipapo_lookup() 563 b = pipapo_refill(res_map, f->bsize, f->rules, fill_map, f->mt, in pipapo_get() 634 if (new_bucket_size == f->bsize) in pipapo_resize() 637 if (new_bucket_size > f->bsize) in pipapo_resize() 638 copy = f->bsize; in pipapo_resize() 658 if (new_bucket_size > f->bsize) in pipapo_resize() 659 new_p += new_bucket_size - f->bsize; in pipapo_resize() 661 old_p += f->bsize - new_bucket_size; in pipapo_resize() 679 f->bsize = new_bucket_size; in pipapo_resize() 703 pos += f->bsize * NFT_PIPAPO_BUCKETS(f->bb) * group; in pipapo_bucket_set() [all …]
|
/linux-6.6.21/crypto/ |
D | cbc.c | 20 unsigned int bsize = crypto_skcipher_blocksize(skcipher); in crypto_cbc_encrypt_segment() local 34 crypto_xor(iv, src, bsize); in crypto_cbc_encrypt_segment() 36 memcpy(iv, dst, bsize); in crypto_cbc_encrypt_segment() 38 src += bsize; in crypto_cbc_encrypt_segment() 39 dst += bsize; in crypto_cbc_encrypt_segment() 40 } while ((nbytes -= bsize) >= bsize); in crypto_cbc_encrypt_segment() 48 unsigned int bsize = crypto_skcipher_blocksize(skcipher); in crypto_cbc_encrypt_inplace() local 61 crypto_xor(src, iv, bsize); in crypto_cbc_encrypt_inplace() 65 src += bsize; in crypto_cbc_encrypt_inplace() 66 } while ((nbytes -= bsize) >= bsize); in crypto_cbc_encrypt_inplace() [all …]
|
D | pcbc.c | 24 int bsize = crypto_cipher_blocksize(tfm); in crypto_pcbc_encrypt_segment() local 31 crypto_xor(iv, src, bsize); in crypto_pcbc_encrypt_segment() 33 crypto_xor_cpy(iv, dst, src, bsize); in crypto_pcbc_encrypt_segment() 35 src += bsize; in crypto_pcbc_encrypt_segment() 36 dst += bsize; in crypto_pcbc_encrypt_segment() 37 } while ((nbytes -= bsize) >= bsize); in crypto_pcbc_encrypt_segment() 46 int bsize = crypto_cipher_blocksize(tfm); in crypto_pcbc_encrypt_inplace() local 53 memcpy(tmpbuf, src, bsize); in crypto_pcbc_encrypt_inplace() 54 crypto_xor(iv, src, bsize); in crypto_pcbc_encrypt_inplace() 56 crypto_xor_cpy(iv, tmpbuf, src, bsize); in crypto_pcbc_encrypt_inplace() [all …]
|
D | cfb.c | 61 const unsigned int bsize = crypto_cfb_bsize(tfm); in crypto_cfb_encrypt_segment() local 69 crypto_xor(dst, src, bsize); in crypto_cfb_encrypt_segment() 72 src += bsize; in crypto_cfb_encrypt_segment() 73 dst += bsize; in crypto_cfb_encrypt_segment() 74 } while ((nbytes -= bsize) >= bsize); in crypto_cfb_encrypt_segment() 76 memcpy(walk->iv, iv, bsize); in crypto_cfb_encrypt_segment() 84 const unsigned int bsize = crypto_cfb_bsize(tfm); in crypto_cfb_encrypt_inplace() local 92 crypto_xor(src, tmp, bsize); in crypto_cfb_encrypt_inplace() 95 src += bsize; in crypto_cfb_encrypt_inplace() 96 } while ((nbytes -= bsize) >= bsize); in crypto_cfb_encrypt_inplace() [all …]
|
D | cts.c | 103 int bsize = crypto_skcipher_blocksize(tfm); in cts_cbc_encrypt() local 112 sg = scatterwalk_ffwd(rctx->sg, req->dst, offset - bsize); in cts_cbc_encrypt() 113 scatterwalk_map_and_copy(d + bsize, sg, 0, bsize, 0); in cts_cbc_encrypt() 115 memset(d, 0, bsize); in cts_cbc_encrypt() 118 scatterwalk_map_and_copy(d, sg, 0, bsize + lastn, 1); in cts_cbc_encrypt() 124 skcipher_request_set_crypt(subreq, sg, sg, bsize, req->iv); in cts_cbc_encrypt() 149 int bsize = crypto_skcipher_blocksize(tfm); in crypto_cts_encrypt() local 155 if (nbytes < bsize) in crypto_cts_encrypt() 158 if (nbytes == bsize) { in crypto_cts_encrypt() 167 offset = rounddown(nbytes - 1, bsize); in crypto_cts_encrypt() [all …]
|
D | ctr.c | 31 unsigned int bsize = crypto_cipher_blocksize(tfm); in crypto_ctr_crypt_final() local 43 crypto_inc(ctrblk, bsize); in crypto_ctr_crypt_final() 51 unsigned int bsize = crypto_cipher_blocksize(tfm); in crypto_ctr_crypt_segment() local 60 crypto_xor(dst, src, bsize); in crypto_ctr_crypt_segment() 63 crypto_inc(ctrblk, bsize); in crypto_ctr_crypt_segment() 65 src += bsize; in crypto_ctr_crypt_segment() 66 dst += bsize; in crypto_ctr_crypt_segment() 67 } while ((nbytes -= bsize) >= bsize); in crypto_ctr_crypt_segment() 77 unsigned int bsize = crypto_cipher_blocksize(tfm); in crypto_ctr_crypt_inplace() local 88 crypto_xor(src, keystream, bsize); in crypto_ctr_crypt_inplace() [all …]
|
D | ofb.c | 22 const unsigned int bsize = crypto_cipher_blocksize(cipher); in crypto_ofb_crypt() local 28 while (walk.nbytes >= bsize) { in crypto_ofb_crypt() 36 crypto_xor_cpy(dst, src, iv, bsize); in crypto_ofb_crypt() 37 dst += bsize; in crypto_ofb_crypt() 38 src += bsize; in crypto_ofb_crypt() 39 } while ((nbytes -= bsize) >= bsize); in crypto_ofb_crypt()
|
D | ecb.c | 20 const unsigned int bsize = crypto_cipher_blocksize(cipher); in crypto_ecb_crypt() local 34 src += bsize; in crypto_ecb_crypt() 35 dst += bsize; in crypto_ecb_crypt() 36 } while ((nbytes -= bsize) >= bsize); in crypto_ecb_crypt()
|
/linux-6.6.21/arch/x86/crypto/ |
D | des3_ede_glue.c | 68 const unsigned int bsize = DES3_EDE_BLOCK_SIZE; in ecb_crypt() local 80 if (nbytes >= bsize * 3) { in ecb_crypt() 85 wsrc += bsize * 3; in ecb_crypt() 86 wdst += bsize * 3; in ecb_crypt() 87 nbytes -= bsize * 3; in ecb_crypt() 88 } while (nbytes >= bsize * 3); in ecb_crypt() 90 if (nbytes < bsize) in ecb_crypt() 98 wsrc += bsize; in ecb_crypt() 99 wdst += bsize; in ecb_crypt() 100 nbytes -= bsize; in ecb_crypt() [all …]
|
D | sm4_aesni_avx_glue.c | 127 unsigned int bsize, sm4_crypt_func func) in sm4_avx_cbc_decrypt() argument 143 while (nbytes >= bsize) { in sm4_avx_cbc_decrypt() 145 dst += bsize; in sm4_avx_cbc_decrypt() 146 src += bsize; in sm4_avx_cbc_decrypt() 147 nbytes -= bsize; in sm4_avx_cbc_decrypt() 233 unsigned int bsize, sm4_crypt_func func) in sm4_avx_cfb_decrypt() argument 249 while (nbytes >= bsize) { in sm4_avx_cfb_decrypt() 251 dst += bsize; in sm4_avx_cfb_decrypt() 252 src += bsize; in sm4_avx_cfb_decrypt() 253 nbytes -= bsize; in sm4_avx_cfb_decrypt() [all …]
|
D | ecb_cbc_helpers.h | 14 #define ECB_WALK_START(req, bsize, fpu_blocks) do { \ argument 17 const int __bsize = (bsize); \ 26 u8 __maybe_unused buf[(bsize)]; \ 29 #define CBC_WALK_START(req, bsize, fpu_blocks) \ argument 30 ECB_WALK_START(req, bsize, fpu_blocks)
|
D | sm4-avx.h | 15 unsigned int bsize, sm4_crypt_func func); 19 unsigned int bsize, sm4_crypt_func func); 22 unsigned int bsize, sm4_crypt_func func);
|
/linux-6.6.21/arch/s390/crypto/ |
D | sha_common.c | 19 unsigned int bsize = crypto_shash_blocksize(desc->tfm); in s390_sha_update() local 23 index = ctx->count % bsize; in s390_sha_update() 26 if ((index + len) < bsize) in s390_sha_update() 31 memcpy(ctx->buf + index, data, bsize - index); in s390_sha_update() 32 cpacf_kimd(ctx->func, ctx->state, ctx->buf, bsize); in s390_sha_update() 33 data += bsize - index; in s390_sha_update() 34 len -= bsize - index; in s390_sha_update() 39 if (len >= bsize) { in s390_sha_update() 40 n = (len / bsize) * bsize; in s390_sha_update() 75 unsigned int bsize = crypto_shash_blocksize(desc->tfm); in s390_sha_final() local [all …]
|
/linux-6.6.21/arch/m68k/emu/ |
D | nfblock.c | 55 u32 blocks, bsize; member 97 static int __init nfhd_init_one(int id, u32 blocks, u32 bsize) in nfhd_init_one() argument 104 blocks, bsize); in nfhd_init_one() 106 if (bsize < 512 || (bsize & (bsize - 1))) { in nfhd_init_one() 117 dev->bsize = bsize; in nfhd_init_one() 118 dev->bshift = ffs(bsize) - 10; in nfhd_init_one() 130 set_capacity(dev->disk, (sector_t)blocks * (bsize / 512)); in nfhd_init_one() 131 blk_queue_logical_block_size(dev->disk->queue, bsize); in nfhd_init_one() 150 u32 blocks, bsize; in nfhd_init() local 168 if (nfhd_get_capacity(i, 0, &blocks, &bsize)) in nfhd_init() [all …]
|
/linux-6.6.21/lib/crypto/mpi/ |
D | mpi-pow.c | 31 mpi_size_t esize, msize, bsize, rsize; in mpi_powm() local 79 bsize = base->nlimbs; in mpi_powm() 81 if (bsize > msize) { /* The base is larger than the module. Reduce it. */ in mpi_powm() 84 bp = bp_marker = mpi_alloc_limb_space(bsize + 1); in mpi_powm() 87 MPN_COPY(bp, base->d, bsize); in mpi_powm() 90 mpihelp_divrem(bp + msize, 0, bp, bsize, mp, msize); in mpi_powm() 91 bsize = msize; in mpi_powm() 94 MPN_NORMALIZE(bp, bsize); in mpi_powm() 98 if (!bsize) { in mpi_powm() 122 bp = bp_marker = mpi_alloc_limb_space(bsize); in mpi_powm() [all …]
|
/linux-6.6.21/include/crypto/ |
D | ctr.h | 48 int bsize = min(nbytes, blocksize); in crypto_ctr_encrypt_walk() local 52 crypto_xor_cpy(dst, src, buf, bsize); in crypto_ctr_encrypt_walk() 55 dst += bsize; in crypto_ctr_encrypt_walk() 56 src += bsize; in crypto_ctr_encrypt_walk() 57 nbytes -= bsize; in crypto_ctr_encrypt_walk()
|
/linux-6.6.21/fs/jfs/ |
D | jfs_mount.c | 297 s32 bsize; in chkSuper() local 313 bsize = le32_to_cpu(j_sb->s_bsize); in chkSuper() 314 if (bsize != PSIZE) { in chkSuper() 328 AIM_bytesize = lengthPXD(&(j_sb->s_aim2)) * bsize; in chkSuper() 330 AIT_bytesize = lengthPXD(&(j_sb->s_ait2)) * bsize; in chkSuper() 331 AIM_byte_addr = addressPXD(&(j_sb->s_aim2)) * bsize; in chkSuper() 332 AIT_byte_addr = addressPXD(&(j_sb->s_ait2)) * bsize; in chkSuper() 334 fsckwsp_addr = addressPXD(&(j_sb->s_fsckpxd)) * bsize; in chkSuper() 362 sbi->bsize = bsize; in chkSuper() 366 if (sbi->l2bsize != ilog2((u32)bsize) || in chkSuper()
|
/linux-6.6.21/fs/freevxfs/ |
D | vxfs_lookup.c | 59 u_long bsize = ip->i_sb->s_blocksize; in vxfs_find_entry() local 80 if ((pos & (bsize - 1)) < 4) { in vxfs_find_entry() 92 pos += bsize - 1; in vxfs_find_entry() 93 pos &= ~(bsize - 1); in vxfs_find_entry() 195 u_long bsize = sbp->s_blocksize; in vxfs_readdir() local 231 if ((pos & (bsize - 1)) < 4) { in vxfs_readdir() 243 pos += bsize - 1; in vxfs_readdir() 244 pos &= ~(bsize - 1); in vxfs_readdir()
|
D | vxfs_olt.c | 33 vxfs_oblock(struct super_block *sbp, daddr_t block, u_long bsize) in vxfs_oblock() argument 35 BUG_ON(sbp->s_blocksize % bsize); in vxfs_oblock() 36 return (block * (sbp->s_blocksize / bsize)); in vxfs_oblock() 53 vxfs_read_olt(struct super_block *sbp, u_long bsize) in vxfs_read_olt() argument 60 bp = sb_bread(sbp, vxfs_oblock(sbp, infp->vsi_oltext, bsize)); in vxfs_read_olt()
|
/linux-6.6.21/fs/affs/ |
D | file.c | 532 size_t bidx, boff, bsize; in affs_do_read_folio_ofs() local 538 bsize = AFFS_SB(sb)->s_data_blksize; in affs_do_read_folio_ofs() 540 bidx = tmp / bsize; in affs_do_read_folio_ofs() 541 boff = tmp % bsize; in affs_do_read_folio_ofs() 547 tmp = min(bsize - boff, to - pos); in affs_do_read_folio_ofs() 548 BUG_ON(pos + tmp > to || tmp > bsize); in affs_do_read_folio_ofs() 564 u32 size, bsize; in affs_extent_file_ofs() local 568 bsize = AFFS_SB(sb)->s_data_blksize; in affs_extent_file_ofs() 571 bidx = size / bsize; in affs_extent_file_ofs() 572 boff = size % bsize; in affs_extent_file_ofs() [all …]
|
/linux-6.6.21/tools/perf/util/ |
D | mem2node.c | 39 phys_entry__init(struct phys_entry *entry, u64 start, u64 bsize, u64 node) in phys_entry__init() argument 42 entry->end = start + bsize; in phys_entry__init() 51 u64 bsize = env->memory_bsize; in mem2node__init() local 77 start = bit * bsize; in mem2node__init() 88 prev->end += bsize; in mem2node__init() 93 phys_entry__init(&entries[j++], start, bsize, n->node); in mem2node__init()
|
/linux-6.6.21/fs/squashfs/ |
D | file_cache.c | 21 int squashfs_readpage_block(struct page *page, u64 block, int bsize, int expected) in squashfs_readpage_block() argument 25 block, bsize); in squashfs_readpage_block() 30 bsize); in squashfs_readpage_block()
|
/linux-6.6.21/arch/powerpc/kernel/ |
D | setup_64.c | 551 u32 bsize, u32 sets) in init_cache_info() argument 556 info->block_size = bsize; in init_cache_info() 557 info->log_block_size = __ilog2(bsize); in init_cache_info() 558 if (bsize) in init_cache_info() 559 info->blocks_per_page = PAGE_SIZE / bsize; in init_cache_info() 587 u32 size, lsize, bsize, sets; in parse_cache_info() local 592 lsize = bsize = cur_cpu_spec->dcache_bsize; in parse_cache_info() 608 bsize = be32_to_cpu(*bsizep); in parse_cache_info() 623 init_cache_info(info, size, lsize, bsize, sets); in parse_cache_info()
|