Home
last modified time | relevance | path

Searched refs:partial (Results 1 – 25 of 234) sorted by relevance

12345678910

/linux-6.1.9/fs/ext4/
Dindirect.c252 Indirect *partial) in ext4_find_goal() argument
260 goal = ext4_find_near(inode, partial); in ext4_find_goal()
538 Indirect *partial; in ext4_ind_map_blocks() local
554 partial = ext4_get_branch(inode, depth, offsets, chain, &err); in ext4_ind_map_blocks()
557 if (!partial) { in ext4_ind_map_blocks()
586 for (i = partial - chain + 1; i < depth; i++) in ext4_ind_map_blocks()
620 ar.goal = ext4_find_goal(inode, map->m_lblk, partial); in ext4_ind_map_blocks()
623 indirect_blks = (chain + depth) - partial - 1; in ext4_ind_map_blocks()
629 ar.len = ext4_blks_to_allocate(partial, indirect_blks, in ext4_ind_map_blocks()
636 offsets + (partial - chain), partial); in ext4_ind_map_blocks()
[all …]
/linux-6.1.9/fs/minix/
Ditree_common.c158 Indirect *partial; in get_block() local
166 partial = get_branch(inode, depth, offsets, chain, &err); in get_block()
169 if (!partial) { in get_block()
173 partial = chain+depth-1; /* the whole chain */ in get_block()
180 while (partial > chain) { in get_block()
181 brelse(partial->bh); in get_block()
182 partial--; in get_block()
196 left = (chain + depth) - partial; in get_block()
197 err = alloc_branch(inode, left, offsets+(partial-chain), partial); in get_block()
201 if (splice_branch(inode, chain, partial, left) < 0) in get_block()
[all …]
/linux-6.1.9/include/crypto/
Dsha1_base.h41 unsigned int partial = sctx->count % SHA1_BLOCK_SIZE; in sha1_base_do_update() local
45 if (unlikely((partial + len) >= SHA1_BLOCK_SIZE)) { in sha1_base_do_update()
48 if (partial) { in sha1_base_do_update()
49 int p = SHA1_BLOCK_SIZE - partial; in sha1_base_do_update()
51 memcpy(sctx->buffer + partial, data, p); in sha1_base_do_update()
65 partial = 0; in sha1_base_do_update()
68 memcpy(sctx->buffer + partial, data, len); in sha1_base_do_update()
79 unsigned int partial = sctx->count % SHA1_BLOCK_SIZE; in sha1_base_do_finalize() local
81 sctx->buffer[partial++] = 0x80; in sha1_base_do_finalize()
82 if (partial > bit_offset) { in sha1_base_do_finalize()
[all …]
Dsha256_base.h44 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in sha256_base_do_update() local
48 if (unlikely((partial + len) >= SHA256_BLOCK_SIZE)) { in sha256_base_do_update()
51 if (partial) { in sha256_base_do_update()
52 int p = SHA256_BLOCK_SIZE - partial; in sha256_base_do_update()
54 memcpy(sctx->buf + partial, data, p); in sha256_base_do_update()
68 partial = 0; in sha256_base_do_update()
71 memcpy(sctx->buf + partial, data, len); in sha256_base_do_update()
82 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in sha256_base_do_finalize() local
84 sctx->buf[partial++] = 0x80; in sha256_base_do_finalize()
85 if (partial > bit_offset) { in sha256_base_do_finalize()
[all …]
Dsm3_base.h44 unsigned int partial = sctx->count % SM3_BLOCK_SIZE; in sm3_base_do_update() local
48 if (unlikely((partial + len) >= SM3_BLOCK_SIZE)) { in sm3_base_do_update()
51 if (partial) { in sm3_base_do_update()
52 int p = SM3_BLOCK_SIZE - partial; in sm3_base_do_update()
54 memcpy(sctx->buffer + partial, data, p); in sm3_base_do_update()
68 partial = 0; in sm3_base_do_update()
71 memcpy(sctx->buffer + partial, data, len); in sm3_base_do_update()
82 unsigned int partial = sctx->count % SM3_BLOCK_SIZE; in sm3_base_do_finalize() local
84 sctx->buffer[partial++] = 0x80; in sm3_base_do_finalize()
85 if (partial > bit_offset) { in sm3_base_do_finalize()
[all …]
Dsha512_base.h62 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_base_do_update() local
68 if (unlikely((partial + len) >= SHA512_BLOCK_SIZE)) { in sha512_base_do_update()
71 if (partial) { in sha512_base_do_update()
72 int p = SHA512_BLOCK_SIZE - partial; in sha512_base_do_update()
74 memcpy(sctx->buf + partial, data, p); in sha512_base_do_update()
88 partial = 0; in sha512_base_do_update()
91 memcpy(sctx->buf + partial, data, len); in sha512_base_do_update()
102 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_base_do_finalize() local
104 sctx->buf[partial++] = 0x80; in sha512_base_do_finalize()
105 if (partial > bit_offset) { in sha512_base_do_finalize()
[all …]
/linux-6.1.9/fs/sysv/
Ditree.c209 Indirect *partial; in get_block() local
218 partial = get_branch(inode, depth, offsets, chain, &err); in get_block()
222 if (!partial) { in get_block()
227 partial = chain+depth-1; /* the whole chain */ in get_block()
234 while (partial > chain) { in get_block()
235 brelse(partial->bh); in get_block()
236 partial--; in get_block()
250 left = (chain + depth) - partial; in get_block()
251 err = alloc_branch(inode, left, offsets+(partial-chain), partial); in get_block()
255 if (splice_branch(inode, chain, partial, left) < 0) in get_block()
[all …]
/linux-6.1.9/drivers/crypto/
Dpadlock-sha.c284 unsigned int partial, done; in padlock_sha1_update_nano() local
291 partial = sctx->count & 0x3f; in padlock_sha1_update_nano()
297 if ((partial + len) >= SHA1_BLOCK_SIZE) { in padlock_sha1_update_nano()
300 if (partial) { in padlock_sha1_update_nano()
301 done = -partial; in padlock_sha1_update_nano()
302 memcpy(sctx->buffer + partial, data, in padlock_sha1_update_nano()
321 partial = 0; in padlock_sha1_update_nano()
324 memcpy(sctx->buffer + partial, src, len - done); in padlock_sha1_update_nano()
332 unsigned int partial, padlen; in padlock_sha1_final_nano() local
339 partial = state->count & 0x3f; in padlock_sha1_final_nano()
[all …]
/linux-6.1.9/arch/arm64/crypto/
Dsha3-ce-glue.c43 if ((sctx->partial + len) >= sctx->rsiz) { in sha3_update()
46 if (sctx->partial) { in sha3_update()
47 int p = sctx->rsiz - sctx->partial; in sha3_update()
49 memcpy(sctx->buf + sctx->partial, data, p); in sha3_update()
56 sctx->partial = 0; in sha3_update()
75 memcpy(sctx->buf + sctx->partial, data, len); in sha3_update()
76 sctx->partial += len; in sha3_update()
91 sctx->buf[sctx->partial++] = 0x06; in sha3_final()
92 memset(sctx->buf + sctx->partial, 0, sctx->rsiz - sctx->partial); in sha3_final()
/linux-6.1.9/arch/powerpc/crypto/
Dsha1.c30 unsigned int partial, done; in powerpc_sha1_update() local
33 partial = sctx->count & 0x3f; in powerpc_sha1_update()
38 if ((partial + len) > 63) { in powerpc_sha1_update()
40 if (partial) { in powerpc_sha1_update()
41 done = -partial; in powerpc_sha1_update()
42 memcpy(sctx->buffer + partial, data, done + 64); in powerpc_sha1_update()
52 partial = 0; in powerpc_sha1_update()
54 memcpy(sctx->buffer + partial, src, len - done); in powerpc_sha1_update()
/linux-6.1.9/fs/ext2/
Dinode.c331 Indirect *partial) in ext2_find_goal() argument
346 return ext2_find_near(inode, partial); in ext2_find_goal()
628 Indirect *partial; in ext2_get_blocks() local
644 partial = ext2_get_branch(inode, depth, offsets, chain, &err); in ext2_get_blocks()
646 if (!partial) { in ext2_get_blocks()
662 partial = chain + depth - 1; in ext2_get_blocks()
692 if (err == -EAGAIN || !verify_chain(chain, partial)) { in ext2_get_blocks()
693 while (partial > chain) { in ext2_get_blocks()
694 brelse(partial->bh); in ext2_get_blocks()
695 partial--; in ext2_get_blocks()
[all …]
/linux-6.1.9/arch/sparc/crypto/
Dsha1_glue.c31 unsigned int len, unsigned int partial) in __sha1_sparc64_update() argument
36 if (partial) { in __sha1_sparc64_update()
37 done = SHA1_BLOCK_SIZE - partial; in __sha1_sparc64_update()
38 memcpy(sctx->buffer + partial, data, done); in __sha1_sparc64_update()
55 unsigned int partial = sctx->count % SHA1_BLOCK_SIZE; in sha1_sparc64_update() local
58 if (partial + len < SHA1_BLOCK_SIZE) { in sha1_sparc64_update()
60 memcpy(sctx->buffer + partial, data, len); in sha1_sparc64_update()
62 __sha1_sparc64_update(sctx, data, len, partial); in sha1_sparc64_update()
Dmd5_glue.c47 unsigned int len, unsigned int partial) in __md5_sparc64_update() argument
52 if (partial) { in __md5_sparc64_update()
53 done = MD5_HMAC_BLOCK_SIZE - partial; in __md5_sparc64_update()
54 memcpy((u8 *)sctx->block + partial, data, done); in __md5_sparc64_update()
71 unsigned int partial = sctx->byte_count % MD5_HMAC_BLOCK_SIZE; in md5_sparc64_update() local
74 if (partial + len < MD5_HMAC_BLOCK_SIZE) { in md5_sparc64_update()
76 memcpy((u8 *)sctx->block + partial, data, len); in md5_sparc64_update()
78 __md5_sparc64_update(sctx, data, len, partial); in md5_sparc64_update()
Dsha512_glue.c30 unsigned int len, unsigned int partial) in __sha512_sparc64_update() argument
36 if (partial) { in __sha512_sparc64_update()
37 done = SHA512_BLOCK_SIZE - partial; in __sha512_sparc64_update()
38 memcpy(sctx->buf + partial, data, done); in __sha512_sparc64_update()
55 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_sparc64_update() local
58 if (partial + len < SHA512_BLOCK_SIZE) { in sha512_sparc64_update()
61 memcpy(sctx->buf + partial, data, len); in sha512_sparc64_update()
63 __sha512_sparc64_update(sctx, data, len, partial); in sha512_sparc64_update()
Dsha256_glue.c31 unsigned int len, unsigned int partial) in __sha256_sparc64_update() argument
36 if (partial) { in __sha256_sparc64_update()
37 done = SHA256_BLOCK_SIZE - partial; in __sha256_sparc64_update()
38 memcpy(sctx->buf + partial, data, done); in __sha256_sparc64_update()
55 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in sha256_sparc64_update() local
58 if (partial + len < SHA256_BLOCK_SIZE) { in sha256_sparc64_update()
60 memcpy(sctx->buf + partial, data, len); in sha256_sparc64_update()
62 __sha256_sparc64_update(sctx, data, len, partial); in sha256_sparc64_update()
/linux-6.1.9/arch/mips/cavium-octeon/crypto/
Docteon-sha1.c78 unsigned int partial; in __octeon_sha1_update() local
82 partial = sctx->count % SHA1_BLOCK_SIZE; in __octeon_sha1_update()
87 if ((partial + len) >= SHA1_BLOCK_SIZE) { in __octeon_sha1_update()
88 if (partial) { in __octeon_sha1_update()
89 done = -partial; in __octeon_sha1_update()
90 memcpy(sctx->buffer + partial, data, in __octeon_sha1_update()
101 partial = 0; in __octeon_sha1_update()
103 memcpy(sctx->buffer + partial, src, len - done); in __octeon_sha1_update()
Docteon-sha256.c70 unsigned int partial; in __octeon_sha256_update() local
74 partial = sctx->count % SHA256_BLOCK_SIZE; in __octeon_sha256_update()
79 if ((partial + len) >= SHA256_BLOCK_SIZE) { in __octeon_sha256_update()
80 if (partial) { in __octeon_sha256_update()
81 done = -partial; in __octeon_sha256_update()
82 memcpy(sctx->buf + partial, data, in __octeon_sha256_update()
93 partial = 0; in __octeon_sha256_update()
95 memcpy(sctx->buf + partial, src, len - done); in __octeon_sha256_update()
/linux-6.1.9/drivers/usb/storage/
Dfreecom.c227 unsigned int partial; in freecom_transport() local
266 FCM_STATUS_PACKET_LENGTH, &partial); in freecom_transport()
267 usb_stor_dbg(us, "foo Status result %d %u\n", result, partial); in freecom_transport()
271 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport()
308 FCM_STATUS_PACKET_LENGTH, &partial); in freecom_transport()
310 usb_stor_dbg(us, "bar Status result %d %u\n", result, partial); in freecom_transport()
314 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport()
317 if (partial != 4) in freecom_transport()
376 FCM_PACKET_LENGTH, &partial); in freecom_transport()
377 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport()
[all …]
/linux-6.1.9/arch/arm/crypto/
Dghash-ce-glue.c98 unsigned int partial = ctx->count % GHASH_BLOCK_SIZE; in ghash_update() local
102 if ((partial + len) >= GHASH_BLOCK_SIZE) { in ghash_update()
106 if (partial) { in ghash_update()
107 int p = GHASH_BLOCK_SIZE - partial; in ghash_update()
109 memcpy(ctx->buf + partial, src, p); in ghash_update()
118 partial ? ctx->buf : NULL); in ghash_update()
120 partial = 0; in ghash_update()
123 memcpy(ctx->buf + partial, src, len); in ghash_update()
130 unsigned int partial = ctx->count % GHASH_BLOCK_SIZE; in ghash_final() local
132 if (partial) { in ghash_final()
[all …]
/linux-6.1.9/net/smc/
Dsmc_rx.c150 struct partial_page *partial; in smc_rx_splice() local
163 partial = kcalloc(nr_pages, sizeof(*partial), GFP_KERNEL); in smc_rx_splice()
164 if (!partial) in smc_rx_splice()
180 partial[0].offset = src - (char *)smc->conn.rmb_desc->cpu_addr; in smc_rx_splice()
181 partial[0].len = len; in smc_rx_splice()
182 partial[0].private = (unsigned long)priv[0]; in smc_rx_splice()
193 partial[i].offset = offset; in smc_rx_splice()
194 partial[i].len = size; in smc_rx_splice()
195 partial[i].private = (unsigned long)priv[i]; in smc_rx_splice()
204 spd.partial = partial; in smc_rx_splice()
[all …]
/linux-6.1.9/arch/x86/include/asm/
Dunwind.h72 bool *partial) in unwind_get_entry_regs() argument
77 if (partial) { in unwind_get_entry_regs()
79 *partial = !state->full_regs; in unwind_get_entry_regs()
81 *partial = false; in unwind_get_entry_regs()
89 bool *partial) in unwind_get_entry_regs() argument
/linux-6.1.9/crypto/
Dsha3_generic.c168 sctx->partial = 0; in crypto_sha3_init()
185 if ((sctx->partial + len) > (sctx->rsiz - 1)) { in crypto_sha3_update()
186 if (sctx->partial) { in crypto_sha3_update()
187 done = -sctx->partial; in crypto_sha3_update()
188 memcpy(sctx->buf + sctx->partial, data, in crypto_sha3_update()
204 sctx->partial = 0; in crypto_sha3_update()
206 memcpy(sctx->buf + sctx->partial, src, len - done); in crypto_sha3_update()
207 sctx->partial += (len - done); in crypto_sha3_update()
216 unsigned int i, inlen = sctx->partial; in crypto_sha3_final()
Dsm3.c180 unsigned int partial = sctx->count % SM3_BLOCK_SIZE; in sm3_update() local
185 if ((partial + len) >= SM3_BLOCK_SIZE) { in sm3_update()
188 if (partial) { in sm3_update()
189 int p = SM3_BLOCK_SIZE - partial; in sm3_update()
191 memcpy(sctx->buffer + partial, data, p); in sm3_update()
208 partial = 0; in sm3_update()
211 memcpy(sctx->buffer + partial, data, len); in sm3_update()
220 unsigned int partial = sctx->count % SM3_BLOCK_SIZE; in sm3_final() local
224 sctx->buffer[partial++] = 0x80; in sm3_final()
225 if (partial > bit_offset) { in sm3_final()
[all …]
/linux-6.1.9/lib/crypto/
Dsha256.c124 unsigned int partial, done; in sha256_update() local
128 partial = sctx->count & 0x3f; in sha256_update()
133 if ((partial + len) > 63) { in sha256_update()
134 if (partial) { in sha256_update()
135 done = -partial; in sha256_update()
136 memcpy(sctx->buf + partial, data, done + 64); in sha256_update()
148 partial = 0; in sha256_update()
150 memcpy(sctx->buf + partial, src, len - done); in sha256_update()
/linux-6.1.9/arch/x86/kernel/
Ddumpstack.c155 bool partial, const char *log_lvl) in show_regs_if_on_stack() argument
166 if (!partial && on_stack(info, regs, sizeof(*regs))) { in show_regs_if_on_stack()
169 } else if (partial && on_stack(info, (void *)regs + IRET_FRAME_OFFSET, in show_regs_if_on_stack()
193 bool partial = false; in show_trace_log_lvl() local
199 regs = unwind_get_entry_regs(&state, &partial); in show_trace_log_lvl()
237 show_regs_if_on_stack(&stack_info, regs, partial, log_lvl); in show_trace_log_lvl()
296 regs = unwind_get_entry_regs(&state, &partial); in show_trace_log_lvl()
298 show_regs_if_on_stack(&stack_info, regs, partial, log_lvl); in show_trace_log_lvl()

12345678910