@@ -129,7 +129,7 @@ static output_t make_output(const uint32_t input_cv[8],
129
129
* bytes. For that reason, chaining values in the CV stack are represented as
130
130
* bytes.
131
131
*/
132
- static void output_chaining_value (const blake3_impl_ops_t * ops ,
132
+ static void output_chaining_value (const blake3_ops_t * ops ,
133
133
const output_t * ctx , uint8_t cv [32 ])
134
134
{
135
135
uint32_t cv_words [8 ];
@@ -139,7 +139,7 @@ static void output_chaining_value(const blake3_impl_ops_t *ops,
139
139
store_cv_words (cv , cv_words );
140
140
}
141
141
142
- static void output_root_bytes (const blake3_impl_ops_t * ops , const output_t * ctx ,
142
+ static void output_root_bytes (const blake3_ops_t * ops , const output_t * ctx ,
143
143
uint64_t seek , uint8_t * out , size_t out_len )
144
144
{
145
145
uint64_t output_block_counter = seek / 64 ;
@@ -163,7 +163,7 @@ static void output_root_bytes(const blake3_impl_ops_t *ops, const output_t *ctx,
163
163
}
164
164
}
165
165
166
- static void chunk_state_update (const blake3_impl_ops_t * ops ,
166
+ static void chunk_state_update (const blake3_ops_t * ops ,
167
167
blake3_chunk_state_t * ctx , const uint8_t * input , size_t input_len )
168
168
{
169
169
if (ctx -> buf_len > 0 ) {
@@ -230,7 +230,7 @@ static size_t left_len(size_t content_len)
230
230
* number of chunks hashed. These chunks are never the root and never empty;
231
231
* those cases use a different codepath.
232
232
*/
233
- static size_t compress_chunks_parallel (const blake3_impl_ops_t * ops ,
233
+ static size_t compress_chunks_parallel (const blake3_ops_t * ops ,
234
234
const uint8_t * input , size_t input_len , const uint32_t key [8 ],
235
235
uint64_t chunk_counter , uint8_t flags , uint8_t * out )
236
236
{
@@ -274,7 +274,7 @@ static size_t compress_chunks_parallel(const blake3_impl_ops_t *ops,
274
274
* return it as an additional output.) These parents are never the root and
275
275
* never empty; those cases use a different codepath.
276
276
*/
277
- static size_t compress_parents_parallel (const blake3_impl_ops_t * ops ,
277
+ static size_t compress_parents_parallel (const blake3_ops_t * ops ,
278
278
const uint8_t * child_chaining_values , size_t num_chaining_values ,
279
279
const uint32_t key [8 ], uint8_t flags , uint8_t * out )
280
280
{
@@ -320,7 +320,7 @@ static size_t compress_parents_parallel(const blake3_impl_ops_t *ops,
320
320
* of implementing this special rule? Because we don't want to limit SIMD or
321
321
* multi-threading parallelism for that update().
322
322
*/
323
- static size_t blake3_compress_subtree_wide (const blake3_impl_ops_t * ops ,
323
+ static size_t blake3_compress_subtree_wide (const blake3_ops_t * ops ,
324
324
const uint8_t * input , size_t input_len , const uint32_t key [8 ],
325
325
uint64_t chunk_counter , uint8_t flags , uint8_t * out )
326
326
{
@@ -406,7 +406,7 @@ static size_t blake3_compress_subtree_wide(const blake3_impl_ops_t *ops,
406
406
* As with compress_subtree_wide(), this function is not used on inputs of 1
407
407
* chunk or less. That's a different codepath.
408
408
*/
409
- static void compress_subtree_to_parent_node (const blake3_impl_ops_t * ops ,
409
+ static void compress_subtree_to_parent_node (const blake3_ops_t * ops ,
410
410
const uint8_t * input , size_t input_len , const uint32_t key [8 ],
411
411
uint64_t chunk_counter , uint8_t flags , uint8_t out [2 * BLAKE3_OUT_LEN ])
412
412
{
@@ -434,7 +434,7 @@ static void hasher_init_base(BLAKE3_CTX *ctx, const uint32_t key[8],
434
434
memcpy (ctx -> key , key , BLAKE3_KEY_LEN );
435
435
chunk_state_init (& ctx -> chunk , key , flags );
436
436
ctx -> cv_stack_len = 0 ;
437
- ctx -> ops = blake3_impl_get_ops ();
437
+ ctx -> ops = blake3_get_ops ();
438
438
}
439
439
440
440
/*
0 commit comments