Fix a handful of spelling mistakes and typos

There are quite a few spelling mistakes and typos in comments and
user facing message literal strings as found using codespell. Fix
these.

Signed-off-by: Colin Ian King <colin.i.king@gmail.com>
Signed-off-by: Pablo de Lara <pablo.de.lara.guarch@intel.com>
This commit is contained in:
Colin Ian King 2024-01-23 11:05:40 +00:00 committed by Tomasz Kantecki
parent ffc16330d8
commit 1500db751d
22 changed files with 40 additions and 40 deletions

View File

@ -139,7 +139,7 @@ v2.31
- Zlib header can be now created on the fly when starting the compression. - Zlib header can be now created on the fly when starting the compression.
- Added isal_zlib_hdr_init() function to initialize the zlib header to 0. - Added isal_zlib_hdr_init() function to initialize the zlib header to 0.
* Zero-memory dectection improvements: * Zero-memory detection improvements:
- Optimized AVX implementation. - Optimized AVX implementation.
- Added new AVX2 and AVX512 implementations. - Added new AVX2 and AVX512 implementations.
@ -209,7 +209,7 @@ v2.25
* Igzip performance improvements and features. * Igzip performance improvements and features.
- Performance improvements for uncompressable files. Random or uncompressable - Performance improvements for uncompressable files. Random or uncompressable
files can be up to 3x faster in level 1 or 2 compression. files can be up to 3x faster in level 1 or 2 compression.
- Additional small file performance improvments. - Additional small file performance improvements.
- New options in igzip cli: use name from header or not, test compressed file. - New options in igzip cli: use name from header or not, test compressed file.
* Multi-arch autoconf script. * Multi-arch autoconf script.
@ -331,7 +331,7 @@ v2.14
v2.13 v2.13
* Erasure code improvments * Erasure code improvements
- 32-bit port of optimized gf_vect_dot_prod() functions. This makes - 32-bit port of optimized gf_vect_dot_prod() functions. This makes
ec_encode_data() functions much faster on 32-bit processors. ec_encode_data() functions much faster on 32-bit processors.
- Avoton performance improvements. Performance on Avoton for - Avoton performance improvements. Performance on Avoton for

View File

@ -185,7 +185,7 @@ full_block:
; ;; branch into array ; ;; branch into array
; jmp CONCAT(crc_,128,) ; jmp CONCAT(crc_,128,)
; Fall thruogh into top of crc array (crc_128) ; Fall through into top of crc array (crc_128)
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;; ;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;; ;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;

View File

@ -61,7 +61,7 @@ void gf_gen_full_pb_cauchy_matrix(u8 * a, int m, int k)
{ {
int i, j, p = m - k; int i, j, p = m - k;
// Identity matrix in top k x k to indicate a symetric code // Identity matrix in top k x k to indicate a symmetric code
memset(a, 0, k * m); memset(a, 0, k * m);
for (i = 0; i < k; i++) for (i = 0; i < k; i++)
a[k * i + i] = 1; a[k * i + i] = 1;
@ -99,7 +99,7 @@ void gf_gen_full_pb_vand_matrix(u8 * a, int m, int k)
int i, j, p = m - k; int i, j, p = m - k;
unsigned char q, gen = 1; unsigned char q, gen = 1;
// Identity matrix in top k x k to indicate a symetric code // Identity matrix in top k x k to indicate a symmetric code
memset(a, 0, k * m); memset(a, 0, k * m);
for (i = 0; i < k; i++) for (i = 0; i < k; i++)
a[k * i + i] = 1; a[k * i + i] = 1;

View File

@ -54,7 +54,7 @@ Macros
uint32_t adler32_neon(uint32_t adler32, uint8_t * start, uint32_t length); uint32_t adler32_neon(uint32_t adler32, uint8_t * start, uint32_t length);
*/ */
/* /*
Arguements list Arguments list
*/ */
adler32 .req w0 adler32 .req w0
start .req x1 start .req x1

View File

@ -106,7 +106,7 @@ void isal_deflate_icf_body_hash_hist_base(struct isal_zstream *stream);
declare_generic_reg param1, 1,x declare_generic_reg param1, 1,x
declare_generic_reg param2, 2,x declare_generic_reg param2, 2,x
/* local varibale */ /* local variable */
declare_generic_reg level_buf, 18,x declare_generic_reg level_buf, 18,x
declare_generic_reg avail_in, 13,w declare_generic_reg avail_in, 13,w
declare_generic_reg end_in, 13,x declare_generic_reg end_in, 13,x

View File

@ -91,7 +91,7 @@ default rel
save_reg r12, 2*16 + 2*8 save_reg r12, 2*16 + 2*8
save_reg r13, 2*16 + 3*8 save_reg r13, 2*16 + 3*8
end_prolog end_prolog
mov init_d, ecx ; initalize init_d from arg1 to keep ecx free mov init_d, ecx ; initialize init_d from arg1 to keep ecx free
%endmacro %endmacro
%macro FUNC_RESTORE 0 %macro FUNC_RESTORE 0

View File

@ -86,7 +86,7 @@ default rel
save_reg r12, 2*8 save_reg r12, 2*8
save_reg r13, 3*8 save_reg r13, 3*8
end_prolog end_prolog
mov init_d, ecx ; initalize init_d from arg1 to keep ecx free mov init_d, ecx ; initialize init_d from arg1 to keep ecx free
%endmacro %endmacro
%macro FUNC_RESTORE 0 %macro FUNC_RESTORE 0

View File

@ -962,7 +962,7 @@ gen_huff_code_lens(struct heap_tree *heap_space, uint32_t heap_size, uint32_t *
* @requires table has been initialized to already contain the code length for each element. * @requires table has been initialized to already contain the code length for each element.
* @param table: A lookup table used to store the codes. * @param table: A lookup table used to store the codes.
* @param table_length: The length of table. * @param table_length: The length of table.
* @param count: a histogram representing the number of occurences of codes of a given length * @param count: a histogram representing the number of occurrences of codes of a given length
*/ */
static inline uint32_t set_huff_codes(struct huff_code *huff_code_table, int table_length, static inline uint32_t set_huff_codes(struct huff_code *huff_code_table, int table_length,
uint32_t * count) uint32_t * count)
@ -1031,7 +1031,7 @@ static inline uint32_t set_dist_huff_codes(struct huff_code *codes, uint32_t * b
* 0 corresponds to not end of block and all other inputs correspond to end of block. * 0 corresponds to not end of block and all other inputs correspond to end of block.
* @param hclen: Length of huffman code for huffman codes minus 4. * @param hclen: Length of huffman code for huffman codes minus 4.
* @param hlit: Length of literal/length table minus 257. * @param hlit: Length of literal/length table minus 257.
* @parm hdist: Length of distance table minus 1. * @param hdist: Length of distance table minus 1.
*/ */
static int create_huffman_header(struct BitBuf2 *header_bitbuf, static int create_huffman_header(struct BitBuf2 *header_bitbuf,
struct huff_code *lookup_table, struct huff_code *lookup_table,
@ -1233,7 +1233,7 @@ static inline uint32_t rl_encode(uint16_t * codes, uint32_t num_codes, uint64_t
* @brief Creates a two table representation of huffman codes. * @brief Creates a two table representation of huffman codes.
* @param code_table: output table containing the code * @param code_table: output table containing the code
* @param code_size_table: output table containing the code length * @param code_size_table: output table containing the code length
* @param length: the lenght of hufftable * @param length: the length of hufftable
* @param hufftable: a huffman lookup table * @param hufftable: a huffman lookup table
*/ */
static void create_code_tables(uint16_t * code_table, uint8_t * code_length_table, static void create_code_tables(uint16_t * code_table, uint8_t * code_length_table,
@ -1262,7 +1262,7 @@ static void create_packed_len_table(uint32_t * packed_table,
uint16_t extra_bits_count = 0; uint16_t extra_bits_count = 0;
/* Gain extra bits is the next place where the number of extra bits in /* Gain extra bits is the next place where the number of extra bits in
* lenght codes increases. */ * length codes increases. */
uint16_t gain_extra_bits = LEN_EXTRA_BITS_START; uint16_t gain_extra_bits = LEN_EXTRA_BITS_START;
for (i = 257; i < LIT_LEN - 1; i++) { for (i = 257; i < LIT_LEN - 1; i++) {
@ -1366,7 +1366,7 @@ static int are_hufftables_useable(struct huff_code *lit_len_hufftable,
max_code_len = max_lit_code_len + max_len_code_len + max_dist_code_len; max_code_len = max_lit_code_len + max_len_code_len + max_dist_code_len;
/* Some versions of igzip can write upto one literal, one length and one /* Some versions of igzip can write up to one literal, one length and one
* distance code at the same time. This checks to make sure that is * distance code at the same time. This checks to make sure that is
* always writeable in bitbuf*/ * always writeable in bitbuf*/
return (max_code_len > MAX_BITBUF_BIT_WRITE); return (max_code_len > MAX_BITBUF_BIT_WRITE);

View File

@ -173,7 +173,7 @@ struct hufftables_icf {
* *
* @param bb: bitbuf structure where the header huffman code header is written * @param bb: bitbuf structure where the header huffman code header is written
* @param hufftables: output huffman code representation * @param hufftables: output huffman code representation
* @param hist: histogram used to generat huffman code * @param hist: histogram used to generate huffman code
* @param end_of_block: flag whether this is the final huffman code * @param end_of_block: flag whether this is the final huffman code
* *
* @returns Returns the length in bits of the block with histogram hist encoded * @returns Returns the length in bits of the block with histogram hist encoded

View File

@ -625,7 +625,7 @@ static void write_constant_compressed_stateless(struct isal_zstream *stream,
/* Add smaller repeat codes as necessary. Code280 can describe repeat /* Add smaller repeat codes as necessary. Code280 can describe repeat
* lengths of 115-130 bits. Code10 can describe repeat lengths of 10 * lengths of 115-130 bits. Code10 can describe repeat lengths of 10
* bits. If more than 230 bits, fill code with two code280s. Else if * bits. If more than 230 bits, fill code with two code280s. Else if
* more than 115 repeates, fill with code10s until one code280 can * more than 115 repeats, fill with code10s until one code280 can
* finish the rest of the repeats. Else, fill with code10s and * finish the rest of the repeats. Else, fill with code10s and
* literals */ * literals */
if (rep_extra > 115) { if (rep_extra > 115) {

View File

@ -442,7 +442,7 @@ align 16
or code4, code or code4, code
add code_len2, rcx add code_len2, rcx
;; Setup for updateing hash ;; Setup for updating hash
lea tmp3, [f_i + 2] ; tmp3 <= k lea tmp3, [f_i + 2] ; tmp3 <= k
add f_i, len add f_i, len

View File

@ -399,7 +399,7 @@ stack_size equ 4 * 8 + 8 * 8
%define %%state_offset %3 ; Type of huff code, should be either LIT or DIST %define %%state_offset %3 ; Type of huff code, should be either LIT or DIST
%define %%read_in %4 ; Bits read in from compressed stream %define %%read_in %4 ; Bits read in from compressed stream
%define %%read_in_length %5 ; Number of valid bits in read_in %define %%read_in_length %5 ; Number of valid bits in read_in
%define %%next_sym %6 ; Returned symobl %define %%next_sym %6 ; Returned symbol
%define %%next_extra_bits %7 %define %%next_extra_bits %7
%define %%next_bits %8 %define %%next_bits %8
@ -447,7 +447,7 @@ stack_size equ 4 * 8 + 8 * 8
%define %%state_offset %3 %define %%state_offset %3
%define %%read_in %4 ; Bits read in from compressed stream %define %%read_in %4 ; Bits read in from compressed stream
%define %%read_in_length %5 ; Number of valid bits in read_in %define %%read_in_length %5 ; Number of valid bits in read_in
%define %%next_sym %6 ; Returned symobl %define %%next_sym %6 ; Returned symbol
%define %%next_extra_bits %7 %define %%next_extra_bits %7
%define %%next_bits %8 %define %%next_bits %8

View File

@ -99,7 +99,7 @@ int usage(void)
" -X use compression level X with 0 <= X <= 1\n" " -X use compression level X with 0 <= X <= 1\n"
" -b <size> input buffer size, 0 buffers all the input\n" " -b <size> input buffer size, 0 buffers all the input\n"
" -i <time> time in seconds to benchmark (at least 0)\n" " -i <time> time in seconds to benchmark (at least 0)\n"
" -o <file> output file for compresed data\n" " -o <file> output file for compressed data\n"
" -d <file> dictionary file used by compression\n" " -d <file> dictionary file used by compression\n"
" -w <size> log base 2 size of history window, between 8 and 15\n"); " -w <size> log base 2 size of history window, between 8 and 15\n");

View File

@ -436,7 +436,7 @@ isal_deflate_icf_body_ %+ METHOD %+ _ %+ ARCH %+ :
.len_dist_huffman: .len_dist_huffman:
dec f_i dec f_i
;; Setup for updateing hash ;; Setup for updating hash
lea tmp3, [f_i + 2] ; tmp3 <= k lea tmp3, [f_i + 2] ; tmp3 <= k
neg dist neg dist

View File

@ -454,7 +454,7 @@ static void make_inflate_huff_code_lit_len(struct inflate_huff_code_large *resul
min_length = last_length; min_length = last_length;
for (; last_length <= ISAL_DECODE_LONG_BITS; last_length++) { for (; last_length <= ISAL_DECODE_LONG_BITS; last_length++) {
/* Copy forward previosly set codes */ /* Copy forward previously set codes */
memcpy(short_code_lookup + copy_size, short_code_lookup, memcpy(short_code_lookup + copy_size, short_code_lookup,
sizeof(*short_code_lookup) * copy_size); sizeof(*short_code_lookup) * copy_size);
copy_size *= 2; copy_size *= 2;
@ -1900,7 +1900,7 @@ static int check_gzip_checksum(struct inflate_state *state)
int ret; int ret;
if (state->read_in_length >= 8 * GZIP_TRAILER_LEN) { if (state->read_in_length >= 8 * GZIP_TRAILER_LEN) {
/* The following is unecessary as state->read_in_length == 64 */ /* The following is unnecessary as state->read_in_length == 64 */
/* bit_count = state->read_in_length % 8; */ /* bit_count = state->read_in_length % 8; */
/* state->read_in >>= bit_count; */ /* state->read_in >>= bit_count; */
/* state->read_in_length -= bit_count; */ /* state->read_in_length -= bit_count; */

View File

@ -626,7 +626,7 @@ int isal_inflate_with_checks(struct inflate_state *state, uint32_t compress_len,
ret = isal_inflate(state); ret = isal_inflate(state);
/* Verify the stream is in a valid state when no errors occured */ /* Verify the stream is in a valid state when no errors occurred */
if (ret >= 0) { if (ret >= 0) {
stream_check = stream_check =
inflate_state_valid_check(state, in_buf, in_size, out_buf, out_size, inflate_state_valid_check(state, in_buf, in_size, out_buf, out_size,
@ -1210,7 +1210,7 @@ int compress_multi_pass(uint8_t * data, uint32_t data_size, uint8_t * compressed
/* Setup out buffer for next round of compression */ /* Setup out buffer for next round of compression */
if (stream->avail_out == 0) { if (stream->avail_out == 0) {
/* Save compressed data inot compressed_buf */ /* Save compressed data into compressed_buf */
if (out_buf != NULL) { if (out_buf != NULL) {
memcpy(compressed_buf + out_processed, out_buf, memcpy(compressed_buf + out_processed, out_buf,
out_size - stream->avail_out); out_size - stream->avail_out);
@ -1450,7 +1450,7 @@ int compress_ver_rep_buf(uint8_t * data, uint32_t data_size, uint64_t data_rep_s
return COMPRESS_GENERAL_ERROR; return COMPRESS_GENERAL_ERROR;
} }
/* Verfiy the compressed buffer */ /* Verify the compressed buffer */
state.next_in = compressed_buf; state.next_in = compressed_buf;
state.avail_in = compressed_size; state.avail_in = compressed_size;
state.next_out = NULL; state.next_out = NULL;

View File

@ -86,7 +86,7 @@ void print_error(int32_t error)
printf("Incorrect os field found\n"); printf("Incorrect os field found\n");
break; break;
case INCORRECT_EXTRA_LEN: case INCORRECT_EXTRA_LEN:
printf("Incorect extra_len field found\n"); printf("Incorrect extra_len field found\n");
break; break;
case INCORRECT_EXTRA_BUF: case INCORRECT_EXTRA_BUF:
printf("Incorrect extra buffer found\n"); printf("Incorrect extra buffer found\n");

View File

@ -46,7 +46,7 @@
#define END_OF_BLOCK 0x7 #define END_OF_BLOCK 0x7
#define END_OF_BLOCK_LEN 0x4 #define END_OF_BLOCK_LEN 0x4
/* MIN_REPEAT_LEN currently optimizes storage space, another possiblity is to /* MIN_REPEAT_LEN currently optimizes storage space, another possibility is to
* find the size which optimizes speed instead.*/ * find the size which optimizes speed instead.*/
#define MIN_REPEAT_LEN 4*1024 #define MIN_REPEAT_LEN 4*1024

View File

@ -56,7 +56,7 @@
;;;; ;;;;
; multibinary macro: ; multibinary macro:
; creates the visable entry point that uses HW optimized call pointer ; creates the visible entry point that uses HW optimized call pointer
; creates the init of the HW optimized call pointer ; creates the init of the HW optimized call pointer
;;;; ;;;;
%macro mbin_interface 1 %macro mbin_interface 1

View File

@ -73,7 +73,7 @@ extern "C" {
# define DEBUG_PRINT(x) do {} while (0) # define DEBUG_PRINT(x) do {} while (0)
#endif #endif
/* Decide wether to use benchmark time as an approximation or a minimum. Fewer /* Decide whether to use benchmark time as an approximation or a minimum. Fewer
* calls to the timer are required for the approximation case.*/ * calls to the timer are required for the approximation case.*/
#define BENCHMARK_MIN_TIME 0 #define BENCHMARK_MIN_TIME 0
#define BENCHMARK_APPROX_TIME 1 #define BENCHMARK_APPROX_TIME 1
@ -100,7 +100,7 @@ extern "C" {
#endif #endif
# define GHZ 1000000000 # define GHZ 1000000000
# define UNIT_SCALE (GHZ) # define UNIT_SCALE (GHZ)
# define CALLIBRATE_TIME (UNIT_SCALE / 2) # define CALIBRATE_TIME (UNIT_SCALE / 2)
static inline long long get_time(void) { static inline long long get_time(void) {
unsigned int dummy; unsigned int dummy;
return __rdtscp(&dummy); return __rdtscp(&dummy);
@ -115,7 +115,7 @@ static inline long long get_res(void) {
#endif #endif
#ifdef _MSC_VER #ifdef _MSC_VER
#define UNIT_SCALE get_res() #define UNIT_SCALE get_res()
#define CALLIBRATE_TIME (UNIT_SCALE / 4) #define CALIBRATE_TIME (UNIT_SCALE / 4)
static inline long long get_time(void) { static inline long long get_time(void) {
long long ret = 0; long long ret = 0;
QueryPerformanceCounter(&ret); QueryPerformanceCounter(&ret);
@ -130,7 +130,7 @@ static inline long long get_res(void) {
#else #else
# define NANO_SCALE 1000000000 # define NANO_SCALE 1000000000
# define UNIT_SCALE NANO_SCALE # define UNIT_SCALE NANO_SCALE
# define CALLIBRATE_TIME (UNIT_SCALE / 4) # define CALIBRATE_TIME (UNIT_SCALE / 4)
#ifdef __FreeBSD__ #ifdef __FreeBSD__
# define CLOCK_ID CLOCK_MONOTONIC_PRECISE # define CLOCK_ID CLOCK_MONOTONIC_PRECISE
#else #else
@ -208,15 +208,15 @@ static inline unsigned long long estimate_perf_iterations(struct perf *p,
return (total + get_res() - 1) / get_res(); return (total + get_res() - 1) / get_res();
} }
#define CALLIBRATE(PERF, FUNC_CALL) { \ #define CALIBRATE(PERF, FUNC_CALL) { \
unsigned long long _i, _iter = 1; \ unsigned long long _i, _iter = 1; \
perf_start(PERF); \ perf_start(PERF); \
FUNC_CALL; \ FUNC_CALL; \
perf_pause(PERF); \ perf_pause(PERF); \
\ \
while (get_base_elapsed(PERF) < CALLIBRATE_TIME) { \ while (get_base_elapsed(PERF) < CALIBRATE_TIME) { \
_iter = estimate_perf_iterations(PERF, _iter, \ _iter = estimate_perf_iterations(PERF, _iter, \
2 * CALLIBRATE_TIME); \ 2 * CALIBRATE_TIME); \
perf_start(PERF); \ perf_start(PERF); \
for (_i = 0; _i < _iter; _i++) { \ for (_i = 0; _i < _iter; _i++) { \
FUNC_CALL; \ FUNC_CALL; \
@ -255,7 +255,7 @@ static inline unsigned long long estimate_perf_iterations(struct perf *p,
#define BENCHMARK(PERF, RUN_TIME, FUNC_CALL) { \ #define BENCHMARK(PERF, RUN_TIME, FUNC_CALL) { \
if((RUN_TIME) > 0) { \ if((RUN_TIME) > 0) { \
CALLIBRATE(PERF, FUNC_CALL); \ CALIBRATE(PERF, FUNC_CALL); \
PERFORMANCE_TEST(PERF, RUN_TIME, FUNC_CALL); \ PERFORMANCE_TEST(PERF, RUN_TIME, FUNC_CALL); \
\ \
} else { \ } else { \

View File

@ -244,7 +244,7 @@ if command -V md5sum >/dev/null 2>&1 && command -V dd >/dev/null 2>&1; then
dd if=<(for i in `seq 1000`; do cat $TEST_FILE; done) iflag=fullblock bs=1M count=201 2> out.stder | tee >(md5sum > out.sum1) \ dd if=<(for i in `seq 1000`; do cat $TEST_FILE; done) iflag=fullblock bs=1M count=201 2> out.stder | tee >(md5sum > out.sum1) \
| $IGZIP -c -T 4 | $IGZIP -d | md5sum > out.sum2 \ | $IGZIP -c -T 4 | $IGZIP -d | md5sum > out.sum2 \
&& $DIFF out.sum1 out.sum2 || ret=1 && $DIFF out.sum1 out.sum2 || ret=1
pass_check $ret "Large stream compresss test" pass_check $ret "Large stream compress test"
clear_dir clear_dir
if test -e /dev/urandom; then if test -e /dev/urandom; then

View File

@ -185,7 +185,7 @@ next_vect:
;;; prefetch [ptr+pos+(8*16)] ;;; prefetch [ptr+pos+(8*16)]
jge next_vect ;Loop for each vect jge next_vect ;Loop for each vect
;; End of vects, chech that all parity regs = 0 ;; End of vects, check that all parity regs = 0
mov tmp, vec ;Back to last vector mov tmp, vec ;Back to last vector
por xmm0, xmm1 por xmm0, xmm1
por xmm2, xmm3 por xmm2, xmm3