@@ -108,7 +108,7 @@ int unpack_exact(const uint8_t* in, Uint* out, int batch_size, int bit_offset) {
108108 // Due to misalignment, on large bit width, the spread can be larger than the maximum
109109 // size integer. For instance a 63 bit width misaligned packed integer can spread over 9
110110 // aligned bytes.
111- constexpr bool kOversized = kBufferSize < kMaxSpreadBytes ;
111+ constexpr bool kLarge = kBufferSize < kMaxSpreadBytes ;
112112 constexpr buffer_uint kLowMask =
113113 bit_util::LeastSignificantBitMask<buffer_uint, true >(kPackedBitWidth );
114114
@@ -126,7 +126,7 @@ int unpack_exact(const uint8_t* in, Uint* out, int batch_size, int bit_offset) {
126126 // Reading the bytes for the current value.
127127 // Must be careful not to read out of input bounds.
128128 buffer_uint buffer = 0 ;
129- if constexpr (kOversized ) {
129+ if constexpr (kLarge ) {
130130 // We read the max possible bytes in the first pass and handle the rest after.
131131 // Even though the worst spread does not happen on all iterations we can still read
132132 // all bytes because we will mask them.
@@ -141,7 +141,7 @@ int unpack_exact(const uint8_t* in, Uint* out, int batch_size, int bit_offset) {
141141 Uint val = static_cast <Uint>(buffer & kLowMask );
142142
143143 // Handle the oversized bytes
144- if constexpr (kOversized ) {
144+ if constexpr (kLarge ) {
145145 // The oversized bytes do not happen at all iterations
146146 if (spread_bytes > kBufferSize ) {
147147 std::memcpy (&buffer, in + start_byte + kBufferSize , spread_bytes - kBufferSize );
0 commit comments