Remove static modifiers
This commit is contained in:
parent
5e5592178d
commit
da0f1cacea
|
@ -1,4 +1,4 @@
|
||||||
/* auto-generated on Fri Aug 23 09:54:21 DST 2019. Do not edit! */
|
/* auto-generated on Fri Aug 23 10:23:28 DST 2019. Do not edit! */
|
||||||
|
|
||||||
#include <iostream>
|
#include <iostream>
|
||||||
#include "simdjson.h"
|
#include "simdjson.h"
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
/* auto-generated on Fri Aug 23 09:54:21 DST 2019. Do not edit! */
|
/* auto-generated on Fri Aug 23 10:23:28 DST 2019. Do not edit! */
|
||||||
#include "simdjson.h"
|
#include "simdjson.h"
|
||||||
|
|
||||||
/* used for http://dmalloc.com/ Dmalloc - Debug Malloc Library */
|
/* used for http://dmalloc.com/ Dmalloc - Debug Malloc Library */
|
||||||
|
@ -1451,7 +1451,7 @@ UNTARGET_REGION // westmere
|
||||||
|
|
||||||
namespace simdjson::arm64 {
|
namespace simdjson::arm64 {
|
||||||
|
|
||||||
static really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
||||||
|
|
||||||
#ifdef __ARM_FEATURE_CRYPTO // some ARM processors lack this extension
|
#ifdef __ARM_FEATURE_CRYPTO // some ARM processors lack this extension
|
||||||
return vmull_p64(-1ULL, quote_bits);
|
return vmull_p64(-1ULL, quote_bits);
|
||||||
|
@ -1460,7 +1460,7 @@ static really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
static really_inline void find_whitespace_and_structurals(
|
really_inline void find_whitespace_and_structurals(
|
||||||
simd_input<ARCHITECTURE> in, uint64_t &whitespace,
|
simd_input<ARCHITECTURE> in, uint64_t &whitespace,
|
||||||
uint64_t &structurals) {
|
uint64_t &structurals) {
|
||||||
const uint8x16_t low_nibble_mask =
|
const uint8x16_t low_nibble_mask =
|
||||||
|
@ -1518,7 +1518,7 @@ static really_inline void find_whitespace_and_structurals(
|
||||||
// This is just a naive implementation. It should be normally
|
// This is just a naive implementation. It should be normally
|
||||||
// disable, but can be used for research purposes to compare
|
// disable, but can be used for research purposes to compare
|
||||||
// again our optimized version.
|
// again our optimized version.
|
||||||
static really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
||||||
uint32_t *out_ptr = base_ptr + base;
|
uint32_t *out_ptr = base_ptr + base;
|
||||||
idx -= 64;
|
idx -= 64;
|
||||||
while (bits != 0) {
|
while (bits != 0) {
|
||||||
|
@ -1536,7 +1536,7 @@ static really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint3
|
||||||
// base_ptr[base] incrementing base as we go
|
// base_ptr[base] incrementing base as we go
|
||||||
// will potentially store extra values beyond end of valid bits, so base_ptr
|
// will potentially store extra values beyond end of valid bits, so base_ptr
|
||||||
// needs to be large enough to handle this
|
// needs to be large enough to handle this
|
||||||
static really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
||||||
// In some instances, the next branch is expensive because it is mispredicted.
|
// In some instances, the next branch is expensive because it is mispredicted.
|
||||||
// Unfortunately, in other cases,
|
// Unfortunately, in other cases,
|
||||||
// it helps tremendously.
|
// it helps tremendously.
|
||||||
|
@ -1864,7 +1864,7 @@ int find_structural_bits<Architecture::ARM64>(const uint8_t *buf, size_t len, si
|
||||||
TARGET_HASWELL
|
TARGET_HASWELL
|
||||||
namespace simdjson::haswell {
|
namespace simdjson::haswell {
|
||||||
|
|
||||||
static really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
||||||
// There should be no such thing with a processing supporting avx2
|
// There should be no such thing with a processing supporting avx2
|
||||||
// but not clmul.
|
// but not clmul.
|
||||||
uint64_t quote_mask = _mm_cvtsi128_si64(_mm_clmulepi64_si128(
|
uint64_t quote_mask = _mm_cvtsi128_si64(_mm_clmulepi64_si128(
|
||||||
|
@ -1872,7 +1872,7 @@ static really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
||||||
return quote_mask;
|
return quote_mask;
|
||||||
}
|
}
|
||||||
|
|
||||||
static really_inline void find_whitespace_and_structurals(simd_input<ARCHITECTURE> in,
|
really_inline void find_whitespace_and_structurals(simd_input<ARCHITECTURE> in,
|
||||||
uint64_t &whitespace, uint64_t &structurals) {
|
uint64_t &whitespace, uint64_t &structurals) {
|
||||||
|
|
||||||
#ifdef SIMDJSON_NAIVE_STRUCTURAL
|
#ifdef SIMDJSON_NAIVE_STRUCTURAL
|
||||||
|
@ -1938,7 +1938,7 @@ static really_inline void find_whitespace_and_structurals(simd_input<ARCHITECTUR
|
||||||
// base_ptr[base] incrementing base as we go
|
// base_ptr[base] incrementing base as we go
|
||||||
// will potentially store extra values beyond end of valid bits, so base_ptr
|
// will potentially store extra values beyond end of valid bits, so base_ptr
|
||||||
// needs to be large enough to handle this
|
// needs to be large enough to handle this
|
||||||
static really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
||||||
// In some instances, the next branch is expensive because it is mispredicted.
|
// In some instances, the next branch is expensive because it is mispredicted.
|
||||||
// Unfortunately, in other cases,
|
// Unfortunately, in other cases,
|
||||||
// it helps tremendously.
|
// it helps tremendously.
|
||||||
|
@ -2269,12 +2269,12 @@ UNTARGET_REGION
|
||||||
TARGET_WESTMERE
|
TARGET_WESTMERE
|
||||||
namespace simdjson::westmere {
|
namespace simdjson::westmere {
|
||||||
|
|
||||||
static really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
||||||
return _mm_cvtsi128_si64(_mm_clmulepi64_si128(
|
return _mm_cvtsi128_si64(_mm_clmulepi64_si128(
|
||||||
_mm_set_epi64x(0ULL, quote_bits), _mm_set1_epi8(0xFFu), 0));
|
_mm_set_epi64x(0ULL, quote_bits), _mm_set1_epi8(0xFFu), 0));
|
||||||
}
|
}
|
||||||
|
|
||||||
static really_inline void find_whitespace_and_structurals(simd_input<ARCHITECTURE> in,
|
really_inline void find_whitespace_and_structurals(simd_input<ARCHITECTURE> in,
|
||||||
uint64_t &whitespace, uint64_t &structurals) {
|
uint64_t &whitespace, uint64_t &structurals) {
|
||||||
|
|
||||||
const __m128i structural_table =
|
const __m128i structural_table =
|
||||||
|
@ -2306,7 +2306,7 @@ static really_inline void find_whitespace_and_structurals(simd_input<ARCHITECTUR
|
||||||
// This is just a naive implementation. It should be normally
|
// This is just a naive implementation. It should be normally
|
||||||
// disable, but can be used for research purposes to compare
|
// disable, but can be used for research purposes to compare
|
||||||
// again our optimized version.
|
// again our optimized version.
|
||||||
static really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
||||||
uint32_t *out_ptr = base_ptr + base;
|
uint32_t *out_ptr = base_ptr + base;
|
||||||
idx -= 64;
|
idx -= 64;
|
||||||
while (bits != 0) {
|
while (bits != 0) {
|
||||||
|
@ -2324,7 +2324,7 @@ static really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint3
|
||||||
// base_ptr[base] incrementing base as we go
|
// base_ptr[base] incrementing base as we go
|
||||||
// will potentially store extra values beyond end of valid bits, so base_ptr
|
// will potentially store extra values beyond end of valid bits, so base_ptr
|
||||||
// needs to be large enough to handle this
|
// needs to be large enough to handle this
|
||||||
static really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
||||||
// In some instances, the next branch is expensive because it is mispredicted.
|
// In some instances, the next branch is expensive because it is mispredicted.
|
||||||
// Unfortunately, in other cases,
|
// Unfortunately, in other cases,
|
||||||
// it helps tremendously.
|
// it helps tremendously.
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
/* auto-generated on Fri Aug 23 09:54:21 DST 2019. Do not edit! */
|
/* auto-generated on Fri Aug 23 10:23:28 DST 2019. Do not edit! */
|
||||||
/* begin file include/simdjson/simdjson_version.h */
|
/* begin file include/simdjson/simdjson_version.h */
|
||||||
// /include/simdjson/simdjson_version.h automatically generated by release.py,
|
// /include/simdjson/simdjson_version.h automatically generated by release.py,
|
||||||
// do not change by hand
|
// do not change by hand
|
||||||
|
|
|
@ -12,7 +12,7 @@
|
||||||
|
|
||||||
namespace simdjson::arm64 {
|
namespace simdjson::arm64 {
|
||||||
|
|
||||||
static really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
||||||
|
|
||||||
#ifdef __ARM_FEATURE_CRYPTO // some ARM processors lack this extension
|
#ifdef __ARM_FEATURE_CRYPTO // some ARM processors lack this extension
|
||||||
return vmull_p64(-1ULL, quote_bits);
|
return vmull_p64(-1ULL, quote_bits);
|
||||||
|
@ -21,7 +21,7 @@ static really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
static really_inline void find_whitespace_and_structurals(
|
really_inline void find_whitespace_and_structurals(
|
||||||
simd_input<ARCHITECTURE> in, uint64_t &whitespace,
|
simd_input<ARCHITECTURE> in, uint64_t &whitespace,
|
||||||
uint64_t &structurals) {
|
uint64_t &structurals) {
|
||||||
const uint8x16_t low_nibble_mask =
|
const uint8x16_t low_nibble_mask =
|
||||||
|
|
|
@ -8,7 +8,7 @@
|
||||||
// This is just a naive implementation. It should be normally
|
// This is just a naive implementation. It should be normally
|
||||||
// disable, but can be used for research purposes to compare
|
// disable, but can be used for research purposes to compare
|
||||||
// again our optimized version.
|
// again our optimized version.
|
||||||
static really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
||||||
uint32_t *out_ptr = base_ptr + base;
|
uint32_t *out_ptr = base_ptr + base;
|
||||||
idx -= 64;
|
idx -= 64;
|
||||||
while (bits != 0) {
|
while (bits != 0) {
|
||||||
|
@ -26,7 +26,7 @@ static really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint3
|
||||||
// base_ptr[base] incrementing base as we go
|
// base_ptr[base] incrementing base as we go
|
||||||
// will potentially store extra values beyond end of valid bits, so base_ptr
|
// will potentially store extra values beyond end of valid bits, so base_ptr
|
||||||
// needs to be large enough to handle this
|
// needs to be large enough to handle this
|
||||||
static really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
||||||
// In some instances, the next branch is expensive because it is mispredicted.
|
// In some instances, the next branch is expensive because it is mispredicted.
|
||||||
// Unfortunately, in other cases,
|
// Unfortunately, in other cases,
|
||||||
// it helps tremendously.
|
// it helps tremendously.
|
||||||
|
|
|
@ -13,7 +13,7 @@
|
||||||
TARGET_HASWELL
|
TARGET_HASWELL
|
||||||
namespace simdjson::haswell {
|
namespace simdjson::haswell {
|
||||||
|
|
||||||
static really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
||||||
// There should be no such thing with a processing supporting avx2
|
// There should be no such thing with a processing supporting avx2
|
||||||
// but not clmul.
|
// but not clmul.
|
||||||
uint64_t quote_mask = _mm_cvtsi128_si64(_mm_clmulepi64_si128(
|
uint64_t quote_mask = _mm_cvtsi128_si64(_mm_clmulepi64_si128(
|
||||||
|
@ -21,7 +21,7 @@ static really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
||||||
return quote_mask;
|
return quote_mask;
|
||||||
}
|
}
|
||||||
|
|
||||||
static really_inline void find_whitespace_and_structurals(simd_input<ARCHITECTURE> in,
|
really_inline void find_whitespace_and_structurals(simd_input<ARCHITECTURE> in,
|
||||||
uint64_t &whitespace, uint64_t &structurals) {
|
uint64_t &whitespace, uint64_t &structurals) {
|
||||||
|
|
||||||
#ifdef SIMDJSON_NAIVE_STRUCTURAL
|
#ifdef SIMDJSON_NAIVE_STRUCTURAL
|
||||||
|
@ -87,7 +87,7 @@ static really_inline void find_whitespace_and_structurals(simd_input<ARCHITECTUR
|
||||||
// base_ptr[base] incrementing base as we go
|
// base_ptr[base] incrementing base as we go
|
||||||
// will potentially store extra values beyond end of valid bits, so base_ptr
|
// will potentially store extra values beyond end of valid bits, so base_ptr
|
||||||
// needs to be large enough to handle this
|
// needs to be large enough to handle this
|
||||||
static really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
really_inline void flatten_bits(uint32_t *base_ptr, uint32_t &base, uint32_t idx, uint64_t bits) {
|
||||||
// In some instances, the next branch is expensive because it is mispredicted.
|
// In some instances, the next branch is expensive because it is mispredicted.
|
||||||
// Unfortunately, in other cases,
|
// Unfortunately, in other cases,
|
||||||
// it helps tremendously.
|
// it helps tremendously.
|
||||||
|
|
|
@ -13,12 +13,12 @@
|
||||||
TARGET_WESTMERE
|
TARGET_WESTMERE
|
||||||
namespace simdjson::westmere {
|
namespace simdjson::westmere {
|
||||||
|
|
||||||
static really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
really_inline uint64_t compute_quote_mask(uint64_t quote_bits) {
|
||||||
return _mm_cvtsi128_si64(_mm_clmulepi64_si128(
|
return _mm_cvtsi128_si64(_mm_clmulepi64_si128(
|
||||||
_mm_set_epi64x(0ULL, quote_bits), _mm_set1_epi8(0xFFu), 0));
|
_mm_set_epi64x(0ULL, quote_bits), _mm_set1_epi8(0xFFu), 0));
|
||||||
}
|
}
|
||||||
|
|
||||||
static really_inline void find_whitespace_and_structurals(simd_input<ARCHITECTURE> in,
|
really_inline void find_whitespace_and_structurals(simd_input<ARCHITECTURE> in,
|
||||||
uint64_t &whitespace, uint64_t &structurals) {
|
uint64_t &whitespace, uint64_t &structurals) {
|
||||||
|
|
||||||
const __m128i structural_table =
|
const __m128i structural_table =
|
||||||
|
|
Loading…
Reference in New Issue