faiss 0.2.5 → 0.2.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +9 -0
- data/LICENSE.txt +1 -1
- data/ext/faiss/extconf.rb +1 -1
- data/ext/faiss/index.cpp +13 -0
- data/lib/faiss/version.rb +1 -1
- data/lib/faiss.rb +2 -2
- data/vendor/faiss/faiss/AutoTune.cpp +15 -4
- data/vendor/faiss/faiss/AutoTune.h +0 -1
- data/vendor/faiss/faiss/Clustering.cpp +1 -5
- data/vendor/faiss/faiss/Clustering.h +0 -2
- data/vendor/faiss/faiss/IVFlib.h +0 -2
- data/vendor/faiss/faiss/Index.h +1 -2
- data/vendor/faiss/faiss/IndexAdditiveQuantizer.cpp +17 -3
- data/vendor/faiss/faiss/IndexAdditiveQuantizer.h +10 -1
- data/vendor/faiss/faiss/IndexBinary.h +0 -1
- data/vendor/faiss/faiss/IndexBinaryFlat.cpp +2 -1
- data/vendor/faiss/faiss/IndexBinaryFlat.h +4 -0
- data/vendor/faiss/faiss/IndexBinaryHash.cpp +1 -3
- data/vendor/faiss/faiss/IndexBinaryIVF.cpp +273 -48
- data/vendor/faiss/faiss/IndexBinaryIVF.h +18 -11
- data/vendor/faiss/faiss/IndexFastScan.cpp +13 -10
- data/vendor/faiss/faiss/IndexFastScan.h +5 -1
- data/vendor/faiss/faiss/IndexFlat.cpp +16 -3
- data/vendor/faiss/faiss/IndexFlat.h +1 -1
- data/vendor/faiss/faiss/IndexFlatCodes.cpp +5 -0
- data/vendor/faiss/faiss/IndexFlatCodes.h +7 -2
- data/vendor/faiss/faiss/IndexHNSW.cpp +3 -6
- data/vendor/faiss/faiss/IndexHNSW.h +0 -1
- data/vendor/faiss/faiss/IndexIDMap.cpp +4 -4
- data/vendor/faiss/faiss/IndexIDMap.h +0 -2
- data/vendor/faiss/faiss/IndexIVF.cpp +155 -129
- data/vendor/faiss/faiss/IndexIVF.h +121 -61
- data/vendor/faiss/faiss/IndexIVFAdditiveQuantizer.cpp +2 -2
- data/vendor/faiss/faiss/IndexIVFFastScan.cpp +12 -11
- data/vendor/faiss/faiss/IndexIVFFastScan.h +6 -1
- data/vendor/faiss/faiss/IndexIVFPQ.cpp +221 -165
- data/vendor/faiss/faiss/IndexIVFPQ.h +1 -0
- data/vendor/faiss/faiss/IndexIVFPQFastScan.cpp +6 -1
- data/vendor/faiss/faiss/IndexIVFSpectralHash.cpp +0 -2
- data/vendor/faiss/faiss/IndexNNDescent.cpp +1 -2
- data/vendor/faiss/faiss/IndexNNDescent.h +0 -1
- data/vendor/faiss/faiss/IndexNSG.cpp +1 -2
- data/vendor/faiss/faiss/IndexPQ.cpp +7 -9
- data/vendor/faiss/faiss/IndexRefine.cpp +1 -1
- data/vendor/faiss/faiss/IndexReplicas.cpp +3 -4
- data/vendor/faiss/faiss/IndexReplicas.h +0 -1
- data/vendor/faiss/faiss/IndexRowwiseMinMax.cpp +8 -1
- data/vendor/faiss/faiss/IndexRowwiseMinMax.h +7 -0
- data/vendor/faiss/faiss/IndexShards.cpp +26 -109
- data/vendor/faiss/faiss/IndexShards.h +2 -3
- data/vendor/faiss/faiss/IndexShardsIVF.cpp +246 -0
- data/vendor/faiss/faiss/IndexShardsIVF.h +42 -0
- data/vendor/faiss/faiss/MetaIndexes.cpp +86 -0
- data/vendor/faiss/faiss/MetaIndexes.h +29 -0
- data/vendor/faiss/faiss/MetricType.h +14 -0
- data/vendor/faiss/faiss/VectorTransform.cpp +8 -10
- data/vendor/faiss/faiss/VectorTransform.h +1 -3
- data/vendor/faiss/faiss/clone_index.cpp +232 -18
- data/vendor/faiss/faiss/cppcontrib/SaDecodeKernels.h +25 -3
- data/vendor/faiss/faiss/cppcontrib/detail/CoarseBitType.h +7 -0
- data/vendor/faiss/faiss/cppcontrib/detail/UintReader.h +78 -0
- data/vendor/faiss/faiss/cppcontrib/sa_decode/Level2-avx2-inl.h +20 -6
- data/vendor/faiss/faiss/cppcontrib/sa_decode/Level2-inl.h +7 -1
- data/vendor/faiss/faiss/cppcontrib/sa_decode/Level2-neon-inl.h +21 -7
- data/vendor/faiss/faiss/cppcontrib/sa_decode/MinMax-inl.h +7 -0
- data/vendor/faiss/faiss/cppcontrib/sa_decode/MinMaxFP16-inl.h +7 -0
- data/vendor/faiss/faiss/cppcontrib/sa_decode/PQ-avx2-inl.h +10 -3
- data/vendor/faiss/faiss/cppcontrib/sa_decode/PQ-inl.h +7 -1
- data/vendor/faiss/faiss/cppcontrib/sa_decode/PQ-neon-inl.h +11 -3
- data/vendor/faiss/faiss/gpu/GpuAutoTune.cpp +25 -2
- data/vendor/faiss/faiss/gpu/GpuCloner.cpp +76 -29
- data/vendor/faiss/faiss/gpu/GpuCloner.h +2 -2
- data/vendor/faiss/faiss/gpu/GpuClonerOptions.h +14 -13
- data/vendor/faiss/faiss/gpu/GpuDistance.h +18 -6
- data/vendor/faiss/faiss/gpu/GpuIndex.h +23 -21
- data/vendor/faiss/faiss/gpu/GpuIndexBinaryFlat.h +10 -10
- data/vendor/faiss/faiss/gpu/GpuIndexFlat.h +11 -12
- data/vendor/faiss/faiss/gpu/GpuIndexIVF.h +29 -50
- data/vendor/faiss/faiss/gpu/GpuIndexIVFFlat.h +3 -3
- data/vendor/faiss/faiss/gpu/GpuIndexIVFPQ.h +8 -8
- data/vendor/faiss/faiss/gpu/GpuIndexIVFScalarQuantizer.h +4 -4
- data/vendor/faiss/faiss/gpu/impl/IndexUtils.h +2 -5
- data/vendor/faiss/faiss/gpu/impl/RemapIndices.cpp +9 -7
- data/vendor/faiss/faiss/gpu/impl/RemapIndices.h +4 -4
- data/vendor/faiss/faiss/gpu/perf/IndexWrapper-inl.h +2 -2
- data/vendor/faiss/faiss/gpu/perf/IndexWrapper.h +1 -1
- data/vendor/faiss/faiss/gpu/test/TestGpuIndexBinaryFlat.cpp +55 -6
- data/vendor/faiss/faiss/gpu/test/TestGpuIndexFlat.cpp +20 -6
- data/vendor/faiss/faiss/gpu/test/TestGpuIndexIVFFlat.cpp +95 -25
- data/vendor/faiss/faiss/gpu/test/TestGpuIndexIVFPQ.cpp +67 -16
- data/vendor/faiss/faiss/gpu/test/TestGpuIndexIVFScalarQuantizer.cpp +4 -4
- data/vendor/faiss/faiss/gpu/test/TestUtils.cpp +7 -7
- data/vendor/faiss/faiss/gpu/test/TestUtils.h +4 -4
- data/vendor/faiss/faiss/gpu/test/demo_ivfpq_indexing_gpu.cpp +1 -1
- data/vendor/faiss/faiss/gpu/utils/DeviceUtils.h +6 -0
- data/vendor/faiss/faiss/impl/AdditiveQuantizer.cpp +0 -7
- data/vendor/faiss/faiss/impl/AdditiveQuantizer.h +9 -9
- data/vendor/faiss/faiss/impl/AuxIndexStructures.cpp +1 -1
- data/vendor/faiss/faiss/impl/AuxIndexStructures.h +2 -7
- data/vendor/faiss/faiss/impl/CodePacker.cpp +67 -0
- data/vendor/faiss/faiss/impl/CodePacker.h +71 -0
- data/vendor/faiss/faiss/impl/DistanceComputer.h +0 -2
- data/vendor/faiss/faiss/impl/HNSW.cpp +3 -7
- data/vendor/faiss/faiss/impl/HNSW.h +6 -9
- data/vendor/faiss/faiss/impl/IDSelector.cpp +1 -1
- data/vendor/faiss/faiss/impl/IDSelector.h +39 -1
- data/vendor/faiss/faiss/impl/LocalSearchQuantizer.cpp +62 -51
- data/vendor/faiss/faiss/impl/LocalSearchQuantizer.h +11 -12
- data/vendor/faiss/faiss/impl/NNDescent.cpp +3 -9
- data/vendor/faiss/faiss/impl/NNDescent.h +10 -10
- data/vendor/faiss/faiss/impl/NSG.cpp +1 -6
- data/vendor/faiss/faiss/impl/NSG.h +4 -7
- data/vendor/faiss/faiss/impl/PolysemousTraining.cpp +1 -15
- data/vendor/faiss/faiss/impl/PolysemousTraining.h +11 -10
- data/vendor/faiss/faiss/impl/ProductAdditiveQuantizer.cpp +0 -7
- data/vendor/faiss/faiss/impl/ProductQuantizer.cpp +25 -12
- data/vendor/faiss/faiss/impl/ProductQuantizer.h +2 -4
- data/vendor/faiss/faiss/impl/Quantizer.h +6 -3
- data/vendor/faiss/faiss/impl/ResidualQuantizer.cpp +796 -174
- data/vendor/faiss/faiss/impl/ResidualQuantizer.h +16 -8
- data/vendor/faiss/faiss/impl/ScalarQuantizer.cpp +3 -5
- data/vendor/faiss/faiss/impl/ScalarQuantizer.h +4 -4
- data/vendor/faiss/faiss/impl/ThreadedIndex-inl.h +3 -3
- data/vendor/faiss/faiss/impl/ThreadedIndex.h +4 -4
- data/vendor/faiss/faiss/impl/code_distance/code_distance-avx2.h +291 -0
- data/vendor/faiss/faiss/impl/code_distance/code_distance-generic.h +74 -0
- data/vendor/faiss/faiss/impl/code_distance/code_distance.h +123 -0
- data/vendor/faiss/faiss/impl/code_distance/code_distance_avx512.h +102 -0
- data/vendor/faiss/faiss/impl/index_read.cpp +13 -10
- data/vendor/faiss/faiss/impl/index_write.cpp +3 -4
- data/vendor/faiss/faiss/impl/kmeans1d.cpp +0 -1
- data/vendor/faiss/faiss/impl/kmeans1d.h +3 -3
- data/vendor/faiss/faiss/impl/lattice_Zn.cpp +1 -1
- data/vendor/faiss/faiss/impl/platform_macros.h +61 -0
- data/vendor/faiss/faiss/impl/pq4_fast_scan.cpp +48 -4
- data/vendor/faiss/faiss/impl/pq4_fast_scan.h +18 -4
- data/vendor/faiss/faiss/impl/pq4_fast_scan_search_qbs.cpp +2 -2
- data/vendor/faiss/faiss/index_factory.cpp +8 -10
- data/vendor/faiss/faiss/invlists/BlockInvertedLists.cpp +29 -12
- data/vendor/faiss/faiss/invlists/BlockInvertedLists.h +8 -2
- data/vendor/faiss/faiss/invlists/DirectMap.cpp +1 -1
- data/vendor/faiss/faiss/invlists/DirectMap.h +2 -4
- data/vendor/faiss/faiss/invlists/InvertedLists.cpp +118 -18
- data/vendor/faiss/faiss/invlists/InvertedLists.h +44 -4
- data/vendor/faiss/faiss/invlists/OnDiskInvertedLists.cpp +3 -3
- data/vendor/faiss/faiss/invlists/OnDiskInvertedLists.h +1 -1
- data/vendor/faiss/faiss/python/python_callbacks.cpp +1 -1
- data/vendor/faiss/faiss/python/python_callbacks.h +1 -1
- data/vendor/faiss/faiss/utils/AlignedTable.h +3 -1
- data/vendor/faiss/faiss/utils/Heap.cpp +139 -3
- data/vendor/faiss/faiss/utils/Heap.h +35 -1
- data/vendor/faiss/faiss/utils/approx_topk/approx_topk.h +84 -0
- data/vendor/faiss/faiss/utils/approx_topk/avx2-inl.h +196 -0
- data/vendor/faiss/faiss/utils/approx_topk/generic.h +138 -0
- data/vendor/faiss/faiss/utils/approx_topk/mode.h +34 -0
- data/vendor/faiss/faiss/utils/approx_topk_hamming/approx_topk_hamming.h +367 -0
- data/vendor/faiss/faiss/utils/distances.cpp +61 -7
- data/vendor/faiss/faiss/utils/distances.h +11 -0
- data/vendor/faiss/faiss/utils/distances_fused/avx512.cpp +346 -0
- data/vendor/faiss/faiss/utils/distances_fused/avx512.h +36 -0
- data/vendor/faiss/faiss/utils/distances_fused/distances_fused.cpp +42 -0
- data/vendor/faiss/faiss/utils/distances_fused/distances_fused.h +40 -0
- data/vendor/faiss/faiss/utils/distances_fused/simdlib_based.cpp +352 -0
- data/vendor/faiss/faiss/utils/distances_fused/simdlib_based.h +32 -0
- data/vendor/faiss/faiss/utils/distances_simd.cpp +515 -327
- data/vendor/faiss/faiss/utils/extra_distances-inl.h +17 -1
- data/vendor/faiss/faiss/utils/extra_distances.cpp +37 -8
- data/vendor/faiss/faiss/utils/extra_distances.h +2 -1
- data/vendor/faiss/faiss/utils/fp16-fp16c.h +7 -0
- data/vendor/faiss/faiss/utils/fp16-inl.h +7 -0
- data/vendor/faiss/faiss/utils/fp16.h +7 -0
- data/vendor/faiss/faiss/utils/hamming-inl.h +0 -456
- data/vendor/faiss/faiss/utils/hamming.cpp +104 -120
- data/vendor/faiss/faiss/utils/hamming.h +21 -10
- data/vendor/faiss/faiss/utils/hamming_distance/avx2-inl.h +535 -0
- data/vendor/faiss/faiss/utils/hamming_distance/common.h +48 -0
- data/vendor/faiss/faiss/utils/hamming_distance/generic-inl.h +519 -0
- data/vendor/faiss/faiss/utils/hamming_distance/hamdis-inl.h +26 -0
- data/vendor/faiss/faiss/utils/hamming_distance/neon-inl.h +614 -0
- data/vendor/faiss/faiss/utils/partitioning.cpp +21 -25
- data/vendor/faiss/faiss/utils/simdlib_avx2.h +344 -3
- data/vendor/faiss/faiss/utils/simdlib_emulated.h +390 -0
- data/vendor/faiss/faiss/utils/simdlib_neon.h +655 -130
- data/vendor/faiss/faiss/utils/sorting.cpp +692 -0
- data/vendor/faiss/faiss/utils/sorting.h +71 -0
- data/vendor/faiss/faiss/utils/transpose/transpose-avx2-inl.h +165 -0
- data/vendor/faiss/faiss/utils/utils.cpp +4 -176
- data/vendor/faiss/faiss/utils/utils.h +2 -9
- metadata +30 -4
- data/vendor/faiss/faiss/gpu/GpuClonerOptions.cpp +0 -26
@@ -0,0 +1,535 @@
|
|
1
|
+
/**
|
2
|
+
* Copyright (c) Facebook, Inc. and its affiliates.
|
3
|
+
*
|
4
|
+
* This source code is licensed under the MIT license found in the
|
5
|
+
* LICENSE file in the root directory of this source tree.
|
6
|
+
*/
|
7
|
+
|
8
|
+
#ifndef HAMMING_AVX2_INL_H
|
9
|
+
#define HAMMING_AVX2_INL_H
|
10
|
+
|
11
|
+
// AVX2 version
|
12
|
+
|
13
|
+
#include <cassert>
|
14
|
+
#include <cstddef>
|
15
|
+
#include <cstdint>
|
16
|
+
|
17
|
+
#include <faiss/impl/platform_macros.h>
|
18
|
+
|
19
|
+
#include <immintrin.h>
|
20
|
+
|
21
|
+
namespace faiss {
|
22
|
+
|
23
|
+
/* Elementary Hamming distance computation: unoptimized */
|
24
|
+
template <size_t nbits, typename T>
|
25
|
+
inline T hamming(const uint8_t* bs1, const uint8_t* bs2) {
|
26
|
+
const size_t nbytes = nbits / 8;
|
27
|
+
size_t i;
|
28
|
+
T h = 0;
|
29
|
+
for (i = 0; i < nbytes; i++) {
|
30
|
+
h += (T)hamdis_tab_ham_bytes[bs1[i] ^ bs2[i]];
|
31
|
+
}
|
32
|
+
return h;
|
33
|
+
}
|
34
|
+
|
35
|
+
/* Hamming distances for multiples of 64 bits */
|
36
|
+
template <size_t nbits>
|
37
|
+
inline hamdis_t hamming(const uint64_t* bs1, const uint64_t* bs2) {
|
38
|
+
const size_t nwords = nbits / 64;
|
39
|
+
size_t i;
|
40
|
+
hamdis_t h = 0;
|
41
|
+
for (i = 0; i < nwords; i++) {
|
42
|
+
h += popcount64(bs1[i] ^ bs2[i]);
|
43
|
+
}
|
44
|
+
return h;
|
45
|
+
}
|
46
|
+
|
47
|
+
/* specialized (optimized) functions */
|
48
|
+
template <>
|
49
|
+
inline hamdis_t hamming<64>(const uint64_t* pa, const uint64_t* pb) {
|
50
|
+
return popcount64(pa[0] ^ pb[0]);
|
51
|
+
}
|
52
|
+
|
53
|
+
template <>
|
54
|
+
inline hamdis_t hamming<128>(const uint64_t* pa, const uint64_t* pb) {
|
55
|
+
return popcount64(pa[0] ^ pb[0]) + popcount64(pa[1] ^ pb[1]);
|
56
|
+
}
|
57
|
+
|
58
|
+
template <>
|
59
|
+
inline hamdis_t hamming<256>(const uint64_t* pa, const uint64_t* pb) {
|
60
|
+
return popcount64(pa[0] ^ pb[0]) + popcount64(pa[1] ^ pb[1]) +
|
61
|
+
popcount64(pa[2] ^ pb[2]) + popcount64(pa[3] ^ pb[3]);
|
62
|
+
}
|
63
|
+
|
64
|
+
/* Hamming distances for multiple of 64 bits */
|
65
|
+
inline hamdis_t hamming(
|
66
|
+
const uint64_t* bs1,
|
67
|
+
const uint64_t* bs2,
|
68
|
+
size_t nwords) {
|
69
|
+
hamdis_t h = 0;
|
70
|
+
for (size_t i = 0; i < nwords; i++) {
|
71
|
+
h += popcount64(bs1[i] ^ bs2[i]);
|
72
|
+
}
|
73
|
+
return h;
|
74
|
+
}
|
75
|
+
|
76
|
+
/******************************************************************
|
77
|
+
* The HammingComputer series of classes compares a single code of
|
78
|
+
* size 4 to 32 to incoming codes. They are intended for use as a
|
79
|
+
* template class where it would be inefficient to switch on the code
|
80
|
+
* size in the inner loop. Hopefully the compiler will inline the
|
81
|
+
* hamming() functions and put the a0, a1, ... in registers.
|
82
|
+
******************************************************************/
|
83
|
+
|
84
|
+
struct HammingComputer4 {
|
85
|
+
uint32_t a0;
|
86
|
+
|
87
|
+
HammingComputer4() {}
|
88
|
+
|
89
|
+
HammingComputer4(const uint8_t* a, int code_size) {
|
90
|
+
set(a, code_size);
|
91
|
+
}
|
92
|
+
|
93
|
+
void set(const uint8_t* a, int code_size) {
|
94
|
+
assert(code_size == 4);
|
95
|
+
a0 = *(uint32_t*)a;
|
96
|
+
}
|
97
|
+
|
98
|
+
inline int hamming(const uint8_t* b) const {
|
99
|
+
return popcount64(*(uint32_t*)b ^ a0);
|
100
|
+
}
|
101
|
+
|
102
|
+
inline static constexpr int get_code_size() {
|
103
|
+
return 4;
|
104
|
+
}
|
105
|
+
};
|
106
|
+
|
107
|
+
struct HammingComputer8 {
|
108
|
+
uint64_t a0;
|
109
|
+
|
110
|
+
HammingComputer8() {}
|
111
|
+
|
112
|
+
HammingComputer8(const uint8_t* a, int code_size) {
|
113
|
+
set(a, code_size);
|
114
|
+
}
|
115
|
+
|
116
|
+
void set(const uint8_t* a, int code_size) {
|
117
|
+
assert(code_size == 8);
|
118
|
+
a0 = *(uint64_t*)a;
|
119
|
+
}
|
120
|
+
|
121
|
+
inline int hamming(const uint8_t* b) const {
|
122
|
+
return popcount64(*(uint64_t*)b ^ a0);
|
123
|
+
}
|
124
|
+
|
125
|
+
inline static constexpr int get_code_size() {
|
126
|
+
return 8;
|
127
|
+
}
|
128
|
+
};
|
129
|
+
|
130
|
+
struct HammingComputer16 {
|
131
|
+
uint64_t a0, a1;
|
132
|
+
|
133
|
+
HammingComputer16() {}
|
134
|
+
|
135
|
+
HammingComputer16(const uint8_t* a8, int code_size) {
|
136
|
+
set(a8, code_size);
|
137
|
+
}
|
138
|
+
|
139
|
+
void set(const uint8_t* a8, int code_size) {
|
140
|
+
assert(code_size == 16);
|
141
|
+
const uint64_t* a = (uint64_t*)a8;
|
142
|
+
a0 = a[0];
|
143
|
+
a1 = a[1];
|
144
|
+
}
|
145
|
+
|
146
|
+
inline int hamming(const uint8_t* b8) const {
|
147
|
+
const uint64_t* b = (uint64_t*)b8;
|
148
|
+
return popcount64(b[0] ^ a0) + popcount64(b[1] ^ a1);
|
149
|
+
}
|
150
|
+
|
151
|
+
inline static constexpr int get_code_size() {
|
152
|
+
return 16;
|
153
|
+
}
|
154
|
+
};
|
155
|
+
|
156
|
+
// when applied to an array, 1/2 of the 64-bit accesses are unaligned.
|
157
|
+
// This incurs a penalty of ~10% wrt. fully aligned accesses.
|
158
|
+
struct HammingComputer20 {
|
159
|
+
uint64_t a0, a1;
|
160
|
+
uint32_t a2;
|
161
|
+
|
162
|
+
HammingComputer20() {}
|
163
|
+
|
164
|
+
HammingComputer20(const uint8_t* a8, int code_size) {
|
165
|
+
set(a8, code_size);
|
166
|
+
}
|
167
|
+
|
168
|
+
void set(const uint8_t* a8, int code_size) {
|
169
|
+
assert(code_size == 20);
|
170
|
+
const uint64_t* a = (uint64_t*)a8;
|
171
|
+
a0 = a[0];
|
172
|
+
a1 = a[1];
|
173
|
+
a2 = a[2];
|
174
|
+
}
|
175
|
+
|
176
|
+
inline int hamming(const uint8_t* b8) const {
|
177
|
+
const uint64_t* b = (uint64_t*)b8;
|
178
|
+
return popcount64(b[0] ^ a0) + popcount64(b[1] ^ a1) +
|
179
|
+
popcount64(*(uint32_t*)(b + 2) ^ a2);
|
180
|
+
}
|
181
|
+
|
182
|
+
inline static constexpr int get_code_size() {
|
183
|
+
return 20;
|
184
|
+
}
|
185
|
+
};
|
186
|
+
|
187
|
+
struct HammingComputer32 {
|
188
|
+
uint64_t a0, a1, a2, a3;
|
189
|
+
|
190
|
+
HammingComputer32() {}
|
191
|
+
|
192
|
+
HammingComputer32(const uint8_t* a8, int code_size) {
|
193
|
+
set(a8, code_size);
|
194
|
+
}
|
195
|
+
|
196
|
+
void set(const uint8_t* a8, int code_size) {
|
197
|
+
assert(code_size == 32);
|
198
|
+
const uint64_t* a = (uint64_t*)a8;
|
199
|
+
a0 = a[0];
|
200
|
+
a1 = a[1];
|
201
|
+
a2 = a[2];
|
202
|
+
a3 = a[3];
|
203
|
+
}
|
204
|
+
|
205
|
+
inline int hamming(const uint8_t* b8) const {
|
206
|
+
const uint64_t* b = (uint64_t*)b8;
|
207
|
+
return popcount64(b[0] ^ a0) + popcount64(b[1] ^ a1) +
|
208
|
+
popcount64(b[2] ^ a2) + popcount64(b[3] ^ a3);
|
209
|
+
}
|
210
|
+
|
211
|
+
inline static constexpr int get_code_size() {
|
212
|
+
return 32;
|
213
|
+
}
|
214
|
+
};
|
215
|
+
|
216
|
+
struct HammingComputer64 {
|
217
|
+
uint64_t a0, a1, a2, a3, a4, a5, a6, a7;
|
218
|
+
|
219
|
+
HammingComputer64() {}
|
220
|
+
|
221
|
+
HammingComputer64(const uint8_t* a8, int code_size) {
|
222
|
+
set(a8, code_size);
|
223
|
+
}
|
224
|
+
|
225
|
+
void set(const uint8_t* a8, int code_size) {
|
226
|
+
assert(code_size == 64);
|
227
|
+
const uint64_t* a = (uint64_t*)a8;
|
228
|
+
a0 = a[0];
|
229
|
+
a1 = a[1];
|
230
|
+
a2 = a[2];
|
231
|
+
a3 = a[3];
|
232
|
+
a4 = a[4];
|
233
|
+
a5 = a[5];
|
234
|
+
a6 = a[6];
|
235
|
+
a7 = a[7];
|
236
|
+
}
|
237
|
+
|
238
|
+
inline int hamming(const uint8_t* b8) const {
|
239
|
+
const uint64_t* b = (uint64_t*)b8;
|
240
|
+
return popcount64(b[0] ^ a0) + popcount64(b[1] ^ a1) +
|
241
|
+
popcount64(b[2] ^ a2) + popcount64(b[3] ^ a3) +
|
242
|
+
popcount64(b[4] ^ a4) + popcount64(b[5] ^ a5) +
|
243
|
+
popcount64(b[6] ^ a6) + popcount64(b[7] ^ a7);
|
244
|
+
}
|
245
|
+
|
246
|
+
inline static constexpr int get_code_size() {
|
247
|
+
return 64;
|
248
|
+
}
|
249
|
+
};
|
250
|
+
|
251
|
+
struct HammingComputerDefault {
|
252
|
+
const uint8_t* a8;
|
253
|
+
int quotient8;
|
254
|
+
int remainder8;
|
255
|
+
|
256
|
+
HammingComputerDefault() {}
|
257
|
+
|
258
|
+
HammingComputerDefault(const uint8_t* a8, int code_size) {
|
259
|
+
set(a8, code_size);
|
260
|
+
}
|
261
|
+
|
262
|
+
void set(const uint8_t* a8, int code_size) {
|
263
|
+
this->a8 = a8;
|
264
|
+
quotient8 = code_size / 8;
|
265
|
+
remainder8 = code_size % 8;
|
266
|
+
}
|
267
|
+
|
268
|
+
int hamming(const uint8_t* b8) const {
|
269
|
+
int accu = 0;
|
270
|
+
|
271
|
+
const uint64_t* a64 = reinterpret_cast<const uint64_t*>(a8);
|
272
|
+
const uint64_t* b64 = reinterpret_cast<const uint64_t*>(b8);
|
273
|
+
int i = 0, len = quotient8;
|
274
|
+
switch (len & 7) {
|
275
|
+
default:
|
276
|
+
while (len > 7) {
|
277
|
+
len -= 8;
|
278
|
+
accu += popcount64(a64[i] ^ b64[i]);
|
279
|
+
i++;
|
280
|
+
case 7:
|
281
|
+
accu += popcount64(a64[i] ^ b64[i]);
|
282
|
+
i++;
|
283
|
+
case 6:
|
284
|
+
accu += popcount64(a64[i] ^ b64[i]);
|
285
|
+
i++;
|
286
|
+
case 5:
|
287
|
+
accu += popcount64(a64[i] ^ b64[i]);
|
288
|
+
i++;
|
289
|
+
case 4:
|
290
|
+
accu += popcount64(a64[i] ^ b64[i]);
|
291
|
+
i++;
|
292
|
+
case 3:
|
293
|
+
accu += popcount64(a64[i] ^ b64[i]);
|
294
|
+
i++;
|
295
|
+
case 2:
|
296
|
+
accu += popcount64(a64[i] ^ b64[i]);
|
297
|
+
i++;
|
298
|
+
case 1:
|
299
|
+
accu += popcount64(a64[i] ^ b64[i]);
|
300
|
+
i++;
|
301
|
+
}
|
302
|
+
}
|
303
|
+
if (remainder8) {
|
304
|
+
const uint8_t* a = a8 + 8 * quotient8;
|
305
|
+
const uint8_t* b = b8 + 8 * quotient8;
|
306
|
+
switch (remainder8) {
|
307
|
+
case 7:
|
308
|
+
accu += hamdis_tab_ham_bytes[a[6] ^ b[6]];
|
309
|
+
case 6:
|
310
|
+
accu += hamdis_tab_ham_bytes[a[5] ^ b[5]];
|
311
|
+
case 5:
|
312
|
+
accu += hamdis_tab_ham_bytes[a[4] ^ b[4]];
|
313
|
+
case 4:
|
314
|
+
accu += hamdis_tab_ham_bytes[a[3] ^ b[3]];
|
315
|
+
case 3:
|
316
|
+
accu += hamdis_tab_ham_bytes[a[2] ^ b[2]];
|
317
|
+
case 2:
|
318
|
+
accu += hamdis_tab_ham_bytes[a[1] ^ b[1]];
|
319
|
+
case 1:
|
320
|
+
accu += hamdis_tab_ham_bytes[a[0] ^ b[0]];
|
321
|
+
default:
|
322
|
+
break;
|
323
|
+
}
|
324
|
+
}
|
325
|
+
|
326
|
+
return accu;
|
327
|
+
}
|
328
|
+
|
329
|
+
inline int get_code_size() const {
|
330
|
+
return quotient8 * 8 + remainder8;
|
331
|
+
}
|
332
|
+
};
|
333
|
+
|
334
|
+
// more inefficient than HammingComputerDefault (obsolete)
|
335
|
+
struct HammingComputerM8 {
|
336
|
+
const uint64_t* a;
|
337
|
+
int n;
|
338
|
+
|
339
|
+
HammingComputerM8() {}
|
340
|
+
|
341
|
+
HammingComputerM8(const uint8_t* a8, int code_size) {
|
342
|
+
set(a8, code_size);
|
343
|
+
}
|
344
|
+
|
345
|
+
void set(const uint8_t* a8, int code_size) {
|
346
|
+
assert(code_size % 8 == 0);
|
347
|
+
a = (uint64_t*)a8;
|
348
|
+
n = code_size / 8;
|
349
|
+
}
|
350
|
+
|
351
|
+
int hamming(const uint8_t* b8) const {
|
352
|
+
const uint64_t* b = (uint64_t*)b8;
|
353
|
+
int accu = 0;
|
354
|
+
for (int i = 0; i < n; i++)
|
355
|
+
accu += popcount64(a[i] ^ b[i]);
|
356
|
+
return accu;
|
357
|
+
}
|
358
|
+
|
359
|
+
inline int get_code_size() const {
|
360
|
+
return n * 8;
|
361
|
+
}
|
362
|
+
};
|
363
|
+
|
364
|
+
// more inefficient than HammingComputerDefault (obsolete)
|
365
|
+
struct HammingComputerM4 {
|
366
|
+
const uint32_t* a;
|
367
|
+
int n;
|
368
|
+
|
369
|
+
HammingComputerM4() {}
|
370
|
+
|
371
|
+
HammingComputerM4(const uint8_t* a4, int code_size) {
|
372
|
+
set(a4, code_size);
|
373
|
+
}
|
374
|
+
|
375
|
+
void set(const uint8_t* a4, int code_size) {
|
376
|
+
assert(code_size % 4 == 0);
|
377
|
+
a = (uint32_t*)a4;
|
378
|
+
n = code_size / 4;
|
379
|
+
}
|
380
|
+
|
381
|
+
int hamming(const uint8_t* b8) const {
|
382
|
+
const uint32_t* b = (uint32_t*)b8;
|
383
|
+
int accu = 0;
|
384
|
+
for (int i = 0; i < n; i++)
|
385
|
+
accu += popcount64(a[i] ^ b[i]);
|
386
|
+
return accu;
|
387
|
+
}
|
388
|
+
|
389
|
+
inline int get_code_size() const {
|
390
|
+
return n * 4;
|
391
|
+
}
|
392
|
+
};
|
393
|
+
|
394
|
+
/***************************************************************************
|
395
|
+
* Equivalence with a template class when code size is known at compile time
|
396
|
+
**************************************************************************/
|
397
|
+
|
398
|
+
// default template
|
399
|
+
template <int CODE_SIZE>
|
400
|
+
struct HammingComputer : HammingComputerDefault {
|
401
|
+
HammingComputer(const uint8_t* a, int code_size)
|
402
|
+
: HammingComputerDefault(a, code_size) {}
|
403
|
+
};
|
404
|
+
|
405
|
+
#define SPECIALIZED_HC(CODE_SIZE) \
|
406
|
+
template <> \
|
407
|
+
struct HammingComputer<CODE_SIZE> : HammingComputer##CODE_SIZE { \
|
408
|
+
HammingComputer(const uint8_t* a) \
|
409
|
+
: HammingComputer##CODE_SIZE(a, CODE_SIZE) {} \
|
410
|
+
}
|
411
|
+
|
412
|
+
SPECIALIZED_HC(4);
|
413
|
+
SPECIALIZED_HC(8);
|
414
|
+
SPECIALIZED_HC(16);
|
415
|
+
SPECIALIZED_HC(20);
|
416
|
+
SPECIALIZED_HC(32);
|
417
|
+
SPECIALIZED_HC(64);
|
418
|
+
|
419
|
+
#undef SPECIALIZED_HC
|
420
|
+
|
421
|
+
/***************************************************************************
|
422
|
+
* generalized Hamming = number of bytes that are different between
|
423
|
+
* two codes.
|
424
|
+
***************************************************************************/
|
425
|
+
|
426
|
+
inline int generalized_hamming_64(uint64_t a) {
|
427
|
+
a |= a >> 1;
|
428
|
+
a |= a >> 2;
|
429
|
+
a |= a >> 4;
|
430
|
+
a &= 0x0101010101010101UL;
|
431
|
+
return popcount64(a);
|
432
|
+
}
|
433
|
+
|
434
|
+
struct GenHammingComputer8 {
|
435
|
+
uint64_t a0;
|
436
|
+
|
437
|
+
GenHammingComputer8(const uint8_t* a, int code_size) {
|
438
|
+
assert(code_size == 8);
|
439
|
+
a0 = *(uint64_t*)a;
|
440
|
+
}
|
441
|
+
|
442
|
+
inline int hamming(const uint8_t* b) const {
|
443
|
+
return generalized_hamming_64(*(uint64_t*)b ^ a0);
|
444
|
+
}
|
445
|
+
|
446
|
+
inline static constexpr int get_code_size() {
|
447
|
+
return 8;
|
448
|
+
}
|
449
|
+
};
|
450
|
+
|
451
|
+
// I'm not sure whether this version is faster of slower, tbh
|
452
|
+
// todo: test on different CPUs
|
453
|
+
struct GenHammingComputer16 {
|
454
|
+
__m128i a;
|
455
|
+
|
456
|
+
GenHammingComputer16(const uint8_t* a8, int code_size) {
|
457
|
+
assert(code_size == 16);
|
458
|
+
a = _mm_loadu_si128((const __m128i_u*)a8);
|
459
|
+
}
|
460
|
+
|
461
|
+
inline int hamming(const uint8_t* b8) const {
|
462
|
+
const __m128i b = _mm_loadu_si128((const __m128i_u*)b8);
|
463
|
+
const __m128i cmp = _mm_cmpeq_epi8(a, b);
|
464
|
+
const auto movemask = _mm_movemask_epi8(cmp);
|
465
|
+
return 16 - popcount32(movemask);
|
466
|
+
}
|
467
|
+
|
468
|
+
inline static constexpr int get_code_size() {
|
469
|
+
return 16;
|
470
|
+
}
|
471
|
+
};
|
472
|
+
|
473
|
+
struct GenHammingComputer32 {
|
474
|
+
__m256i a;
|
475
|
+
|
476
|
+
GenHammingComputer32(const uint8_t* a8, int code_size) {
|
477
|
+
assert(code_size == 32);
|
478
|
+
a = _mm256_loadu_si256((const __m256i_u*)a8);
|
479
|
+
}
|
480
|
+
|
481
|
+
inline int hamming(const uint8_t* b8) const {
|
482
|
+
const __m256i b = _mm256_loadu_si256((const __m256i_u*)b8);
|
483
|
+
const __m256i cmp = _mm256_cmpeq_epi8(a, b);
|
484
|
+
const uint32_t movemask = _mm256_movemask_epi8(cmp);
|
485
|
+
return 32 - popcount32(movemask);
|
486
|
+
}
|
487
|
+
|
488
|
+
inline static constexpr int get_code_size() {
|
489
|
+
return 32;
|
490
|
+
}
|
491
|
+
};
|
492
|
+
|
493
|
+
// A specialized version might be needed for the very long
|
494
|
+
// GenHamming code_size. In such a case, one may accumulate
|
495
|
+
// counts using _mm256_sub_epi8 and then compute a horizontal
|
496
|
+
// sum (using _mm256_sad_epu8, maybe, in blocks of no larger
|
497
|
+
// than 256 * 32 bytes).
|
498
|
+
|
499
|
+
struct GenHammingComputerM8 {
|
500
|
+
const uint64_t* a;
|
501
|
+
int n;
|
502
|
+
|
503
|
+
GenHammingComputerM8(const uint8_t* a8, int code_size) {
|
504
|
+
assert(code_size % 8 == 0);
|
505
|
+
a = (uint64_t*)a8;
|
506
|
+
n = code_size / 8;
|
507
|
+
}
|
508
|
+
|
509
|
+
int hamming(const uint8_t* b8) const {
|
510
|
+
const uint64_t* b = (uint64_t*)b8;
|
511
|
+
int accu = 0;
|
512
|
+
|
513
|
+
int i = 0;
|
514
|
+
int n4 = (n / 4) * 4;
|
515
|
+
for (; i < n4; i += 4) {
|
516
|
+
const __m256i av = _mm256_loadu_si256((const __m256i_u*)(a + i));
|
517
|
+
const __m256i bv = _mm256_loadu_si256((const __m256i_u*)(b + i));
|
518
|
+
const __m256i cmp = _mm256_cmpeq_epi8(av, bv);
|
519
|
+
const uint32_t movemask = _mm256_movemask_epi8(cmp);
|
520
|
+
accu += 32 - popcount32(movemask);
|
521
|
+
}
|
522
|
+
|
523
|
+
for (; i < n; i++)
|
524
|
+
accu += generalized_hamming_64(a[i] ^ b[i]);
|
525
|
+
return accu;
|
526
|
+
}
|
527
|
+
|
528
|
+
inline int get_code_size() const {
|
529
|
+
return n * 8;
|
530
|
+
}
|
531
|
+
};
|
532
|
+
|
533
|
+
} // namespace faiss
|
534
|
+
|
535
|
+
#endif
|
@@ -0,0 +1,48 @@
|
|
1
|
+
/**
|
2
|
+
* Copyright (c) Facebook, Inc. and its affiliates.
|
3
|
+
*
|
4
|
+
* This source code is licensed under the MIT license found in the
|
5
|
+
* LICENSE file in the root directory of this source tree.
|
6
|
+
*/
|
7
|
+
|
8
|
+
#ifndef FAISS_hamming_common_h
|
9
|
+
#define FAISS_hamming_common_h
|
10
|
+
|
11
|
+
#include <cstdint>
|
12
|
+
|
13
|
+
#include <faiss/impl/platform_macros.h>
|
14
|
+
|
15
|
+
/* The Hamming distance type */
|
16
|
+
using hamdis_t = int32_t;
|
17
|
+
|
18
|
+
namespace faiss {
|
19
|
+
|
20
|
+
inline int popcount32(uint32_t x) {
|
21
|
+
return __builtin_popcount(x);
|
22
|
+
}
|
23
|
+
|
24
|
+
// popcount
|
25
|
+
inline int popcount64(uint64_t x) {
|
26
|
+
return __builtin_popcountl(x);
|
27
|
+
}
|
28
|
+
|
29
|
+
// This table was moved from .cpp to .h file, because
|
30
|
+
// otherwise it was causing compilation errors while trying to
|
31
|
+
// compile swig modules on Windows.
|
32
|
+
// todo for C++17: switch to 'inline constexpr'
|
33
|
+
static constexpr uint8_t hamdis_tab_ham_bytes[256] = {
|
34
|
+
0, 1, 1, 2, 1, 2, 2, 3, 1, 2, 2, 3, 2, 3, 3, 4, 1, 2, 2, 3, 2, 3, 3, 4,
|
35
|
+
2, 3, 3, 4, 3, 4, 4, 5, 1, 2, 2, 3, 2, 3, 3, 4, 2, 3, 3, 4, 3, 4, 4, 5,
|
36
|
+
2, 3, 3, 4, 3, 4, 4, 5, 3, 4, 4, 5, 4, 5, 5, 6, 1, 2, 2, 3, 2, 3, 3, 4,
|
37
|
+
2, 3, 3, 4, 3, 4, 4, 5, 2, 3, 3, 4, 3, 4, 4, 5, 3, 4, 4, 5, 4, 5, 5, 6,
|
38
|
+
2, 3, 3, 4, 3, 4, 4, 5, 3, 4, 4, 5, 4, 5, 5, 6, 3, 4, 4, 5, 4, 5, 5, 6,
|
39
|
+
4, 5, 5, 6, 5, 6, 6, 7, 1, 2, 2, 3, 2, 3, 3, 4, 2, 3, 3, 4, 3, 4, 4, 5,
|
40
|
+
2, 3, 3, 4, 3, 4, 4, 5, 3, 4, 4, 5, 4, 5, 5, 6, 2, 3, 3, 4, 3, 4, 4, 5,
|
41
|
+
3, 4, 4, 5, 4, 5, 5, 6, 3, 4, 4, 5, 4, 5, 5, 6, 4, 5, 5, 6, 5, 6, 6, 7,
|
42
|
+
2, 3, 3, 4, 3, 4, 4, 5, 3, 4, 4, 5, 4, 5, 5, 6, 3, 4, 4, 5, 4, 5, 5, 6,
|
43
|
+
4, 5, 5, 6, 5, 6, 6, 7, 3, 4, 4, 5, 4, 5, 5, 6, 4, 5, 5, 6, 5, 6, 6, 7,
|
44
|
+
4, 5, 5, 6, 5, 6, 6, 7, 5, 6, 6, 7, 6, 7, 7, 8};
|
45
|
+
|
46
|
+
} // namespace faiss
|
47
|
+
|
48
|
+
#endif
|