faiss 0.2.0 → 0.2.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +4 -0
- data/lib/faiss/version.rb +1 -1
- data/vendor/faiss/faiss/AutoTune.cpp +292 -291
- data/vendor/faiss/faiss/AutoTune.h +55 -56
- data/vendor/faiss/faiss/Clustering.cpp +334 -195
- data/vendor/faiss/faiss/Clustering.h +88 -35
- data/vendor/faiss/faiss/IVFlib.cpp +171 -195
- data/vendor/faiss/faiss/IVFlib.h +48 -51
- data/vendor/faiss/faiss/Index.cpp +85 -103
- data/vendor/faiss/faiss/Index.h +54 -48
- data/vendor/faiss/faiss/Index2Layer.cpp +139 -164
- data/vendor/faiss/faiss/Index2Layer.h +22 -22
- data/vendor/faiss/faiss/IndexBinary.cpp +45 -37
- data/vendor/faiss/faiss/IndexBinary.h +140 -132
- data/vendor/faiss/faiss/IndexBinaryFlat.cpp +73 -53
- data/vendor/faiss/faiss/IndexBinaryFlat.h +29 -24
- data/vendor/faiss/faiss/IndexBinaryFromFloat.cpp +46 -43
- data/vendor/faiss/faiss/IndexBinaryFromFloat.h +16 -15
- data/vendor/faiss/faiss/IndexBinaryHNSW.cpp +215 -232
- data/vendor/faiss/faiss/IndexBinaryHNSW.h +25 -24
- data/vendor/faiss/faiss/IndexBinaryHash.cpp +182 -177
- data/vendor/faiss/faiss/IndexBinaryHash.h +41 -34
- data/vendor/faiss/faiss/IndexBinaryIVF.cpp +489 -461
- data/vendor/faiss/faiss/IndexBinaryIVF.h +97 -68
- data/vendor/faiss/faiss/IndexFlat.cpp +116 -147
- data/vendor/faiss/faiss/IndexFlat.h +35 -46
- data/vendor/faiss/faiss/IndexHNSW.cpp +372 -348
- data/vendor/faiss/faiss/IndexHNSW.h +57 -41
- data/vendor/faiss/faiss/IndexIVF.cpp +474 -454
- data/vendor/faiss/faiss/IndexIVF.h +146 -113
- data/vendor/faiss/faiss/IndexIVFFlat.cpp +248 -250
- data/vendor/faiss/faiss/IndexIVFFlat.h +48 -51
- data/vendor/faiss/faiss/IndexIVFPQ.cpp +457 -516
- data/vendor/faiss/faiss/IndexIVFPQ.h +74 -66
- data/vendor/faiss/faiss/IndexIVFPQFastScan.cpp +406 -372
- data/vendor/faiss/faiss/IndexIVFPQFastScan.h +82 -57
- data/vendor/faiss/faiss/IndexIVFPQR.cpp +104 -102
- data/vendor/faiss/faiss/IndexIVFPQR.h +33 -28
- data/vendor/faiss/faiss/IndexIVFSpectralHash.cpp +125 -133
- data/vendor/faiss/faiss/IndexIVFSpectralHash.h +19 -21
- data/vendor/faiss/faiss/IndexLSH.cpp +75 -96
- data/vendor/faiss/faiss/IndexLSH.h +21 -26
- data/vendor/faiss/faiss/IndexLattice.cpp +42 -56
- data/vendor/faiss/faiss/IndexLattice.h +11 -16
- data/vendor/faiss/faiss/IndexNNDescent.cpp +231 -0
- data/vendor/faiss/faiss/IndexNNDescent.h +72 -0
- data/vendor/faiss/faiss/IndexNSG.cpp +303 -0
- data/vendor/faiss/faiss/IndexNSG.h +85 -0
- data/vendor/faiss/faiss/IndexPQ.cpp +405 -464
- data/vendor/faiss/faiss/IndexPQ.h +64 -67
- data/vendor/faiss/faiss/IndexPQFastScan.cpp +143 -170
- data/vendor/faiss/faiss/IndexPQFastScan.h +46 -32
- data/vendor/faiss/faiss/IndexPreTransform.cpp +120 -150
- data/vendor/faiss/faiss/IndexPreTransform.h +33 -36
- data/vendor/faiss/faiss/IndexRefine.cpp +115 -131
- data/vendor/faiss/faiss/IndexRefine.h +22 -23
- data/vendor/faiss/faiss/IndexReplicas.cpp +147 -153
- data/vendor/faiss/faiss/IndexReplicas.h +62 -56
- data/vendor/faiss/faiss/IndexResidual.cpp +291 -0
- data/vendor/faiss/faiss/IndexResidual.h +152 -0
- data/vendor/faiss/faiss/IndexScalarQuantizer.cpp +120 -155
- data/vendor/faiss/faiss/IndexScalarQuantizer.h +41 -45
- data/vendor/faiss/faiss/IndexShards.cpp +256 -240
- data/vendor/faiss/faiss/IndexShards.h +85 -73
- data/vendor/faiss/faiss/MatrixStats.cpp +112 -97
- data/vendor/faiss/faiss/MatrixStats.h +7 -10
- data/vendor/faiss/faiss/MetaIndexes.cpp +135 -157
- data/vendor/faiss/faiss/MetaIndexes.h +40 -34
- data/vendor/faiss/faiss/MetricType.h +7 -7
- data/vendor/faiss/faiss/VectorTransform.cpp +652 -474
- data/vendor/faiss/faiss/VectorTransform.h +61 -89
- data/vendor/faiss/faiss/clone_index.cpp +77 -73
- data/vendor/faiss/faiss/clone_index.h +4 -9
- data/vendor/faiss/faiss/gpu/GpuAutoTune.cpp +33 -38
- data/vendor/faiss/faiss/gpu/GpuAutoTune.h +11 -9
- data/vendor/faiss/faiss/gpu/GpuCloner.cpp +197 -170
- data/vendor/faiss/faiss/gpu/GpuCloner.h +53 -35
- data/vendor/faiss/faiss/gpu/GpuClonerOptions.cpp +12 -14
- data/vendor/faiss/faiss/gpu/GpuClonerOptions.h +27 -25
- data/vendor/faiss/faiss/gpu/GpuDistance.h +116 -112
- data/vendor/faiss/faiss/gpu/GpuFaissAssert.h +1 -2
- data/vendor/faiss/faiss/gpu/GpuIndex.h +134 -137
- data/vendor/faiss/faiss/gpu/GpuIndexBinaryFlat.h +76 -73
- data/vendor/faiss/faiss/gpu/GpuIndexFlat.h +173 -162
- data/vendor/faiss/faiss/gpu/GpuIndexIVF.h +67 -64
- data/vendor/faiss/faiss/gpu/GpuIndexIVFFlat.h +89 -86
- data/vendor/faiss/faiss/gpu/GpuIndexIVFPQ.h +150 -141
- data/vendor/faiss/faiss/gpu/GpuIndexIVFScalarQuantizer.h +101 -103
- data/vendor/faiss/faiss/gpu/GpuIndicesOptions.h +17 -16
- data/vendor/faiss/faiss/gpu/GpuResources.cpp +116 -128
- data/vendor/faiss/faiss/gpu/GpuResources.h +182 -186
- data/vendor/faiss/faiss/gpu/StandardGpuResources.cpp +433 -422
- data/vendor/faiss/faiss/gpu/StandardGpuResources.h +131 -130
- data/vendor/faiss/faiss/gpu/impl/InterleavedCodes.cpp +468 -456
- data/vendor/faiss/faiss/gpu/impl/InterleavedCodes.h +25 -19
- data/vendor/faiss/faiss/gpu/impl/RemapIndices.cpp +22 -20
- data/vendor/faiss/faiss/gpu/impl/RemapIndices.h +9 -8
- data/vendor/faiss/faiss/gpu/perf/IndexWrapper-inl.h +39 -44
- data/vendor/faiss/faiss/gpu/perf/IndexWrapper.h +16 -14
- data/vendor/faiss/faiss/gpu/perf/PerfClustering.cpp +77 -71
- data/vendor/faiss/faiss/gpu/perf/PerfIVFPQAdd.cpp +109 -88
- data/vendor/faiss/faiss/gpu/perf/WriteIndex.cpp +75 -64
- data/vendor/faiss/faiss/gpu/test/TestCodePacking.cpp +230 -215
- data/vendor/faiss/faiss/gpu/test/TestGpuIndexBinaryFlat.cpp +80 -86
- data/vendor/faiss/faiss/gpu/test/TestGpuIndexFlat.cpp +284 -277
- data/vendor/faiss/faiss/gpu/test/TestGpuIndexIVFFlat.cpp +416 -416
- data/vendor/faiss/faiss/gpu/test/TestGpuIndexIVFPQ.cpp +611 -517
- data/vendor/faiss/faiss/gpu/test/TestGpuIndexIVFScalarQuantizer.cpp +166 -164
- data/vendor/faiss/faiss/gpu/test/TestGpuMemoryException.cpp +61 -53
- data/vendor/faiss/faiss/gpu/test/TestUtils.cpp +274 -238
- data/vendor/faiss/faiss/gpu/test/TestUtils.h +73 -57
- data/vendor/faiss/faiss/gpu/test/demo_ivfpq_indexing_gpu.cpp +47 -50
- data/vendor/faiss/faiss/gpu/utils/DeviceUtils.h +79 -72
- data/vendor/faiss/faiss/gpu/utils/StackDeviceMemory.cpp +140 -146
- data/vendor/faiss/faiss/gpu/utils/StackDeviceMemory.h +69 -71
- data/vendor/faiss/faiss/gpu/utils/StaticUtils.h +21 -16
- data/vendor/faiss/faiss/gpu/utils/Timer.cpp +25 -29
- data/vendor/faiss/faiss/gpu/utils/Timer.h +30 -29
- data/vendor/faiss/faiss/impl/AdditiveQuantizer.cpp +270 -0
- data/vendor/faiss/faiss/impl/AdditiveQuantizer.h +115 -0
- data/vendor/faiss/faiss/impl/AuxIndexStructures.cpp +90 -120
- data/vendor/faiss/faiss/impl/AuxIndexStructures.h +81 -65
- data/vendor/faiss/faiss/impl/FaissAssert.h +73 -58
- data/vendor/faiss/faiss/impl/FaissException.cpp +56 -48
- data/vendor/faiss/faiss/impl/FaissException.h +41 -29
- data/vendor/faiss/faiss/impl/HNSW.cpp +595 -611
- data/vendor/faiss/faiss/impl/HNSW.h +179 -200
- data/vendor/faiss/faiss/impl/LocalSearchQuantizer.cpp +672 -0
- data/vendor/faiss/faiss/impl/LocalSearchQuantizer.h +172 -0
- data/vendor/faiss/faiss/impl/NNDescent.cpp +487 -0
- data/vendor/faiss/faiss/impl/NNDescent.h +154 -0
- data/vendor/faiss/faiss/impl/NSG.cpp +682 -0
- data/vendor/faiss/faiss/impl/NSG.h +199 -0
- data/vendor/faiss/faiss/impl/PolysemousTraining.cpp +484 -454
- data/vendor/faiss/faiss/impl/PolysemousTraining.h +52 -55
- data/vendor/faiss/faiss/impl/ProductQuantizer-inl.h +26 -47
- data/vendor/faiss/faiss/impl/ProductQuantizer.cpp +469 -459
- data/vendor/faiss/faiss/impl/ProductQuantizer.h +76 -87
- data/vendor/faiss/faiss/impl/ResidualQuantizer.cpp +448 -0
- data/vendor/faiss/faiss/impl/ResidualQuantizer.h +130 -0
- data/vendor/faiss/faiss/impl/ResultHandler.h +96 -132
- data/vendor/faiss/faiss/impl/ScalarQuantizer.cpp +648 -701
- data/vendor/faiss/faiss/impl/ScalarQuantizer.h +48 -46
- data/vendor/faiss/faiss/impl/ThreadedIndex-inl.h +129 -131
- data/vendor/faiss/faiss/impl/ThreadedIndex.h +61 -55
- data/vendor/faiss/faiss/impl/index_read.cpp +547 -479
- data/vendor/faiss/faiss/impl/index_write.cpp +497 -407
- data/vendor/faiss/faiss/impl/io.cpp +75 -94
- data/vendor/faiss/faiss/impl/io.h +31 -41
- data/vendor/faiss/faiss/impl/io_macros.h +40 -29
- data/vendor/faiss/faiss/impl/lattice_Zn.cpp +137 -186
- data/vendor/faiss/faiss/impl/lattice_Zn.h +40 -51
- data/vendor/faiss/faiss/impl/platform_macros.h +29 -8
- data/vendor/faiss/faiss/impl/pq4_fast_scan.cpp +77 -124
- data/vendor/faiss/faiss/impl/pq4_fast_scan.h +39 -48
- data/vendor/faiss/faiss/impl/pq4_fast_scan_search_1.cpp +41 -52
- data/vendor/faiss/faiss/impl/pq4_fast_scan_search_qbs.cpp +80 -117
- data/vendor/faiss/faiss/impl/simd_result_handlers.h +109 -137
- data/vendor/faiss/faiss/index_factory.cpp +269 -218
- data/vendor/faiss/faiss/index_factory.h +6 -7
- data/vendor/faiss/faiss/index_io.h +23 -26
- data/vendor/faiss/faiss/invlists/BlockInvertedLists.cpp +67 -75
- data/vendor/faiss/faiss/invlists/BlockInvertedLists.h +22 -24
- data/vendor/faiss/faiss/invlists/DirectMap.cpp +96 -112
- data/vendor/faiss/faiss/invlists/DirectMap.h +29 -33
- data/vendor/faiss/faiss/invlists/InvertedLists.cpp +307 -364
- data/vendor/faiss/faiss/invlists/InvertedLists.h +151 -151
- data/vendor/faiss/faiss/invlists/InvertedListsIOHook.cpp +29 -34
- data/vendor/faiss/faiss/invlists/InvertedListsIOHook.h +17 -18
- data/vendor/faiss/faiss/invlists/OnDiskInvertedLists.cpp +257 -293
- data/vendor/faiss/faiss/invlists/OnDiskInvertedLists.h +50 -45
- data/vendor/faiss/faiss/python/python_callbacks.cpp +23 -26
- data/vendor/faiss/faiss/python/python_callbacks.h +9 -16
- data/vendor/faiss/faiss/utils/AlignedTable.h +79 -44
- data/vendor/faiss/faiss/utils/Heap.cpp +40 -48
- data/vendor/faiss/faiss/utils/Heap.h +186 -209
- data/vendor/faiss/faiss/utils/WorkerThread.cpp +67 -76
- data/vendor/faiss/faiss/utils/WorkerThread.h +32 -33
- data/vendor/faiss/faiss/utils/distances.cpp +301 -310
- data/vendor/faiss/faiss/utils/distances.h +133 -118
- data/vendor/faiss/faiss/utils/distances_simd.cpp +456 -516
- data/vendor/faiss/faiss/utils/extra_distances-inl.h +117 -0
- data/vendor/faiss/faiss/utils/extra_distances.cpp +113 -232
- data/vendor/faiss/faiss/utils/extra_distances.h +30 -29
- data/vendor/faiss/faiss/utils/hamming-inl.h +260 -209
- data/vendor/faiss/faiss/utils/hamming.cpp +375 -469
- data/vendor/faiss/faiss/utils/hamming.h +62 -85
- data/vendor/faiss/faiss/utils/ordered_key_value.h +16 -18
- data/vendor/faiss/faiss/utils/partitioning.cpp +393 -318
- data/vendor/faiss/faiss/utils/partitioning.h +26 -21
- data/vendor/faiss/faiss/utils/quantize_lut.cpp +78 -66
- data/vendor/faiss/faiss/utils/quantize_lut.h +22 -20
- data/vendor/faiss/faiss/utils/random.cpp +39 -63
- data/vendor/faiss/faiss/utils/random.h +13 -16
- data/vendor/faiss/faiss/utils/simdlib.h +4 -2
- data/vendor/faiss/faiss/utils/simdlib_avx2.h +88 -85
- data/vendor/faiss/faiss/utils/simdlib_emulated.h +226 -165
- data/vendor/faiss/faiss/utils/simdlib_neon.h +832 -0
- data/vendor/faiss/faiss/utils/utils.cpp +304 -287
- data/vendor/faiss/faiss/utils/utils.h +53 -48
- metadata +20 -2
|
@@ -10,30 +10,27 @@
|
|
|
10
10
|
#ifndef FAISS_INDEX_IVF_H
|
|
11
11
|
#define FAISS_INDEX_IVF_H
|
|
12
12
|
|
|
13
|
-
|
|
14
|
-
#include <vector>
|
|
15
|
-
#include <unordered_map>
|
|
16
13
|
#include <stdint.h>
|
|
14
|
+
#include <unordered_map>
|
|
15
|
+
#include <vector>
|
|
17
16
|
|
|
18
|
-
#include <faiss/Index.h>
|
|
19
|
-
#include <faiss/invlists/InvertedLists.h>
|
|
20
|
-
#include <faiss/invlists/DirectMap.h>
|
|
21
17
|
#include <faiss/Clustering.h>
|
|
18
|
+
#include <faiss/Index.h>
|
|
22
19
|
#include <faiss/impl/platform_macros.h>
|
|
20
|
+
#include <faiss/invlists/DirectMap.h>
|
|
21
|
+
#include <faiss/invlists/InvertedLists.h>
|
|
23
22
|
#include <faiss/utils/Heap.h>
|
|
24
23
|
|
|
25
|
-
|
|
26
24
|
namespace faiss {
|
|
27
25
|
|
|
28
|
-
|
|
29
26
|
/** Encapsulates a quantizer object for the IndexIVF
|
|
30
27
|
*
|
|
31
28
|
* The class isolates the fields that are independent of the storage
|
|
32
29
|
* of the lists (especially training)
|
|
33
30
|
*/
|
|
34
31
|
struct Level1Quantizer {
|
|
35
|
-
Index
|
|
36
|
-
size_t nlist;
|
|
32
|
+
Index* quantizer; ///< quantizer that maps vectors to inverted lists
|
|
33
|
+
size_t nlist; ///< number of possible key values
|
|
37
34
|
|
|
38
35
|
/**
|
|
39
36
|
* = 0: use the quantizer as index in a kmeans training
|
|
@@ -41,40 +38,37 @@ struct Level1Quantizer {
|
|
|
41
38
|
* = 2: kmeans training on a flat index + add the centroids to the quantizer
|
|
42
39
|
*/
|
|
43
40
|
char quantizer_trains_alone;
|
|
44
|
-
bool own_fields;
|
|
41
|
+
bool own_fields; ///< whether object owns the quantizer
|
|
45
42
|
|
|
46
43
|
ClusteringParameters cp; ///< to override default clustering params
|
|
47
|
-
Index
|
|
44
|
+
Index* clustering_index; ///< to override index used during clustering
|
|
48
45
|
|
|
49
46
|
/// Trains the quantizer and calls train_residual to train sub-quantizers
|
|
50
|
-
void train_q1
|
|
51
|
-
|
|
52
|
-
|
|
47
|
+
void train_q1(
|
|
48
|
+
size_t n,
|
|
49
|
+
const float* x,
|
|
50
|
+
bool verbose,
|
|
51
|
+
MetricType metric_type);
|
|
53
52
|
|
|
54
53
|
/// compute the number of bytes required to store list ids
|
|
55
|
-
size_t coarse_code_size
|
|
56
|
-
void encode_listno
|
|
57
|
-
Index::idx_t decode_listno
|
|
54
|
+
size_t coarse_code_size() const;
|
|
55
|
+
void encode_listno(Index::idx_t list_no, uint8_t* code) const;
|
|
56
|
+
Index::idx_t decode_listno(const uint8_t* code) const;
|
|
58
57
|
|
|
59
|
-
Level1Quantizer
|
|
58
|
+
Level1Quantizer(Index* quantizer, size_t nlist);
|
|
60
59
|
|
|
61
|
-
Level1Quantizer
|
|
62
|
-
|
|
63
|
-
~Level1Quantizer ();
|
|
60
|
+
Level1Quantizer();
|
|
64
61
|
|
|
62
|
+
~Level1Quantizer();
|
|
65
63
|
};
|
|
66
64
|
|
|
67
|
-
|
|
68
|
-
|
|
69
65
|
struct IVFSearchParameters {
|
|
70
|
-
size_t nprobe;
|
|
71
|
-
size_t max_codes;
|
|
72
|
-
IVFSearchParameters(): nprobe(1), max_codes(0) {}
|
|
73
|
-
virtual ~IVFSearchParameters
|
|
66
|
+
size_t nprobe; ///< number of probes at query time
|
|
67
|
+
size_t max_codes; ///< max nb of codes to visit to do a query
|
|
68
|
+
IVFSearchParameters() : nprobe(1), max_codes(0) {}
|
|
69
|
+
virtual ~IVFSearchParameters() {}
|
|
74
70
|
};
|
|
75
71
|
|
|
76
|
-
|
|
77
|
-
|
|
78
72
|
struct InvertedListScanner;
|
|
79
73
|
struct IndexIVFStats;
|
|
80
74
|
|
|
@@ -98,15 +92,15 @@ struct IndexIVFStats;
|
|
|
98
92
|
* Sub-classes implement a post-filtering of the index that refines
|
|
99
93
|
* the distance estimation from the query to databse vectors.
|
|
100
94
|
*/
|
|
101
|
-
struct IndexIVF: Index, Level1Quantizer {
|
|
95
|
+
struct IndexIVF : Index, Level1Quantizer {
|
|
102
96
|
/// Access to the actual data
|
|
103
|
-
InvertedLists
|
|
97
|
+
InvertedLists* invlists;
|
|
104
98
|
bool own_invlists;
|
|
105
99
|
|
|
106
|
-
size_t code_size;
|
|
100
|
+
size_t code_size; ///< code size per vector in bytes
|
|
107
101
|
|
|
108
|
-
size_t nprobe;
|
|
109
|
-
size_t max_codes;
|
|
102
|
+
size_t nprobe; ///< number of probes at query time
|
|
103
|
+
size_t max_codes; ///< max nb of codes to visit to do a query
|
|
110
104
|
|
|
111
105
|
/** Parallel mode determines how queries are parallelized with OpenMP
|
|
112
106
|
*
|
|
@@ -130,9 +124,12 @@ struct IndexIVF: Index, Level1Quantizer {
|
|
|
130
124
|
* identifier. The pointer is borrowed: the quantizer should not
|
|
131
125
|
* be deleted while the IndexIVF is in use.
|
|
132
126
|
*/
|
|
133
|
-
IndexIVF
|
|
134
|
-
|
|
135
|
-
|
|
127
|
+
IndexIVF(
|
|
128
|
+
Index* quantizer,
|
|
129
|
+
size_t d,
|
|
130
|
+
size_t nlist,
|
|
131
|
+
size_t code_size,
|
|
132
|
+
MetricType metric = METRIC_L2);
|
|
136
133
|
|
|
137
134
|
void reset() override;
|
|
138
135
|
|
|
@@ -145,6 +142,19 @@ struct IndexIVF: Index, Level1Quantizer {
|
|
|
145
142
|
/// default implementation that calls encode_vectors
|
|
146
143
|
void add_with_ids(idx_t n, const float* x, const idx_t* xids) override;
|
|
147
144
|
|
|
145
|
+
/** Implementation of vector addition where the vector assignments are
|
|
146
|
+
* predefined. The default implementation hands over the code extraction to
|
|
147
|
+
* encode_vectors.
|
|
148
|
+
*
|
|
149
|
+
* @param precomputed_idx quantization indices for the input vectors
|
|
150
|
+
* (size n)
|
|
151
|
+
*/
|
|
152
|
+
virtual void add_core(
|
|
153
|
+
idx_t n,
|
|
154
|
+
const float* x,
|
|
155
|
+
const idx_t* xids,
|
|
156
|
+
const idx_t* precomputed_idx);
|
|
157
|
+
|
|
148
158
|
/** Encodes a set of vectors as they would appear in the inverted lists
|
|
149
159
|
*
|
|
150
160
|
* @param list_nos inverted list ids as returned by the
|
|
@@ -154,14 +164,16 @@ struct IndexIVF: Index, Level1Quantizer {
|
|
|
154
164
|
* include the list ids in the code (in this case add
|
|
155
165
|
* ceil(log8(nlist)) to the code size)
|
|
156
166
|
*/
|
|
157
|
-
virtual void encode_vectors(
|
|
158
|
-
|
|
159
|
-
|
|
160
|
-
|
|
167
|
+
virtual void encode_vectors(
|
|
168
|
+
idx_t n,
|
|
169
|
+
const float* x,
|
|
170
|
+
const idx_t* list_nos,
|
|
171
|
+
uint8_t* codes,
|
|
172
|
+
bool include_listno = false) const = 0;
|
|
161
173
|
|
|
162
174
|
/// Sub-classes that encode the residuals can train their encoders here
|
|
163
175
|
/// does nothing by default
|
|
164
|
-
virtual void train_residual
|
|
176
|
+
virtual void train_residual(idx_t n, const float* x);
|
|
165
177
|
|
|
166
178
|
/** search a set of vectors, that are pre-quantized by the IVF
|
|
167
179
|
* quantizer. Fill in the corresponding heaps with the query
|
|
@@ -182,36 +194,50 @@ struct IndexIVF: Index, Level1Quantizer {
|
|
|
182
194
|
* @param params used to override the object's search parameters
|
|
183
195
|
* @param stats search stats to be updated (can be null)
|
|
184
196
|
*/
|
|
185
|
-
virtual void search_preassigned
|
|
186
|
-
idx_t n,
|
|
187
|
-
const
|
|
188
|
-
|
|
197
|
+
virtual void search_preassigned(
|
|
198
|
+
idx_t n,
|
|
199
|
+
const float* x,
|
|
200
|
+
idx_t k,
|
|
201
|
+
const idx_t* assign,
|
|
202
|
+
const float* centroid_dis,
|
|
203
|
+
float* distances,
|
|
204
|
+
idx_t* labels,
|
|
189
205
|
bool store_pairs,
|
|
190
|
-
const IVFSearchParameters
|
|
191
|
-
IndexIVFStats
|
|
192
|
-
) const;
|
|
206
|
+
const IVFSearchParameters* params = nullptr,
|
|
207
|
+
IndexIVFStats* stats = nullptr) const;
|
|
193
208
|
|
|
194
209
|
/** assign the vectors, then call search_preassign */
|
|
195
|
-
void search
|
|
196
|
-
|
|
197
|
-
|
|
198
|
-
|
|
199
|
-
|
|
210
|
+
void search(
|
|
211
|
+
idx_t n,
|
|
212
|
+
const float* x,
|
|
213
|
+
idx_t k,
|
|
214
|
+
float* distances,
|
|
215
|
+
idx_t* labels) const override;
|
|
216
|
+
|
|
217
|
+
void range_search(
|
|
218
|
+
idx_t n,
|
|
219
|
+
const float* x,
|
|
220
|
+
float radius,
|
|
221
|
+
RangeSearchResult* result) const override;
|
|
200
222
|
|
|
201
223
|
void range_search_preassigned(
|
|
202
|
-
idx_t nx,
|
|
203
|
-
const
|
|
204
|
-
|
|
205
|
-
|
|
206
|
-
const
|
|
207
|
-
|
|
224
|
+
idx_t nx,
|
|
225
|
+
const float* x,
|
|
226
|
+
float radius,
|
|
227
|
+
const idx_t* keys,
|
|
228
|
+
const float* coarse_dis,
|
|
229
|
+
RangeSearchResult* result,
|
|
230
|
+
bool store_pairs = false,
|
|
231
|
+
const IVFSearchParameters* params = nullptr,
|
|
232
|
+
IndexIVFStats* stats = nullptr) const;
|
|
208
233
|
|
|
209
234
|
/// get a scanner for this index (store_pairs means ignore labels)
|
|
210
|
-
virtual InvertedListScanner
|
|
211
|
-
|
|
235
|
+
virtual InvertedListScanner* get_InvertedListScanner(
|
|
236
|
+
bool store_pairs = false) const;
|
|
212
237
|
|
|
213
|
-
/** reconstruct a vector. Works only if maintain_direct_map is set to 1 or 2
|
|
214
|
-
|
|
238
|
+
/** reconstruct a vector. Works only if maintain_direct_map is set to 1 or 2
|
|
239
|
+
*/
|
|
240
|
+
void reconstruct(idx_t key, float* recons) const override;
|
|
215
241
|
|
|
216
242
|
/** Update a subset of vectors.
|
|
217
243
|
*
|
|
@@ -221,7 +247,7 @@ struct IndexIVF: Index, Level1Quantizer {
|
|
|
221
247
|
* @param idx vector indices to update, size nv
|
|
222
248
|
* @param v vectors of new values, size nv*d
|
|
223
249
|
*/
|
|
224
|
-
virtual void update_vectors
|
|
250
|
+
virtual void update_vectors(int nv, const idx_t* idx, const float* v);
|
|
225
251
|
|
|
226
252
|
/** Reconstruct a subset of the indexed vectors.
|
|
227
253
|
*
|
|
@@ -243,9 +269,13 @@ struct IndexIVF: Index, Level1Quantizer {
|
|
|
243
269
|
*
|
|
244
270
|
* @param recons reconstructed vectors size (n, k, d)
|
|
245
271
|
*/
|
|
246
|
-
void search_and_reconstruct
|
|
247
|
-
|
|
248
|
-
|
|
272
|
+
void search_and_reconstruct(
|
|
273
|
+
idx_t n,
|
|
274
|
+
const float* x,
|
|
275
|
+
idx_t k,
|
|
276
|
+
float* distances,
|
|
277
|
+
idx_t* labels,
|
|
278
|
+
float* recons) const override;
|
|
249
279
|
|
|
250
280
|
/** Reconstruct a vector given the location in terms of (inv list index +
|
|
251
281
|
* inv list offset) instead of the id.
|
|
@@ -254,9 +284,10 @@ struct IndexIVF: Index, Level1Quantizer {
|
|
|
254
284
|
* the inv list offset is computed by search_preassigned() with
|
|
255
285
|
* `store_pairs` set.
|
|
256
286
|
*/
|
|
257
|
-
virtual void reconstruct_from_offset
|
|
258
|
-
|
|
259
|
-
|
|
287
|
+
virtual void reconstruct_from_offset(
|
|
288
|
+
int64_t list_no,
|
|
289
|
+
int64_t offset,
|
|
290
|
+
float* recons) const;
|
|
260
291
|
|
|
261
292
|
/// Dataset manipulation functions
|
|
262
293
|
|
|
@@ -265,12 +296,12 @@ struct IndexIVF: Index, Level1Quantizer {
|
|
|
265
296
|
/** check that the two indexes are compatible (ie, they are
|
|
266
297
|
* trained in the same way and have the same
|
|
267
298
|
* parameters). Otherwise throw. */
|
|
268
|
-
void check_compatible_for_merge
|
|
299
|
+
void check_compatible_for_merge(const IndexIVF& other) const;
|
|
269
300
|
|
|
270
301
|
/** moves the entries from another dataset to self. On output,
|
|
271
302
|
* other is empty. add_id is added to all moved ids (for
|
|
272
303
|
* sequential ids, this would be this->ntotal */
|
|
273
|
-
virtual void merge_from
|
|
304
|
+
virtual void merge_from(IndexIVF& other, idx_t add_id);
|
|
274
305
|
|
|
275
306
|
/** copy a subset of the entries index to the other index
|
|
276
307
|
*
|
|
@@ -279,34 +310,36 @@ struct IndexIVF: Index, Level1Quantizer {
|
|
|
279
310
|
* if subset_type == 2: copies inverted lists such that a1
|
|
280
311
|
* elements are left before and a2 elements are after
|
|
281
312
|
*/
|
|
282
|
-
virtual void copy_subset_to
|
|
283
|
-
|
|
313
|
+
virtual void copy_subset_to(
|
|
314
|
+
IndexIVF& other,
|
|
315
|
+
int subset_type,
|
|
316
|
+
idx_t a1,
|
|
317
|
+
idx_t a2) const;
|
|
284
318
|
|
|
285
319
|
~IndexIVF() override;
|
|
286
320
|
|
|
287
|
-
size_t get_list_size
|
|
288
|
-
|
|
321
|
+
size_t get_list_size(size_t list_no) const {
|
|
322
|
+
return invlists->list_size(list_no);
|
|
323
|
+
}
|
|
289
324
|
|
|
290
325
|
/** intialize a direct map
|
|
291
326
|
*
|
|
292
327
|
* @param new_maintain_direct_map if true, create a direct map,
|
|
293
328
|
* else clear it
|
|
294
329
|
*/
|
|
295
|
-
void make_direct_map
|
|
296
|
-
|
|
297
|
-
void set_direct_map_type (DirectMap::Type type);
|
|
330
|
+
void make_direct_map(bool new_maintain_direct_map = true);
|
|
298
331
|
|
|
332
|
+
void set_direct_map_type(DirectMap::Type type);
|
|
299
333
|
|
|
300
334
|
/// replace the inverted lists, old one is deallocated if own_invlists
|
|
301
|
-
void replace_invlists
|
|
335
|
+
void replace_invlists(InvertedLists* il, bool own = false);
|
|
302
336
|
|
|
303
337
|
/* The standalone codec interface (except sa_decode that is specific) */
|
|
304
|
-
size_t sa_code_size
|
|
338
|
+
size_t sa_code_size() const override;
|
|
305
339
|
|
|
306
|
-
void sa_encode
|
|
307
|
-
uint8_t *bytes) const override;
|
|
340
|
+
void sa_encode(idx_t n, const float* x, uint8_t* bytes) const override;
|
|
308
341
|
|
|
309
|
-
IndexIVF
|
|
342
|
+
IndexIVF();
|
|
310
343
|
};
|
|
311
344
|
|
|
312
345
|
struct RangeQueryResult;
|
|
@@ -316,17 +349,16 @@ struct RangeQueryResult;
|
|
|
316
349
|
* distance_to_code and scan_codes can be called in multiple
|
|
317
350
|
* threads */
|
|
318
351
|
struct InvertedListScanner {
|
|
319
|
-
|
|
320
352
|
using idx_t = Index::idx_t;
|
|
321
353
|
|
|
322
354
|
/// from now on we handle this query.
|
|
323
|
-
virtual void set_query
|
|
355
|
+
virtual void set_query(const float* query_vector) = 0;
|
|
324
356
|
|
|
325
357
|
/// following codes come from this inverted list
|
|
326
|
-
virtual void set_list
|
|
358
|
+
virtual void set_list(idx_t list_no, float coarse_dis) = 0;
|
|
327
359
|
|
|
328
360
|
/// compute a single query-to-code distance
|
|
329
|
-
virtual float distance_to_code
|
|
361
|
+
virtual float distance_to_code(const uint8_t* code) const = 0;
|
|
330
362
|
|
|
331
363
|
/** scan a set of codes, compute distances to current query and
|
|
332
364
|
* update heap of results if necessary.
|
|
@@ -339,45 +371,46 @@ struct InvertedListScanner {
|
|
|
339
371
|
* @param k heap size
|
|
340
372
|
* @return number of heap updates performed
|
|
341
373
|
*/
|
|
342
|
-
virtual size_t scan_codes
|
|
343
|
-
|
|
344
|
-
|
|
345
|
-
|
|
346
|
-
|
|
374
|
+
virtual size_t scan_codes(
|
|
375
|
+
size_t n,
|
|
376
|
+
const uint8_t* codes,
|
|
377
|
+
const idx_t* ids,
|
|
378
|
+
float* distances,
|
|
379
|
+
idx_t* labels,
|
|
380
|
+
size_t k) const = 0;
|
|
347
381
|
|
|
348
382
|
/** scan a set of codes, compute distances to current query and
|
|
349
383
|
* update results if distances are below radius
|
|
350
384
|
*
|
|
351
385
|
* (default implementation fails) */
|
|
352
|
-
virtual void scan_codes_range
|
|
353
|
-
|
|
354
|
-
|
|
355
|
-
|
|
356
|
-
|
|
357
|
-
|
|
358
|
-
|
|
359
|
-
|
|
386
|
+
virtual void scan_codes_range(
|
|
387
|
+
size_t n,
|
|
388
|
+
const uint8_t* codes,
|
|
389
|
+
const idx_t* ids,
|
|
390
|
+
float radius,
|
|
391
|
+
RangeQueryResult& result) const;
|
|
392
|
+
|
|
393
|
+
virtual ~InvertedListScanner() {}
|
|
360
394
|
};
|
|
361
395
|
|
|
362
|
-
|
|
363
396
|
struct IndexIVFStats {
|
|
364
|
-
size_t nq;
|
|
365
|
-
size_t nlist;
|
|
366
|
-
size_t ndis;
|
|
367
|
-
size_t nheap_updates;
|
|
397
|
+
size_t nq; // nb of queries run
|
|
398
|
+
size_t nlist; // nb of inverted lists scanned
|
|
399
|
+
size_t ndis; // nb of distancs computed
|
|
400
|
+
size_t nheap_updates; // nb of times the heap was updated
|
|
368
401
|
double quantization_time; // time spent quantizing vectors (in ms)
|
|
369
402
|
double search_time; // time spent searching lists (in ms)
|
|
370
403
|
|
|
371
|
-
IndexIVFStats
|
|
372
|
-
|
|
373
|
-
|
|
404
|
+
IndexIVFStats() {
|
|
405
|
+
reset();
|
|
406
|
+
}
|
|
407
|
+
void reset();
|
|
408
|
+
void add(const IndexIVFStats& other);
|
|
374
409
|
};
|
|
375
410
|
|
|
376
411
|
// global var that collects them all
|
|
377
412
|
FAISS_API extern IndexIVFStats indexIVF_stats;
|
|
378
413
|
|
|
379
|
-
|
|
380
414
|
} // namespace faiss
|
|
381
415
|
|
|
382
|
-
|
|
383
416
|
#endif
|
|
@@ -9,332 +9,340 @@
|
|
|
9
9
|
|
|
10
10
|
#include <faiss/IndexIVFFlat.h>
|
|
11
11
|
|
|
12
|
+
#include <omp.h>
|
|
13
|
+
|
|
12
14
|
#include <cinttypes>
|
|
13
15
|
#include <cstdio>
|
|
14
16
|
|
|
15
17
|
#include <faiss/IndexFlat.h>
|
|
16
18
|
|
|
19
|
+
#include <faiss/impl/AuxIndexStructures.h>
|
|
20
|
+
#include <faiss/impl/FaissAssert.h>
|
|
17
21
|
#include <faiss/utils/distances.h>
|
|
18
22
|
#include <faiss/utils/utils.h>
|
|
19
|
-
#include <faiss/impl/FaissAssert.h>
|
|
20
|
-
#include <faiss/impl/AuxIndexStructures.h>
|
|
21
|
-
|
|
22
23
|
|
|
23
24
|
namespace faiss {
|
|
24
25
|
|
|
25
|
-
|
|
26
26
|
/*****************************************
|
|
27
27
|
* IndexIVFFlat implementation
|
|
28
28
|
******************************************/
|
|
29
29
|
|
|
30
|
-
IndexIVFFlat::IndexIVFFlat
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
|
|
30
|
+
IndexIVFFlat::IndexIVFFlat(
|
|
31
|
+
Index* quantizer,
|
|
32
|
+
size_t d,
|
|
33
|
+
size_t nlist,
|
|
34
|
+
MetricType metric)
|
|
35
|
+
: IndexIVF(quantizer, d, nlist, sizeof(float) * d, metric) {
|
|
34
36
|
code_size = sizeof(float) * d;
|
|
35
37
|
}
|
|
36
38
|
|
|
39
|
+
void IndexIVFFlat::add_core(
|
|
40
|
+
idx_t n,
|
|
41
|
+
const float* x,
|
|
42
|
+
const int64_t* xids,
|
|
43
|
+
const int64_t* coarse_idx)
|
|
37
44
|
|
|
38
|
-
void IndexIVFFlat::add_with_ids (idx_t n, const float * x, const idx_t *xids)
|
|
39
45
|
{
|
|
40
|
-
|
|
41
|
-
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
const int64_t *precomputed_idx)
|
|
46
|
+
FAISS_THROW_IF_NOT(is_trained);
|
|
47
|
+
FAISS_THROW_IF_NOT(coarse_idx);
|
|
48
|
+
assert(invlists);
|
|
49
|
+
direct_map.check_can_add(xids);
|
|
45
50
|
|
|
46
|
-
{
|
|
47
|
-
FAISS_THROW_IF_NOT (is_trained);
|
|
48
|
-
assert (invlists);
|
|
49
|
-
direct_map.check_can_add (xids);
|
|
50
|
-
const int64_t * idx;
|
|
51
|
-
ScopeDeleter<int64_t> del;
|
|
52
|
-
|
|
53
|
-
if (precomputed_idx) {
|
|
54
|
-
idx = precomputed_idx;
|
|
55
|
-
} else {
|
|
56
|
-
int64_t * idx0 = new int64_t [n];
|
|
57
|
-
del.set (idx0);
|
|
58
|
-
quantizer->assign (n, x, idx0);
|
|
59
|
-
idx = idx0;
|
|
60
|
-
}
|
|
61
51
|
int64_t n_add = 0;
|
|
62
|
-
|
|
63
|
-
|
|
64
|
-
|
|
65
|
-
|
|
66
|
-
|
|
67
|
-
|
|
68
|
-
|
|
69
|
-
|
|
70
|
-
|
|
71
|
-
|
|
72
|
-
|
|
73
|
-
|
|
52
|
+
|
|
53
|
+
DirectMapAdd dm_adder(direct_map, n, xids);
|
|
54
|
+
|
|
55
|
+
#pragma omp parallel reduction(+ : n_add)
|
|
56
|
+
{
|
|
57
|
+
int nt = omp_get_num_threads();
|
|
58
|
+
int rank = omp_get_thread_num();
|
|
59
|
+
|
|
60
|
+
// each thread takes care of a subset of lists
|
|
61
|
+
for (size_t i = 0; i < n; i++) {
|
|
62
|
+
idx_t list_no = coarse_idx[i];
|
|
63
|
+
|
|
64
|
+
if (list_no >= 0 && list_no % nt == rank) {
|
|
65
|
+
idx_t id = xids ? xids[i] : ntotal + i;
|
|
66
|
+
const float* xi = x + i * d;
|
|
67
|
+
size_t offset =
|
|
68
|
+
invlists->add_entry(list_no, id, (const uint8_t*)xi);
|
|
69
|
+
dm_adder.add(i, list_no, offset);
|
|
70
|
+
n_add++;
|
|
71
|
+
} else if (rank == 0 && list_no == -1) {
|
|
72
|
+
dm_adder.add(i, -1, 0);
|
|
73
|
+
}
|
|
74
74
|
}
|
|
75
|
-
direct_map.add_single_id (id, list_no, offset);
|
|
76
75
|
}
|
|
77
76
|
|
|
78
77
|
if (verbose) {
|
|
79
|
-
printf("IndexIVFFlat::add_core: added %" PRId64 " / %" PRId64
|
|
80
|
-
|
|
78
|
+
printf("IndexIVFFlat::add_core: added %" PRId64 " / %" PRId64
|
|
79
|
+
" vectors\n",
|
|
80
|
+
n_add,
|
|
81
|
+
n);
|
|
81
82
|
}
|
|
82
83
|
ntotal += n;
|
|
83
84
|
}
|
|
84
85
|
|
|
85
|
-
void IndexIVFFlat::encode_vectors(
|
|
86
|
-
|
|
87
|
-
|
|
88
|
-
|
|
89
|
-
|
|
86
|
+
void IndexIVFFlat::encode_vectors(
|
|
87
|
+
idx_t n,
|
|
88
|
+
const float* x,
|
|
89
|
+
const idx_t* list_nos,
|
|
90
|
+
uint8_t* codes,
|
|
91
|
+
bool include_listnos) const {
|
|
90
92
|
if (!include_listnos) {
|
|
91
|
-
memcpy
|
|
93
|
+
memcpy(codes, x, code_size * n);
|
|
92
94
|
} else {
|
|
93
|
-
size_t coarse_size = coarse_code_size
|
|
95
|
+
size_t coarse_size = coarse_code_size();
|
|
94
96
|
for (size_t i = 0; i < n; i++) {
|
|
95
|
-
int64_t list_no = list_nos
|
|
96
|
-
uint8_t
|
|
97
|
-
const float
|
|
97
|
+
int64_t list_no = list_nos[i];
|
|
98
|
+
uint8_t* code = codes + i * (code_size + coarse_size);
|
|
99
|
+
const float* xi = x + i * d;
|
|
98
100
|
if (list_no >= 0) {
|
|
99
|
-
encode_listno
|
|
100
|
-
memcpy
|
|
101
|
+
encode_listno(list_no, code);
|
|
102
|
+
memcpy(code + coarse_size, xi, code_size);
|
|
101
103
|
} else {
|
|
102
|
-
memset
|
|
104
|
+
memset(code, 0, code_size + coarse_size);
|
|
103
105
|
}
|
|
104
|
-
|
|
105
106
|
}
|
|
106
107
|
}
|
|
107
108
|
}
|
|
108
109
|
|
|
109
|
-
void IndexIVFFlat::sa_decode
|
|
110
|
-
|
|
111
|
-
{
|
|
112
|
-
size_t coarse_size = coarse_code_size ();
|
|
110
|
+
void IndexIVFFlat::sa_decode(idx_t n, const uint8_t* bytes, float* x) const {
|
|
111
|
+
size_t coarse_size = coarse_code_size();
|
|
113
112
|
for (size_t i = 0; i < n; i++) {
|
|
114
|
-
const uint8_t
|
|
115
|
-
float
|
|
116
|
-
memcpy
|
|
113
|
+
const uint8_t* code = bytes + i * (code_size + coarse_size);
|
|
114
|
+
float* xi = x + i * d;
|
|
115
|
+
memcpy(xi, code + coarse_size, code_size);
|
|
117
116
|
}
|
|
118
117
|
}
|
|
119
118
|
|
|
120
|
-
|
|
121
119
|
namespace {
|
|
122
120
|
|
|
123
|
-
|
|
124
|
-
|
|
125
|
-
struct IVFFlatScanner: InvertedListScanner {
|
|
121
|
+
template <MetricType metric, class C>
|
|
122
|
+
struct IVFFlatScanner : InvertedListScanner {
|
|
126
123
|
size_t d;
|
|
127
124
|
bool store_pairs;
|
|
128
125
|
|
|
129
|
-
IVFFlatScanner(size_t d, bool store_pairs)
|
|
130
|
-
|
|
126
|
+
IVFFlatScanner(size_t d, bool store_pairs)
|
|
127
|
+
: d(d), store_pairs(store_pairs) {}
|
|
131
128
|
|
|
132
|
-
const float
|
|
133
|
-
void set_query
|
|
129
|
+
const float* xi;
|
|
130
|
+
void set_query(const float* query) override {
|
|
134
131
|
this->xi = query;
|
|
135
132
|
}
|
|
136
133
|
|
|
137
134
|
idx_t list_no;
|
|
138
|
-
void set_list
|
|
135
|
+
void set_list(idx_t list_no, float /* coarse_dis */) override {
|
|
139
136
|
this->list_no = list_no;
|
|
140
137
|
}
|
|
141
138
|
|
|
142
|
-
float distance_to_code
|
|
143
|
-
const float
|
|
144
|
-
float dis = metric == METRIC_INNER_PRODUCT
|
|
145
|
-
|
|
139
|
+
float distance_to_code(const uint8_t* code) const override {
|
|
140
|
+
const float* yj = (float*)code;
|
|
141
|
+
float dis = metric == METRIC_INNER_PRODUCT
|
|
142
|
+
? fvec_inner_product(xi, yj, d)
|
|
143
|
+
: fvec_L2sqr(xi, yj, d);
|
|
146
144
|
return dis;
|
|
147
145
|
}
|
|
148
146
|
|
|
149
|
-
size_t scan_codes
|
|
150
|
-
|
|
151
|
-
|
|
152
|
-
|
|
153
|
-
|
|
154
|
-
|
|
155
|
-
|
|
147
|
+
size_t scan_codes(
|
|
148
|
+
size_t list_size,
|
|
149
|
+
const uint8_t* codes,
|
|
150
|
+
const idx_t* ids,
|
|
151
|
+
float* simi,
|
|
152
|
+
idx_t* idxi,
|
|
153
|
+
size_t k) const override {
|
|
154
|
+
const float* list_vecs = (const float*)codes;
|
|
156
155
|
size_t nup = 0;
|
|
157
156
|
for (size_t j = 0; j < list_size; j++) {
|
|
158
|
-
const float
|
|
159
|
-
float dis = metric == METRIC_INNER_PRODUCT
|
|
160
|
-
|
|
161
|
-
|
|
162
|
-
|
|
163
|
-
|
|
157
|
+
const float* yj = list_vecs + d * j;
|
|
158
|
+
float dis = metric == METRIC_INNER_PRODUCT
|
|
159
|
+
? fvec_inner_product(xi, yj, d)
|
|
160
|
+
: fvec_L2sqr(xi, yj, d);
|
|
161
|
+
if (C::cmp(simi[0], dis)) {
|
|
162
|
+
int64_t id = store_pairs ? lo_build(list_no, j) : ids[j];
|
|
163
|
+
heap_replace_top<C>(k, simi, idxi, dis, id);
|
|
164
164
|
nup++;
|
|
165
165
|
}
|
|
166
166
|
}
|
|
167
167
|
return nup;
|
|
168
168
|
}
|
|
169
169
|
|
|
170
|
-
void scan_codes_range
|
|
171
|
-
|
|
172
|
-
|
|
173
|
-
|
|
174
|
-
|
|
175
|
-
|
|
176
|
-
const float
|
|
170
|
+
void scan_codes_range(
|
|
171
|
+
size_t list_size,
|
|
172
|
+
const uint8_t* codes,
|
|
173
|
+
const idx_t* ids,
|
|
174
|
+
float radius,
|
|
175
|
+
RangeQueryResult& res) const override {
|
|
176
|
+
const float* list_vecs = (const float*)codes;
|
|
177
177
|
for (size_t j = 0; j < list_size; j++) {
|
|
178
|
-
const float
|
|
179
|
-
float dis = metric == METRIC_INNER_PRODUCT
|
|
180
|
-
|
|
181
|
-
|
|
182
|
-
|
|
183
|
-
|
|
178
|
+
const float* yj = list_vecs + d * j;
|
|
179
|
+
float dis = metric == METRIC_INNER_PRODUCT
|
|
180
|
+
? fvec_inner_product(xi, yj, d)
|
|
181
|
+
: fvec_L2sqr(xi, yj, d);
|
|
182
|
+
if (C::cmp(radius, dis)) {
|
|
183
|
+
int64_t id = store_pairs ? lo_build(list_no, j) : ids[j];
|
|
184
|
+
res.add(dis, id);
|
|
184
185
|
}
|
|
185
186
|
}
|
|
186
187
|
}
|
|
187
|
-
|
|
188
|
-
|
|
189
188
|
};
|
|
190
189
|
|
|
191
|
-
|
|
192
190
|
} // anonymous namespace
|
|
193
191
|
|
|
194
|
-
|
|
195
|
-
|
|
196
|
-
InvertedListScanner* IndexIVFFlat::get_InvertedListScanner
|
|
197
|
-
(bool store_pairs) const
|
|
198
|
-
{
|
|
192
|
+
InvertedListScanner* IndexIVFFlat::get_InvertedListScanner(
|
|
193
|
+
bool store_pairs) const {
|
|
199
194
|
if (metric_type == METRIC_INNER_PRODUCT) {
|
|
200
|
-
return new IVFFlatScanner<
|
|
201
|
-
|
|
195
|
+
return new IVFFlatScanner<METRIC_INNER_PRODUCT, CMin<float, int64_t>>(
|
|
196
|
+
d, store_pairs);
|
|
202
197
|
} else if (metric_type == METRIC_L2) {
|
|
203
|
-
return new IVFFlatScanner<
|
|
204
|
-
|
|
198
|
+
return new IVFFlatScanner<METRIC_L2, CMax<float, int64_t>>(
|
|
199
|
+
d, store_pairs);
|
|
205
200
|
} else {
|
|
206
201
|
FAISS_THROW_MSG("metric type not supported");
|
|
207
202
|
}
|
|
208
203
|
return nullptr;
|
|
209
204
|
}
|
|
210
205
|
|
|
211
|
-
|
|
212
|
-
|
|
213
|
-
|
|
214
|
-
|
|
215
|
-
|
|
216
|
-
{
|
|
217
|
-
memcpy (recons, invlists->get_single_code (list_no, offset), code_size);
|
|
206
|
+
void IndexIVFFlat::reconstruct_from_offset(
|
|
207
|
+
int64_t list_no,
|
|
208
|
+
int64_t offset,
|
|
209
|
+
float* recons) const {
|
|
210
|
+
memcpy(recons, invlists->get_single_code(list_no, offset), code_size);
|
|
218
211
|
}
|
|
219
212
|
|
|
220
213
|
/*****************************************
|
|
221
214
|
* IndexIVFFlatDedup implementation
|
|
222
215
|
******************************************/
|
|
223
216
|
|
|
224
|
-
IndexIVFFlatDedup::IndexIVFFlatDedup
|
|
225
|
-
|
|
226
|
-
|
|
227
|
-
|
|
228
|
-
|
|
229
|
-
|
|
217
|
+
IndexIVFFlatDedup::IndexIVFFlatDedup(
|
|
218
|
+
Index* quantizer,
|
|
219
|
+
size_t d,
|
|
220
|
+
size_t nlist_,
|
|
221
|
+
MetricType metric_type)
|
|
222
|
+
: IndexIVFFlat(quantizer, d, nlist_, metric_type) {}
|
|
230
223
|
|
|
231
|
-
void IndexIVFFlatDedup::train(idx_t n, const float* x)
|
|
232
|
-
{
|
|
224
|
+
void IndexIVFFlatDedup::train(idx_t n, const float* x) {
|
|
233
225
|
std::unordered_map<uint64_t, idx_t> map;
|
|
234
|
-
float
|
|
235
|
-
ScopeDeleter<float> del
|
|
226
|
+
float* x2 = new float[n * d];
|
|
227
|
+
ScopeDeleter<float> del(x2);
|
|
236
228
|
|
|
237
229
|
int64_t n2 = 0;
|
|
238
230
|
for (int64_t i = 0; i < n; i++) {
|
|
239
|
-
uint64_t hash = hash_bytes((uint8_t
|
|
231
|
+
uint64_t hash = hash_bytes((uint8_t*)(x + i * d), code_size);
|
|
240
232
|
if (map.count(hash) &&
|
|
241
|
-
!memcmp
|
|
233
|
+
!memcmp(x2 + map[hash] * d, x + i * d, code_size)) {
|
|
242
234
|
// is duplicate, skip
|
|
243
235
|
} else {
|
|
244
|
-
map
|
|
245
|
-
memcpy
|
|
246
|
-
n2
|
|
236
|
+
map[hash] = n2;
|
|
237
|
+
memcpy(x2 + n2 * d, x + i * d, code_size);
|
|
238
|
+
n2++;
|
|
247
239
|
}
|
|
248
240
|
}
|
|
249
241
|
if (verbose) {
|
|
250
|
-
printf
|
|
251
|
-
|
|
242
|
+
printf("IndexIVFFlatDedup::train: train on %" PRId64
|
|
243
|
+
" points after dedup "
|
|
244
|
+
"(was %" PRId64 " points)\n",
|
|
245
|
+
n2,
|
|
246
|
+
n);
|
|
252
247
|
}
|
|
253
|
-
IndexIVFFlat::train
|
|
248
|
+
IndexIVFFlat::train(n2, x2);
|
|
254
249
|
}
|
|
255
250
|
|
|
251
|
+
void IndexIVFFlatDedup::add_with_ids(
|
|
252
|
+
idx_t na,
|
|
253
|
+
const float* x,
|
|
254
|
+
const idx_t* xids) {
|
|
255
|
+
FAISS_THROW_IF_NOT(is_trained);
|
|
256
|
+
assert(invlists);
|
|
257
|
+
FAISS_THROW_IF_NOT_MSG(
|
|
258
|
+
direct_map.no(), "IVFFlatDedup not implemented with direct_map");
|
|
259
|
+
int64_t* idx = new int64_t[na];
|
|
260
|
+
ScopeDeleter<int64_t> del(idx);
|
|
261
|
+
quantizer->assign(na, x, idx);
|
|
256
262
|
|
|
263
|
+
int64_t n_add = 0, n_dup = 0;
|
|
257
264
|
|
|
258
|
-
|
|
259
|
-
|
|
260
|
-
|
|
265
|
+
#pragma omp parallel reduction(+ : n_add, n_dup)
|
|
266
|
+
{
|
|
267
|
+
int nt = omp_get_num_threads();
|
|
268
|
+
int rank = omp_get_thread_num();
|
|
261
269
|
|
|
262
|
-
|
|
263
|
-
|
|
264
|
-
|
|
265
|
-
"IVFFlatDedup not implemented with direct_map");
|
|
266
|
-
int64_t * idx = new int64_t [na];
|
|
267
|
-
ScopeDeleter<int64_t> del (idx);
|
|
268
|
-
quantizer->assign (na, x, idx);
|
|
270
|
+
// each thread takes care of a subset of lists
|
|
271
|
+
for (size_t i = 0; i < na; i++) {
|
|
272
|
+
int64_t list_no = idx[i];
|
|
269
273
|
|
|
270
|
-
|
|
271
|
-
|
|
272
|
-
|
|
273
|
-
idx_t id = xids ? xids[i] : ntotal + i;
|
|
274
|
-
int64_t list_no = idx [i];
|
|
274
|
+
if (list_no < 0 || list_no % nt != rank) {
|
|
275
|
+
continue;
|
|
276
|
+
}
|
|
275
277
|
|
|
276
|
-
|
|
277
|
-
|
|
278
|
-
}
|
|
279
|
-
const float *xi = x + i * d;
|
|
278
|
+
idx_t id = xids ? xids[i] : ntotal + i;
|
|
279
|
+
const float* xi = x + i * d;
|
|
280
280
|
|
|
281
|
-
|
|
282
|
-
|
|
281
|
+
// search if there is already an entry with that id
|
|
282
|
+
InvertedLists::ScopedCodes codes(invlists, list_no);
|
|
283
283
|
|
|
284
|
-
|
|
285
|
-
|
|
286
|
-
|
|
287
|
-
|
|
288
|
-
|
|
289
|
-
|
|
290
|
-
|
|
284
|
+
int64_t n = invlists->list_size(list_no);
|
|
285
|
+
int64_t offset = -1;
|
|
286
|
+
for (int64_t o = 0; o < n; o++) {
|
|
287
|
+
if (!memcmp(codes.get() + o * code_size, xi, code_size)) {
|
|
288
|
+
offset = o;
|
|
289
|
+
break;
|
|
290
|
+
}
|
|
291
291
|
}
|
|
292
|
-
}
|
|
293
292
|
|
|
294
|
-
|
|
295
|
-
|
|
296
|
-
|
|
297
|
-
|
|
298
|
-
|
|
299
|
-
|
|
300
|
-
|
|
301
|
-
|
|
293
|
+
if (offset == -1) { // not found
|
|
294
|
+
invlists->add_entry(list_no, id, (const uint8_t*)xi);
|
|
295
|
+
} else {
|
|
296
|
+
// mark equivalence
|
|
297
|
+
idx_t id2 = invlists->get_single_id(list_no, offset);
|
|
298
|
+
std::pair<idx_t, idx_t> pair(id2, id);
|
|
299
|
+
|
|
300
|
+
#pragma omp critical
|
|
301
|
+
// executed by one thread at a time
|
|
302
|
+
instances.insert(pair);
|
|
303
|
+
|
|
304
|
+
n_dup++;
|
|
305
|
+
}
|
|
306
|
+
n_add++;
|
|
302
307
|
}
|
|
303
|
-
n_add++;
|
|
304
308
|
}
|
|
305
309
|
if (verbose) {
|
|
306
|
-
printf("IndexIVFFlat::add_with_ids: added %" PRId64 " / %" PRId64
|
|
310
|
+
printf("IndexIVFFlat::add_with_ids: added %" PRId64 " / %" PRId64
|
|
311
|
+
" vectors"
|
|
307
312
|
" (out of which %" PRId64 " are duplicates)\n",
|
|
308
|
-
n_add,
|
|
313
|
+
n_add,
|
|
314
|
+
na,
|
|
315
|
+
n_dup);
|
|
309
316
|
}
|
|
310
317
|
ntotal += n_add;
|
|
311
318
|
}
|
|
312
319
|
|
|
313
|
-
void IndexIVFFlatDedup::search_preassigned
|
|
314
|
-
|
|
315
|
-
|
|
316
|
-
|
|
317
|
-
|
|
318
|
-
|
|
319
|
-
|
|
320
|
-
|
|
321
|
-
|
|
322
|
-
|
|
323
|
-
|
|
324
|
-
|
|
325
|
-
|
|
326
|
-
|
|
327
|
-
|
|
328
|
-
|
|
329
|
-
|
|
330
|
-
std::vector
|
|
320
|
+
void IndexIVFFlatDedup::search_preassigned(
|
|
321
|
+
idx_t n,
|
|
322
|
+
const float* x,
|
|
323
|
+
idx_t k,
|
|
324
|
+
const idx_t* assign,
|
|
325
|
+
const float* centroid_dis,
|
|
326
|
+
float* distances,
|
|
327
|
+
idx_t* labels,
|
|
328
|
+
bool store_pairs,
|
|
329
|
+
const IVFSearchParameters* params,
|
|
330
|
+
IndexIVFStats* stats) const {
|
|
331
|
+
FAISS_THROW_IF_NOT_MSG(
|
|
332
|
+
!store_pairs, "store_pairs not supported in IVFDedup");
|
|
333
|
+
|
|
334
|
+
IndexIVFFlat::search_preassigned(
|
|
335
|
+
n, x, k, assign, centroid_dis, distances, labels, false, params);
|
|
336
|
+
|
|
337
|
+
std::vector<idx_t> labels2(k);
|
|
338
|
+
std::vector<float> dis2(k);
|
|
331
339
|
|
|
332
340
|
for (int64_t i = 0; i < n; i++) {
|
|
333
|
-
idx_t
|
|
334
|
-
float
|
|
341
|
+
idx_t* labels1 = labels + i * k;
|
|
342
|
+
float* dis1 = distances + i * k;
|
|
335
343
|
int64_t j = 0;
|
|
336
344
|
for (; j < k; j++) {
|
|
337
|
-
if (instances.find
|
|
345
|
+
if (instances.find(labels1[j]) != instances.end()) {
|
|
338
346
|
// a duplicate: special handling
|
|
339
347
|
break;
|
|
340
348
|
}
|
|
@@ -344,11 +352,11 @@ void IndexIVFFlatDedup::search_preassigned (
|
|
|
344
352
|
int64_t j0 = j;
|
|
345
353
|
int64_t rp = j;
|
|
346
354
|
while (j < k) {
|
|
347
|
-
auto range = instances.equal_range
|
|
355
|
+
auto range = instances.equal_range(labels1[rp]);
|
|
348
356
|
float dis = dis1[rp];
|
|
349
357
|
labels2[j] = labels1[rp];
|
|
350
358
|
dis2[j] = dis;
|
|
351
|
-
j
|
|
359
|
+
j++;
|
|
352
360
|
for (auto it = range.first; j < k && it != range.second; ++it) {
|
|
353
361
|
labels2[j] = it->second;
|
|
354
362
|
dis2[j] = dis;
|
|
@@ -356,21 +364,18 @@ void IndexIVFFlatDedup::search_preassigned (
|
|
|
356
364
|
}
|
|
357
365
|
rp++;
|
|
358
366
|
}
|
|
359
|
-
memcpy
|
|
360
|
-
|
|
361
|
-
|
|
362
|
-
|
|
367
|
+
memcpy(labels1 + j0,
|
|
368
|
+
labels2.data() + j0,
|
|
369
|
+
sizeof(labels1[0]) * (k - j0));
|
|
370
|
+
memcpy(dis1 + j0, dis2.data() + j0, sizeof(dis2[0]) * (k - j0));
|
|
363
371
|
}
|
|
364
372
|
}
|
|
365
|
-
|
|
366
373
|
}
|
|
367
374
|
|
|
368
|
-
|
|
369
|
-
size_t IndexIVFFlatDedup::remove_ids(const IDSelector& sel)
|
|
370
|
-
{
|
|
375
|
+
size_t IndexIVFFlatDedup::remove_ids(const IDSelector& sel) {
|
|
371
376
|
std::unordered_map<idx_t, idx_t> replace;
|
|
372
|
-
std::vector<std::pair<idx_t, idx_t
|
|
373
|
-
for (auto it = instances.begin(); it != instances.end();
|
|
377
|
+
std::vector<std::pair<idx_t, idx_t>> toadd;
|
|
378
|
+
for (auto it = instances.begin(); it != instances.end();) {
|
|
374
379
|
if (sel.is_member(it->first)) {
|
|
375
380
|
// then we erase this entry
|
|
376
381
|
if (!sel.is_member(it->second)) {
|
|
@@ -378,8 +383,8 @@ size_t IndexIVFFlatDedup::remove_ids(const IDSelector& sel)
|
|
|
378
383
|
if (replace.count(it->first) == 0) {
|
|
379
384
|
replace[it->first] = it->second;
|
|
380
385
|
} else { // remember we should add an element
|
|
381
|
-
std::pair<idx_t, idx_t> new_entry
|
|
382
|
-
|
|
386
|
+
std::pair<idx_t, idx_t> new_entry(
|
|
387
|
+
replace[it->first], it->second);
|
|
383
388
|
toadd.push_back(new_entry);
|
|
384
389
|
}
|
|
385
390
|
}
|
|
@@ -393,32 +398,34 @@ size_t IndexIVFFlatDedup::remove_ids(const IDSelector& sel)
|
|
|
393
398
|
}
|
|
394
399
|
}
|
|
395
400
|
|
|
396
|
-
instances.insert
|
|
401
|
+
instances.insert(toadd.begin(), toadd.end());
|
|
397
402
|
|
|
398
403
|
// mostly copied from IndexIVF.cpp
|
|
399
404
|
|
|
400
|
-
FAISS_THROW_IF_NOT_MSG
|
|
401
|
-
|
|
405
|
+
FAISS_THROW_IF_NOT_MSG(
|
|
406
|
+
direct_map.no(), "direct map remove not implemented");
|
|
402
407
|
|
|
403
408
|
std::vector<int64_t> toremove(nlist);
|
|
404
409
|
|
|
405
410
|
#pragma omp parallel for
|
|
406
411
|
for (int64_t i = 0; i < nlist; i++) {
|
|
407
|
-
int64_t l0 = invlists->list_size
|
|
408
|
-
InvertedLists::ScopedIds idsi
|
|
412
|
+
int64_t l0 = invlists->list_size(i), l = l0, j = 0;
|
|
413
|
+
InvertedLists::ScopedIds idsi(invlists, i);
|
|
409
414
|
while (j < l) {
|
|
410
|
-
if (sel.is_member
|
|
415
|
+
if (sel.is_member(idsi[j])) {
|
|
411
416
|
if (replace.count(idsi[j]) == 0) {
|
|
412
417
|
l--;
|
|
413
|
-
invlists->update_entry
|
|
414
|
-
|
|
415
|
-
|
|
416
|
-
|
|
418
|
+
invlists->update_entry(
|
|
419
|
+
i,
|
|
420
|
+
j,
|
|
421
|
+
invlists->get_single_id(i, l),
|
|
422
|
+
InvertedLists::ScopedCodes(invlists, i, l).get());
|
|
417
423
|
} else {
|
|
418
|
-
invlists->update_entry
|
|
419
|
-
|
|
420
|
-
|
|
421
|
-
|
|
424
|
+
invlists->update_entry(
|
|
425
|
+
i,
|
|
426
|
+
j,
|
|
427
|
+
replace[idsi[j]],
|
|
428
|
+
InvertedLists::ScopedCodes(invlists, i, j).get());
|
|
422
429
|
j++;
|
|
423
430
|
}
|
|
424
431
|
} else {
|
|
@@ -432,37 +439,28 @@ size_t IndexIVFFlatDedup::remove_ids(const IDSelector& sel)
|
|
|
432
439
|
for (int64_t i = 0; i < nlist; i++) {
|
|
433
440
|
if (toremove[i] > 0) {
|
|
434
441
|
nremove += toremove[i];
|
|
435
|
-
invlists->resize(
|
|
436
|
-
i, invlists->list_size(i) - toremove[i]);
|
|
442
|
+
invlists->resize(i, invlists->list_size(i) - toremove[i]);
|
|
437
443
|
}
|
|
438
444
|
}
|
|
439
445
|
ntotal -= nremove;
|
|
440
446
|
return nremove;
|
|
441
447
|
}
|
|
442
448
|
|
|
443
|
-
|
|
444
449
|
void IndexIVFFlatDedup::range_search(
|
|
445
|
-
idx_t
|
|
446
|
-
const float
|
|
447
|
-
float
|
|
448
|
-
RangeSearchResult*
|
|
449
|
-
|
|
450
|
-
FAISS_THROW_MSG ("not implemented");
|
|
450
|
+
idx_t,
|
|
451
|
+
const float*,
|
|
452
|
+
float,
|
|
453
|
+
RangeSearchResult*) const {
|
|
454
|
+
FAISS_THROW_MSG("not implemented");
|
|
451
455
|
}
|
|
452
456
|
|
|
453
|
-
void IndexIVFFlatDedup::update_vectors
|
|
454
|
-
|
|
455
|
-
FAISS_THROW_MSG ("not implemented");
|
|
457
|
+
void IndexIVFFlatDedup::update_vectors(int, const idx_t*, const float*) {
|
|
458
|
+
FAISS_THROW_MSG("not implemented");
|
|
456
459
|
}
|
|
457
460
|
|
|
458
|
-
|
|
459
|
-
|
|
460
|
-
|
|
461
|
-
{
|
|
462
|
-
FAISS_THROW_MSG ("not implemented");
|
|
461
|
+
void IndexIVFFlatDedup::reconstruct_from_offset(int64_t, int64_t, float*)
|
|
462
|
+
const {
|
|
463
|
+
FAISS_THROW_MSG("not implemented");
|
|
463
464
|
}
|
|
464
465
|
|
|
465
|
-
|
|
466
|
-
|
|
467
|
-
|
|
468
466
|
} // namespace faiss
|