8
8
"testing"
9
9
"time"
10
10
11
- "github.com/prometheus/prometheus/model/labels"
12
-
13
11
"github.com/grafana/dskit/multierror"
14
12
15
13
"github.com/grafana/loki/pkg/push"
@@ -101,8 +99,6 @@ func TestTokenizerPopulate(t *testing.T) {
101
99
bt := NewBloomTokenizer (DefaultNGramLength , DefaultNGramSkip , 0 , metrics )
102
100
103
101
sbf := filter .NewScalableBloomFilter (1024 , 0.01 , 0.8 )
104
- var lbsList []labels.Labels
105
- lbsList = append (lbsList , labels .FromStrings ("foo" , "bar" ))
106
102
107
103
memChunk := chunkenc .NewMemChunk (chunkenc .ChunkFormatV4 , chunkenc .EncSnappy , chunkenc .ChunkHeadFormatFor (chunkenc .ChunkFormatV4 ), 256000 , 1500000 )
108
104
_ = memChunk .Append (& push.Entry {
@@ -121,13 +117,9 @@ func TestTokenizerPopulate(t *testing.T) {
121
117
bloom := Bloom {
122
118
ScalableBloomFilter : * sbf ,
123
119
}
124
- series := Series {
125
- Fingerprint : model .Fingerprint (lbsList [0 ].Hash ()),
126
- }
127
120
128
121
blooms , err := populateAndConsumeBloom (
129
122
bt ,
130
- series ,
131
123
NewSliceIter ([]* Bloom {& bloom }),
132
124
NewSliceIter ([]ChunkRefWithIter {{Ref : ChunkRef {},
133
125
Itr : itr }}),
@@ -147,9 +139,6 @@ func TestBloomTokenizerPopulateWithoutPreexistingBloom(t *testing.T) {
147
139
var testLine = "this is a log line"
148
140
bt := NewBloomTokenizer (DefaultNGramLength , DefaultNGramSkip , 0 , metrics )
149
141
150
- var lbsList []labels.Labels
151
- lbsList = append (lbsList , labels .FromStrings ("foo" , "bar" ))
152
-
153
142
memChunk := chunkenc .NewMemChunk (chunkenc .ChunkFormatV4 , chunkenc .EncSnappy , chunkenc .ChunkHeadFormatFor (chunkenc .ChunkFormatV4 ), 256000 , 1500000 )
154
143
_ = memChunk .Append (& push.Entry {
155
144
Timestamp : time .Unix (0 , 1 ),
@@ -164,13 +153,8 @@ func TestBloomTokenizerPopulateWithoutPreexistingBloom(t *testing.T) {
164
153
)
165
154
require .Nil (t , err )
166
155
167
- series := Series {
168
- Fingerprint : model .Fingerprint (lbsList [0 ].Hash ()),
169
- }
170
-
171
156
blooms , err := populateAndConsumeBloom (
172
157
bt ,
173
- series ,
174
158
NewEmptyIter [* Bloom ](),
175
159
NewSliceIter ([]ChunkRefWithIter {{Ref : ChunkRef {},
176
160
Itr : itr }}),
@@ -227,11 +211,6 @@ func TestTokenizerPopulateWontExceedMaxSize(t *testing.T) {
227
211
itr , err := chunkRefItrFromLines (line )
228
212
require .NoError (t , err )
229
213
go bt .Populate (
230
- & Series {
231
- Chunks : ChunkRefs {
232
- {},
233
- },
234
- },
235
214
NewSliceIter ([]* Bloom {
236
215
{
237
216
* filter .NewScalableBloomFilter (1024 , 0.01 , 0.8 ),
@@ -258,13 +237,12 @@ func TestTokenizerPopulateWontExceedMaxSize(t *testing.T) {
258
237
259
238
func populateAndConsumeBloom (
260
239
bt * BloomTokenizer ,
261
- s Series ,
262
240
blooms SizedIterator [* Bloom ],
263
241
chks Iterator [ChunkRefWithIter ],
264
242
) (res []* Bloom , err error ) {
265
243
var e multierror.MultiError
266
244
ch := make (chan * BloomCreation )
267
- go bt .Populate (& s , blooms , chks , ch )
245
+ go bt .Populate (blooms , chks , ch )
268
246
for x := range ch {
269
247
if x .Err != nil {
270
248
e = append (e , x .Err )
@@ -281,8 +259,6 @@ func BenchmarkPopulateSeriesWithBloom(b *testing.B) {
281
259
bt := NewBloomTokenizer (DefaultNGramLength , DefaultNGramSkip , 0 , metrics )
282
260
283
261
sbf := filter .NewScalableBloomFilter (1024 , 0.01 , 0.8 )
284
- var lbsList []labels.Labels
285
- lbsList = append (lbsList , labels .FromStrings ("foo" , "bar" ))
286
262
287
263
memChunk := chunkenc .NewMemChunk (chunkenc .ChunkFormatV4 , chunkenc .EncSnappy , chunkenc .ChunkHeadFormatFor (chunkenc .ChunkFormatV4 ), 256000 , 1500000 )
288
264
_ = memChunk .Append (& push.Entry {
@@ -301,12 +277,9 @@ func BenchmarkPopulateSeriesWithBloom(b *testing.B) {
301
277
bloom := Bloom {
302
278
ScalableBloomFilter : * sbf ,
303
279
}
304
- series := Series {
305
- Fingerprint : model .Fingerprint (lbsList [0 ].Hash ()),
306
- }
280
+
307
281
_ , err = populateAndConsumeBloom (
308
282
bt ,
309
- series ,
310
283
NewSliceIter ([]* Bloom {& bloom }),
311
284
NewSliceIter ([]ChunkRefWithIter {{Ref : ChunkRef {},
312
285
Itr : itr }}),
0 commit comments