package iter import ( "context" "io" "math" "sync" "time" "github.com/grafana/loki/pkg/logproto" "github.com/grafana/loki/pkg/logqlmodel/stats" "github.com/grafana/loki/pkg/util" "github.com/grafana/loki/pkg/util/loser" ) // EntryIterator iterates over entries in time-order. type EntryIterator interface { Iterator Entry() logproto.Entry } // streamIterator iterates over entries in a stream. type streamIterator struct { i int stream logproto.Stream } // NewStreamIterator iterates over entries in a stream. func NewStreamIterator(stream logproto.Stream) EntryIterator { return &streamIterator{ i: -1, stream: stream, } } func (i *streamIterator) Next() bool { i.i++ return i.i < len(i.stream.Entries) } func (i *streamIterator) Error() error { return nil } func (i *streamIterator) Labels() string { return i.stream.Labels } func (i *streamIterator) StreamHash() uint64 { return i.stream.Hash } func (i *streamIterator) Entry() logproto.Entry { return i.stream.Entries[i.i] } func (i *streamIterator) Close() error { return nil } // HeapIterator iterates over a heap of iterators with ability to push new iterators and get some properties like time of entry at peek and len // Not safe for concurrent use type HeapIterator interface { EntryIterator Peek() time.Time IsEmpty() bool Push(EntryIterator) } // mergeEntryIterator iterates over a heap of iterators and merge duplicate entries. type mergeEntryIterator struct { tree *loser.Tree[sortFields, EntryIterator] stats *stats.Context // buffer of entries to be returned by Next() // We buffer entries with the same timestamp to correctly dedupe them. buffer []entryWithLabels currEntry entryWithLabels errs []error } // NewMergeEntryIterator returns a new iterator which uses a heap to merge together entries for multiple iterators and deduplicate entries if any. // The iterator only order and merge entries across given `is` iterators, it does not merge entries within individual iterator. // This means using this iterator with a single iterator will result in the same result as the input iterator. // If you don't need to deduplicate entries, use `NewSortEntryIterator` instead. func NewMergeEntryIterator(ctx context.Context, is []EntryIterator, direction logproto.Direction) HeapIterator { maxVal, less := treeLess(direction) result := &mergeEntryIterator{stats: stats.FromContext(ctx)} result.tree = loser.New(is, maxVal, sortFieldsAt, less, result.closeEntry) result.buffer = make([]entryWithLabels, 0, len(is)) return result } func (i *mergeEntryIterator) closeEntry(e EntryIterator) { if err := e.Error(); err != nil { i.errs = append(i.errs, err) } util.LogError("closing iterator", e.Close) } func (i *mergeEntryIterator) Push(ei EntryIterator) { i.tree.Push(ei) } // Next fetches entries from the tree until it finds an entry with a different timestamp or stream hash. // Generally i.buffer has one or more entries with the same timestamp and stream hash, // followed by one more item where the timestamp or stream hash was different. func (i *mergeEntryIterator) Next() bool { if len(i.buffer) < 2 { i.fillBuffer() } if len(i.buffer) == 0 { return false } i.nextFromBuffer() return true } func (i *mergeEntryIterator) fillBuffer() { if !i.tree.Next() { return } // At this point we have zero or one items in i.buffer, and the next item is available from i.tree. // We support multiple entries with the same timestamp, and we want to // preserve their original order. // Entries with identical timestamp and line are removed as duplicates. for { next := i.tree.Winner() entry := next.Entry() i.buffer = append(i.buffer, entryWithLabels{ Entry: entry, labels: next.Labels(), streamHash: next.StreamHash(), }) if len(i.buffer) > 1 && (i.buffer[0].streamHash != next.StreamHash() || !i.buffer[0].Entry.Timestamp.Equal(entry.Timestamp)) { break } previous := i.buffer[:len(i.buffer)-1] var dupe bool for _, t := range previous { if t.Entry.Line == entry.Line { i.stats.AddDuplicates(1) dupe = true break } } if dupe { i.buffer = previous } if !i.tree.Next() { break } } } func (i *mergeEntryIterator) nextFromBuffer() { i.currEntry.Entry = i.buffer[0].Entry i.currEntry.labels = i.buffer[0].labels i.currEntry.streamHash = i.buffer[0].streamHash if len(i.buffer) == 2 { i.buffer[0] = i.buffer[1] i.buffer = i.buffer[:1] return } if len(i.buffer) == 1 { i.buffer = i.buffer[:0] return } i.buffer = i.buffer[1:] } func (i *mergeEntryIterator) Entry() logproto.Entry { return i.currEntry.Entry } func (i *mergeEntryIterator) Labels() string { return i.currEntry.labels } func (i *mergeEntryIterator) StreamHash() uint64 { return i.currEntry.streamHash } func (i *mergeEntryIterator) Error() error { switch len(i.errs) { case 0: return nil case 1: return i.errs[0] default: return util.MultiError(i.errs) } } func (i *mergeEntryIterator) Close() error { i.tree.Close() i.buffer = nil return i.Error() } func (i *mergeEntryIterator) Peek() time.Time { if len(i.buffer) == 0 { i.fillBuffer() } if len(i.buffer) == 0 { return time.Time{} } return i.buffer[0].Timestamp } // IsEmpty returns true if there are no more entries to pull. func (i *mergeEntryIterator) IsEmpty() bool { if len(i.buffer) == 0 { i.fillBuffer() } return len(i.buffer) == 0 } type entrySortIterator struct { tree *loser.Tree[sortFields, EntryIterator] currEntry entryWithLabels errs []error } // NewSortEntryIterator returns a new EntryIterator that sorts entries by timestamp (depending on the direction) the input iterators. // The iterator only order entries across given `is` iterators, it does not sort entries within individual iterator. // This means using this iterator with a single iterator will result in the same result as the input iterator. // When timestamp is equal, the iterator sorts samples by their label alphabetically. func NewSortEntryIterator(is []EntryIterator, direction logproto.Direction) EntryIterator { if len(is) == 0 { return NoopIterator } if len(is) == 1 { return is[0] } maxVal, less := treeLess(direction) result := &entrySortIterator{} result.tree = loser.New(is, maxVal, sortFieldsAt, less, result.closeEntry) return result } func treeLess(direction logproto.Direction) (maxVal sortFields, less func(a, b sortFields) bool) { switch direction { case logproto.BACKWARD: maxVal = sortFields{timeNanos: math.MinInt64} less = lessDescending case logproto.FORWARD: maxVal = sortFields{timeNanos: math.MaxInt64} less = lessAscending default: panic("bad direction") } return } type sortFields struct { labels string timeNanos int64 streamHash uint64 } func sortFieldsAt(i EntryIterator) sortFields { return sortFields{ timeNanos: i.Entry().Timestamp.UnixNano(), labels: i.Labels(), streamHash: i.StreamHash(), } } func lessAscending(e1, e2 sortFields) bool { if e1.timeNanos == e2.timeNanos { // The underlying stream hash may not be available, such as when merging LokiResponses in the // frontend which were sharded. Prefer to use the underlying stream hash when available, // which is needed in deduping code, but defer to label sorting when it's not present. if e1.streamHash == 0 { return e1.labels < e2.labels } return e1.streamHash < e2.streamHash } return e1.timeNanos < e2.timeNanos } func lessDescending(e1, e2 sortFields) bool { if e1.timeNanos == e2.timeNanos { if e1.streamHash == 0 { return e1.labels < e2.labels } return e1.streamHash < e2.streamHash } return e1.timeNanos > e2.timeNanos } func (i *entrySortIterator) closeEntry(e EntryIterator) { if err := e.Error(); err != nil { i.errs = append(i.errs, err) } util.LogError("closing iterator", e.Close) } func (i *entrySortIterator) Next() bool { ret := i.tree.Next() if !ret { return false } next := i.tree.Winner() i.currEntry.Entry = next.Entry() i.currEntry.labels = next.Labels() i.currEntry.streamHash = next.StreamHash() return true } func (i *entrySortIterator) Entry() logproto.Entry { return i.currEntry.Entry } func (i *entrySortIterator) Labels() string { return i.currEntry.labels } func (i *entrySortIterator) StreamHash() uint64 { return i.currEntry.streamHash } func (i *entrySortIterator) Error() error { switch len(i.errs) { case 0: return nil case 1: return i.errs[0] default: return util.MultiError(i.errs) } } func (i *entrySortIterator) Close() error { i.tree.Close() return i.Error() } // NewStreamsIterator returns an iterator over logproto.Stream func NewStreamsIterator(streams []logproto.Stream, direction logproto.Direction) EntryIterator { is := make([]EntryIterator, 0, len(streams)) for i := range streams { is = append(is, NewStreamIterator(streams[i])) } return NewSortEntryIterator(is, direction) } // NewQueryResponseIterator returns an iterator over a QueryResponse. func NewQueryResponseIterator(resp *logproto.QueryResponse, direction logproto.Direction) EntryIterator { return NewStreamsIterator(resp.Streams, direction) } type queryClientIterator struct { client logproto.Querier_QueryClient direction logproto.Direction err error curr EntryIterator } // NewQueryClientIterator returns an iterator over a QueryClient. func NewQueryClientIterator(client logproto.Querier_QueryClient, direction logproto.Direction) EntryIterator { return &queryClientIterator{ client: client, direction: direction, } } func (i *queryClientIterator) Next() bool { ctx := i.client.Context() for i.curr == nil || !i.curr.Next() { batch, err := i.client.Recv() if err == io.EOF { return false } else if err != nil { i.err = err return false } stats.JoinIngesters(ctx, batch.Stats) i.curr = NewQueryResponseIterator(batch, i.direction) } return true } func (i *queryClientIterator) Entry() logproto.Entry { return i.curr.Entry() } func (i *queryClientIterator) Labels() string { return i.curr.Labels() } func (i *queryClientIterator) StreamHash() uint64 { return i.curr.StreamHash() } func (i *queryClientIterator) Error() error { return i.err } func (i *queryClientIterator) Close() error { return i.client.CloseSend() } type nonOverlappingIterator struct { iterators []EntryIterator curr EntryIterator } // NewNonOverlappingIterator gives a chained iterator over a list of iterators. func NewNonOverlappingIterator(iterators []EntryIterator) EntryIterator { return &nonOverlappingIterator{ iterators: iterators, } } func (i *nonOverlappingIterator) Next() bool { for i.curr == nil || !i.curr.Next() { if len(i.iterators) == 0 { if i.curr != nil { i.curr.Close() } return false } if i.curr != nil { i.curr.Close() } i.curr, i.iterators = i.iterators[0], i.iterators[1:] } return true } func (i *nonOverlappingIterator) Entry() logproto.Entry { return i.curr.Entry() } func (i *nonOverlappingIterator) Labels() string { if i.curr == nil { return "" } return i.curr.Labels() } func (i *nonOverlappingIterator) StreamHash() uint64 { if i.curr == nil { return 0 } return i.curr.StreamHash() } func (i *nonOverlappingIterator) Error() error { if i.curr == nil { return nil } return i.curr.Error() } func (i *nonOverlappingIterator) Close() error { if i.curr != nil { i.curr.Close() } for _, iter := range i.iterators { iter.Close() } i.iterators = nil return nil } type timeRangedIterator struct { EntryIterator mint, maxt time.Time } // NewTimeRangedIterator returns an iterator which filters entries by time range. // Note: Only works with iterators that go forwards. func NewTimeRangedIterator(it EntryIterator, mint, maxt time.Time) EntryIterator { return &timeRangedIterator{ EntryIterator: it, mint: mint, maxt: maxt, } } func (i *timeRangedIterator) Next() bool { ok := i.EntryIterator.Next() if !ok { i.EntryIterator.Close() return ok } ts := i.EntryIterator.Entry().Timestamp for ok && i.mint.After(ts) { ok = i.EntryIterator.Next() if !ok { continue } ts = i.EntryIterator.Entry().Timestamp } if ok { if ts.Equal(i.mint) { // The mint is inclusive return true } if i.maxt.Before(ts) || i.maxt.Equal(ts) { // The maxt is exclusive. ok = false } } if !ok { i.EntryIterator.Close() } return ok } type entryWithLabels struct { logproto.Entry labels string streamHash uint64 } type reverseIterator struct { iter EntryIterator cur entryWithLabels entriesWithLabels []entryWithLabels loaded bool limit uint32 } // NewReversedIter returns an iterator which loads all or up to N entries // of an existing iterator, and then iterates over them backward. // Preload entries when they are being queried with a timeout. func NewReversedIter(it EntryIterator, limit uint32, preload bool) (EntryIterator, error) { iter, err := &reverseIterator{ iter: it, entriesWithLabels: make([]entryWithLabels, 0, 1024), limit: limit, }, it.Error() if err != nil { return nil, err } if preload { iter.load() } return iter, nil } func (i *reverseIterator) load() { if !i.loaded { i.loaded = true for count := uint32(0); (i.limit == 0 || count < i.limit) && i.iter.Next(); count++ { i.entriesWithLabels = append(i.entriesWithLabels, entryWithLabels{i.iter.Entry(), i.iter.Labels(), i.iter.StreamHash()}) } i.iter.Close() } } func (i *reverseIterator) Next() bool { i.load() if len(i.entriesWithLabels) == 0 { i.entriesWithLabels = nil return false } i.cur, i.entriesWithLabels = i.entriesWithLabels[len(i.entriesWithLabels)-1], i.entriesWithLabels[:len(i.entriesWithLabels)-1] return true } func (i *reverseIterator) Entry() logproto.Entry { return i.cur.Entry } func (i *reverseIterator) Labels() string { return i.cur.labels } func (i *reverseIterator) StreamHash() uint64 { return i.cur.streamHash } func (i *reverseIterator) Error() error { return nil } func (i *reverseIterator) Close() error { if !i.loaded { return i.iter.Close() } return nil } var entryBufferPool = sync.Pool{ New: func() interface{} { return &entryBuffer{ entries: make([]entryWithLabels, 0, 1024), } }, } type entryBuffer struct { entries []entryWithLabels } type reverseEntryIterator struct { iter EntryIterator cur entryWithLabels buf *entryBuffer loaded bool } // NewEntryReversedIter returns an iterator which loads all entries and iterates backward. // The labels of entries is always empty. func NewEntryReversedIter(it EntryIterator) (EntryIterator, error) { iter, err := &reverseEntryIterator{ iter: it, buf: entryBufferPool.Get().(*entryBuffer), }, it.Error() if err != nil { return nil, err } return iter, nil } func (i *reverseEntryIterator) load() { if !i.loaded { i.loaded = true for i.iter.Next() { i.buf.entries = append(i.buf.entries, entryWithLabels{i.iter.Entry(), i.iter.Labels(), i.iter.StreamHash()}) } i.iter.Close() } } func (i *reverseEntryIterator) Next() bool { i.load() if i.buf == nil || len(i.buf.entries) == 0 { i.release() return false } i.cur, i.buf.entries = i.buf.entries[len(i.buf.entries)-1], i.buf.entries[:len(i.buf.entries)-1] return true } func (i *reverseEntryIterator) Entry() logproto.Entry { return i.cur.Entry } func (i *reverseEntryIterator) Labels() string { return i.cur.labels } func (i *reverseEntryIterator) StreamHash() uint64 { return i.cur.streamHash } func (i *reverseEntryIterator) Error() error { return nil } func (i *reverseEntryIterator) release() { if i.buf == nil { return } if i.buf.entries != nil { // preserve the underlying slice before releasing to pool i.buf.entries = i.buf.entries[:0] } entryBufferPool.Put(i.buf) i.buf = nil } func (i *reverseEntryIterator) Close() error { i.release() if !i.loaded { return i.iter.Close() } return nil } // ReadBatch reads a set of entries off an iterator. func ReadBatch(i EntryIterator, size uint32) (*logproto.QueryResponse, uint32, error) { var ( streams = map[uint64]map[string]*logproto.Stream{} respSize uint32 streamsCount int ) for ; respSize < size && i.Next(); respSize++ { labels, hash, entry := i.Labels(), i.StreamHash(), i.Entry() mutatedStreams, ok := streams[hash] if !ok { mutatedStreams = map[string]*logproto.Stream{} streams[hash] = mutatedStreams } mutatedStream, ok := mutatedStreams[labels] if !ok { streamsCount++ mutatedStream = &logproto.Stream{ Labels: labels, Hash: hash, } mutatedStreams[labels] = mutatedStream } mutatedStream.Entries = append(mutatedStream.Entries, entry) } result := logproto.QueryResponse{ Streams: make([]logproto.Stream, 0, streamsCount), } for _, mutatedStreams := range streams { for _, s := range mutatedStreams { result.Streams = append(result.Streams, *s) } } return &result, respSize, i.Error() } type peekingEntryIterator struct { iter EntryIterator cache *entryWithLabels next *entryWithLabels } // PeekingEntryIterator is an entry iterator that can look ahead an entry // using `Peek` without advancing its cursor. type PeekingEntryIterator interface { EntryIterator Peek() (string, logproto.Entry, bool) } // NewPeekingIterator creates a new peeking iterator. func NewPeekingIterator(iter EntryIterator) PeekingEntryIterator { // initialize the next entry so we can peek right from the start. var cache *entryWithLabels next := &entryWithLabels{} if iter.Next() { cache = &entryWithLabels{ Entry: iter.Entry(), labels: iter.Labels(), streamHash: iter.StreamHash(), } next.Entry = cache.Entry next.labels = cache.labels } return &peekingEntryIterator{ iter: iter, cache: cache, next: next, } } // Next implements `EntryIterator` func (it *peekingEntryIterator) Next() bool { if it.cache != nil { it.next.Entry = it.cache.Entry it.next.labels = it.cache.labels it.next.streamHash = it.cache.streamHash it.cacheNext() return true } return false } // cacheNext caches the next element if it exists. func (it *peekingEntryIterator) cacheNext() { if it.iter.Next() { it.cache.Entry = it.iter.Entry() it.cache.labels = it.iter.Labels() it.cache.streamHash = it.iter.StreamHash() return } // nothing left removes the cached entry it.cache = nil } // Peek implements `PeekingEntryIterator` func (it *peekingEntryIterator) Peek() (string, logproto.Entry, bool) { if it.cache != nil { return it.cache.labels, it.cache.Entry, true } return "", logproto.Entry{}, false } // Labels implements `EntryIterator` func (it *peekingEntryIterator) Labels() string { if it.next != nil { return it.next.labels } return "" } func (it *peekingEntryIterator) StreamHash() uint64 { if it.next != nil { return it.next.streamHash } return 0 } // Entry implements `EntryIterator` func (it *peekingEntryIterator) Entry() logproto.Entry { if it.next != nil { return it.next.Entry } return logproto.Entry{} } // Error implements `EntryIterator` func (it *peekingEntryIterator) Error() error { return it.iter.Error() } // Close implements `EntryIterator` func (it *peekingEntryIterator) Close() error { return it.iter.Close() }