Like Prometheus, but for logs.
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
 
 
 
loki/pkg/queue/dequeue_qos_test.go

178 lines
5.0 KiB

package queue
import (
"context"
"fmt"
"strings"
"sync"
"testing"
"time"
"github.com/stretchr/testify/require"
"go.uber.org/atomic"
"github.com/grafana/loki/v3/pkg/util/constants"
)
const (
numRequestsPerActor = 1
numQueriers = 5
)
type req struct {
duration time.Duration
tenant string
actor string
queryID int
subQueryID int
}
func enqueueRequestsForActor(t testing.TB, actor []string, useActor bool, queue *RequestQueue, numSubRequests int, d time.Duration) {
tenant := "tenant"
serializedActor := strings.Join(actor, "|")
for x := 0; x < numRequestsPerActor; x++ {
for y := 0; y < numSubRequests; y++ {
r := &req{
duration: d,
queryID: x,
subQueryID: y,
tenant: tenant,
actor: serializedActor,
}
if !useActor {
actor = nil
}
err := queue.Enqueue("tenant", actor, r, nil)
if err != nil {
t.Fatal(err)
}
}
}
}
func BenchmarkQueryFairness(t *testing.B) {
numSubRequestsActorA, numSubRequestsActorB := 123, 45
total := int64((numSubRequestsActorA + numSubRequestsActorA + numSubRequestsActorB) * numRequestsPerActor)
for _, useActor := range []bool{false, true} {
t.Run(fmt.Sprintf("use hierarchical queues = %v", useActor), func(t *testing.B) {
requestQueue := NewRequestQueue(1024, 0, noQueueLimits, NewMetrics(nil, constants.Loki, "query_scheduler"))
enqueueRequestsForActor(t, []string{}, useActor, requestQueue, numSubRequestsActorA, 50*time.Millisecond)
enqueueRequestsForActor(t, []string{"a"}, useActor, requestQueue, numSubRequestsActorA, 100*time.Millisecond)
enqueueRequestsForActor(t, []string{"b"}, useActor, requestQueue, numSubRequestsActorB, 50*time.Millisecond)
requestQueue.queues.recomputeUserConsumers()
// set timeout to minize impact on overall test run duration in case something goes wrong
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
defer cancel()
start := time.Now()
durations := &sync.Map{}
var wg sync.WaitGroup
var responseCount atomic.Int64
// Simulate querier loop
for q := 0; q < numQueriers; q++ {
wg.Add(1)
go func(id string) {
defer wg.Done()
requestQueue.RegisterConsumerConnection(id)
defer requestQueue.UnregisterConsumerConnection(id)
idx := StartIndex
for ctx.Err() == nil {
r, newIdx, err := requestQueue.Dequeue(ctx, idx, id)
if err != nil {
if err != context.Canceled {
t.Log("Dequeue() returned error:", err)
}
break
}
if r == nil {
t.Log("Dequeue() returned nil response")
break
}
res, _ := r.(*req)
idx = newIdx
time.Sleep(res.duration)
count := responseCount.Add(1)
durations.Store(res.actor, time.Since(start))
if count == total {
t.Log("count", count, "total", total)
cancel()
}
}
}(fmt.Sprintf("querier-%d", q))
}
wg.Wait()
require.Equal(t, total, responseCount.Load())
durations.Range(func(k, v any) bool {
t.Log("duration actor", k, v)
return true
})
t.Log("total duration", time.Since(start))
})
}
}
func TestQueryFairnessAcrossSameLevel(t *testing.T) {
/**
`tenant1`, `tenant1|abc`, and `tenant1|xyz` have equal preference
`tenant1|xyz|123` and `tenant1|xyz|456` have equal preference
root:
tenant1: [0, 1, 2]
abc: [10, 11, 12]
xyz: [20, 21, 22]
123: [200]
456: [210]
**/
requestQueue := NewRequestQueue(1024, 0, noQueueLimits, NewMetrics(nil, constants.Loki, "query_scheduler"))
_ = requestQueue.Enqueue("tenant1", []string{}, r(0), nil)
_ = requestQueue.Enqueue("tenant1", []string{}, r(1), nil)
_ = requestQueue.Enqueue("tenant1", []string{}, r(2), nil)
_ = requestQueue.Enqueue("tenant1", []string{"abc"}, r(10), nil)
_ = requestQueue.Enqueue("tenant1", []string{"abc"}, r(11), nil)
_ = requestQueue.Enqueue("tenant1", []string{"abc"}, r(12), nil)
_ = requestQueue.Enqueue("tenant1", []string{"xyz"}, r(20), nil)
_ = requestQueue.Enqueue("tenant1", []string{"xyz"}, r(21), nil)
_ = requestQueue.Enqueue("tenant1", []string{"xyz"}, r(22), nil)
_ = requestQueue.Enqueue("tenant1", []string{"xyz", "123"}, r(200), nil)
_ = requestQueue.Enqueue("tenant1", []string{"xyz", "456"}, r(210), nil)
requestQueue.queues.recomputeUserConsumers()
items := make([]int, 0)
// set timeout to minize impact on overall test run duration in case something goes wrong
ctx, cancel := context.WithTimeout(context.Background(), 3*time.Second)
defer cancel()
requestQueue.RegisterConsumerConnection("querier")
defer requestQueue.UnregisterConsumerConnection("querier")
idx := StartIndexWithLocalQueue
for ctx.Err() == nil {
r, newIdx, err := requestQueue.Dequeue(ctx, idx, "querier")
if err != nil {
if err != context.Canceled {
t.Log("Dequeue() returned error:", err)
}
break
}
if r == nil {
t.Log("Dequeue() returned nil response")
break
}
res, _ := r.(*dummyRequest)
idx = newIdx
items = append(items, res.id)
}
require.Equal(t, []int{0, 10, 20, 1, 11, 200, 2, 12, 210, 21, 22}, items)
}