Like Prometheus, but for logs.
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
 
 
 
loki/pkg/loghttp/query_test.go

394 lines
11 KiB

package loghttp
import (
"net/http"
"net/url"
"reflect"
"testing"
"time"
jsoniter "github.com/json-iterator/go"
"github.com/prometheus/prometheus/model/labels"
"github.com/stretchr/testify/require"
"github.com/grafana/loki/v3/pkg/logproto"
"github.com/grafana/loki/v3/pkg/logqlmodel/stats"
)
func TestParseRangeQuery(t *testing.T) {
t.Parallel()
tests := []struct {
name string
r *http.Request
want *RangeQuery
wantErr bool
}{
{"bad start", &http.Request{URL: mustParseURL(`?query={foo="bar"}&start=t`)}, nil, true},
{"bad end", &http.Request{URL: mustParseURL(`?query={foo="bar"}&end=t`)}, nil, true},
{"end before start", &http.Request{URL: mustParseURL(`?query={foo="bar"}&start=2016-06-10T21:42:24.760738998Z&end=2015-06-10T21:42:24.760738998Z`)}, nil, true},
{"bad limit", &http.Request{URL: mustParseURL(`?query={foo="bar"}&start=2016-06-10T21:42:24.760738998Z&end=2017-06-10T21:42:24.760738998Z&limit=h`)}, nil, true},
{
"bad direction",
&http.Request{
URL: mustParseURL(`?query={foo="bar"}&start=2016-06-10T21:42:24.760738998Z&end=2017-06-10T21:42:24.760738998Z&limit=100&direction=fw`),
}, nil, true,
},
{
"bad step",
&http.Request{
URL: mustParseURL(`?query={foo="bar"}&start=2016-06-10T21:42:24.760738998Z&end=2017-06-10T21:42:24.760738998Z&limit=100&direction=FORWARD&step=h`),
}, nil, true,
},
{
"negative step",
&http.Request{
URL: mustParseURL(`?query={foo="bar"}&start=2016-06-10T21:42:24.760738998Z&end=2017-06-10T21:42:24.760738998Z&limit=100&direction=BACKWARD&step=-1`),
}, nil, true,
},
{
"too small step",
&http.Request{
URL: mustParseURL(`?query={foo="bar"}&start=2016-06-10T21:42:24.760738998Z&end=2017-06-10T21:42:24.760738998Z&limit=100&direction=BACKWARD&step=1`),
}, nil, true,
},
{
"negative interval",
&http.Request{
URL: mustParseURL(`?query={foo="bar"}&start=2016-06-10T21:42:24.760738998Z&end=2017-06-10T21:42:24.760738998Z&limit=100&direction=BACKWARD&step=1,interval=-1`),
}, nil, true,
},
{
"good",
&http.Request{
URL: mustParseURL(`?query={foo="bar"}&start=2017-06-10T21:42:24.760738998Z&end=2017-07-10T21:42:24.760738998Z&limit=1000&direction=BACKWARD&step=3600`),
}, &RangeQuery{
Step: time.Hour,
Query: `{foo="bar"}`,
Direction: logproto.BACKWARD,
Start: time.Date(2017, 06, 10, 21, 42, 24, 760738998, time.UTC),
End: time.Date(2017, 07, 10, 21, 42, 24, 760738998, time.UTC),
Limit: 1000,
}, false,
},
}
for _, tt := range tests {
t.Run(tt.name, func(t *testing.T) {
err := tt.r.ParseForm()
require.Nil(t, err)
got, err := ParseRangeQuery(tt.r)
if (err != nil) != tt.wantErr {
t.Errorf("ParseRangeQuery() error = %v, wantErr %v", err, tt.wantErr)
return
}
if !reflect.DeepEqual(got, tt.want) {
t.Errorf("ParseRangeQuery() = %v, want %v", got, tt.want)
}
})
}
}
func TestParseInstantQuery(t *testing.T) {
tests := []struct {
name string
r *http.Request
want *InstantQuery
wantErr bool
}{
{"bad time", &http.Request{URL: mustParseURL(`?query={foo="bar"}&time=t`)}, nil, true},
{"bad limit", &http.Request{URL: mustParseURL(`?query={foo="bar"}&time=2016-06-10T21:42:24.760738998Z&limit=h`)}, nil, true},
{
"bad direction",
&http.Request{
URL: mustParseURL(`?query={foo="bar"}&time=2016-06-10T21:42:24.760738998Z&limit=100&direction=fw`),
}, nil, true,
},
{
"good",
&http.Request{
URL: mustParseURL(`?query={foo="bar"}&time=2017-06-10T21:42:24.760738998Z&limit=1000&direction=BACKWARD`),
}, &InstantQuery{
Query: `{foo="bar"}`,
Direction: logproto.BACKWARD,
Ts: time.Date(2017, 06, 10, 21, 42, 24, 760738998, time.UTC),
Limit: 1000,
}, false,
},
}
for _, tt := range tests {
t.Run(tt.name, func(t *testing.T) {
err := tt.r.ParseForm()
require.Nil(t, err)
got, err := ParseInstantQuery(tt.r)
if (err != nil) != tt.wantErr {
t.Errorf("ParseInstantQuery() error = %v, wantErr %v", err, tt.wantErr)
return
}
if !reflect.DeepEqual(got, tt.want) {
t.Errorf("ParseInstantQuery() = %v, want %v", got, tt.want)
}
})
}
}
func mustParseURL(u string) *url.URL {
url, err := url.Parse(u)
if err != nil {
panic(err)
}
return url
}
func TestStreams_ToProto(t *testing.T) {
tests := []struct {
name string
s Streams
want []logproto.Stream
}{
{"empty", nil, nil},
{
"some",
[]Stream{
{
Labels: map[string]string{"job": "fake"},
Entries: []Entry{
{Timestamp: time.Unix(0, 1), Line: "1"},
{Timestamp: time.Unix(0, 2), Line: "2", StructuredMetadata: labels.FromStrings("foo", "a", "bar", "b")},
},
},
{
Labels: map[string]string{"job": "fake", "lvl": "error"},
Entries: []Entry{
{Timestamp: time.Unix(0, 3), Line: "3"},
{Timestamp: time.Unix(0, 4), Line: "4", StructuredMetadata: labels.FromStrings("foo", "a", "bar", "b")},
},
},
},
[]logproto.Stream{
{
Labels: `{job="fake"}`,
Entries: []logproto.Entry{
{Timestamp: time.Unix(0, 1), Line: "1"},
{Timestamp: time.Unix(0, 2), Line: "2", StructuredMetadata: []logproto.LabelAdapter{
{Name: "bar", Value: "b"},
{Name: "foo", Value: "a"},
}},
},
},
{
Labels: `{job="fake", lvl="error"}`,
Entries: []logproto.Entry{
{Timestamp: time.Unix(0, 3), Line: "3"},
{Timestamp: time.Unix(0, 4), Line: "4", StructuredMetadata: []logproto.LabelAdapter{
{Name: "bar", Value: "b"},
{Name: "foo", Value: "a"},
}},
},
},
},
},
}
for _, tt := range tests {
t.Run(tt.name, func(t *testing.T) {
if got := tt.s.ToProto(); !reflect.DeepEqual(got, tt.want) {
t.Errorf("Streams.ToProto() = %v, want %v", got, tt.want)
}
})
}
}
func Test_QueryResponseUnmarshal(t *testing.T) {
for _, tt := range []QueryResponse{
{
Status: "ok",
Data: QueryResponseData{
ResultType: "streams",
Result: Streams{},
Statistics: stats.Result{},
},
},
{
Status: "ok",
Data: QueryResponseData{
ResultType: "streams",
Result: Streams{
Stream{
Labels: LabelSet{"foo": "bar"},
Entries: []Entry{
{Timestamp: time.Unix(0, 1), Line: "1"},
{Timestamp: time.Unix(0, 2), Line: "2", StructuredMetadata: labels.FromStrings("foo", "a", "bar", "b")},
},
},
},
Statistics: stats.Result{
Summary: stats.Summary{
BytesProcessedPerSecond: 1238,
},
},
},
},
{
Status: "ok",
Data: QueryResponseData{
ResultType: "streams",
Result: Streams{
Stream{
Labels: LabelSet{"foo": "bar"},
Entries: []Entry{
{Timestamp: time.Unix(0, 1), Line: "log line 1"},
{Timestamp: time.Unix(0, 2), Line: "some log line 2", StructuredMetadata: labels.FromStrings("foo", "a", "bar", "b")},
},
},
Stream{
Labels: LabelSet{"bar": "buzz", "level": "err", "foo": "bar"},
Entries: []Entry{
{Timestamp: time.Unix(0, 1), Line: "1"},
{Timestamp: time.Unix(0, 2), Line: "2"},
{Timestamp: time.Unix(0, 2), Line: "2"},
{Timestamp: time.Unix(0, 2), Line: "2"},
{Timestamp: time.Unix(0, 2), Line: "2", StructuredMetadata: labels.FromStrings("foo", "a", "bar", "b")},
},
},
},
Statistics: stats.Result{
Summary: stats.Summary{
BytesProcessedPerSecond: 1238,
},
Querier: stats.Querier{
Store: stats.Store{
TotalChunksRef: 1234123123,
},
},
},
},
},
} {
t.Run("", func(t *testing.T) {
b, err := jsoniter.Marshal(tt)
require.Nil(t, err)
var actual QueryResponse
err = actual.UnmarshalJSON(b)
require.Nil(t, err)
require.Equal(t, tt, actual)
})
}
}
func Test_ParseVolumeInstantQuery(t *testing.T) {
req := &http.Request{
URL: mustParseURL(`?query={foo="bar"}` +
`&start=2017-06-10T21:42:24.760738998Z` +
`&end=2017-07-10T21:42:24.760738998Z` +
`&limit=1000` +
`&targetLabels=foo,bar`,
),
}
err := req.ParseForm()
require.NoError(t, err)
actual, err := ParseVolumeInstantQuery(req)
require.NoError(t, err)
expected := &VolumeInstantQuery{
Start: time.Date(2017, 06, 10, 21, 42, 24, 760738998, time.UTC),
End: time.Date(2017, 07, 10, 21, 42, 24, 760738998, time.UTC),
Query: `{foo="bar"}`,
Limit: 1000,
TargetLabels: []string{"foo", "bar"},
AggregateBy: "series",
}
require.Equal(t, expected, actual)
t.Run("aggregate by", func(t *testing.T) {
url := `?query={foo="bar"}` +
`&start=2017-06-10T21:42:24.760738998Z` +
`&end=2017-07-10T21:42:24.760738998Z` +
`&limit=1000` +
`&step=3600` +
`&targetLabels=foo,bar`
t.Run("labels", func(t *testing.T) {
req := &http.Request{URL: mustParseURL(url + `&aggregateBy=labels`)}
err := req.ParseForm()
require.NoError(t, err)
actual, err := ParseVolumeInstantQuery(req)
require.NoError(t, err)
require.Equal(t, "labels", actual.AggregateBy)
})
t.Run("invalid", func(t *testing.T) {
req := &http.Request{URL: mustParseURL(url + `&aggregateBy=invalid`)}
err := req.ParseForm()
require.NoError(t, err)
_, err = ParseVolumeInstantQuery(req)
require.EqualError(t, err, "invalid aggregation option")
})
})
}
func Test_ParseVolumeRangeQuery(t *testing.T) {
req := &http.Request{
URL: mustParseURL(`?query={foo="bar"}` +
`&start=2017-06-10T21:42:24.760738998Z` +
`&end=2017-07-10T21:42:24.760738998Z` +
`&limit=1000` +
`&step=3600` +
`&targetLabels=foo,bar`,
),
}
err := req.ParseForm()
require.NoError(t, err)
actual, err := ParseVolumeRangeQuery(req)
require.NoError(t, err)
expected := &VolumeRangeQuery{
Start: time.Date(2017, 06, 10, 21, 42, 24, 760738998, time.UTC),
End: time.Date(2017, 07, 10, 21, 42, 24, 760738998, time.UTC),
Query: `{foo="bar"}`,
Limit: 1000,
Step: time.Hour,
TargetLabels: []string{"foo", "bar"},
AggregateBy: "series",
}
require.Equal(t, expected, actual)
t.Run("aggregate by", func(t *testing.T) {
url := `?query={foo="bar"}` +
`&start=2017-06-10T21:42:24.760738998Z` +
`&end=2017-07-10T21:42:24.760738998Z` +
`&limit=1000` +
`&step=3600` +
`&targetLabels=foo,bar`
t.Run("labels", func(t *testing.T) {
req := &http.Request{URL: mustParseURL(url + `&aggregateBy=labels`)}
err := req.ParseForm()
require.NoError(t, err)
actual, err := ParseVolumeRangeQuery(req)
require.NoError(t, err)
require.Equal(t, "labels", actual.AggregateBy)
})
t.Run("invalid", func(t *testing.T) {
req := &http.Request{URL: mustParseURL(url + `&aggregateBy=invalid`)}
err := req.ParseForm()
require.NoError(t, err)
_, err = ParseVolumeRangeQuery(req)
require.EqualError(t, err, "invalid aggregation option")
})
})
}