forked from grafana/tempo
-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathtraceid_sharder.go
More file actions
132 lines (115 loc) · 4.54 KB
/
traceid_sharder.go
File metadata and controls
132 lines (115 loc) · 4.54 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
package frontend
import (
"encoding/hex"
"net/http"
"github.com/go-kit/log" //nolint:all //deprecated
"github.com/grafana/tempo/modules/frontend/combiner"
"github.com/grafana/tempo/modules/frontend/pipeline"
"github.com/grafana/tempo/modules/querier"
"github.com/grafana/tempo/pkg/api"
"github.com/grafana/tempo/pkg/blockboundary"
"github.com/grafana/tempo/pkg/validation"
"github.com/prometheus/client_golang/prometheus"
)
const (
minQueryShards = 2
maxQueryShards = 100_000
)
type asyncTraceSharder struct {
next pipeline.AsyncRoundTripper[combiner.PipelineResponse]
cfg *TraceByIDConfig
logger log.Logger
blockBoundaries [][]byte
jobsPerQuery *prometheus.HistogramVec
}
func newAsyncTraceIDSharder(cfg *TraceByIDConfig, jobsPerQuery *prometheus.HistogramVec, logger log.Logger) pipeline.AsyncMiddleware[combiner.PipelineResponse] {
return pipeline.AsyncMiddlewareFunc[combiner.PipelineResponse](func(next pipeline.AsyncRoundTripper[combiner.PipelineResponse]) pipeline.AsyncRoundTripper[combiner.PipelineResponse] {
// Calculate block boundaries:
// - If external is enabled: N-2 block shards (1 ingester + 1 external + N-2 blocks = N total)
// - If external is disabled: N-1 block shards (1 ingester + N-1 blocks = N total)
numBlockShards := cfg.QueryShards - 1
if cfg.ExternalEnabled {
numBlockShards = cfg.QueryShards - 2
}
return asyncTraceSharder{
next: next,
cfg: cfg,
logger: logger,
blockBoundaries: blockboundary.CreateBlockBoundaries(numBlockShards),
jobsPerQuery: jobsPerQuery,
}
})
}
// RoundTrip implements http.RoundTripper
func (s asyncTraceSharder) RoundTrip(pipelineRequest pipeline.Request) (pipeline.Responses[combiner.PipelineResponse], error) {
ctx, span := tracer.Start(pipelineRequest.Context(), "frontend.ShardQuery")
defer span.End()
pipelineRequest.SetContext(ctx)
reqs, err := s.buildShardedRequests(pipelineRequest)
if err != nil {
return nil, err
}
s.jobsPerQuery.WithLabelValues(traceByIDOp).Observe(float64(len(reqs)))
// execute requests
concurrentShards := uint(s.cfg.QueryShards)
// if concurrent shards is set, respect that value
if s.cfg.ConcurrentShards > 0 {
concurrentShards = uint(s.cfg.ConcurrentShards)
}
// concurrent_shards grater then query_shards should not be allowed because it would create
// more goroutines then the jobs to send these jobs to queriers.
if concurrentShards > uint(s.cfg.QueryShards) {
// set the concurrent shards to the total shards
concurrentShards = uint(s.cfg.QueryShards)
}
return pipeline.NewAsyncSharderFunc(ctx, int(concurrentShards), len(reqs), func(i int) pipeline.Request {
pipelineReq := reqs[i]
return pipelineReq
}, s.next), nil
}
// buildShardedRequests returns a slice of requests sharded on the precalculated
// block boundaries
func (s *asyncTraceSharder) buildShardedRequests(parent pipeline.Request) ([]pipeline.Request, error) {
userID, err := validation.ExtractValidTenantID(parent.Context())
if err != nil {
return nil, err
}
reqs := make([]pipeline.Request, 0, s.cfg.QueryShards)
params := map[string]string{}
// Job 0: ingester job
req, err := cloneRequestforQueriers(parent, userID, func(r *http.Request) (*http.Request, error) {
params[querier.QueryModeKey] = querier.QueryModeIngesters
return api.BuildQueryRequest(r, params), nil
})
if err != nil {
return nil, err
}
reqs = append(reqs, req)
// Job 1: external job (if enabled)
if s.cfg.ExternalEnabled {
req, err = cloneRequestforQueriers(parent, userID, func(r *http.Request) (*http.Request, error) {
params[querier.QueryModeKey] = querier.QueryModeExternal
return api.BuildQueryRequest(r, params), nil
})
if err != nil {
return nil, err
}
reqs = append(reqs, req)
}
// Jobs 2 to N-1: block queries
// When external is enabled, we have N-2 block shards
// When external is disabled, we have N-1 block shards
// blockBoundaries has length equal to numBlockShards, and we create shards between boundaries
for i := 1; i < len(s.blockBoundaries); i++ {
i := i // save the loop variable locally to make sure the closure grabs the correct var.
pipelineR, _ := cloneRequestforQueriers(parent, userID, func(r *http.Request) (*http.Request, error) {
// block queries
params[querier.BlockStartKey] = hex.EncodeToString(s.blockBoundaries[i-1])
params[querier.BlockEndKey] = hex.EncodeToString(s.blockBoundaries[i])
params[querier.QueryModeKey] = querier.QueryModeBlocks
return api.BuildQueryRequest(r, params), nil
})
reqs = append(reqs, pipelineR)
}
return reqs, nil
}