package graphite import ( "fmt" "net/http" "regexp" "sort" "strings" "sync" "time" "github.com/VictoriaMetrics/VictoriaMetrics/app/vmselect/bufferedwriter" "github.com/VictoriaMetrics/VictoriaMetrics/app/vmselect/netstorage" "github.com/VictoriaMetrics/VictoriaMetrics/app/vmselect/searchutils" "github.com/VictoriaMetrics/VictoriaMetrics/lib/auth" "github.com/VictoriaMetrics/VictoriaMetrics/lib/storage" "github.com/VictoriaMetrics/metrics" ) // MetricsFindHandler implements /metrics/find handler. // // See https://graphite-api.readthedocs.io/en/latest/api.html#metrics-find func MetricsFindHandler(startTime time.Time, at *auth.Token, w http.ResponseWriter, r *http.Request) error { deadline := searchutils.GetDeadlineForQuery(r, startTime) if err := r.ParseForm(); err != nil { return fmt.Errorf("cannot parse form values: %w", err) } format := r.FormValue("format") if format == "" { format = "treejson" } switch format { case "treejson", "completer": default: return fmt.Errorf(`unexpected "format" query arg: %q; expecting "treejson" or "completer"`, format) } query := r.FormValue("query") if len(query) == 0 { return fmt.Errorf("expecting non-empty `query` arg") } delimiter := r.FormValue("delimiter") if delimiter == "" { delimiter = "." } if len(delimiter) > 1 { return fmt.Errorf("`delimiter` query arg must contain only a single char") } if searchutils.GetBool(r, "automatic_variants") { // See https://github.com/graphite-project/graphite-web/blob/bb9feb0e6815faa73f538af6ed35adea0fb273fd/webapp/graphite/metrics/views.py#L152 query = addAutomaticVariants(query, delimiter) } if format == "completer" { // See https://github.com/graphite-project/graphite-web/blob/bb9feb0e6815faa73f538af6ed35adea0fb273fd/webapp/graphite/metrics/views.py#L148 query = strings.ReplaceAll(query, "..", ".*") if !strings.HasSuffix(query, "*") { query += "*" } } leavesOnly := searchutils.GetBool(r, "leavesOnly") wildcards := searchutils.GetBool(r, "wildcards") label := r.FormValue("label") if label == "__name__" { label = "" } jsonp := r.FormValue("jsonp") from, err := searchutils.GetTime(r, "from", 0) if err != nil { return err } ct := startTime.UnixNano() / 1e6 until, err := searchutils.GetTime(r, "until", ct) if err != nil { return err } tr := storage.TimeRange{ MinTimestamp: from, MaxTimestamp: until, } paths, isPartial, err := metricsFind(at, tr, label, query, delimiter[0], false, deadline) if err != nil { return err } if isPartial && searchutils.GetDenyPartialResponse(r) { return fmt.Errorf("cannot return full response, since some of vmstorage nodes are unavailable") } if leavesOnly { paths = filterLeaves(paths, delimiter) } paths = deduplicatePaths(paths, delimiter) sortPaths(paths, delimiter) contentType := "application/json" if jsonp != "" { contentType = "text/javascript" } w.Header().Set("Content-Type", contentType) bw := bufferedwriter.Get(w) defer bufferedwriter.Put(bw) WriteMetricsFindResponse(bw, paths, delimiter, format, wildcards, jsonp) if err := bw.Flush(); err != nil { return err } metricsFindDuration.UpdateDuration(startTime) return nil } func deduplicatePaths(paths []string, delimiter string) []string { if len(paths) == 0 { return nil } sort.Strings(paths) dst := paths[:1] for _, path := range paths[1:] { prevPath := dst[len(dst)-1] if path == prevPath { // Skip duplicate path. continue } dst = append(dst, path) } return dst } // MetricsExpandHandler implements /metrics/expand handler. // // See https://graphite-api.readthedocs.io/en/latest/api.html#metrics-expand func MetricsExpandHandler(startTime time.Time, at *auth.Token, w http.ResponseWriter, r *http.Request) error { deadline := searchutils.GetDeadlineForQuery(r, startTime) if err := r.ParseForm(); err != nil { return fmt.Errorf("cannot parse form values: %w", err) } queries := r.Form["query"] if len(queries) == 0 { return fmt.Errorf("missing `query` arg") } groupByExpr := searchutils.GetBool(r, "groupByExpr") leavesOnly := searchutils.GetBool(r, "leavesOnly") label := r.FormValue("label") if label == "__name__" { label = "" } delimiter := r.FormValue("delimiter") if delimiter == "" { delimiter = "." } if len(delimiter) > 1 { return fmt.Errorf("`delimiter` query arg must contain only a single char") } jsonp := r.FormValue("jsonp") from, err := searchutils.GetTime(r, "from", 0) if err != nil { return err } ct := startTime.UnixNano() / 1e6 until, err := searchutils.GetTime(r, "until", ct) if err != nil { return err } tr := storage.TimeRange{ MinTimestamp: from, MaxTimestamp: until, } m := make(map[string][]string, len(queries)) for _, query := range queries { paths, isPartial, err := metricsFind(at, tr, label, query, delimiter[0], true, deadline) if err != nil { return err } if isPartial && searchutils.GetDenyPartialResponse(r) { return fmt.Errorf("cannot return full response, since some of vmstorage nodes are unavailable") } if leavesOnly { paths = filterLeaves(paths, delimiter) } m[query] = paths } contentType := "application/json" if jsonp != "" { contentType = "text/javascript" } w.Header().Set("Content-Type", contentType) if groupByExpr { for _, paths := range m { sortPaths(paths, delimiter) } WriteMetricsExpandResponseByQuery(w, m, jsonp) return nil } paths := m[queries[0]] if len(m) > 1 { pathsSet := make(map[string]struct{}) for _, paths := range m { for _, path := range paths { pathsSet[path] = struct{}{} } } paths = make([]string, 0, len(pathsSet)) for path := range pathsSet { paths = append(paths, path) } } sortPaths(paths, delimiter) bw := bufferedwriter.Get(w) defer bufferedwriter.Put(bw) WriteMetricsExpandResponseFlat(bw, paths, jsonp) if err := bw.Flush(); err != nil { return err } metricsExpandDuration.UpdateDuration(startTime) return nil } // MetricsIndexHandler implements /metrics/index.json handler. // // See https://graphite-api.readthedocs.io/en/latest/api.html#metrics-index-json func MetricsIndexHandler(startTime time.Time, at *auth.Token, w http.ResponseWriter, r *http.Request) error { deadline := searchutils.GetDeadlineForQuery(r, startTime) if err := r.ParseForm(); err != nil { return fmt.Errorf("cannot parse form values: %w", err) } jsonp := r.FormValue("jsonp") metricNames, isPartial, err := netstorage.GetLabelValues(at, "__name__", deadline) if err != nil { return fmt.Errorf(`cannot obtain metric names: %w`, err) } if isPartial && searchutils.GetDenyPartialResponse(r) { return fmt.Errorf("cannot return full response, since some of vmstorage nodes are unavailable") } contentType := "application/json" if jsonp != "" { contentType = "text/javascript" } w.Header().Set("Content-Type", contentType) bw := bufferedwriter.Get(w) defer bufferedwriter.Put(bw) WriteMetricsIndexResponse(bw, metricNames, jsonp) if err := bw.Flush(); err != nil { return err } metricsIndexDuration.UpdateDuration(startTime) return nil } // metricsFind searches for label values that match the given query. func metricsFind(at *auth.Token, tr storage.TimeRange, label, query string, delimiter byte, isExpand bool, deadline searchutils.Deadline) ([]string, bool, error) { n := strings.IndexAny(query, "*{[") if n < 0 || n == len(query)-1 && strings.HasSuffix(query, "*") { expandTail := n >= 0 if expandTail { query = query[:len(query)-1] } suffixes, isPartial, err := netstorage.GetTagValueSuffixes(at, tr, label, query, delimiter, deadline) if err != nil { return nil, false, err } if len(suffixes) == 0 { return nil, false, nil } if !expandTail && len(query) > 0 && query[len(query)-1] == delimiter { return []string{query}, false, nil } results := make([]string, 0, len(suffixes)) for _, suffix := range suffixes { if expandTail || len(suffix) == 0 || len(suffix) == 1 && suffix[0] == delimiter { results = append(results, query+suffix) } } return results, isPartial, nil } subquery := query[:n] + "*" paths, isPartial, err := metricsFind(at, tr, label, subquery, delimiter, isExpand, deadline) if err != nil { return nil, false, err } tail := "" suffix := query[n:] if m := strings.IndexByte(suffix, delimiter); m >= 0 { tail = suffix[m+1:] suffix = suffix[:m+1] } qPrefix := query[:n] + suffix rePrefix, err := getRegexpForQuery(qPrefix, delimiter) if err != nil { return nil, false, fmt.Errorf("cannot convert query %q to regexp: %w", qPrefix, err) } results := make([]string, 0, len(paths)) for _, path := range paths { if !rePrefix.MatchString(path) { continue } if tail == "" { results = append(results, path) continue } subquery := path + tail fullPaths, isPartialLocal, err := metricsFind(at, tr, label, subquery, delimiter, isExpand, deadline) if err != nil { return nil, false, err } if isPartialLocal { isPartial = true } if isExpand { results = append(results, fullPaths...) } else { for _, fullPath := range fullPaths { results = append(results, qPrefix+fullPath[len(path):]) } } } return results, isPartial, nil } var ( metricsFindDuration = metrics.NewSummary(`vm_request_duration_seconds{path="/select/{}/graphite/metrics/find"}`) metricsExpandDuration = metrics.NewSummary(`vm_request_duration_seconds{path="/select/{}/graphite/metrics/expand"}`) metricsIndexDuration = metrics.NewSummary(`vm_request_duration_seconds{path="/select/{}/graphite/metrics/index.json"}`) ) func addAutomaticVariants(query, delimiter string) string { // See https://github.com/graphite-project/graphite-web/blob/bb9feb0e6815faa73f538af6ed35adea0fb273fd/webapp/graphite/metrics/views.py#L152 parts := strings.Split(query, delimiter) for i, part := range parts { if strings.Contains(part, ",") && !strings.Contains(part, "{") { parts[i] = "{" + part + "}" } } return strings.Join(parts, delimiter) } func filterLeaves(paths []string, delimiter string) []string { leaves := paths[:0] for _, path := range paths { if !strings.HasSuffix(path, delimiter) { leaves = append(leaves, path) } } return leaves } func sortPaths(paths []string, delimiter string) { sort.Slice(paths, func(i, j int) bool { a, b := paths[i], paths[j] isNodeA := strings.HasSuffix(a, delimiter) isNodeB := strings.HasSuffix(b, delimiter) if isNodeA == isNodeB { return a < b } return isNodeA }) } func getRegexpForQuery(query string, delimiter byte) (*regexp.Regexp, error) { regexpCacheLock.Lock() defer regexpCacheLock.Unlock() k := regexpCacheKey{ query: query, delimiter: delimiter, } if re := regexpCache[k]; re != nil { return re.re, re.err } a := make([]string, 0, len(query)) quotedDelimiter := regexp.QuoteMeta(string([]byte{delimiter})) tillNextDelimiter := "[^" + quotedDelimiter + "]*" for i := 0; i < len(query); i++ { switch query[i] { case '*': a = append(a, tillNextDelimiter) case '{': tmp := query[i+1:] if n := strings.IndexByte(tmp, '}'); n < 0 { a = append(a, regexp.QuoteMeta(query[i:])) i = len(query) } else { a = append(a, "(?:") opts := strings.Split(tmp[:n], ",") for j, opt := range opts { opts[j] = regexp.QuoteMeta(opt) } a = append(a, strings.Join(opts, "|")) a = append(a, ")") i += n + 1 } case '[': tmp := query[i:] if n := strings.IndexByte(tmp, ']'); n < 0 { a = append(a, regexp.QuoteMeta(query[i:])) i = len(query) } else { a = append(a, tmp[:n+1]) i += n } default: a = append(a, regexp.QuoteMeta(query[i:i+1])) } } s := strings.Join(a, "") if !strings.HasSuffix(s, quotedDelimiter) { s += quotedDelimiter + "?" } s = "^(?:" + s + ")$" re, err := regexp.Compile(s) regexpCache[k] = ®expCacheEntry{ re: re, err: err, } if len(regexpCache) >= maxRegexpCacheSize { for k := range regexpCache { if len(regexpCache) < maxRegexpCacheSize { break } delete(regexpCache, k) } } return re, err } type regexpCacheEntry struct { re *regexp.Regexp err error } type regexpCacheKey struct { query string delimiter byte } var regexpCache = make(map[regexpCacheKey]*regexpCacheEntry) var regexpCacheLock sync.Mutex const maxRegexpCacheSize = 10000