package cloudmonitoring import ( "context" "encoding/json" "errors" "fmt" "io/ioutil" "math" "net/http" "net/url" "path" "regexp" "strconv" "strings" "time" "github.com/grafana/grafana/pkg/api/pluginproxy" "github.com/grafana/grafana/pkg/components/null" "github.com/grafana/grafana/pkg/components/simplejson" "github.com/grafana/grafana/pkg/infra/log" "github.com/grafana/grafana/pkg/models" "github.com/grafana/grafana/pkg/plugins" "github.com/grafana/grafana/pkg/setting" "github.com/grafana/grafana/pkg/tsdb" "github.com/grafana/grafana/pkg/tsdb/sqleng" "github.com/opentracing/opentracing-go" "golang.org/x/net/context/ctxhttp" "golang.org/x/oauth2/google" ) var ( slog log.Logger ) var ( matchAllCap = regexp.MustCompile("(.)([A-Z][a-z]*)") legendKeyFormat = regexp.MustCompile(`\{\{\s*(.+?)\s*\}\}`) metricNameFormat = regexp.MustCompile(`([\w\d_]+)\.(googleapis\.com|io)/(.+)`) wildcardRegexRe = regexp.MustCompile(`[-\/^$+?.()|[\]{}]`) alignmentPeriodRe = regexp.MustCompile("[0-9]+") ) const ( gceAuthentication string = "gce" jwtAuthentication string = "jwt" metricQueryType string = "metrics" sloQueryType string = "slo" ) // CloudMonitoringExecutor executes queries for the CloudMonitoring datasource type CloudMonitoringExecutor struct { httpClient *http.Client dsInfo *models.DataSource } // NewCloudMonitoringExecutor initializes a http client func NewCloudMonitoringExecutor(dsInfo *models.DataSource) (tsdb.TsdbQueryEndpoint, error) { httpClient, err := dsInfo.GetHttpClient() if err != nil { return nil, err } return &CloudMonitoringExecutor{ httpClient: httpClient, dsInfo: dsInfo, }, nil } func init() { slog = log.New("tsdb.cloudMonitoring") tsdb.RegisterTsdbQueryEndpoint("stackdriver", NewCloudMonitoringExecutor) } // Query takes in the frontend queries, parses them into the CloudMonitoring query format // executes the queries against the CloudMonitoring API and parses the response into // the time series or table format func (e *CloudMonitoringExecutor) Query(ctx context.Context, dsInfo *models.DataSource, tsdbQuery *tsdb.TsdbQuery) (*tsdb.Response, error) { var result *tsdb.Response var err error queryType := tsdbQuery.Queries[0].Model.Get("type").MustString("") switch queryType { case "annotationQuery": result, err = e.executeAnnotationQuery(ctx, tsdbQuery) case "getGCEDefaultProject": result, err = e.getGCEDefaultProject(ctx, tsdbQuery) case "timeSeriesQuery": fallthrough default: result, err = e.executeTimeSeriesQuery(ctx, tsdbQuery) } return result, err } func (e *CloudMonitoringExecutor) getGCEDefaultProject(ctx context.Context, tsdbQuery *tsdb.TsdbQuery) (*tsdb.Response, error) { result := &tsdb.Response{ Results: make(map[string]*tsdb.QueryResult), } refId := tsdbQuery.Queries[0].RefId queryResult := &tsdb.QueryResult{Meta: simplejson.New(), RefId: refId} gceDefaultProject, err := e.getDefaultProject(ctx) if err != nil { return nil, fmt.Errorf("failed to retrieve default project from GCE metadata server, error: %w", err) } queryResult.Meta.Set("defaultProject", gceDefaultProject) result.Results[refId] = queryResult return result, nil } func (query *cloudMonitoringQuery) isSLO() bool { return query.Slo != "" } func (query *cloudMonitoringQuery) buildDeepLink() string { if query.isSLO() { return "" } filter := query.Params.Get("filter") if !strings.Contains(filter, "resource.type=") { resourceType := query.Params.Get("resourceType") if resourceType == "" { slog.Error("Failed to generate deep link: no resource type found", "ProjectName", query.ProjectName, "query", query.RefID) return "" } filter = fmt.Sprintf(`resource.type="%s" %s`, resourceType, filter) } u, err := url.Parse("https://console.cloud.google.com/monitoring/metrics-explorer") if err != nil { slog.Error("Failed to generate deep link: unable to parse metrics explorer URL", "ProjectName", query.ProjectName, "query", query.RefID) return "" } q := u.Query() q.Set("project", query.ProjectName) q.Set("Grafana_deeplink", "true") pageState := map[string]interface{}{ "xyChart": map[string]interface{}{ "constantLines": []string{}, "dataSets": []map[string]interface{}{ { "timeSeriesFilter": map[string]interface{}{ "aggregations": []string{}, "crossSeriesReducer": query.Params.Get("aggregation.crossSeriesReducer"), "filter": filter, "groupByFields": query.Params["aggregation.groupByFields"], "minAlignmentPeriod": strings.TrimPrefix(query.Params.Get("aggregation.alignmentPeriod"), "+"), // get rid of leading + "perSeriesAligner": query.Params.Get("aggregation.perSeriesAligner"), "secondaryGroupByFields": []string{}, "unitOverride": "1", }, }, }, "timeshiftDuration": "0s", "y1Axis": map[string]string{ "label": "y1Axis", "scale": "LINEAR", }, }, "timeSelection": map[string]string{ "timeRange": "custom", "start": query.Params.Get("interval.startTime"), "end": query.Params.Get("interval.endTime"), }, } blob, err := json.Marshal(pageState) if err != nil { slog.Error("Failed to generate deep link", "pageState", pageState, "ProjectName", query.ProjectName, "query", query.RefID) return "" } q.Set("pageState", string(blob)) u.RawQuery = q.Encode() accountChooserURL, err := url.Parse("https://accounts.google.com/AccountChooser") if err != nil { slog.Error("Failed to generate deep link: unable to parse account chooser URL", "ProjectName", query.ProjectName, "query", query.RefID) return "" } accountChooserQuery := accountChooserURL.Query() accountChooserQuery.Set("continue", u.String()) accountChooserURL.RawQuery = accountChooserQuery.Encode() return accountChooserURL.String() } func (e *CloudMonitoringExecutor) executeTimeSeriesQuery(ctx context.Context, tsdbQuery *tsdb.TsdbQuery) (*tsdb.Response, error) { result := &tsdb.Response{ Results: make(map[string]*tsdb.QueryResult), } queries, err := e.buildQueries(tsdbQuery) if err != nil { return nil, err } for _, query := range queries { queryRes, resp, err := e.executeQuery(ctx, query, tsdbQuery) if err != nil { return nil, err } err = e.parseResponse(queryRes, resp, query) if err != nil { queryRes.Error = err } result.Results[query.RefID] = queryRes resourceType := "" for _, s := range resp.TimeSeries { resourceType = s.Resource.Type // set the first resource type found break } query.Params.Set("resourceType", resourceType) dl := "" if len(resp.TimeSeries) > 0 { dl = query.buildDeepLink() } queryRes.Meta.Set("deepLink", dl) } return result, nil } func (e *CloudMonitoringExecutor) buildQueries(tsdbQuery *tsdb.TsdbQuery) ([]*cloudMonitoringQuery, error) { cloudMonitoringQueries := []*cloudMonitoringQuery{} startTime, err := tsdbQuery.TimeRange.ParseFrom() if err != nil { return nil, err } endTime, err := tsdbQuery.TimeRange.ParseTo() if err != nil { return nil, err } durationSeconds := int(endTime.Sub(startTime).Seconds()) for _, query := range tsdbQuery.Queries { migrateLegacyQueryModel(query) q := grafanaQuery{} model, _ := query.Model.MarshalJSON() if err := json.Unmarshal(model, &q); err != nil { return nil, fmt.Errorf("could not unmarshal CloudMonitoringQuery json: %w", err) } var target string params := url.Values{} params.Add("interval.startTime", startTime.UTC().Format(time.RFC3339)) params.Add("interval.endTime", endTime.UTC().Format(time.RFC3339)) sq := &cloudMonitoringQuery{ RefID: query.RefId, GroupBys: []string{}, } if q.QueryType == metricQueryType { sq.AliasBy = q.MetricQuery.AliasBy sq.GroupBys = append(sq.GroupBys, q.MetricQuery.GroupBys...) sq.ProjectName = q.MetricQuery.ProjectName if q.MetricQuery.View == "" { q.MetricQuery.View = "FULL" } params.Add("filter", buildFilterString(q.MetricQuery.MetricType, q.MetricQuery.Filters)) params.Add("view", q.MetricQuery.View) setMetricAggParams(¶ms, &q.MetricQuery, durationSeconds, query.IntervalMs) } else if q.QueryType == sloQueryType { sq.AliasBy = q.SloQuery.AliasBy sq.ProjectName = q.SloQuery.ProjectName sq.Selector = q.SloQuery.SelectorName sq.Service = q.SloQuery.ServiceId sq.Slo = q.SloQuery.SloId params.Add("filter", buildSLOFilterExpression(q.SloQuery)) setSloAggParams(¶ms, &q.SloQuery, durationSeconds, query.IntervalMs) } target = params.Encode() sq.Target = target sq.Params = params if setting.Env == setting.Dev { slog.Debug("CloudMonitoring request", "params", params) } cloudMonitoringQueries = append(cloudMonitoringQueries, sq) } return cloudMonitoringQueries, nil } func migrateLegacyQueryModel(query *tsdb.Query) { mq := query.Model.Get("metricQuery").MustMap() if mq == nil { migratedModel := simplejson.NewFromAny(map[string]interface{}{ "queryType": metricQueryType, "metricQuery": query.Model.MustMap(), }) query.Model = migratedModel } } func reverse(s string) string { chars := []rune(s) for i, j := 0, len(chars)-1; i < j; i, j = i+1, j-1 { chars[i], chars[j] = chars[j], chars[i] } return string(chars) } func interpolateFilterWildcards(value string) string { matches := strings.Count(value, "*") switch { case matches == 2 && strings.HasSuffix(value, "*") && strings.HasPrefix(value, "*"): value = strings.ReplaceAll(value, "*", "") value = fmt.Sprintf(`has_substring("%s")`, value) case matches == 1 && strings.HasPrefix(value, "*"): value = strings.Replace(value, "*", "", 1) value = fmt.Sprintf(`ends_with("%s")`, value) case matches == 1 && strings.HasSuffix(value, "*"): value = reverse(strings.Replace(reverse(value), "*", "", 1)) value = fmt.Sprintf(`starts_with("%s")`, value) case matches != 0: value = string(wildcardRegexRe.ReplaceAllFunc([]byte(value), func(in []byte) []byte { return []byte(strings.Replace(string(in), string(in), `\\`+string(in), 1)) })) value = strings.ReplaceAll(value, "*", ".*") value = strings.ReplaceAll(value, `"`, `\\"`) value = fmt.Sprintf(`monitoring.regex.full_match("^%s$")`, value) } return value } func buildFilterString(metricType string, filterParts []string) string { filterString := "" for i, part := range filterParts { mod := i % 4 switch { case part == "AND": filterString += " " case mod == 2: operator := filterParts[i-1] switch { case operator == "=~" || operator == "!=~": filterString = reverse(strings.Replace(reverse(filterString), "~", "", 1)) filterString += fmt.Sprintf(`monitoring.regex.full_match("%s")`, part) case strings.Contains(part, "*"): filterString += interpolateFilterWildcards(part) default: filterString += fmt.Sprintf(`"%s"`, part) } default: filterString += part } } return strings.Trim(fmt.Sprintf(`metric.type="%s" %s`, metricType, filterString), " ") } func buildSLOFilterExpression(q sloQuery) string { return fmt.Sprintf(`%s("projects/%s/services/%s/serviceLevelObjectives/%s")`, q.SelectorName, q.ProjectName, q.ServiceId, q.SloId) } func setMetricAggParams(params *url.Values, query *metricQuery, durationSeconds int, intervalMs int64) { if query.CrossSeriesReducer == "" { query.CrossSeriesReducer = "REDUCE_NONE" } if query.PerSeriesAligner == "" { query.PerSeriesAligner = "ALIGN_MEAN" } params.Add("aggregation.crossSeriesReducer", query.CrossSeriesReducer) params.Add("aggregation.perSeriesAligner", query.PerSeriesAligner) params.Add("aggregation.alignmentPeriod", calculateAlignmentPeriod(query.AlignmentPeriod, intervalMs, durationSeconds)) for _, groupBy := range query.GroupBys { params.Add("aggregation.groupByFields", groupBy) } } func setSloAggParams(params *url.Values, query *sloQuery, durationSeconds int, intervalMs int64) { params.Add("aggregation.alignmentPeriod", calculateAlignmentPeriod(query.AlignmentPeriod, intervalMs, durationSeconds)) if query.SelectorName == "select_slo_health" { params.Add("aggregation.perSeriesAligner", "ALIGN_MEAN") } else { params.Add("aggregation.perSeriesAligner", "ALIGN_NEXT_OLDER") } } func calculateAlignmentPeriod(alignmentPeriod string, intervalMs int64, durationSeconds int) string { if alignmentPeriod == "grafana-auto" || alignmentPeriod == "" { alignmentPeriodValue := int(math.Max(float64(intervalMs)/1000, 60.0)) alignmentPeriod = "+" + strconv.Itoa(alignmentPeriodValue) + "s" } if alignmentPeriod == "cloud-monitoring-auto" || alignmentPeriod == "stackdriver-auto" { // legacy alignmentPeriodValue := int(math.Max(float64(durationSeconds), 60.0)) switch { case alignmentPeriodValue < 60*60*23: alignmentPeriod = "+60s" case alignmentPeriodValue < 60*60*24*6: alignmentPeriod = "+300s" default: alignmentPeriod = "+3600s" } } return alignmentPeriod } func (e *CloudMonitoringExecutor) executeQuery(ctx context.Context, query *cloudMonitoringQuery, tsdbQuery *tsdb.TsdbQuery) (*tsdb.QueryResult, cloudMonitoringResponse, error) { queryResult := &tsdb.QueryResult{Meta: simplejson.New(), RefId: query.RefID} projectName := query.ProjectName if projectName == "" { defaultProject, err := e.getDefaultProject(ctx) if err != nil { queryResult.Error = err return queryResult, cloudMonitoringResponse{}, nil } projectName = defaultProject slog.Info("No project name set on query, using project name from datasource", "projectName", projectName) } req, err := e.createRequest(ctx, e.dsInfo, query, fmt.Sprintf("cloudmonitoring%s", "v3/projects/"+projectName+"/timeSeries")) if err != nil { queryResult.Error = err return queryResult, cloudMonitoringResponse{}, nil } req.URL.RawQuery = query.Params.Encode() queryResult.Meta.Set(sqleng.MetaKeyExecutedQueryString, req.URL.RawQuery) alignmentPeriod, ok := req.URL.Query()["aggregation.alignmentPeriod"] if ok { seconds, err := strconv.ParseInt(alignmentPeriodRe.FindString(alignmentPeriod[0]), 10, 64) if err == nil { queryResult.Meta.Set("alignmentPeriod", seconds) } } span, ctx := opentracing.StartSpanFromContext(ctx, "cloudMonitoring query") span.SetTag("target", query.Target) span.SetTag("from", tsdbQuery.TimeRange.From) span.SetTag("until", tsdbQuery.TimeRange.To) span.SetTag("datasource_id", e.dsInfo.Id) span.SetTag("org_id", e.dsInfo.OrgId) defer span.Finish() if err := opentracing.GlobalTracer().Inject( span.Context(), opentracing.HTTPHeaders, opentracing.HTTPHeadersCarrier(req.Header)); err != nil { queryResult.Error = err return queryResult, cloudMonitoringResponse{}, nil } res, err := ctxhttp.Do(ctx, e.httpClient, req) if err != nil { queryResult.Error = err return queryResult, cloudMonitoringResponse{}, nil } data, err := e.unmarshalResponse(res) if err != nil { queryResult.Error = err return queryResult, cloudMonitoringResponse{}, nil } return queryResult, data, nil } func (e *CloudMonitoringExecutor) unmarshalResponse(res *http.Response) (cloudMonitoringResponse, error) { body, err := ioutil.ReadAll(res.Body) defer res.Body.Close() if err != nil { return cloudMonitoringResponse{}, err } if res.StatusCode/100 != 2 { slog.Error("Request failed", "status", res.Status, "body", string(body)) return cloudMonitoringResponse{}, fmt.Errorf(string(body)) } var data cloudMonitoringResponse err = json.Unmarshal(body, &data) if err != nil { slog.Error("Failed to unmarshal CloudMonitoring response", "error", err, "status", res.Status, "body", string(body)) return cloudMonitoringResponse{}, err } return data, nil } func handleDistributionSeries(series timeSeries, defaultMetricName string, seriesLabels map[string]string, query *cloudMonitoringQuery, queryRes *tsdb.QueryResult) { points := make([]tsdb.TimePoint, 0) for i := len(series.Points) - 1; i >= 0; i-- { point := series.Points[i] value := point.Value.DoubleValue if series.ValueType == "INT64" { parsedValue, err := strconv.ParseFloat(point.Value.IntValue, 64) if err == nil { value = parsedValue } } if series.ValueType == "BOOL" { if point.Value.BoolValue { value = 1 } else { value = 0 } } points = append(points, tsdb.NewTimePoint(null.FloatFrom(value), float64((point.Interval.EndTime).Unix())*1000)) } metricName := formatLegendKeys(series.Metric.Type, defaultMetricName, seriesLabels, nil, query) queryRes.Series = append(queryRes.Series, &tsdb.TimeSeries{ Name: metricName, Points: points, }) } func (e *CloudMonitoringExecutor) parseResponse(queryRes *tsdb.QueryResult, data cloudMonitoringResponse, query *cloudMonitoringQuery) error { labels := make(map[string]map[string]bool) for _, series := range data.TimeSeries { seriesLabels := make(map[string]string) defaultMetricName := series.Metric.Type labels["resource.type"] = map[string]bool{series.Resource.Type: true} seriesLabels["resource.type"] = series.Resource.Type for key, value := range series.Metric.Labels { if _, ok := labels["metric.label."+key]; !ok { labels["metric.label."+key] = map[string]bool{} } labels["metric.label."+key][value] = true seriesLabels["metric.label."+key] = value if len(query.GroupBys) == 0 || containsLabel(query.GroupBys, "metric.label."+key) { defaultMetricName += " " + value } } for key, value := range series.Resource.Labels { if _, ok := labels["resource.label."+key]; !ok { labels["resource.label."+key] = map[string]bool{} } labels["resource.label."+key][value] = true seriesLabels["resource.label."+key] = value if containsLabel(query.GroupBys, "resource.label."+key) { defaultMetricName += " " + value } } for labelType, labelTypeValues := range series.MetaData { for labelKey, labelValue := range labelTypeValues { key := toSnakeCase(fmt.Sprintf("metadata.%s.%s", labelType, labelKey)) if _, ok := labels[key]; !ok { labels[key] = map[string]bool{} } switch v := labelValue.(type) { case string: labels[key][v] = true seriesLabels[key] = v case bool: strVal := strconv.FormatBool(v) labels[key][strVal] = true seriesLabels[key] = strVal case []interface{}: for _, v := range v { strVal := v.(string) labels[key][strVal] = true if len(seriesLabels[key]) > 0 { strVal = fmt.Sprintf("%s, %s", seriesLabels[key], strVal) } seriesLabels[key] = strVal } } } } // reverse the order to be ascending if series.ValueType != "DISTRIBUTION" { handleDistributionSeries(series, defaultMetricName, seriesLabels, query, queryRes) } else { buckets := make(map[int]*tsdb.TimeSeries) for i := len(series.Points) - 1; i >= 0; i-- { point := series.Points[i] if len(point.Value.DistributionValue.BucketCounts) == 0 { continue } maxKey := 0 for i := 0; i < len(point.Value.DistributionValue.BucketCounts); i++ { value, err := strconv.ParseFloat(point.Value.DistributionValue.BucketCounts[i], 64) if err != nil { continue } if _, ok := buckets[i]; !ok { // set lower bounds // https://cloud.google.com/monitoring/api/ref_v3/rest/v3/TimeSeries#Distribution bucketBound := calcBucketBound(point.Value.DistributionValue.BucketOptions, i) additionalLabels := map[string]string{"bucket": bucketBound} buckets[i] = &tsdb.TimeSeries{ Name: formatLegendKeys(series.Metric.Type, defaultMetricName, nil, additionalLabels, query), Points: make([]tsdb.TimePoint, 0), } if maxKey < i { maxKey = i } } buckets[i].Points = append(buckets[i].Points, tsdb.NewTimePoint(null.FloatFrom(value), float64((point.Interval.EndTime).Unix())*1000)) } // fill empty bucket for i := 0; i < maxKey; i++ { if _, ok := buckets[i]; !ok { bucketBound := calcBucketBound(point.Value.DistributionValue.BucketOptions, i) additionalLabels := map[string]string{"bucket": bucketBound} buckets[i] = &tsdb.TimeSeries{ Name: formatLegendKeys(series.Metric.Type, defaultMetricName, seriesLabels, additionalLabels, query), Points: make([]tsdb.TimePoint, 0), } } } } for i := 0; i < len(buckets); i++ { queryRes.Series = append(queryRes.Series, buckets[i]) } } } labelsByKey := make(map[string][]string) for key, values := range labels { for value := range values { labelsByKey[key] = append(labelsByKey[key], value) } } queryRes.Meta.Set("labels", labelsByKey) queryRes.Meta.Set("groupBys", query.GroupBys) return nil } func toSnakeCase(str string) string { return strings.ToLower(matchAllCap.ReplaceAllString(str, "${1}_${2}")) } func containsLabel(labels []string, newLabel string) bool { for _, val := range labels { if val == newLabel { return true } } return false } func formatLegendKeys(metricType string, defaultMetricName string, labels map[string]string, additionalLabels map[string]string, query *cloudMonitoringQuery) string { if query.AliasBy == "" { return defaultMetricName } result := legendKeyFormat.ReplaceAllFunc([]byte(query.AliasBy), func(in []byte) []byte { metaPartName := strings.Replace(string(in), "{{", "", 1) metaPartName = strings.Replace(metaPartName, "}}", "", 1) metaPartName = strings.TrimSpace(metaPartName) if metaPartName == "metric.type" { return []byte(metricType) } metricPart := replaceWithMetricPart(metaPartName, metricType) if metricPart != nil { return metricPart } if val, exists := labels[metaPartName]; exists { return []byte(val) } if val, exists := additionalLabels[metaPartName]; exists { return []byte(val) } if metaPartName == "project" && query.ProjectName != "" { return []byte(query.ProjectName) } if metaPartName == "service" && query.Service != "" { return []byte(query.Service) } if metaPartName == "slo" && query.Slo != "" { return []byte(query.Slo) } if metaPartName == "selector" && query.Selector != "" { return []byte(query.Selector) } return in }) return string(result) } func replaceWithMetricPart(metaPartName string, metricType string) []byte { // https://cloud.google.com/monitoring/api/v3/metrics-details#label_names shortMatches := metricNameFormat.FindStringSubmatch(metricType) if metaPartName == "metric.name" { if len(shortMatches) > 2 { return []byte(shortMatches[3]) } } if metaPartName == "metric.service" { if len(shortMatches) > 0 { return []byte(shortMatches[1]) } } return nil } func calcBucketBound(bucketOptions cloudMonitoringBucketOptions, n int) string { bucketBound := "0" if n == 0 { return bucketBound } switch { case bucketOptions.LinearBuckets != nil: bucketBound = strconv.FormatInt(bucketOptions.LinearBuckets.Offset+(bucketOptions.LinearBuckets.Width*int64(n-1)), 10) case bucketOptions.ExponentialBuckets != nil: bucketBound = strconv.FormatInt(int64(bucketOptions.ExponentialBuckets.Scale*math.Pow(bucketOptions.ExponentialBuckets.GrowthFactor, float64(n-1))), 10) case bucketOptions.ExplicitBuckets != nil: bucketBound = fmt.Sprintf("%g", bucketOptions.ExplicitBuckets.Bounds[n]) } return bucketBound } func (e *CloudMonitoringExecutor) createRequest(ctx context.Context, dsInfo *models.DataSource, query *cloudMonitoringQuery, proxyPass string) (*http.Request, error) { u, err := url.Parse(dsInfo.Url) if err != nil { return nil, err } u.Path = path.Join(u.Path, "render") req, err := http.NewRequest(http.MethodGet, "https://monitoring.googleapis.com/", nil) if err != nil { slog.Error("Failed to create request", "error", err) return nil, fmt.Errorf("failed to create request: %w", err) } req.Header.Set("Content-Type", "application/json") req.Header.Set("User-Agent", fmt.Sprintf("Grafana/%s", setting.BuildVersion)) // find plugin plugin, ok := plugins.DataSources[dsInfo.Type] if !ok { return nil, errors.New("unable to find datasource plugin CloudMonitoring") } var cloudMonitoringRoute *plugins.AppPluginRoute for _, route := range plugin.Routes { if route.Path == "cloudmonitoring" { cloudMonitoringRoute = route break } } pluginproxy.ApplyRoute(ctx, req, proxyPass, cloudMonitoringRoute, dsInfo) return req, nil } func (e *CloudMonitoringExecutor) getDefaultProject(ctx context.Context) (string, error) { authenticationType := e.dsInfo.JsonData.Get("authenticationType").MustString(jwtAuthentication) if authenticationType == gceAuthentication { defaultCredentials, err := google.FindDefaultCredentials(ctx, "https://www.googleapis.com/auth/monitoring.read") if err != nil { return "", fmt.Errorf("failed to retrieve default project from GCE metadata server: %w", err) } token, err := defaultCredentials.TokenSource.Token() if err != nil { return "", fmt.Errorf("failed to retrieve GCP credential token: %w", err) } if !token.Valid() { return "", errors.New("failed to validate GCP credentials") } return defaultCredentials.ProjectID, nil } return e.dsInfo.JsonData.Get("defaultProject").MustString(), nil }