mirror of
https://github.com/grafana/grafana.git
synced 2025-07-29 11:42:36 +08:00
296 lines
8.8 KiB
Go
296 lines
8.8 KiB
Go
package cloudwatch
|
|
|
|
import (
|
|
"fmt"
|
|
"slices"
|
|
"sort"
|
|
"strconv"
|
|
"time"
|
|
|
|
"github.com/aws/aws-sdk-go-v2/service/cloudwatchlogs"
|
|
cloudwatchlogstypes "github.com/aws/aws-sdk-go-v2/service/cloudwatchlogs/types"
|
|
|
|
"github.com/grafana/grafana-plugin-sdk-go/data"
|
|
)
|
|
|
|
const cloudWatchTSFormat = "2006-01-02 15:04:05.000"
|
|
|
|
func logsResultsToDataframes(response *cloudwatchlogs.GetQueryResultsOutput, groupingFieldNames []string) (*data.Frame, error) {
|
|
if response == nil {
|
|
return nil, fmt.Errorf("response is nil, cannot convert log results to data frames")
|
|
}
|
|
|
|
nonEmptyRows := make([][]cloudwatchlogstypes.ResultField, 0)
|
|
for _, row := range response.Results {
|
|
// Sometimes CloudWatch can send empty rows
|
|
if len(row) == 0 {
|
|
continue
|
|
}
|
|
if len(row) == 1 {
|
|
if row[0].Value == nil {
|
|
continue
|
|
}
|
|
// Sometimes it sends rows with only timestamp
|
|
if _, err := time.Parse(cloudWatchTSFormat, *row[0].Value); err == nil {
|
|
continue
|
|
}
|
|
}
|
|
nonEmptyRows = append(nonEmptyRows, row)
|
|
}
|
|
|
|
rowCount := len(nonEmptyRows)
|
|
|
|
fieldValues := make(map[string]any)
|
|
|
|
// Maintaining a list of field names in the order returned from CloudWatch
|
|
// as just iterating over fieldValues would not give a consistent order
|
|
fieldNames := make([]string, 0)
|
|
|
|
for i, row := range nonEmptyRows {
|
|
for _, resultField := range row {
|
|
// Strip @ptr field from results as it's not needed
|
|
if *resultField.Field == "@ptr" {
|
|
continue
|
|
}
|
|
|
|
if _, exists := fieldValues[*resultField.Field]; !exists {
|
|
fieldNames = append(fieldNames, *resultField.Field)
|
|
|
|
// Check if it's a cloudWatchTSFormat field or one of the known timestamp fields:
|
|
// https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CWL_AnalyzeLogData-discoverable-fields.html
|
|
// which can be in a millisecond format as well as cloudWatchTSFormat string format
|
|
if _, err := time.Parse(cloudWatchTSFormat, *resultField.Value); err == nil || isTimestampField(*resultField.Field) {
|
|
fieldValues[*resultField.Field] = make([]*time.Time, rowCount)
|
|
} else if slices.Contains[[]string, string](groupingFieldNames, *resultField.Field) {
|
|
fieldValues[*resultField.Field] = make([]*string, rowCount)
|
|
} else if _, err := strconv.ParseFloat(*resultField.Value, 64); err == nil {
|
|
fieldValues[*resultField.Field] = make([]*float64, rowCount)
|
|
} else {
|
|
fieldValues[*resultField.Field] = make([]*string, rowCount)
|
|
}
|
|
}
|
|
|
|
if timeField, ok := fieldValues[*resultField.Field].([]*time.Time); ok {
|
|
parsedTime, err := time.Parse(cloudWatchTSFormat, *resultField.Value)
|
|
if err != nil {
|
|
unixTimeMs, err := strconv.ParseInt(*resultField.Value, 10, 64)
|
|
if err == nil {
|
|
parsedTime = time.Unix(unixTimeMs/1000, (unixTimeMs%1000)*int64(time.Millisecond))
|
|
} else {
|
|
return nil, err
|
|
}
|
|
}
|
|
timeField[i] = &parsedTime
|
|
} else if numericField, ok := fieldValues[*resultField.Field].([]*float64); ok {
|
|
parsedFloat, err := strconv.ParseFloat(*resultField.Value, 64)
|
|
|
|
if err != nil {
|
|
// This can happen if a field has a mix of numeric and non-numeric values.
|
|
// In that case, we change the field from a numeric field to a string field.
|
|
fieldValues[*resultField.Field] = changeToStringField(rowCount, nonEmptyRows[:i+1], *resultField.Field)
|
|
continue
|
|
}
|
|
|
|
numericField[i] = &parsedFloat
|
|
} else {
|
|
fieldValues[*resultField.Field].([]*string)[i] = resultField.Value
|
|
}
|
|
}
|
|
}
|
|
|
|
newFields := make([]*data.Field, 0, len(fieldNames))
|
|
for _, fieldName := range fieldNames {
|
|
newFields = append(newFields, data.NewField(fieldName, nil, fieldValues[fieldName]))
|
|
|
|
switch fieldName {
|
|
case "@timestamp":
|
|
newFields[len(newFields)-1].SetConfig(&data.FieldConfig{DisplayName: "Time"})
|
|
case logStreamIdentifierInternal, logIdentifierInternal:
|
|
newFields[len(newFields)-1].SetConfig(
|
|
&data.FieldConfig{
|
|
Custom: map[string]any{
|
|
"hidden": true,
|
|
},
|
|
},
|
|
)
|
|
}
|
|
}
|
|
|
|
queryStats := make([]data.QueryStat, 0)
|
|
if response.Statistics != nil {
|
|
queryStats = append(queryStats, data.QueryStat{
|
|
FieldConfig: data.FieldConfig{DisplayName: "Bytes scanned"},
|
|
Value: response.Statistics.BytesScanned,
|
|
})
|
|
|
|
queryStats = append(queryStats, data.QueryStat{
|
|
FieldConfig: data.FieldConfig{DisplayName: "Records scanned"},
|
|
Value: response.Statistics.RecordsScanned,
|
|
})
|
|
|
|
queryStats = append(queryStats, data.QueryStat{
|
|
FieldConfig: data.FieldConfig{DisplayName: "Records matched"},
|
|
Value: response.Statistics.RecordsMatched,
|
|
})
|
|
}
|
|
|
|
frame := data.NewFrame("CloudWatchLogsResponse", newFields...)
|
|
frame.Meta = &data.FrameMeta{
|
|
Stats: nil,
|
|
Custom: nil,
|
|
}
|
|
|
|
if len(queryStats) > 0 {
|
|
frame.Meta.Stats = queryStats
|
|
}
|
|
|
|
frame.Meta.Custom = map[string]any{
|
|
"Status": string(response.Status),
|
|
}
|
|
|
|
// Results aren't guaranteed to come ordered by time (ascending), so we need to sort
|
|
sort.Sort(ByTime(*frame))
|
|
return frame, nil
|
|
}
|
|
|
|
func changeToStringField(lengthOfValues int, rows [][]cloudwatchlogstypes.ResultField, logEventField string) []*string {
|
|
fieldValuesAsStrings := make([]*string, lengthOfValues)
|
|
for i, resultFields := range rows {
|
|
for _, field := range resultFields {
|
|
if *field.Field == logEventField {
|
|
fieldValuesAsStrings[i] = field.Value
|
|
}
|
|
}
|
|
}
|
|
|
|
return fieldValuesAsStrings
|
|
}
|
|
|
|
func groupResults(results *data.Frame, groupingFieldNames []string, fromSyncQuery bool) ([]*data.Frame, error) {
|
|
groupingFields := make([]*data.Field, 0)
|
|
removeFieldIndices := make([]int, 0)
|
|
|
|
for i, field := range results.Fields {
|
|
for _, groupingField := range groupingFieldNames {
|
|
if field.Name == groupingField {
|
|
// For expressions and alerts to work properly we need to remove non-time grouping fields
|
|
if fromSyncQuery && !field.Type().Time() {
|
|
removeFieldIndices = append(removeFieldIndices, i)
|
|
}
|
|
|
|
groupingFields = append(groupingFields, field)
|
|
}
|
|
}
|
|
}
|
|
|
|
rowLength, err := results.RowLen()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
groupedDataFrames := make(map[string]*data.Frame)
|
|
for i := 0; i < rowLength; i++ {
|
|
groupKey := generateGroupKey(groupingFields, i)
|
|
// if group key doesn't exist create it
|
|
if _, exists := groupedDataFrames[groupKey]; !exists {
|
|
newFrame := results.EmptyCopy()
|
|
newFrame.Name = groupKey
|
|
newFrame.Meta = results.Meta
|
|
if fromSyncQuery {
|
|
// remove grouping indices
|
|
newFrame.Fields = removeFieldsByIndex(newFrame.Fields, removeFieldIndices)
|
|
groupLabels := generateLabels(groupingFields, i)
|
|
|
|
// set the group key as the display name for sync queries
|
|
for j := 0; j < len(newFrame.Fields); j++ {
|
|
valueField := newFrame.Fields[j]
|
|
// the time field might not be the first field so we check it here and skip the field if it is
|
|
if valueField.Type().Time() {
|
|
continue
|
|
}
|
|
if valueField.Config == nil {
|
|
valueField.Config = &data.FieldConfig{}
|
|
}
|
|
valueField.Config.DisplayNameFromDS = groupKey
|
|
valueField.Labels = groupLabels
|
|
}
|
|
}
|
|
|
|
groupedDataFrames[groupKey] = newFrame
|
|
}
|
|
|
|
// add row to frame
|
|
row := copyRowWithoutValues(results, i, removeFieldIndices)
|
|
groupedDataFrames[groupKey].AppendRow(row...)
|
|
}
|
|
|
|
newDataFrames := make([]*data.Frame, 0, len(groupedDataFrames))
|
|
for _, dataFrame := range groupedDataFrames {
|
|
newDataFrames = append(newDataFrames, dataFrame)
|
|
}
|
|
|
|
return newDataFrames, nil
|
|
}
|
|
|
|
// remove fields at the listed indices
|
|
func removeFieldsByIndex(fields []*data.Field, removeIndices []int) []*data.Field {
|
|
newGroupingFields := make([]*data.Field, 0)
|
|
removeIndicesIndex := 0
|
|
for i, field := range fields {
|
|
if removeIndicesIndex < len(removeIndices) && i == removeIndices[removeIndicesIndex] {
|
|
removeIndicesIndex++
|
|
if removeIndicesIndex > len(removeIndices) {
|
|
newGroupingFields = append(newGroupingFields, fields[i+1:]...)
|
|
break
|
|
}
|
|
continue
|
|
}
|
|
newGroupingFields = append(newGroupingFields, field)
|
|
}
|
|
return newGroupingFields
|
|
}
|
|
|
|
// copy a row without the listed values
|
|
func copyRowWithoutValues(f *data.Frame, rowIdx int, removeIndices []int) []any {
|
|
vals := make([]any, len(f.Fields)-len(removeIndices))
|
|
valsIdx := 0
|
|
removeIndicesIndex := 0
|
|
for i := range f.Fields {
|
|
if removeIndicesIndex < len(removeIndices) && i == removeIndices[removeIndicesIndex] {
|
|
removeIndicesIndex++
|
|
continue
|
|
}
|
|
vals[valsIdx] = f.CopyAt(i, rowIdx)
|
|
valsIdx++
|
|
}
|
|
return vals
|
|
}
|
|
|
|
func generateGroupKey(fields []*data.Field, row int) string {
|
|
groupKey := ""
|
|
for _, field := range fields {
|
|
if strField, ok := field.At(row).(*string); ok {
|
|
if strField != nil {
|
|
groupKey += *strField
|
|
}
|
|
}
|
|
}
|
|
return groupKey
|
|
}
|
|
|
|
func generateLabels(fields []*data.Field, row int) data.Labels {
|
|
labels := data.Labels{}
|
|
for _, field := range fields {
|
|
if strField, ok := field.At(row).(*string); ok {
|
|
if strField != nil {
|
|
labels[field.Name] = *strField
|
|
}
|
|
}
|
|
}
|
|
return labels
|
|
}
|
|
|
|
func isTimestampField(fieldName string) bool {
|
|
return fieldName == "@timestamp" || fieldName == "@ingestionTime"
|
|
}
|