|
|
|
@ -31,14 +31,14 @@ func (sr *scrutinyRepository) SaveSmartAttributes(ctx context.Context, wwn strin
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// GetSmartAttributeHistory MUST return in sorted order, where newest entries are at the beginning of the list, and oldest are at the end.
|
|
|
|
|
func (sr *scrutinyRepository) GetSmartAttributeHistory(ctx context.Context, wwn string, durationKey string, attributes []string) ([]measurements.Smart, error) {
|
|
|
|
|
func (sr *scrutinyRepository) GetSmartAttributeHistory(ctx context.Context, wwn string, durationKey string, n int, offset int, attributes []string) ([]measurements.Smart, error) {
|
|
|
|
|
// Get SMartResults from InfluxDB
|
|
|
|
|
|
|
|
|
|
//TODO: change the filter startrange to a real number.
|
|
|
|
|
|
|
|
|
|
// Get parser flux query result
|
|
|
|
|
//appConfig.GetString("web.influxdb.bucket")
|
|
|
|
|
queryStr := sr.aggregateSmartAttributesQuery(wwn, durationKey)
|
|
|
|
|
queryStr := sr.aggregateSmartAttributesQuery(wwn, durationKey, n, offset, attributes)
|
|
|
|
|
log.Infoln(queryStr)
|
|
|
|
|
|
|
|
|
|
smartResults := []measurements.Smart{}
|
|
|
|
@ -66,9 +66,6 @@ func (sr *scrutinyRepository) GetSmartAttributeHistory(ctx context.Context, wwn
|
|
|
|
|
return nil, err
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
//we have to sort the smartResults again, because the `union` command will return multiple 'tables' and only sort the records in each table.
|
|
|
|
|
sortSmartMeasurementsDesc(smartResults)
|
|
|
|
|
|
|
|
|
|
return smartResults, nil
|
|
|
|
|
|
|
|
|
|
//if err := device.SquashHistory(); err != nil {
|
|
|
|
@ -85,73 +82,6 @@ func (sr *scrutinyRepository) GetSmartAttributeHistory(ctx context.Context, wwn
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// GetSmartAttributeHistory MUST return in sorted order, where newest entries are at the beginning of the list, and oldest are at the end.
|
|
|
|
|
func (sr *scrutinyRepository) GetSmartAttributeHistoryTail(ctx context.Context, wwn string, n int, offset int, attributes []string) ([]measurements.Smart, error) {
|
|
|
|
|
partialQueryStr := []string{
|
|
|
|
|
`import "influxdata/influxdb/schema"`,
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
nestedDurationKeys := sr.lookupNestedDurationKeys(DURATION_KEY_FOREVER)
|
|
|
|
|
subQueryNames := []string{}
|
|
|
|
|
for _, nestedDurationKey := range nestedDurationKeys {
|
|
|
|
|
bucketName := sr.lookupBucketName(nestedDurationKey)
|
|
|
|
|
durationRange := sr.lookupDuration(nestedDurationKey)
|
|
|
|
|
|
|
|
|
|
subQueryNames = append(subQueryNames, fmt.Sprintf(`%sData`, nestedDurationKey))
|
|
|
|
|
partialQueryStr = append(partialQueryStr, []string{
|
|
|
|
|
fmt.Sprintf(`%sData = from(bucket: "%s")`, nestedDurationKey, bucketName),
|
|
|
|
|
fmt.Sprintf(`|> range(start: %s, stop: %s)`, durationRange[0], durationRange[1]),
|
|
|
|
|
`|> filter(fn: (r) => r["_measurement"] == "smart" )`,
|
|
|
|
|
fmt.Sprintf(`|> filter(fn: (r) => r["device_wwn"] == "%s" )`, wwn),
|
|
|
|
|
// We only need the last `offset` # of entries from each table to guarantee we can
|
|
|
|
|
// get the last `n` # of entries starting from `offset` of the union
|
|
|
|
|
fmt.Sprintf(`|> tail(n: %d)`, offset),
|
|
|
|
|
"|> schema.fieldsAsCols()",
|
|
|
|
|
}...)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
partialQueryStr = append(partialQueryStr, []string{
|
|
|
|
|
fmt.Sprintf("union(tables: [%s])", strings.Join(subQueryNames, ", ")),
|
|
|
|
|
"|> group()",
|
|
|
|
|
`|> sort(columns: ["_time"], desc: false)`,
|
|
|
|
|
fmt.Sprintf(`|> tail(n: %d, offset: %d)`, n, offset),
|
|
|
|
|
`|> yield(name: "last")`,
|
|
|
|
|
}...)
|
|
|
|
|
|
|
|
|
|
queryStr := strings.Join(partialQueryStr, "\n")
|
|
|
|
|
log.Infoln(queryStr)
|
|
|
|
|
|
|
|
|
|
smartResults := []measurements.Smart{}
|
|
|
|
|
|
|
|
|
|
result, err := sr.influxQueryApi.Query(ctx, queryStr)
|
|
|
|
|
if err == nil {
|
|
|
|
|
// Use Next() to iterate over query result lines
|
|
|
|
|
for result.Next() {
|
|
|
|
|
// Observe when there is new grouping key producing new table
|
|
|
|
|
if result.TableChanged() {
|
|
|
|
|
//fmt.Printf("table: %s\n", result.TableMetadata().String())
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
smartData, err := measurements.NewSmartFromInfluxDB(result.Record().Values())
|
|
|
|
|
if err != nil {
|
|
|
|
|
return nil, err
|
|
|
|
|
}
|
|
|
|
|
smartResults = append(smartResults, *smartData)
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
if result.Err() != nil {
|
|
|
|
|
fmt.Printf("Query error: %s\n", result.Err().Error())
|
|
|
|
|
}
|
|
|
|
|
} else {
|
|
|
|
|
return nil, err
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
//we have to sort the smartResults again, because the `union` command will return multiple 'tables' and only sort the records in each table.
|
|
|
|
|
sortSmartMeasurementsDesc(smartResults)
|
|
|
|
|
|
|
|
|
|
return smartResults, nil
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
////////////////////////////////////////////////////////////////////////////////////////////////////////////////////
|
|
|
|
|
// Helper Methods
|
|
|
|
|
////////////////////////////////////////////////////////////////////////////////////////////////////////////////////
|
|
|
|
@ -167,7 +97,7 @@ func (sr *scrutinyRepository) saveDatapoint(influxWriteApi api.WriteAPIBlocking,
|
|
|
|
|
return influxWriteApi.WritePoint(ctx, p)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func (sr *scrutinyRepository) aggregateSmartAttributesQuery(wwn string, durationKey string) string {
|
|
|
|
|
func (sr *scrutinyRepository) aggregateSmartAttributesQuery(wwn string, durationKey string, n int, offset int, attributes []string) string {
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
|
|
|
|
@ -176,28 +106,34 @@ func (sr *scrutinyRepository) aggregateSmartAttributesQuery(wwn string, duration
|
|
|
|
|
|> range(start: -1w, stop: now())
|
|
|
|
|
|> filter(fn: (r) => r["_measurement"] == "smart" )
|
|
|
|
|
|> filter(fn: (r) => r["device_wwn"] == "0x5000c5002df89099" )
|
|
|
|
|
|> tail(n: 10, offset: 0)
|
|
|
|
|
|> schema.fieldsAsCols()
|
|
|
|
|
|
|
|
|
|
monthData = from(bucket: "metrics_weekly")
|
|
|
|
|
|> range(start: -1mo, stop: -1w)
|
|
|
|
|
|> filter(fn: (r) => r["_measurement"] == "smart" )
|
|
|
|
|
|> filter(fn: (r) => r["device_wwn"] == "0x5000c5002df89099" )
|
|
|
|
|
|> tail(n: 10, offset: 0)
|
|
|
|
|
|> schema.fieldsAsCols()
|
|
|
|
|
|
|
|
|
|
yearData = from(bucket: "metrics_monthly")
|
|
|
|
|
|> range(start: -1y, stop: -1mo)
|
|
|
|
|
|> filter(fn: (r) => r["_measurement"] == "smart" )
|
|
|
|
|
|> filter(fn: (r) => r["device_wwn"] == "0x5000c5002df89099" )
|
|
|
|
|
|> tail(n: 10, offset: 0)
|
|
|
|
|
|> schema.fieldsAsCols()
|
|
|
|
|
|
|
|
|
|
foreverData = from(bucket: "metrics_yearly")
|
|
|
|
|
|> range(start: -10y, stop: -1y)
|
|
|
|
|
|> filter(fn: (r) => r["_measurement"] == "smart" )
|
|
|
|
|
|> filter(fn: (r) => r["device_wwn"] == "0x5000c5002df89099" )
|
|
|
|
|
|> tail(n: 10, offset: 0)
|
|
|
|
|
|> schema.fieldsAsCols()
|
|
|
|
|
|
|
|
|
|
union(tables: [weekData, monthData, yearData, foreverData])
|
|
|
|
|
|> sort(columns: ["_time"], desc: false)
|
|
|
|
|
|> group()
|
|
|
|
|
|> sort(columns: ["_time"], desc: true)
|
|
|
|
|
|> tail(n: 6, offset: 4)
|
|
|
|
|
|> yield(name: "last")
|
|
|
|
|
|
|
|
|
|
*/
|
|
|
|
@ -208,34 +144,57 @@ func (sr *scrutinyRepository) aggregateSmartAttributesQuery(wwn string, duration
|
|
|
|
|
|
|
|
|
|
nestedDurationKeys := sr.lookupNestedDurationKeys(durationKey)
|
|
|
|
|
|
|
|
|
|
if len(nestedDurationKeys) == 1 {
|
|
|
|
|
//there's only one bucket being queried, no need to union, just aggregate the dataset and return
|
|
|
|
|
partialQueryStr = append(partialQueryStr, []string{
|
|
|
|
|
sr.generateSmartAttributesSubquery(wwn, nestedDurationKeys[0], n, offset, attributes),
|
|
|
|
|
fmt.Sprintf(`%sData`, nestedDurationKeys[0]),
|
|
|
|
|
`|> sort(columns: ["_time"], desc: true)`,
|
|
|
|
|
`|> yield()`,
|
|
|
|
|
}...)
|
|
|
|
|
return strings.Join(partialQueryStr, "\n")
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
subQueries := []string{}
|
|
|
|
|
subQueryNames := []string{}
|
|
|
|
|
for _, nestedDurationKey := range nestedDurationKeys {
|
|
|
|
|
bucketName := sr.lookupBucketName(nestedDurationKey)
|
|
|
|
|
durationRange := sr.lookupDuration(nestedDurationKey)
|
|
|
|
|
|
|
|
|
|
subQueryNames = append(subQueryNames, fmt.Sprintf(`%sData`, nestedDurationKey))
|
|
|
|
|
partialQueryStr = append(partialQueryStr, []string{
|
|
|
|
|
fmt.Sprintf(`%sData = from(bucket: "%s")`, nestedDurationKey, bucketName),
|
|
|
|
|
fmt.Sprintf(`|> range(start: %s, stop: %s)`, durationRange[0], durationRange[1]),
|
|
|
|
|
`|> filter(fn: (r) => r["_measurement"] == "smart" )`,
|
|
|
|
|
fmt.Sprintf(`|> filter(fn: (r) => r["device_wwn"] == "%s" )`, wwn),
|
|
|
|
|
"|> schema.fieldsAsCols()",
|
|
|
|
|
}...)
|
|
|
|
|
if n > 0 {
|
|
|
|
|
// We only need the last `n + offset` # of entries from each table to guarantee we can
|
|
|
|
|
// get the last `n` # of entries starting from `offset` of the union
|
|
|
|
|
subQueries = append(subQueries, sr.generateSmartAttributesSubquery(wwn, nestedDurationKey, n+offset, 0, attributes))
|
|
|
|
|
} else {
|
|
|
|
|
subQueries = append(subQueries, sr.generateSmartAttributesSubquery(wwn, nestedDurationKey, 0, 0, attributes))
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
partialQueryStr = append(partialQueryStr, subQueries...)
|
|
|
|
|
partialQueryStr = append(partialQueryStr, []string{
|
|
|
|
|
fmt.Sprintf("union(tables: [%s])", strings.Join(subQueryNames, ", ")),
|
|
|
|
|
`|> group()`,
|
|
|
|
|
`|> sort(columns: ["_time"], desc: true)`,
|
|
|
|
|
}...)
|
|
|
|
|
if n > 0 {
|
|
|
|
|
partialQueryStr = append(partialQueryStr, fmt.Sprintf(`|> tail(n: %d, offset: %d)`, n, offset))
|
|
|
|
|
}
|
|
|
|
|
partialQueryStr = append(partialQueryStr, `|> yield(name: "last")`)
|
|
|
|
|
|
|
|
|
|
if len(subQueryNames) == 1 {
|
|
|
|
|
//there's only one bucket being queried, no need to union, just aggregate the dataset and return
|
|
|
|
|
partialQueryStr = append(partialQueryStr, []string{
|
|
|
|
|
subQueryNames[0],
|
|
|
|
|
`|> yield()`,
|
|
|
|
|
}...)
|
|
|
|
|
} else {
|
|
|
|
|
partialQueryStr = append(partialQueryStr, []string{
|
|
|
|
|
fmt.Sprintf("union(tables: [%s])", strings.Join(subQueryNames, ", ")),
|
|
|
|
|
`|> sort(columns: ["_time"], desc: false)`,
|
|
|
|
|
`|> yield(name: "last")`,
|
|
|
|
|
}...)
|
|
|
|
|
return strings.Join(partialQueryStr, "\n")
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func (sr *scrutinyRepository) generateSmartAttributesSubquery(wwn string, durationKey string, n int, offset int, attributes []string) string {
|
|
|
|
|
bucketName := sr.lookupBucketName(durationKey)
|
|
|
|
|
durationRange := sr.lookupDuration(durationKey)
|
|
|
|
|
|
|
|
|
|
partialQueryStr := []string{
|
|
|
|
|
fmt.Sprintf(`%sData = from(bucket: "%s")`, durationKey, bucketName),
|
|
|
|
|
fmt.Sprintf(`|> range(start: %s, stop: %s)`, durationRange[0], durationRange[1]),
|
|
|
|
|
`|> filter(fn: (r) => r["_measurement"] == "smart" )`,
|
|
|
|
|
fmt.Sprintf(`|> filter(fn: (r) => r["device_wwn"] == "%s" )`, wwn),
|
|
|
|
|
}
|
|
|
|
|
if n > 0 {
|
|
|
|
|
partialQueryStr = append(partialQueryStr, fmt.Sprintf(`|> tail(n: %d, offset: %d)`, n, offset))
|
|
|
|
|
}
|
|
|
|
|
partialQueryStr = append(partialQueryStr, "|> schema.fieldsAsCols()")
|
|
|
|
|
|
|
|
|
|
return strings.Join(partialQueryStr, "\n")
|
|
|
|
|
}
|
|
|
|
|