Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
A
asapo
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Wiki
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Build
Pipelines
Jobs
Pipeline schedules
Artifacts
Deploy
Releases
Package registry
Container Registry
Model registry
Operate
Terraform modules
Analyze
Contributor analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
Joao Alvim Oliveira Dias De Almeida
asapo
Commits
607d8585
Commit
607d8585
authored
2 years ago
by
Lars Janssen
Browse files
Options
Downloads
Patches
Plain Diff
Reverted GetTopology and GetGroupDelay
parent
531318af
No related branches found
Branches containing commit
No related tags found
Tags containing commit
No related merge requests found
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
monitoring/monitoring_server/src/asapo_monitoring_server/server/QueryServer.go
+229
-251
229 additions, 251 deletions
..._server/src/asapo_monitoring_server/server/QueryServer.go
with
229 additions
and
251 deletions
monitoring/monitoring_server/src/asapo_monitoring_server/server/QueryServer.go
+
229
−
251
View file @
607d8585
...
@@ -5,6 +5,7 @@ import (
...
@@ -5,6 +5,7 @@ import (
log
"asapo_common/logger"
log
"asapo_common/logger"
"context"
"context"
"errors"
"errors"
"fmt"
"github.com/influxdata/influxdb-client-go/v2/api"
"github.com/influxdata/influxdb-client-go/v2/api"
"google.golang.org/grpc/codes"
"google.golang.org/grpc/codes"
"google.golang.org/grpc/status"
"google.golang.org/grpc/status"
...
@@ -500,13 +501,18 @@ func (s *QueryServer) GetTopology(ctx context.Context, query *pb.ToplogyQuery) (
...
@@ -500,13 +501,18 @@ func (s *QueryServer) GetTopology(ctx context.Context, query *pb.ToplogyQuery) (
startTime
:=
strconv
.
FormatUint
(
query
.
FromTimestamp
,
10
)
startTime
:=
strconv
.
FormatUint
(
query
.
FromTimestamp
,
10
)
endTime
:=
strconv
.
FormatUint
(
query
.
ToTimestamp
,
10
)
endTime
:=
strconv
.
FormatUint
(
query
.
ToTimestamp
,
10
)
queryString
:=
"from(bucket:
\"
"
+
s
.
settings
.
InfluxDbDatabase
+
"
\"
)"
+
// Normal querystring
result
,
err
:=
s
.
dbQueryApi
.
Query
(
ctx
,
"from(bucket:
\"
"
+
s
.
settings
.
InfluxDbDatabase
+
"
\"
)"
+
" |> range(start: "
+
startTime
+
", stop: "
+
endTime
+
")"
+
" |> range(start: "
+
startTime
+
", stop: "
+
endTime
+
")"
+
" |> filter(fn: (r) => r._measurement ==
\"
"
+
dbMeasurementFileInput
+
"
\"
or r._measurement ==
\"
"
+
dbMeasurementBrokerFileRequests
+
"
\"
)"
+
" |> filter(fn: (r) => r._measurement ==
\"
"
+
dbMeasurementFileInput
+
"
\"
or r._measurement ==
\"
"
+
dbMeasurementBrokerFileRequests
+
"
\"
)"
+
" |> filter(fn: (r) => "
+
createFilterElement
(
"beamtime"
,
query
.
BeamtimeFilter
)
+
")"
+
" |> filter(fn: (r) => "
+
createFilterElement
(
"beamtime"
,
query
.
BeamtimeFilter
)
+
")"
+
" |> keep(columns: [
\"
receiverName
\"
,
\"
brokerName
\"
,
\"
pipelineStepId
\"
,
\"
source
\"
,
\"
producerInstanceId
\"
,
\"
consumerInstanceId
\"
])"
+
" |> keep(columns: [
\"
receiverName
\"
,
\"
brokerName
\"
,
\"
pipelineStepId
\"
,
\"
source
\"
,
\"
producerInstanceId
\"
,
\"
consumerInstanceId
\"
])"
+
" |> group()"
" |> group()"
,
queryForOldData
:=
generateFilterForAVGData
(
s
,
queryString
)
// Querystring for averaged data
)
if
err
!=
nil
{
fmt
.
Printf
(
"Error: %v
\n
"
,
err
)
return
nil
,
err
}
type
PipelineStep
struct
{
// these maps here are replacements for HashSets
type
PipelineStep
struct
{
// these maps here are replacements for HashSets
producesSourceId
map
[
string
]
/*produces sourceId*/
bool
producesSourceId
map
[
string
]
/*produces sourceId*/
bool
...
@@ -518,259 +524,244 @@ func (s *QueryServer) GetTopology(ctx context.Context, query *pb.ToplogyQuery) (
...
@@ -518,259 +524,244 @@ func (s *QueryServer) GetTopology(ctx context.Context, query *pb.ToplogyQuery) (
involvedReceivers
map
[
string
]
/*receiver id*/
bool
involvedReceivers
map
[
string
]
/*receiver id*/
bool
}
}
var
responseFun
pb
.
TopologyResponse
pipelineSteps
:=
map
[
string
]
/*pipelineStepId*/
PipelineStep
{}
// Queryfunction that appends the results to a pb.TopologyResponse
queryfunction
:=
func
(
s
*
QueryServer
,
ctx
context
.
Context
,
query
string
,
response
pb
.
TopologyResponse
)
(
*
pb
.
TopologyResponse
,
error
)
{
result
,
err
:=
s
.
dbQueryApi
.
Query
(
ctx
,
query
)
if
err
!=
nil
{
return
nil
,
err
}
pipelineSteps
:=
map
[
string
]
/*pipelineStepId*/
PipelineStep
{}
getOrCreateStep
:=
func
(
stepId
string
)
PipelineStep
{
getOrCreateStep
:=
func
(
stepId
string
)
PipelineStep
{
if
_
,
containsKey
:=
pipelineSteps
[
stepId
];
!
containsKey
{
if
_
,
containsKey
:=
pipelineSteps
[
stepId
];
!
containsKey
{
pipelineSteps
[
stepId
]
=
PipelineStep
{
pipelineSteps
[
stepId
]
=
PipelineStep
{
producesSourceId
:
map
[
string
]
/*produces sourceId*/
bool
{},
producesSourceId
:
map
[
string
]
/*produces sourceId*/
bool
{},
consumesSourceId
:
map
[
string
]
/*consumers sourceId*/
bool
{},
consumesSourceId
:
map
[
string
]
/*consumers sourceId*/
bool
{},
producerInstances
:
map
[
string
]
/*produces instanceId*/
bool
{},
producerInstances
:
map
[
string
]
/*produces instanceId*/
bool
{},
consumerInstances
:
map
[
string
]
/*consumers instanceId*/
bool
{},
consumerInstances
:
map
[
string
]
/*consumers instanceId*/
bool
{},
involvedReceivers
:
map
[
string
]
/*receiver id*/
bool
{},
involvedReceivers
:
map
[
string
]
/*receiver id*/
bool
{},
}
}
}
return
pipelineSteps
[
stepId
]
}
}
return
pipelineSteps
[
stepId
]
}
for
result
.
Next
()
{
for
result
.
Next
()
{
if
result
.
Record
()
.
Values
()[
"receiverName"
]
!=
nil
{
// data is coming from receiver => means it must be a producer
if
result
.
Record
()
.
Values
()[
"receiverName"
]
!=
nil
{
// data is coming from receiver => means it must be a producer
stepId
,
ok
:=
result
.
Record
()
.
Values
()[
"pipelineStepId"
]
.
(
string
)
stepId
,
ok
:=
result
.
Record
()
.
Values
()[
"pipelineStepId"
]
.
(
string
)
if
!
ok
{
if
!
ok
{
stepId
=
"defaultStep"
stepId
=
"defaultStep"
}
source
:=
result
.
Record
()
.
Values
()[
"source"
]
.
(
string
)
producerInstanceId
:=
result
.
Record
()
.
Values
()[
"producerInstanceId"
]
.
(
string
)
var
step
=
getOrCreateStep
(
stepId
)
step
.
producesSourceId
[
source
]
=
true
step
.
producerInstances
[
producerInstanceId
]
=
true
var
receiverName
=
result
.
Record
()
.
Values
()[
"receiverName"
]
.
(
string
)
step
.
involvedReceivers
[
receiverName
]
=
true
}
else
if
result
.
Record
()
.
Values
()[
"brokerName"
]
!=
nil
{
// data is coming from broker => means it must be a consumer
stepId
:=
result
.
Record
()
.
Values
()[
"pipelineStepId"
]
.
(
string
)
source
:=
result
.
Record
()
.
Values
()[
"source"
]
.
(
string
)
consumerInstanceId
:=
result
.
Record
()
.
Values
()[
"consumerInstanceId"
]
.
(
string
)
var
step
=
getOrCreateStep
(
stepId
)
step
.
consumesSourceId
[
source
]
=
true
step
.
consumerInstances
[
consumerInstanceId
]
=
true
}
else
{
log
.
Debug
(
"Got an entry without receiverName or brokerName"
)
}
}
source
:=
result
.
Record
()
.
Values
()[
"source"
]
.
(
string
)
producerInstanceId
:=
result
.
Record
()
.
Values
()[
"producerInstanceId"
]
.
(
string
)
var
step
=
getOrCreateStep
(
stepId
)
step
.
producesSourceId
[
source
]
=
true
step
.
producerInstances
[
producerInstanceId
]
=
true
var
receiverName
=
result
.
Record
()
.
Values
()[
"receiverName"
]
.
(
string
)
step
.
involvedReceivers
[
receiverName
]
=
true
}
else
if
result
.
Record
()
.
Values
()[
"brokerName"
]
!=
nil
{
// data is coming from broker => means it must be a consumer
stepId
:=
result
.
Record
()
.
Values
()[
"pipelineStepId"
]
.
(
string
)
source
:=
result
.
Record
()
.
Values
()[
"source"
]
.
(
string
)
consumerInstanceId
:=
result
.
Record
()
.
Values
()[
"consumerInstanceId"
]
.
(
string
)
var
step
=
getOrCreateStep
(
stepId
)
step
.
consumesSourceId
[
source
]
=
true
step
.
consumerInstances
[
consumerInstanceId
]
=
true
}
else
{
log
.
Debug
(
"Got an entry without receiverName or brokerName"
)
}
}
}
type
PipelineLevelStepInfo
struct
{
type
PipelineLevelStepInfo
struct
{
stepId
string
stepId
string
}
}
type
PipelineEdgeInfo
struct
{
fromStepId
string
toStepId
string
sourceId
string
involvedReceivers
map
[
string
]
/*receiver id*/
bool
}
// Simple algorithm
// Goes through each remainingStepIds and checks if all consumesSourceIds of this step is available
// If all consuming sources are available
// add it to the current level
// add all connecting edges
// and producing sources to availableSourcesInNextLevel
// If not go to next step, which might produces this
// When an iteration of remainingStepIds is done:
// Add all new producing sources to availableSources
// Go to next level
remainingStepIds
:=
map
[
string
]
/*pipelineStepId*/
bool
{}
availableSources
:=
map
[
string
]
/*sourceId*/
[]
/* available from pipelineStepId */
string
{}
for
stepId
:=
range
pipelineSteps
{
remainingStepIds
[
stepId
]
=
true
}
var
levels
[]
/*levelDepth*/
[]
/*verticalIdx*/
PipelineLevelStepInfo
var
edges
[]
PipelineEdgeInfo
var
currentLevel
=
-
1
for
len
(
remainingStepIds
)
>
0
{
type
PipelineEdgeInfo
struct
{
currentLevel
++
fromStepId
string
levels
=
append
(
levels
,
[]
PipelineLevelStepInfo
{})
toStepId
string
sourceId
string
involvedReceivers
map
[
string
]
/*receiver id*/
bool
}
var
foundAtLeastOne
=
false
// Simple algorithm
availableSourcesInNextLevel
:=
map
[
string
]
/*sourceId*/
[]
/* available from pipelineStepId */
string
{}
// Goes through each remainingStepIds and checks if all consumesSourceIds of this step is available
for
stepId
:=
range
remainingStepIds
{
// If all consuming sources are available
var
allConsumingSourcesAvailable
=
true
// add it to the current level
for
requiredSourceId
:=
range
pipelineSteps
[
stepId
]
.
consumesSourceId
{
// add all connecting edges
if
_
,
sourceIsAvailable
:=
availableSources
[
requiredSourceId
];
!
sourceIsAvailable
{
// Oh source is unavailable
// and producing sources to availableSourcesInNextLevel
allConsumingSourcesAvailable
=
false
// If not go to next step, which might produces this
}
else
{
// When an iteration of remainingStepIds is done:
// Verify that no other producer can create this source.
// Add all new producing sources to availableSources
// Maybe we have to wait one or two steps until all producers are available
// Go to next level
for
stepId2
:=
range
remainingStepIds
{
remainingStepIds
:=
map
[
string
]
/*pipelineStepId*/
bool
{}
if
_
,
thereIsAnotherProducer
:=
pipelineSteps
[
stepId2
]
.
producesSourceId
[
requiredSourceId
];
thereIsAnotherProducer
{
availableSources
:=
map
[
string
]
/*sourceId*/
[]
/* available from pipelineStepId */
string
{}
if
stepId
==
stepId2
{
continue
// The pipeline has self reference?! allow it
for
stepId
,
_
:=
range
pipelineSteps
{
}
remainingStepIds
[
stepId
]
=
true
allConsumingSourcesAvailable
=
false
}
var
levels
[]
/*levelDepth*/
[]
/*verticalIdx*/
PipelineLevelStepInfo
var
edges
[]
PipelineEdgeInfo
var
currentLevel
=
-
1
for
len
(
remainingStepIds
)
>
0
{
currentLevel
++
levels
=
append
(
levels
,
[]
PipelineLevelStepInfo
{})
var
foundAtLeastOne
=
false
availableSourcesInNextLevel
:=
map
[
string
]
/*sourceId*/
[]
/* available from pipelineStepId */
string
{}
for
stepId
:=
range
remainingStepIds
{
var
allConsumingSourcesAvailable
=
true
for
requiredSourceId
:=
range
pipelineSteps
[
stepId
]
.
consumesSourceId
{
if
_
,
sourceIsAvailable
:=
availableSources
[
requiredSourceId
];
!
sourceIsAvailable
{
// Oh source is unavailable
allConsumingSourcesAvailable
=
false
}
else
{
// Verify that no other producer can create this source.
// Maybe we have to wait one or two steps until all producers are available
for
stepId2
:=
range
remainingStepIds
{
if
_
,
thereIsAnotherProducer
:=
pipelineSteps
[
stepId2
]
.
producesSourceId
[
requiredSourceId
];
thereIsAnotherProducer
{
if
stepId
==
stepId2
{
continue
// The pipeline has self reference?! allow it
}
}
allConsumingSourcesAvailable
=
false
}
}
}
}
if
!
allConsumingSourcesAvailable
{
break
// We don't even need to try the other options
}
}
}
if
!
allConsumingSourcesAvailable
{
break
// We don't even need to try the other options
}
}
if
allConsumingSourcesAvailable
{
if
allConsumingSourcesAvailable
{
foundAtLeastOne
=
true
foundAtLeastOne
=
true
stepInfo
:=
pipelineSteps
[
stepId
]
stepInfo
:=
pipelineSteps
[
stepId
]
// Add edge connection
for
consumingSourceId
:=
range
stepInfo
.
consumesSourceId
{
if
stepInfo
.
producesSourceId
[
consumingSourceId
]
{
// Add self reference edge
edges
=
append
(
edges
,
PipelineEdgeInfo
{
fromStepId
:
stepId
,
toStepId
:
stepId
,
sourceId
:
consumingSourceId
,
involvedReceivers
:
stepInfo
.
involvedReceivers
,
})
}
for
_
,
sourceIdAvailableFromStepId
:=
range
availableSources
[
consumingSourceId
]
{
// Add all the others edges
var
producingSource
=
pipelineSteps
[
sourceIdAvailableFromStepId
]
edges
=
append
(
edges
,
PipelineEdgeInfo
{
fromStepId
:
sourceIdAvailableFromStepId
,
toStepId
:
stepId
,
sourceId
:
consumingSourceId
,
involvedReceivers
:
producingSource
.
involvedReceivers
,
})
}
}
// prepare sources for next level
for
sourceId
:=
range
stepInfo
.
producesSourceId
{
if
_
,
sourceIsAvailable
:=
availableSourcesInNextLevel
[
sourceId
];
!
sourceIsAvailable
{
availableSourcesInNextLevel
[
sourceId
]
=
nil
}
availableSourcesInNextLevel
[
sourceId
]
=
append
(
availableSourcesInNextLevel
[
sourceId
],
stepId
)
// Add edge connection
for
consumingSourceId
:=
range
stepInfo
.
consumesSourceId
{
if
stepInfo
.
producesSourceId
[
consumingSourceId
]
{
// Add self reference edge
edges
=
append
(
edges
,
PipelineEdgeInfo
{
fromStepId
:
stepId
,
toStepId
:
stepId
,
sourceId
:
consumingSourceId
,
involvedReceivers
:
stepInfo
.
involvedReceivers
,
})
}
for
_
,
sourceIdAvailableFromStepId
:=
range
availableSources
[
consumingSourceId
]
{
// Add all the others edges
var
producingSource
=
pipelineSteps
[
sourceIdAvailableFromStepId
]
edges
=
append
(
edges
,
PipelineEdgeInfo
{
fromStepId
:
sourceIdAvailableFromStepId
,
toStepId
:
stepId
,
sourceId
:
consumingSourceId
,
involvedReceivers
:
producingSource
.
involvedReceivers
,
})
}
}
levels
[
currentLevel
]
=
append
(
levels
[
currentLevel
],
PipelineLevelStepInfo
{
stepId
:
stepId
,
})
delete
(
remainingStepIds
,
stepId
)
}
}
}
if
!
foundAtLeastOne
{
// prepare sources for next level
// probably only requests of files came, but no receiver registered a producer
for
sourceId
:=
range
stepInfo
.
producesSourceId
{
// log.Error("infinite loop while building topology tree; Still has pipeline steps but found no way how to connect them")
if
_
,
sourceIsAvailable
:=
availableSourcesInNextLevel
[
sourceId
];
!
sourceIsAvailable
{
// return nil, errors.New("infinite loop while building topology tree; Still has pipeline steps but found no way how to connect them")
availableSourcesInNextLevel
[
sourceId
]
=
nil
}
var
unkStep
=
"Unknown"
pipelineSteps
[
unkStep
]
=
PipelineStep
{
availableSourcesInNextLevel
[
sourceId
]
=
append
(
availableSourcesInNextLevel
[
sourceId
],
stepId
)
producesSourceId
:
nil
,
consumesSourceId
:
nil
,
producerInstances
:
nil
,
consumerInstances
:
nil
,
involvedReceivers
:
nil
,
}
}
levels
[
0
]
=
append
(
levels
[
0
],
PipelineLevelStepInfo
{
levels
[
currentLevel
]
=
append
(
levels
[
currentLevel
],
PipelineLevelStepInfo
{
stepId
:
unkS
tep
,
stepId
:
s
tep
Id
,
})
})
if
len
(
levels
)
<
2
{
delete
(
remainingStepIds
,
stepId
)
levels
=
append
(
levels
,
[]
PipelineLevelStepInfo
{})
}
}
}
for
stepId
:=
range
remainingStepIds
{
if
!
foundAtLeastOne
{
var
step
=
pipelineSteps
[
stepId
]
// probably only requests of files came, but no receiver registered a producer
levels
[
1
]
=
append
(
levels
[
1
],
PipelineLevelStepInfo
{
// log.Error("infinite loop while building topology tree; Still has pipeline steps but found no way how to connect them")
stepId
:
stepId
,
// return nil, errors.New("infinite loop while building topology tree; Still has pipeline steps but found no way how to connect them")
})
for
sourceId
:=
range
step
.
consumesSourceId
{
var
unkStep
=
"Unknown"
edges
=
append
(
edges
,
PipelineEdgeInfo
{
fromStepId
:
unkStep
,
toStepId
:
stepId
,
sourceId
:
sourceId
,
involvedReceivers
:
nil
,
})
}
}
break
// Go to response building
pipelineSteps
[
unkStep
]
=
PipelineStep
{
producesSourceId
:
nil
,
consumesSourceId
:
nil
,
producerInstances
:
nil
,
consumerInstances
:
nil
,
involvedReceivers
:
nil
,
}
}
for
sourceId
,
element
:=
range
availableSourcesInNextLevel
{
levels
[
0
]
=
append
(
levels
[
0
],
PipelineLevelStepInfo
{
if
_
,
sourceIsAvailable
:=
availableSources
[
sourceId
];
!
sourceIsAvailable
{
stepId
:
unkStep
,
availableSources
[
sourceId
]
=
nil
})
}
for
_
,
newSource
:=
range
element
{
if
len
(
levels
)
<
2
{
availableSources
[
sourceId
]
=
append
(
availableSources
[
sourceId
],
newSource
)
levels
=
append
(
levels
,
[]
PipelineLevelStepInfo
{})
}
}
}
}
// Response building
for
levelIndex
,
level
:=
range
levels
{
for
stepId
:=
range
remainingStepIds
{
for
_
,
node
:=
range
level
{
var
step
=
pipelineSteps
[
stepId
]
var
producers
[]
string
=
nil
levels
[
1
]
=
append
(
levels
[
1
],
PipelineLevelStepInfo
{
var
consumers
[]
string
=
nil
stepId
:
stepId
,
})
for
producerInstanceId
:=
range
pipelineSteps
[
node
.
stepId
]
.
producerInstances
{
for
sourceId
:=
range
step
.
consumesSourceId
{
producers
=
append
(
producers
,
producerInstanceId
)
edges
=
append
(
edges
,
PipelineEdgeInfo
{
}
fromStepId
:
unkStep
,
for
consumerInstanceId
:=
range
pipelineSteps
[
node
.
stepId
]
.
consumerInstances
{
toStepId
:
stepId
,
consumers
=
append
(
consumers
,
consumerInstanceId
)
sourceId
:
sourceId
,
involvedReceivers
:
nil
,
})
}
}
response
.
Nodes
=
append
(
response
.
Nodes
,
&
pb
.
TopologyResponseNode
{
NodeId
:
node
.
stepId
,
Level
:
uint32
(
levelIndex
),
ProducerInstances
:
producers
,
ConsumerInstances
:
consumers
,
})
}
}
break
// Go to response building
}
}
for
_
,
edge
:=
range
edges
{
for
sourceId
,
element
:=
range
availableSourcesInNextLevel
{
newEdge
:=
pb
.
TopologyResponseEdge
{
if
_
,
sourceIsAvailable
:=
availableSources
[
sourceId
];
!
sourceIsAvailable
{
FromNodeId
:
edge
.
fromStepId
,
availableSources
[
sourceId
]
=
nil
ToNodeId
:
edge
.
toStepId
,
SourceName
:
edge
.
sourceId
,
InvolvedReceivers
:
nil
,
}
}
for
receiverId
:=
range
edge
.
involvedReceivers
{
for
_
,
newSource
:=
range
element
{
newEdge
.
InvolvedReceivers
=
append
(
newEdge
.
InvolvedReceivers
,
receiverId
)
availableSources
[
sourceId
]
=
append
(
availableSources
[
sourceId
],
newSource
)
}
}
response
.
Edges
=
append
(
response
.
Edges
,
&
newEdge
)
}
}
return
&
response
,
nil
}
}
firstQuery
,
err
:=
queryfunction
(
s
,
ctx
,
queryString
,
responseFun
)
// Query newer data
if
err
!=
nil
{
// Response building
return
nil
,
err
var
response
pb
.
TopologyResponse
for
levelIndex
,
level
:=
range
levels
{
for
_
,
node
:=
range
level
{
var
producers
[]
string
=
nil
var
consumers
[]
string
=
nil
for
producerInstanceId
:=
range
pipelineSteps
[
node
.
stepId
]
.
producerInstances
{
producers
=
append
(
producers
,
producerInstanceId
)
}
for
consumerInstanceId
:=
range
pipelineSteps
[
node
.
stepId
]
.
consumerInstances
{
consumers
=
append
(
consumers
,
consumerInstanceId
)
}
response
.
Nodes
=
append
(
response
.
Nodes
,
&
pb
.
TopologyResponseNode
{
NodeId
:
node
.
stepId
,
Level
:
uint32
(
levelIndex
),
ProducerInstances
:
producers
,
ConsumerInstances
:
consumers
,
})
}
}
}
secondQuery
,
err
:=
queryfunction
(
s
,
ctx
,
queryForOldData
,
*
firstQuery
)
// Query averaged data and append it to the first query
if
err
!=
nil
{
for
_
,
edge
:=
range
edges
{
return
nil
,
err
newEdge
:=
pb
.
TopologyResponseEdge
{
FromNodeId
:
edge
.
fromStepId
,
ToNodeId
:
edge
.
toStepId
,
SourceName
:
edge
.
sourceId
,
InvolvedReceivers
:
nil
,
}
for
receiverId
:=
range
edge
.
involvedReceivers
{
newEdge
.
InvolvedReceivers
=
append
(
newEdge
.
InvolvedReceivers
,
receiverId
)
}
response
.
Edges
=
append
(
response
.
Edges
,
&
newEdge
)
}
}
return
secondQuery
,
nil
return
&
response
,
nil
}
}
func
(
s
*
QueryServer
)
GetGroupDelay
(
ctx
context
.
Context
,
query
*
pb
.
GroupDelayQuery
)
(
*
pb
.
GroupDelayResponse
,
error
)
{
func
(
s
*
QueryServer
)
GetGroupDelay
(
ctx
context
.
Context
,
query
*
pb
.
GroupDelayQuery
)
(
*
pb
.
GroupDelayResponse
,
error
)
{
if
query
.
BeamtimeFilter
==
""
{
if
query
.
BeamtimeFilter
==
""
{
return
nil
,
status
.
Errorf
(
codes
.
InvalidArgument
,
"Beamtime is required"
)
return
nil
,
status
.
Errorf
(
codes
.
InvalidArgument
,
"Beamtime is required"
)
}
}
...
@@ -782,46 +773,33 @@ func (s *QueryServer) GetGroupDelay(ctx context.Context, query *pb.GroupDelayQue
...
@@ -782,46 +773,33 @@ func (s *QueryServer) GetGroupDelay(ctx context.Context, query *pb.GroupDelayQue
startTime
:=
strconv
.
FormatUint
(
query
.
FromTimestamp
,
10
)
startTime
:=
strconv
.
FormatUint
(
query
.
FromTimestamp
,
10
)
endTime
:=
strconv
.
FormatUint
(
query
.
ToTimestamp
,
10
)
endTime
:=
strconv
.
FormatUint
(
query
.
ToTimestamp
,
10
)
queryString
:=
"from(bucket:
\"
"
+
s
.
settings
.
InfluxDbDatabase
+
"
\"
)"
+
// Normal querystring
result
,
err
:=
s
.
dbQueryApi
.
Query
(
ctx
,
"from(bucket:
\"
"
+
s
.
settings
.
InfluxDbDatabase
+
"
\"
)"
+
" |> range(start: "
+
startTime
+
", stop: "
+
endTime
+
")"
+
" |> range(start: "
+
startTime
+
", stop: "
+
endTime
+
")"
+
" |> filter(fn: (r) => r._measurement ==
\"
"
+
dbMeasurementBrokerFileRequests
+
"
\"
"
+
" and r.brokerCommand ==
\"
next
\"
)"
+
" |> filter(fn: (r) => r._measurement ==
\"
"
+
dbMeasurementBrokerFileRequests
+
"
\"
"
+
" and r.brokerCommand ==
\"
next
\"
)"
+
" |> filter(fn: (r) => r.beamtime ==
\"
"
+
query
.
BeamtimeFilter
+
"
\"
and r._field ==
\"
delayMs
\"
)"
+
" |> filter(fn: (r) => r.beamtime ==
\"
"
+
query
.
BeamtimeFilter
+
"
\"
and r._field ==
\"
delayMs
\"
)"
+
" |> group(columns: [
\"
groupId
\"
])"
+
" |> group(columns: [
\"
groupId
\"
])"
+
" |> last()"
+
" |> last()"
+
" |> keep(columns: [
\"
groupId
\"
,
\"
_value
\"
])"
+
" |> keep(columns: [
\"
groupId
\"
,
\"
_value
\"
])"
+
" |> group()"
" |> group()"
,
queryForOldData
:=
generateFilterForAVGData
(
s
,
queryString
)
// Querystring for averaged data
)
response
:=
pb
.
GroupDelayResponse
{}
if
err
!=
nil
{
// Queryfunction that appends the results to a pb.GroupDelayResponse
return
nil
,
err
queryfunction
:=
func
(
s
*
QueryServer
,
ctx
context
.
Context
,
query
string
,
response
pb
.
GroupDelayResponse
)
(
*
pb
.
GroupDelayResponse
,
error
)
{
}
result
,
err
:=
s
.
dbQueryApi
.
Query
(
ctx
,
query
)
if
err
!=
nil
{
return
nil
,
err
}
for
result
.
Next
()
{
response
:=
pb
.
GroupDelayResponse
{}
if
result
.
Record
()
.
Values
()[
"groupId"
]
!=
nil
&&
result
.
Record
()
.
Value
()
!=
nil
{
for
result
.
Next
()
{
item
:=
pb
.
GroupDelayResponseItem
{
if
result
.
Record
()
.
Values
()[
"groupId"
]
!=
nil
&&
result
.
Record
()
.
Value
()
!=
nil
{
GroupId
:
result
.
Record
()
.
Values
()[
"groupId"
]
.
(
string
),
item
:=
pb
.
GroupDelayResponseItem
{
DelayMs
:
(
uint32
)(
result
.
Record
()
.
Value
()
.
(
int64
)),
GroupId
:
result
.
Record
()
.
Values
()[
"groupId"
]
.
(
string
),
}
DelayMs
:
(
uint32
)(
result
.
Record
()
.
Value
()
.
(
int64
)),
response
.
GroupIds
=
append
(
response
.
GroupIds
,
&
item
)
}
}
response
.
GroupIds
=
append
(
response
.
GroupIds
,
&
item
)
}
}
}
sort
.
SliceStable
(
response
.
GroupIds
,
func
(
i
,
j
int
)
bool
{
sort
.
SliceStable
(
response
.
GroupIds
,
func
(
i
,
j
int
)
bool
{
return
response
.
GroupIds
[
i
]
.
DelayMs
>
response
.
GroupIds
[
j
]
.
DelayMs
return
response
.
GroupIds
[
i
]
.
DelayMs
>
response
.
GroupIds
[
j
]
.
DelayMs
})
})
return
&
response
,
nil
return
&
response
,
nil
}
firstQuery
,
err
:=
queryfunction
(
s
,
ctx
,
queryString
,
response
)
// Query newer data
if
err
!=
nil
{
return
nil
,
err
}
secondQuery
,
err
:=
queryfunction
(
s
,
ctx
,
queryForOldData
,
*
firstQuery
)
// Query averaged data
if
err
!=
nil
{
return
nil
,
err
}
return
secondQuery
,
err
}
}
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment