mirror of
https://github.com/opentffoundation/opentf.git
synced 2026-02-14 10:01:05 -05:00
There is currently a bug in the apply engine that's causing a self-dependency error during promise resolution, which was a good reminder that we'd not previously finished connecting all of the parts to be able to unwind such problems into a useful error message. Due to how the work is split between the apply engine, the evaluator, and the execgraph package it takes some awkward back-and-forth to get all of the needed information together into one place. This compromise aims to do as little work as possible in the happy path and defer more expensive analysis until we actually know we're going to report an error message. In this case we can't really avoid proactively collecting the request IDs because we don't know ahead of time what (if anything) will be involved in a promise error, but only when actually generating an error message will we dig into the original source execution graph to find out what each of the affected requests was actually trying to do and construct a human-friendly summary of each one. This was a bit of a side-quest relative to the current goal of just getting things basically working with a simple configuration, but this is useful in figuring out what's going on with the current bug (which will be fixed in a future commit) and will probably be useful when dealing with future bugs too. Signed-off-by: Martin Atkins <mart@degeneration.co.uk>
89 lines
3.1 KiB
Go
89 lines
3.1 KiB
Go
// Copyright (c) The OpenTofu Authors
|
|
// SPDX-License-Identifier: MPL-2.0
|
|
// Copyright (c) 2023 HashiCorp, Inc.
|
|
// SPDX-License-Identifier: MPL-2.0
|
|
|
|
package applying
|
|
|
|
import (
|
|
"context"
|
|
"iter"
|
|
"maps"
|
|
"sync"
|
|
|
|
"github.com/apparentlymart/go-workgraph/workgraph"
|
|
"github.com/opentofu/opentofu/internal/engine/internal/execgraph"
|
|
"github.com/opentofu/opentofu/internal/lang/eval"
|
|
"github.com/opentofu/opentofu/internal/lang/grapheval"
|
|
)
|
|
|
|
// execRequestTracker is our implementation of [grapheval.RequestTracker] for
|
|
// use when processing an execution graph.
|
|
//
|
|
// An instance of this should be associated with the context passed to the
|
|
// graph execution function so that we can use it to produce helpful error
|
|
// messages if promise-related errors occur during execution.
|
|
type execRequestTracker struct {
|
|
configOracle *eval.ApplyOracle
|
|
graph *execgraph.Graph
|
|
|
|
promiseReqsMu sync.Mutex
|
|
promiseReqs map[execgraph.PromiseDrivenResultKey]workgraph.RequestID
|
|
}
|
|
|
|
var _ grapheval.RequestTracker = (*execRequestTracker)(nil)
|
|
var _ execgraph.RequestTrackerWithNotify = (*execRequestTracker)(nil)
|
|
|
|
func newRequestTracker(graph *execgraph.Graph, ops *execOperations) *execRequestTracker {
|
|
return &execRequestTracker{
|
|
configOracle: ops.configOracle,
|
|
graph: graph,
|
|
promiseReqs: make(map[execgraph.PromiseDrivenResultKey]workgraph.RequestID),
|
|
}
|
|
}
|
|
|
|
// ActiveRequests implements [grapheval.RequestTracker].
|
|
func (e *execRequestTracker) ActiveRequests() iter.Seq2[workgraph.RequestID, grapheval.RequestInfo] {
|
|
return func(yield func(workgraph.RequestID, grapheval.RequestInfo) bool) {
|
|
// To make the downstream implementations of this simpler we
|
|
// always visit all of the requests known to the evaluator and just
|
|
// discard them if the caller has stopped consuming our sequence.
|
|
// In practice we should always read the whole sequence to completion
|
|
// in the error reporting path anyway, so this is really just to
|
|
// honor the general expectations of [iter.Seq2].
|
|
keepGoing := true
|
|
e.configOracle.AnnounceAllGraphevalRequests(func(reqID workgraph.RequestID, info grapheval.RequestInfo) {
|
|
if !keepGoing {
|
|
return
|
|
}
|
|
keepGoing = yield(reqID, info)
|
|
})
|
|
if !keepGoing {
|
|
return
|
|
}
|
|
// We'll also report any requests that execgraph has announced to us
|
|
// using [execRequestTracker.TrackExecutionGraphRequest].
|
|
//
|
|
// We copy the map of requests here just so we don't hold this lock
|
|
// across yields to the caller. We should only end up in here when
|
|
// we're handling an error so it's okay to spend this extra time.
|
|
e.promiseReqsMu.Lock()
|
|
promiseReqs := make(map[execgraph.PromiseDrivenResultKey]workgraph.RequestID, len(e.promiseReqs))
|
|
maps.Copy(promiseReqs, e.promiseReqs)
|
|
e.promiseReqsMu.Unlock()
|
|
for key, reqID := range promiseReqs {
|
|
info := e.graph.PromiseDrivenRequestInfo(key)
|
|
if !yield(reqID, info) {
|
|
return
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// TrackExecutionGraphRequest implements [execgraph.RequestTrackerWithNotify].
|
|
func (e *execRequestTracker) TrackExecutionGraphRequest(ctx context.Context, key execgraph.PromiseDrivenResultKey, reqID workgraph.RequestID) {
|
|
e.promiseReqsMu.Lock()
|
|
e.promiseReqs[key] = reqID
|
|
e.promiseReqsMu.Unlock()
|
|
}
|