Merge pull request #288 from dagger/dagger-up-store-outputs

up: store outputs and merge in query
This commit is contained in:
Andrea Luzzardi 2021-04-08 18:56:15 -07:00 committed by GitHub
commit 92ef1abd28
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
17 changed files with 355 additions and 347 deletions

View File

@ -2,7 +2,6 @@ package common
import ( import (
"context" "context"
"fmt"
"os" "os"
"dagger.io/go/dagger" "dagger.io/go/dagger"
@ -61,22 +60,19 @@ func GetCurrentDeploymentState(ctx context.Context, store *dagger.Store) *dagger
} }
// Re-compute a deployment (equivalent to `dagger up`). // Re-compute a deployment (equivalent to `dagger up`).
// If printOutput is true, print the JSON-encoded computed state to standard output func DeploymentUp(ctx context.Context, state *dagger.DeploymentState) *dagger.Deployment {
func DeploymentUp(ctx context.Context, state *dagger.DeploymentState, printOutput bool) {
lg := log.Ctx(ctx) lg := log.Ctx(ctx)
c, err := dagger.NewClient(ctx, "") c, err := dagger.NewClient(ctx, "")
if err != nil { if err != nil {
lg.Fatal().Err(err).Msg("unable to create client") lg.Fatal().Err(err).Msg("unable to create client")
} }
output, err := c.Do(ctx, state, func(ctx context.Context, deployment *dagger.Deployment, s dagger.Solver) error { result, err := c.Do(ctx, state, func(ctx context.Context, deployment *dagger.Deployment, s dagger.Solver) error {
log.Ctx(ctx).Debug().Msg("bringing deployment up") log.Ctx(ctx).Debug().Msg("bringing deployment up")
return deployment.Up(ctx, s, nil) return deployment.Up(ctx, s)
}) })
if err != nil { if err != nil {
lg.Fatal().Err(err).Msg("failed to up deployment") lg.Fatal().Err(err).Msg("failed to up deployment")
} }
if printOutput { return result
fmt.Println(output.JSON())
}
} }

View File

@ -3,12 +3,15 @@ package cmd
import ( import (
"encoding/json" "encoding/json"
"errors" "errors"
"fmt"
"os" "os"
"strings" "strings"
"cuelang.org/go/cue"
"dagger.io/go/cmd/dagger/cmd/common" "dagger.io/go/cmd/dagger/cmd/common"
"dagger.io/go/cmd/dagger/logger" "dagger.io/go/cmd/dagger/logger"
"dagger.io/go/dagger" "dagger.io/go/dagger"
"dagger.io/go/dagger/compiler"
"go.mozilla.org/sops/v3" "go.mozilla.org/sops/v3"
"go.mozilla.org/sops/v3/decrypt" "go.mozilla.org/sops/v3/decrypt"
@ -127,7 +130,20 @@ var computeCmd = &cobra.Command{
} }
} }
common.DeploymentUp(ctx, st, true) deployment := common.DeploymentUp(ctx, st)
v := compiler.NewValue()
if err := v.FillPath(cue.MakePath(), deployment.Plan()); err != nil {
lg.Fatal().Err(err).Msg("failed to merge")
}
if err := v.FillPath(cue.MakePath(), deployment.Input()); err != nil {
lg.Fatal().Err(err).Msg("failed to merge")
}
if err := v.FillPath(cue.MakePath(), deployment.Computed()); err != nil {
lg.Fatal().Err(err).Msg("failed to merge")
}
fmt.Println(v.JSON())
}, },
} }

View File

@ -63,7 +63,7 @@ var newCmd = &cobra.Command{
Msg("deployment created") Msg("deployment created")
if viper.GetBool("up") { if viper.GetBool("up") {
common.DeploymentUp(ctx, st, false) common.DeploymentUp(ctx, st)
} }
}, },
} }

View File

@ -51,12 +51,37 @@ var queryCmd = &cobra.Command{
if err != nil { if err != nil {
lg.Fatal().Err(err).Msg("unable to create client") lg.Fatal().Err(err).Msg("unable to create client")
} }
output, err := c.Do(ctx, state, nil)
deployment, err := c.Do(ctx, state, nil)
if err != nil { if err != nil {
lg.Fatal().Err(err).Msg("failed to query deployment") lg.Fatal().Err(err).Msg("failed to query deployment")
} }
cueVal := output.LookupPath(cuePath) cueVal := compiler.NewValue()
if !viper.GetBool("no-plan") {
if err := cueVal.FillPath(cue.MakePath(), deployment.Plan()); err != nil {
lg.Fatal().Err(err).Msg("failed to merge plan")
}
}
if !viper.GetBool("no-input") {
if err := cueVal.FillPath(cue.MakePath(), deployment.Input()); err != nil {
lg.Fatal().Err(err).Msg("failed to merge plan with output")
}
}
if !viper.GetBool("no-computed") && state.Computed != "" {
computed, err := compiler.DecodeJSON("", []byte(state.Computed))
if err != nil {
lg.Fatal().Err(err).Msg("failed to decode json")
}
if err := cueVal.FillPath(cue.MakePath(), computed); err != nil {
lg.Fatal().Err(err).Msg("failed to merge plan with computed")
}
}
cueVal = cueVal.LookupPath(cuePath)
if viper.GetBool("concrete") { if viper.GetBool("concrete") {
if err := cueVal.IsConcreteR(); err != nil { if err := cueVal.IsConcreteR(); err != nil {
@ -116,9 +141,9 @@ func init() {
// FIXME: implement the flags below // FIXME: implement the flags below
// queryCmd.Flags().String("revision", "latest", "Query a specific version of the deployment") // queryCmd.Flags().String("revision", "latest", "Query a specific version of the deployment")
queryCmd.Flags().StringP("format", "f", "json", "Output format (json|yaml|cue|text|env)") queryCmd.Flags().StringP("format", "f", "json", "Output format (json|yaml|cue|text|env)")
// queryCmd.Flags().BoolP("no-input", "I", false, "Exclude inputs from query") queryCmd.Flags().BoolP("no-plan", "P", false, "Exclude plan from query")
// queryCmd.Flags().BoolP("no-output", "O", false, "Exclude outputs from query") queryCmd.Flags().BoolP("no-input", "I", false, "Exclude inputs from query")
// queryCmd.Flags().BoolP("no-plan", "P", false, "Exclude outputs from query") queryCmd.Flags().BoolP("no-computed", "C", false, "Exclude computed values from query")
if err := viper.BindPFlags(queryCmd.Flags()); err != nil { if err := viper.BindPFlags(queryCmd.Flags()); err != nil {
panic(err) panic(err)

View File

@ -31,7 +31,11 @@ var upCmd = &cobra.Command{
state := common.GetCurrentDeploymentState(ctx, store) state := common.GetCurrentDeploymentState(ctx, store)
// TODO: Implement options: --no-cache // TODO: Implement options: --no-cache
common.DeploymentUp(ctx, state, true) result := common.DeploymentUp(ctx, state)
state.Computed = result.Computed().JSON().String()
if err := store.UpdateDeployment(ctx, state, nil); err != nil {
lg.Fatal().Err(err).Msg("failed to update deployment")
}
}, },
} }

View File

@ -7,6 +7,7 @@ import (
"hash/adler32" "hash/adler32"
"io" "io"
"strings" "strings"
"sync"
"time" "time"
"github.com/mitchellh/colorstring" "github.com/mitchellh/colorstring"
@ -21,6 +22,7 @@ var colorize = colorstring.Colorize{
type Console struct { type Console struct {
Out io.Writer Out io.Writer
maxLength int maxLength int
l sync.Mutex
} }
func (c *Console) Write(p []byte) (n int, err error) { func (c *Console) Write(p []byte) (n int, err error) {
@ -31,9 +33,12 @@ func (c *Console) Write(p []byte) (n int, err error) {
} }
source := c.parseSource(event) source := c.parseSource(event)
c.l.Lock()
if len(source) > c.maxLength { if len(source) > c.maxLength {
c.maxLength = len(source) c.maxLength = len(source)
} }
c.l.Unlock()
return fmt.Fprintln(c.Out, return fmt.Fprintln(c.Out,
colorize.Color(fmt.Sprintf("%s %s %s%s%s", colorize.Color(fmt.Sprintf("%s %s %s%s%s",

View File

@ -1,16 +1,12 @@
package dagger package dagger
import ( import (
"archive/tar"
"context" "context"
"errors"
"fmt" "fmt"
"io"
"os" "os"
"path/filepath" "path/filepath"
"strings" "strings"
"cuelang.org/go/cue"
"golang.org/x/sync/errgroup" "golang.org/x/sync/errgroup"
"github.com/opentracing/opentracing-go" "github.com/opentracing/opentracing-go"
@ -64,7 +60,7 @@ func NewClient(ctx context.Context, host string) (*Client, error) {
type ClientDoFunc func(context.Context, *Deployment, Solver) error type ClientDoFunc func(context.Context, *Deployment, Solver) error
// FIXME: return completed *Route, instead of *compiler.Value // FIXME: return completed *Route, instead of *compiler.Value
func (c *Client) Do(ctx context.Context, state *DeploymentState, fn ClientDoFunc) (*compiler.Value, error) { func (c *Client) Do(ctx context.Context, state *DeploymentState, fn ClientDoFunc) (*Deployment, error) {
lg := log.Ctx(ctx) lg := log.Ctx(ctx)
eg, gctx := errgroup.WithContext(ctx) eg, gctx := errgroup.WithContext(ctx)
@ -83,24 +79,14 @@ func (c *Client) Do(ctx context.Context, state *DeploymentState, fn ClientDoFunc
}) })
// Spawn build function // Spawn build function
outr, outw := io.Pipe()
eg.Go(func() error { eg.Go(func() error {
defer outw.Close() return c.buildfn(gctx, deployment, fn, events)
return c.buildfn(gctx, deployment, fn, events, outw)
}) })
// Spawn output retriever return deployment, eg.Wait()
var out *compiler.Value
eg.Go(func() error {
defer outr.Close()
out, err = c.outputfn(gctx, outr)
return err
})
return out, eg.Wait()
} }
func (c *Client) buildfn(ctx context.Context, deployment *Deployment, fn ClientDoFunc, ch chan *bk.SolveStatus, w io.WriteCloser) error { func (c *Client) buildfn(ctx context.Context, deployment *Deployment, fn ClientDoFunc, ch chan *bk.SolveStatus) error {
lg := log.Ctx(ctx) lg := log.Ctx(ctx)
// Scan local dirs to grant access // Scan local dirs to grant access
@ -116,15 +102,6 @@ func (c *Client) buildfn(ctx context.Context, deployment *Deployment, fn ClientD
// Setup solve options // Setup solve options
opts := bk.SolveOpt{ opts := bk.SolveOpt{
LocalDirs: localdirs, LocalDirs: localdirs,
// FIXME: catch output & return as cue value
Exports: []bk.ExportEntry{
{
Type: bk.ExporterTar,
Output: func(m map[string]string) (io.WriteCloser, error) {
return w, nil
},
},
},
} }
// Call buildkit solver // Call buildkit solver
@ -154,15 +131,14 @@ func (c *Client) buildfn(ctx context.Context, deployment *Deployment, fn ClientD
span, _ := opentracing.StartSpanFromContext(ctx, "Deployment.Export") span, _ := opentracing.StartSpanFromContext(ctx, "Deployment.Export")
defer span.Finish() defer span.Finish()
stateSource, err := deployment.State().Source() computed := deployment.Computed().JSON().PrettyString()
if err != nil { st := llb.
return nil, compiler.Err(err) Scratch().
} File(
llb.Mkfile("computed.json", 0600, []byte(computed)),
st := llb.Scratch().File( llb.WithCustomName("[internal] serializing computed values"),
llb.Mkfile("state.cue", 0600, stateSource),
llb.WithCustomName("[internal] serializing state to CUE"),
) )
ref, err := s.Solve(ctx, st) ref, err := s.Solve(ctx, st)
if err != nil { if err != nil {
return nil, err return nil, err
@ -185,45 +161,6 @@ func (c *Client) buildfn(ctx context.Context, deployment *Deployment, fn ClientD
return nil return nil
} }
// Read tar export stream from buildkit Build(), and extract cue output
func (c *Client) outputfn(ctx context.Context, r io.Reader) (*compiler.Value, error) {
lg := log.Ctx(ctx)
// FIXME: merge this into deployment output.
out := compiler.EmptyStruct()
tr := tar.NewReader(r)
for {
h, err := tr.Next()
if errors.Is(err, io.EOF) {
break
}
if err != nil {
return nil, fmt.Errorf("read tar stream: %w", err)
}
lg := lg.
With().
Str("file", h.Name).
Logger()
if !strings.HasSuffix(h.Name, ".cue") {
lg.Debug().Msg("skipping non-cue file from exporter tar stream")
continue
}
lg.Debug().Msg("outputfn: compiling & merging")
v, err := compiler.Compile(h.Name, tr)
if err != nil {
return nil, err
}
if err := out.FillPath(cue.MakePath(), v); err != nil {
return nil, fmt.Errorf("%s: %w", h.Name, compiler.Err(err))
}
}
return out, nil
}
func (c *Client) logSolveStatus(ctx context.Context, ch chan *bk.SolveStatus) error { func (c *Client) logSolveStatus(ctx context.Context, ch chan *bk.SolveStatus) error {
parseName := func(v *bk.Vertex) (string, string) { parseName := func(v *bk.Vertex) (string, string) {
// Pattern: `@name@ message`. Minimal length is len("@X@ ") // Pattern: `@name@ message`. Minimal length is len("@X@ ")

View File

@ -2,6 +2,7 @@ package compiler
import ( import (
"errors" "errors"
"fmt"
"sync" "sync"
"cuelang.org/go/cue" "cuelang.org/go/cue"
@ -19,8 +20,8 @@ func Compile(name string, src interface{}) (*Value, error) {
return DefaultCompiler.Compile(name, src) return DefaultCompiler.Compile(name, src)
} }
func EmptyStruct() *Value { func NewValue() *Value {
return DefaultCompiler.EmptyStruct() return DefaultCompiler.NewValue()
} }
// FIXME can be refactored away now? // FIXME can be refactored away now?
@ -28,6 +29,10 @@ func Wrap(v cue.Value, inst *cue.Instance) *Value {
return DefaultCompiler.Wrap(v, inst) return DefaultCompiler.Wrap(v, inst)
} }
func InstanceMerge(src ...*Value) (*Value, error) {
return DefaultCompiler.InstanceMerge(src...)
}
func Cue() *cue.Runtime { func Cue() *cue.Runtime {
return DefaultCompiler.Cue() return DefaultCompiler.Cue()
} }
@ -72,9 +77,9 @@ func (c *Compiler) Cue() *cue.Runtime {
return &(c.Runtime) return &(c.Runtime)
} }
// Compile an empty struct // Compile an empty value
func (c *Compiler) EmptyStruct() *Value { func (c *Compiler) NewValue() *Value {
empty, err := c.Compile("", "") empty, err := c.Compile("", "_")
if err != nil { if err != nil {
panic(err) panic(err)
} }
@ -93,6 +98,34 @@ func (c *Compiler) Compile(name string, src interface{}) (*Value, error) {
return c.Wrap(inst.Value(), inst), nil return c.Wrap(inst.Value(), inst), nil
} }
// InstanceMerge merges multiple values and mirrors the value in the cue.Instance.
// FIXME: AVOID THIS AT ALL COST
// Special case: we must return an instance with the same
// contents as v, for the purposes of cueflow.
func (c *Compiler) InstanceMerge(src ...*Value) (*Value, error) {
var (
v = c.NewValue()
inst = v.CueInst()
err error
)
c.lock()
defer c.unlock()
for _, s := range src {
inst, err = inst.Fill(s.val)
if err != nil {
return nil, fmt.Errorf("merge failed: %w", err)
}
if err := inst.Value().Err(); err != nil {
return nil, err
}
}
v = c.Wrap(inst.Value(), inst)
return v, nil
}
func (c *Compiler) DecodeJSON(path string, data []byte) (*Value, error) { func (c *Compiler) DecodeJSON(path string, data []byte) (*Value, error) {
inst, err := cuejson.Decode(c.Cue(), path, data) inst, err := cuejson.Decode(c.Cue(), path, data)
if err != nil { if err != nil {

View File

@ -18,59 +18,8 @@ import (
"github.com/rs/zerolog/log" "github.com/rs/zerolog/log"
) )
// Contents of a deployment serialized to a file
type DeploymentState struct {
// Globally unique deployment ID
ID string `json:"id,omitempty"`
// Human-friendly deployment name.
// A deployment may have more than one name.
// FIXME: store multiple names?
Name string `json:"name,omitempty"`
// Cue module containing the deployment plan
// The input's top-level artifact is used as a module directory.
PlanSource Input `json:"plan,omitempty"`
Inputs []inputKV `json:"inputs,omitempty"`
}
type inputKV struct {
Key string `json:"key,omitempty"`
Value Input `json:"value,omitempty"`
}
func (s *DeploymentState) SetInput(key string, value Input) error {
for i, inp := range s.Inputs {
if inp.Key != key {
continue
}
// Remove existing inputs with the same key
s.Inputs = append(s.Inputs[:i], s.Inputs[i+1:]...)
}
s.Inputs = append(s.Inputs, inputKV{Key: key, Value: value})
return nil
}
// Remove all inputs at the given key, including sub-keys.
// For example RemoveInputs("foo.bar") will remove all inputs
// at foo.bar, foo.bar.baz, etc.
func (s *DeploymentState) RemoveInputs(key string) error {
newInputs := make([]inputKV, 0, len(s.Inputs))
for _, i := range s.Inputs {
if i.Key == key {
continue
}
newInputs = append(newInputs, i)
}
s.Inputs = newInputs
return nil
}
type Deployment struct { type Deployment struct {
st *DeploymentState state *DeploymentState
// Layer 1: plan configuration // Layer 1: plan configuration
plan *compiler.Value plan *compiler.Value
@ -79,19 +28,16 @@ type Deployment struct {
input *compiler.Value input *compiler.Value
// Layer 3: computed values // Layer 3: computed values
output *compiler.Value computed *compiler.Value
// All layers merged together: plan + input + output
state *compiler.Value
} }
func NewDeployment(st *DeploymentState) (*Deployment, error) { func NewDeployment(st *DeploymentState) (*Deployment, error) {
empty := compiler.EmptyStruct()
d := &Deployment{ d := &Deployment{
st: st, state: st,
plan: empty,
input: empty, plan: compiler.NewValue(),
output: empty, input: compiler.NewValue(),
computed: compiler.NewValue(),
} }
// Prepare inputs // Prepare inputs
@ -109,23 +55,20 @@ func NewDeployment(st *DeploymentState) (*Deployment, error) {
return nil, err return nil, err
} }
} }
if err := d.mergeState(); err != nil {
return nil, err
}
return d, nil return d, nil
} }
func (d *Deployment) ID() string { func (d *Deployment) ID() string {
return d.st.ID return d.state.ID
} }
func (d *Deployment) Name() string { func (d *Deployment) Name() string {
return d.st.Name return d.state.Name
} }
func (d *Deployment) PlanSource() Input { func (d *Deployment) PlanSource() Input {
return d.st.PlanSource return d.state.PlanSource
} }
func (d *Deployment) Plan() *compiler.Value { func (d *Deployment) Plan() *compiler.Value {
@ -136,12 +79,8 @@ func (d *Deployment) Input() *compiler.Value {
return d.input return d.input
} }
func (d *Deployment) Output() *compiler.Value { func (d *Deployment) Computed() *compiler.Value {
return d.output return d.computed
}
func (d *Deployment) State() *compiler.Value {
return d.state
} }
// LoadPlan loads the plan // LoadPlan loads the plan
@ -149,12 +88,12 @@ func (d *Deployment) LoadPlan(ctx context.Context, s Solver) error {
span, ctx := opentracing.StartSpanFromContext(ctx, "deployment.LoadPlan") span, ctx := opentracing.StartSpanFromContext(ctx, "deployment.LoadPlan")
defer span.Finish() defer span.Finish()
planSource, err := d.st.PlanSource.Compile() planSource, err := d.state.PlanSource.Compile()
if err != nil { if err != nil {
return err return err
} }
p := NewPipeline("[internal] source", s, nil) p := NewPipeline("[internal] source", s)
// execute updater script // execute updater script
if err := p.Do(ctx, planSource); err != nil { if err := p.Do(ctx, planSource); err != nil {
return err return err
@ -171,8 +110,7 @@ func (d *Deployment) LoadPlan(ctx context.Context, s Solver) error {
} }
d.plan = plan d.plan = plan
// Commit return nil
return d.mergeState()
} }
// Scan all scripts in the deployment for references to local directories (do:"local"), // Scan all scripts in the deployment for references to local directories (do:"local"),
@ -203,15 +141,22 @@ func (d *Deployment) LocalDirs() map[string]string {
} }
// 1. Scan the deployment state // 1. Scan the deployment state
// FIXME: use a common `flow` instance to avoid rescanning the tree. // FIXME: use a common `flow` instance to avoid rescanning the tree.
inst := d.state.CueInst() src, err := compiler.InstanceMerge(d.plan, d.input)
flow := cueflow.New(&cueflow.Config{}, inst, newTaskFunc(inst, noOpRunner)) if err != nil {
panic(err)
}
flow := cueflow.New(
&cueflow.Config{},
src.CueInst(),
newTaskFunc(src.CueInst(), noOpRunner),
)
for _, t := range flow.Tasks() { for _, t := range flow.Tasks() {
v := compiler.Wrap(t.Value(), inst) v := compiler.Wrap(t.Value(), src.CueInst())
localdirs(v.Lookup("#up")) localdirs(v.Lookup("#up"))
} }
// 2. Scan the plan // 2. Scan the plan
plan, err := d.st.PlanSource.Compile() plan, err := d.state.PlanSource.Compile()
if err != nil { if err != nil {
panic(err) panic(err)
} }
@ -219,95 +164,31 @@ func (d *Deployment) LocalDirs() map[string]string {
return dirs return dirs
} }
// FIXME: this is just a 3-way merge. Add var args to compiler.Value.Merge.
func (d *Deployment) mergeState() error {
// FIXME: make this cleaner in *compiler.Value by keeping intermediary instances
// FIXME: state.CueInst() must return an instance with the same
// contents as state.v, for the purposes of cueflow.
// That is not currently how *compiler.Value works, so we prepare the cue
// instance manually.
// --> refactor the compiler.Value API to do this for us.
var (
state = compiler.EmptyStruct()
stateInst = state.CueInst()
err error
)
stateInst, err = stateInst.Fill(d.plan.Cue())
if err != nil {
return fmt.Errorf("merge base & input: %w", err)
}
stateInst, err = stateInst.Fill(d.input.Cue())
if err != nil {
return fmt.Errorf("merge base & input: %w", err)
}
stateInst, err = stateInst.Fill(d.output.Cue())
if err != nil {
return fmt.Errorf("merge output with base & input: %w", err)
}
state = compiler.Wrap(stateInst.Value(), stateInst)
// commit
d.state = state
return nil
}
type UpOpts struct{}
// Up missing values in deployment configuration, and write them to state. // Up missing values in deployment configuration, and write them to state.
func (d *Deployment) Up(ctx context.Context, s Solver, _ *UpOpts) error { func (d *Deployment) Up(ctx context.Context, s Solver) error {
span, ctx := opentracing.StartSpanFromContext(ctx, "r.Compute") span, ctx := opentracing.StartSpanFromContext(ctx, "deployment.Up")
defer span.Finish() defer span.Finish()
lg := log.Ctx(ctx) // Reset the computed values
d.computed = compiler.NewValue()
// Cueflow cue instance // Cueflow cue instance
inst := d.state.CueInst() src, err := compiler.InstanceMerge(d.plan, d.input)
// Reset the output
d.output = compiler.EmptyStruct()
// Cueflow config
flowCfg := &cueflow.Config{
UpdateFunc: func(c *cueflow.Controller, t *cueflow.Task) error {
if t == nil {
return nil
}
lg := lg.
With().
Str("component", t.Path().String()).
Str("state", t.State().String()).
Logger()
if t.State() != cueflow.Terminated {
return nil
}
// Merge task value into output
err := d.output.FillPath(t.Path(), t.Value())
if err != nil { if err != nil {
lg.
Error().
Err(err).
Msg("failed to fill task result")
return err return err
} }
return nil
},
}
// Orchestrate execution with cueflow // Orchestrate execution with cueflow
flow := cueflow.New(flowCfg, inst, newTaskFunc(inst, newPipelineRunner(inst, s))) flow := cueflow.New(
&cueflow.Config{},
src.CueInst(),
newTaskFunc(src.CueInst(), newPipelineRunner(src.CueInst(), d.computed, s)),
)
if err := flow.Run(ctx); err != nil { if err := flow.Run(ctx); err != nil {
return err return err
} }
{ return nil
span, _ := opentracing.StartSpanFromContext(ctx, "merge state")
defer span.Finish()
return d.mergeState()
}
} }
type DownOpts struct{} type DownOpts struct{}
@ -333,7 +214,7 @@ func noOpRunner(t *cueflow.Task) error {
return nil return nil
} }
func newPipelineRunner(inst *cue.Instance, s Solver) cueflow.RunnerFunc { func newPipelineRunner(inst *cue.Instance, computed *compiler.Value, s Solver) cueflow.RunnerFunc {
return cueflow.RunnerFunc(func(t *cueflow.Task) error { return cueflow.RunnerFunc(func(t *cueflow.Task) error {
ctx := t.Context() ctx := t.Context()
lg := log. lg := log.
@ -358,7 +239,7 @@ func newPipelineRunner(inst *cue.Instance, s Solver) cueflow.RunnerFunc {
Msg("dependency detected") Msg("dependency detected")
} }
v := compiler.Wrap(t.Value(), inst) v := compiler.Wrap(t.Value(), inst)
p := NewPipeline(t.Path().String(), s, NewFillable(t)) p := NewPipeline(t.Path().String(), s)
err := p.Do(ctx, v) err := p.Do(ctx, v)
if err != nil { if err != nil {
span.LogFields(otlog.String("error", err.Error())) span.LogFields(otlog.String("error", err.Error()))
@ -379,6 +260,29 @@ func newPipelineRunner(inst *cue.Instance, s Solver) cueflow.RunnerFunc {
Msg("failed") Msg("failed")
return err return err
} }
// Mirror the computed values in both `Task` and `Result`
if p.Computed().IsEmptyStruct() {
return nil
}
if err := t.Fill(p.Computed().Cue()); err != nil {
lg.
Error().
Err(err).
Msg("failed to fill task")
return err
}
// Merge task value into output
if err := computed.FillPath(t.Path(), p.Computed()); err != nil {
lg.
Error().
Err(err).
Msg("failed to fill task result")
return err
}
lg. lg.
Info(). Info().
Dur("duration", time.Since(start)). Dur("duration", time.Since(start)).

View File

@ -10,6 +10,7 @@ import (
"path" "path"
"strings" "strings"
"cuelang.org/go/cue"
"github.com/docker/distribution/reference" "github.com/docker/distribution/reference"
bk "github.com/moby/buildkit/client" bk "github.com/moby/buildkit/client"
"github.com/moby/buildkit/client/llb" "github.com/moby/buildkit/client/llb"
@ -33,15 +34,15 @@ type Pipeline struct {
s Solver s Solver
state llb.State state llb.State
result bkgw.Reference result bkgw.Reference
out *Fillable computed *compiler.Value
} }
func NewPipeline(name string, s Solver, out *Fillable) *Pipeline { func NewPipeline(name string, s Solver) *Pipeline {
return &Pipeline{ return &Pipeline{
name: name, name: name,
s: s, s: s,
state: llb.Scratch(), state: llb.Scratch(),
out: out, computed: compiler.NewValue(),
} }
} }
@ -60,6 +61,10 @@ func (p *Pipeline) FS() fs.FS {
return NewBuildkitFS(p.result) return NewBuildkitFS(p.result)
} }
func (p *Pipeline) Computed() *compiler.Value {
return p.computed
}
func isComponent(v *compiler.Value) bool { func isComponent(v *compiler.Value) bool {
return v.Lookup("#up").Exists() return v.Lookup("#up").Exists()
} }
@ -246,7 +251,7 @@ func (p *Pipeline) Copy(ctx context.Context, op *compiler.Value, st llb.State) (
return st, err return st, err
} }
// Execute 'from' in a tmp pipeline, and use the resulting fs // Execute 'from' in a tmp pipeline, and use the resulting fs
from := NewPipeline(op.Lookup("from").Path().String(), p.s, nil) from := NewPipeline(op.Lookup("from").Path().String(), p.s)
if err := from.Do(ctx, op.Lookup("from")); err != nil { if err := from.Do(ctx, op.Lookup("from")); err != nil {
return st, err return st, err
} }
@ -446,7 +451,7 @@ func (p *Pipeline) mount(ctx context.Context, dest string, mnt *compiler.Value)
} }
} }
// eg. mount: "/foo": { from: www.source } // eg. mount: "/foo": { from: www.source }
from := NewPipeline(mnt.Lookup("from").Path().String(), p.s, nil) from := NewPipeline(mnt.Lookup("from").Path().String(), p.s)
if err := from.Do(ctx, mnt.Lookup("from")); err != nil { if err := from.Do(ctx, mnt.Lookup("from")); err != nil {
return nil, err return nil, err
} }
@ -488,7 +493,7 @@ func (p *Pipeline) Export(ctx context.Context, op *compiler.Value, st llb.State)
Bytes("contents", contents). Bytes("contents", contents).
Msg("exporting string") Msg("exporting string")
if err := p.out.Fill(string(contents)); err != nil { if err := p.computed.FillPath(cue.MakePath(), string(contents)); err != nil {
return st, err return st, err
} }
case "json": case "json":
@ -504,7 +509,7 @@ func (p *Pipeline) Export(ctx context.Context, op *compiler.Value, st llb.State)
Interface("contents", o). Interface("contents", o).
Msg("exporting json") Msg("exporting json")
if err := p.out.Fill(o); err != nil { if err := p.computed.FillPath(cue.MakePath(), o); err != nil {
return st, err return st, err
} }
case "yaml": case "yaml":
@ -520,7 +525,7 @@ func (p *Pipeline) Export(ctx context.Context, op *compiler.Value, st llb.State)
Interface("contents", o). Interface("contents", o).
Msg("exporting yaml") Msg("exporting yaml")
if err := p.out.Fill(o); err != nil { if err := p.computed.FillPath(cue.MakePath(), o); err != nil {
return st, err return st, err
} }
default: default:
@ -550,7 +555,7 @@ func unmarshalAnything(data []byte, fn unmarshaller) (interface{}, error) {
func (p *Pipeline) Load(ctx context.Context, op *compiler.Value, st llb.State) (llb.State, error) { func (p *Pipeline) Load(ctx context.Context, op *compiler.Value, st llb.State) (llb.State, error) {
// Execute 'from' in a tmp pipeline, and use the resulting fs // Execute 'from' in a tmp pipeline, and use the resulting fs
from := NewPipeline(op.Lookup("from").Path().String(), p.s, nil) from := NewPipeline(op.Lookup("from").Path().String(), p.s)
if err := from.Do(ctx, op.Lookup("from")); err != nil { if err := from.Do(ctx, op.Lookup("from")); err != nil {
return st, err return st, err
} }
@ -683,7 +688,7 @@ func (p *Pipeline) DockerBuild(ctx context.Context, op *compiler.Value, st llb.S
// docker build context. This can come from another component, so we need to // docker build context. This can come from another component, so we need to
// compute it first. // compute it first.
if context.Exists() { if context.Exists() {
from := NewPipeline(op.Lookup("context").Path().String(), p.s, nil) from := NewPipeline(op.Lookup("context").Path().String(), p.s)
if err := from.Do(ctx, context); err != nil { if err := from.Do(ctx, context); err != nil {
return st, err return st, err
} }

56
dagger/state.go Normal file
View File

@ -0,0 +1,56 @@
package dagger
// Contents of a deployment serialized to a file
type DeploymentState struct {
// Globally unique deployment ID
ID string `json:"id,omitempty"`
// Human-friendly deployment name.
// A deployment may have more than one name.
// FIXME: store multiple names?
Name string `json:"name,omitempty"`
// Cue module containing the deployment plan
// The input's top-level artifact is used as a module directory.
PlanSource Input `json:"plan,omitempty"`
// User Inputs
Inputs []inputKV `json:"inputs,omitempty"`
// Computed values
Computed string `json:"output,omitempty"`
}
type inputKV struct {
Key string `json:"key,omitempty"`
Value Input `json:"value,omitempty"`
}
func (s *DeploymentState) SetInput(key string, value Input) error {
for i, inp := range s.Inputs {
if inp.Key != key {
continue
}
// Remove existing inputs with the same key
s.Inputs = append(s.Inputs[:i], s.Inputs[i+1:]...)
}
s.Inputs = append(s.Inputs, inputKV{Key: key, Value: value})
return nil
}
// Remove all inputs at the given key, including sub-keys.
// For example RemoveInputs("foo.bar") will remove all inputs
// at foo.bar, foo.bar.baz, etc.
func (s *DeploymentState) RemoveInputs(key string) error {
newInputs := make([]inputKV, 0, len(s.Inputs))
for _, i := range s.Inputs {
if i.Key == key {
continue
}
newInputs = append(newInputs, i)
}
s.Inputs = newInputs
return nil
}

View File

@ -1,28 +0,0 @@
package dagger
import (
"os"
cueflow "cuelang.org/go/tools/flow"
)
var ErrNotExist = os.ErrNotExist
// Something which can be filled in-place with a cue value
type Fillable struct {
t *cueflow.Task
}
func NewFillable(t *cueflow.Task) *Fillable {
return &Fillable{
t: t,
}
}
func (f *Fillable) Fill(x interface{}) error {
// Use a nil pointer receiver to discard all values
if f == nil {
return nil
}
return f.t.Fill(x)
}

View File

@ -15,7 +15,7 @@ repo: git.#Repository & {
// Host the application with Netlify // Host the application with Netlify
www: netlify.#Site & { www: netlify.#Site & {
// Site name can be overridden // Site name can be overridden
name: string | *"dagger-example-react" name: string | *"dagger-examples-react"
// Deploy the output of yarn build // Deploy the output of yarn build
// (Netlify build feature is not used, to avoid extra cost). // (Netlify build feature is not used, to avoid extra cost).

View File

@ -1,3 +0,0 @@
package testing
nonConcrete: string

View File

@ -4,8 +4,18 @@ import "dagger.io/dagger/op"
foo: "value" foo: "value"
bar: "another value" bar: "another value"
computed: {
#up: [ string
#up: [
op.#FetchContainer & {ref: "busybox"}, op.#FetchContainer & {ref: "busybox"},
op.#Exec & {args: ["true"]}, op.#Exec & {
] args: ["sh", "-c", """
printf test > /export
"""]
},
op.#Export & {
source: "/export"
format: "string"
},
]
}

View File

@ -1,14 +1,14 @@
todoApp: www:
account: account:
name: ENC[AES256_GCM,data:/T5E/8rU3YVxpA==,iv:dscf+kcVg0jYd5ICyJJL/f80BHyUDtfHWW9cRSP7aCU=,tag:xHZpdqNoYa58NbAX691vhg==,type:str] name: ENC[AES256_GCM,data:EsPTWeiDCrVeUQ==,iv:9/tZQOrrjQejsK6NFcgQO9HaAnjIUv1Qc+S0slds+4o=,tag:n3NYautI94ilmEWG9UeFzA==,type:str]
token: ENC[AES256_GCM,data:6RBg3KafhrJYS8hbACxNRtnkpSNeablIiqHhUZTjz04X3cPyP4igPn7cnw==,iv:kj0vn83GaAtocUBdacO5D+6SJu6sVxfs58LYvUdpivA=,tag:CzZFkr9sXFHOOQihqYBdTA==,type:str] token: ENC[AES256_GCM,data:Jx7oVJXcMX3hBmC6Kld7jxOOH/3CGSAzC7rRhHgs25iLFZG+F3iN5fYYVA==,iv:9SVNNv5CTM0AZns0x7x5bSI6jW93jSh8Xt21hXN1g28=,tag:QKBf7OsKs1TxvMA4gdA53Q==,type:str]
sops: sops:
kms: [] kms: []
gcp_kms: [] gcp_kms: []
azure_kv: [] azure_kv: []
hc_vault: [] hc_vault: []
lastmodified: '2021-03-19T00:41:47Z' lastmodified: '2021-04-09T01:47:48Z'
mac: ENC[AES256_GCM,data:eaDHLQHEH8wtCiDhcl9wWIC+WwitEBRr4K/p7chOKJ8sBc3wlH+dT3HVX5hTYX5rwU0ai2wgfcOg6mw4mDksIJvU4JpZV/AIo16ETjZDhRSc5hU1hUAiOQCYuJPqyGa12HzxzA81mcFm8kJETnoTYiTo7ApG9ShmOCp6XPezjPk=,iv:5Pd9eGTrVnr0ITNj/TrgJz9UE2SjVqqmOxxZOw0Spus=,tag:VGxPwp7hXHtz9Zvk+6WsqA==,type:str] mac: ENC[AES256_GCM,data:UqOr8wGUwf6iVnurG/dvpiZVN0k9NrLTaB5CQVn+QTRQybgYuLZLTJuTFNlAqFGvNO07OoGkDx/Vmhw9F6nJS0qUcHC5iWg+Bxaa7anHwer7fkA/xTjKpqJnE6iveq+hzumDgeFbGL7+EJvbyxtJioF/LnWa5gnduBVacizjbKc=,iv:x4ulAfd2R0BdkGF4zkMn+wX+Y8wF3jEeFOu3+5t+wz4=,tag:e5uOip0iAt9xJa6RJGukCA==,type:str]
pgp: pgp:
- created_at: '2021-03-18T22:02:51Z' - created_at: '2021-03-18T22:02:51Z'
enc: | enc: |

View File

@ -51,10 +51,11 @@ test::cli::newdir() {
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "simple" "$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "simple"
test::one "CLI: new: verify we have the right plan" --stdout='{ test::one "CLI: new: verify we have the right plan" --stdout='{
foo: "value" "bar": "another value",
bar: "another value" "computed": "test",
"foo": "value"
}' \ }' \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -f cue -d "simple" -c "$dagger" "${DAGGER_BINARY_ARGS[@]}" query -f cue -d "simple" -c -f json
} }
test::cli::newgit() { test::cli::newgit() {
@ -65,13 +66,13 @@ test::cli::newgit() {
DAGGER_STORE="$(mktemp -d -t dagger-store-XXXXXX)" DAGGER_STORE="$(mktemp -d -t dagger-store-XXXXXX)"
export DAGGER_STORE export DAGGER_STORE
disable test::one "CLI: new git: --plan-git" \ test::one "CLI: new git: --plan-git" \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-git https://github.com/samalba/dagger-test.git simple "$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-git https://github.com/samalba/dagger-test.git simple
disable test::one "CLI: new git: verify plan can be upped" \ test::one "CLI: new git: verify plan can be upped" \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "simple" "$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "simple"
disable test::one "CLI: new git: verify we have the right plan" --stdout='{ test::one "CLI: new git: verify we have the right plan" --stdout='{
foo: "value" foo: "value"
bar: "another value" bar: "another value"
}' \ }' \
@ -89,20 +90,45 @@ test::cli::query() {
test::one "CLI: query: initialize simple" \ test::one "CLI: query: initialize simple" \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-dir "$d"/cli/simple simple "$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-dir "$d"/cli/simple simple
test::one "CLI: query: concrete" --stdout='{ test::one "CLI: query: before up" --stdout='{
foo: "value" "bar": "another value",
bar: "another value" "foo": "value"
}' \ }' \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -f cue -d "simple" -c "$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple"
test::one "CLI: query: concrete should fail" --exit=1 \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple" -c
test::one "CLI: query: target" --stdout='"value"' \ test::one "CLI: query: target" --stdout='"value"' \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -f cue -d "simple" foo "$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple" foo
test::one "CLI: query: initialize nonconcrete" \ test::one "CLI: query: compute missing values" \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-dir "$d"/cli/nonconcrete nonconcrete "$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "simple"
test::one "CLI: query: non concrete" --exit=1 \ test::one "CLI: query: all values" --stdout='{
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -f cue -d "nonconcrete" -c "bar": "another value",
"computed": "test",
"foo": "value"
}' \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple"
test::one "CLI: query: concrete should work" --stdout='{
"bar": "another value",
"computed": "test",
"foo": "value"
}' \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple" -c
test::one "CLI: query --no-computed" --stdout='{
"bar": "another value",
"foo": "value"
}' \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple" --no-computed
test::one "CLI: query: --no-plan" --stdout='{
"computed": "test"
}' \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple" -c --no-plan
} }
test::cli::plan() { test::cli::plan() {
@ -117,19 +143,22 @@ test::cli::plan() {
"$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-dir "$d"/cli/simple simple "$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-dir "$d"/cli/simple simple
test::one "CLI: plan dir" \ test::one "CLI: plan dir" \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" -d "simple" plan dir "$d"/cli/nonconcrete "$dagger" "${DAGGER_BINARY_ARGS[@]}" -d "simple" plan dir "$d"/cli/simple
test::one "CLI: plan dir: query non-concrete" --exit=1 \ test::one "CLI: plan dir: verify we have the right plan" --stdout='{
"$dagger" "${DAGGER_BINARY_ARGS[@]}" -d "simple" query -c "bar": "another value",
"foo": "value"
}' \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" -d "simple" query
disable test::one "CLI: plan git (Disabled: external dependency)" \ test::one "CLI: plan git" \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" -d "simple" plan git https://github.com/samalba/dagger-test.git "$dagger" "${DAGGER_BINARY_ARGS[@]}" -d "simple" plan git https://github.com/samalba/dagger-test.git
disable test::one "CLI: plan git: verify we have the right plan" --stdout='{ test::one "CLI: plan git: verify we have the right plan" --stdout='{
foo: "value" "bar": "another value",
bar: "another value" "foo": "value"
}' \ }' \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -f cue -d "simple" -c "$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple" -c
} }
test::cli::input() { test::cli::input() {
@ -140,21 +169,40 @@ test::cli::input() {
DAGGER_STORE="$(mktemp -d -t dagger-store-XXXXXX)" DAGGER_STORE="$(mktemp -d -t dagger-store-XXXXXX)"
export DAGGER_STORE export DAGGER_STORE
test::one "CLI: new input" \ test::one "CLI: input: new" \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-dir "$d"/cli/input "input" "$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-dir "$d"/cli/input "input"
test::one "CLI: up: missing input" \ test::one "CLI: input: up missing input" \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "input" --stdout='{"foo":"bar"}' "$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "input"
test::one "CLI: input dir" \ test::one "CLI: input: query missing input" \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "input" --stdout='{
"foo": "bar"
}'
test::one "CLI: input: set dir" \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" input -d "input" dir "source" ./tests/cli/input/testdata "$dagger" "${DAGGER_BINARY_ARGS[@]}" input -d "input" dir "source" ./tests/cli/input/testdata
test::one "CLI: up: input is set with input dir" \ test::one "CLI: input: up with input dir" \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "input" --stdout='{"bar":"thisisatest\n","foo":"bar","source":{}}' "$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "input"
disable test::one "CLI: input git" \ test::one "CLI: input: query with input dir" \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "input" --stdout='{
"bar": "thisisatest\n",
"foo": "bar",
"source": {}
}'
test::one "CLI: input: set git" \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" input -d "input" git "source" https://github.com/samalba/dagger-test-simple.git "$dagger" "${DAGGER_BINARY_ARGS[@]}" input -d "input" git "source" https://github.com/samalba/dagger-test-simple.git
disable test::one "CLI: up: input is set with input git" \ test::one "CLI: input: up with input git" \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "input" --stdout='{"bar":"testgit\n","foo":"bar","source":{}}' "$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "input"
test::one "CLI: query with input git" \
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "input" --stdout='{
"bar": "testgit\n",
"foo": "bar",
"source": {}
}'
} }