Merge pull request #288 from dagger/dagger-up-store-outputs
up: store outputs and merge in query
This commit is contained in:
commit
92ef1abd28
@ -2,7 +2,6 @@ package common
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
|
||||||
"os"
|
"os"
|
||||||
|
|
||||||
"dagger.io/go/dagger"
|
"dagger.io/go/dagger"
|
||||||
@ -61,22 +60,19 @@ func GetCurrentDeploymentState(ctx context.Context, store *dagger.Store) *dagger
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Re-compute a deployment (equivalent to `dagger up`).
|
// Re-compute a deployment (equivalent to `dagger up`).
|
||||||
// If printOutput is true, print the JSON-encoded computed state to standard output
|
func DeploymentUp(ctx context.Context, state *dagger.DeploymentState) *dagger.Deployment {
|
||||||
func DeploymentUp(ctx context.Context, state *dagger.DeploymentState, printOutput bool) {
|
|
||||||
lg := log.Ctx(ctx)
|
lg := log.Ctx(ctx)
|
||||||
|
|
||||||
c, err := dagger.NewClient(ctx, "")
|
c, err := dagger.NewClient(ctx, "")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
lg.Fatal().Err(err).Msg("unable to create client")
|
lg.Fatal().Err(err).Msg("unable to create client")
|
||||||
}
|
}
|
||||||
output, err := c.Do(ctx, state, func(ctx context.Context, deployment *dagger.Deployment, s dagger.Solver) error {
|
result, err := c.Do(ctx, state, func(ctx context.Context, deployment *dagger.Deployment, s dagger.Solver) error {
|
||||||
log.Ctx(ctx).Debug().Msg("bringing deployment up")
|
log.Ctx(ctx).Debug().Msg("bringing deployment up")
|
||||||
return deployment.Up(ctx, s, nil)
|
return deployment.Up(ctx, s)
|
||||||
})
|
})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
lg.Fatal().Err(err).Msg("failed to up deployment")
|
lg.Fatal().Err(err).Msg("failed to up deployment")
|
||||||
}
|
}
|
||||||
if printOutput {
|
return result
|
||||||
fmt.Println(output.JSON())
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
@ -3,12 +3,15 @@ package cmd
|
|||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"errors"
|
"errors"
|
||||||
|
"fmt"
|
||||||
"os"
|
"os"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
|
"cuelang.org/go/cue"
|
||||||
"dagger.io/go/cmd/dagger/cmd/common"
|
"dagger.io/go/cmd/dagger/cmd/common"
|
||||||
"dagger.io/go/cmd/dagger/logger"
|
"dagger.io/go/cmd/dagger/logger"
|
||||||
"dagger.io/go/dagger"
|
"dagger.io/go/dagger"
|
||||||
|
"dagger.io/go/dagger/compiler"
|
||||||
"go.mozilla.org/sops/v3"
|
"go.mozilla.org/sops/v3"
|
||||||
"go.mozilla.org/sops/v3/decrypt"
|
"go.mozilla.org/sops/v3/decrypt"
|
||||||
|
|
||||||
@ -127,7 +130,20 @@ var computeCmd = &cobra.Command{
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
common.DeploymentUp(ctx, st, true)
|
deployment := common.DeploymentUp(ctx, st)
|
||||||
|
|
||||||
|
v := compiler.NewValue()
|
||||||
|
if err := v.FillPath(cue.MakePath(), deployment.Plan()); err != nil {
|
||||||
|
lg.Fatal().Err(err).Msg("failed to merge")
|
||||||
|
}
|
||||||
|
if err := v.FillPath(cue.MakePath(), deployment.Input()); err != nil {
|
||||||
|
lg.Fatal().Err(err).Msg("failed to merge")
|
||||||
|
}
|
||||||
|
if err := v.FillPath(cue.MakePath(), deployment.Computed()); err != nil {
|
||||||
|
lg.Fatal().Err(err).Msg("failed to merge")
|
||||||
|
}
|
||||||
|
|
||||||
|
fmt.Println(v.JSON())
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -63,7 +63,7 @@ var newCmd = &cobra.Command{
|
|||||||
Msg("deployment created")
|
Msg("deployment created")
|
||||||
|
|
||||||
if viper.GetBool("up") {
|
if viper.GetBool("up") {
|
||||||
common.DeploymentUp(ctx, st, false)
|
common.DeploymentUp(ctx, st)
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
@ -51,12 +51,37 @@ var queryCmd = &cobra.Command{
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
lg.Fatal().Err(err).Msg("unable to create client")
|
lg.Fatal().Err(err).Msg("unable to create client")
|
||||||
}
|
}
|
||||||
output, err := c.Do(ctx, state, nil)
|
|
||||||
|
deployment, err := c.Do(ctx, state, nil)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
lg.Fatal().Err(err).Msg("failed to query deployment")
|
lg.Fatal().Err(err).Msg("failed to query deployment")
|
||||||
}
|
}
|
||||||
|
|
||||||
cueVal := output.LookupPath(cuePath)
|
cueVal := compiler.NewValue()
|
||||||
|
|
||||||
|
if !viper.GetBool("no-plan") {
|
||||||
|
if err := cueVal.FillPath(cue.MakePath(), deployment.Plan()); err != nil {
|
||||||
|
lg.Fatal().Err(err).Msg("failed to merge plan")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if !viper.GetBool("no-input") {
|
||||||
|
if err := cueVal.FillPath(cue.MakePath(), deployment.Input()); err != nil {
|
||||||
|
lg.Fatal().Err(err).Msg("failed to merge plan with output")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if !viper.GetBool("no-computed") && state.Computed != "" {
|
||||||
|
computed, err := compiler.DecodeJSON("", []byte(state.Computed))
|
||||||
|
if err != nil {
|
||||||
|
lg.Fatal().Err(err).Msg("failed to decode json")
|
||||||
|
}
|
||||||
|
if err := cueVal.FillPath(cue.MakePath(), computed); err != nil {
|
||||||
|
lg.Fatal().Err(err).Msg("failed to merge plan with computed")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
cueVal = cueVal.LookupPath(cuePath)
|
||||||
|
|
||||||
if viper.GetBool("concrete") {
|
if viper.GetBool("concrete") {
|
||||||
if err := cueVal.IsConcreteR(); err != nil {
|
if err := cueVal.IsConcreteR(); err != nil {
|
||||||
@ -116,9 +141,9 @@ func init() {
|
|||||||
// FIXME: implement the flags below
|
// FIXME: implement the flags below
|
||||||
// queryCmd.Flags().String("revision", "latest", "Query a specific version of the deployment")
|
// queryCmd.Flags().String("revision", "latest", "Query a specific version of the deployment")
|
||||||
queryCmd.Flags().StringP("format", "f", "json", "Output format (json|yaml|cue|text|env)")
|
queryCmd.Flags().StringP("format", "f", "json", "Output format (json|yaml|cue|text|env)")
|
||||||
// queryCmd.Flags().BoolP("no-input", "I", false, "Exclude inputs from query")
|
queryCmd.Flags().BoolP("no-plan", "P", false, "Exclude plan from query")
|
||||||
// queryCmd.Flags().BoolP("no-output", "O", false, "Exclude outputs from query")
|
queryCmd.Flags().BoolP("no-input", "I", false, "Exclude inputs from query")
|
||||||
// queryCmd.Flags().BoolP("no-plan", "P", false, "Exclude outputs from query")
|
queryCmd.Flags().BoolP("no-computed", "C", false, "Exclude computed values from query")
|
||||||
|
|
||||||
if err := viper.BindPFlags(queryCmd.Flags()); err != nil {
|
if err := viper.BindPFlags(queryCmd.Flags()); err != nil {
|
||||||
panic(err)
|
panic(err)
|
||||||
|
@ -31,7 +31,11 @@ var upCmd = &cobra.Command{
|
|||||||
state := common.GetCurrentDeploymentState(ctx, store)
|
state := common.GetCurrentDeploymentState(ctx, store)
|
||||||
|
|
||||||
// TODO: Implement options: --no-cache
|
// TODO: Implement options: --no-cache
|
||||||
common.DeploymentUp(ctx, state, true)
|
result := common.DeploymentUp(ctx, state)
|
||||||
|
state.Computed = result.Computed().JSON().String()
|
||||||
|
if err := store.UpdateDeployment(ctx, state, nil); err != nil {
|
||||||
|
lg.Fatal().Err(err).Msg("failed to update deployment")
|
||||||
|
}
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -7,6 +7,7 @@ import (
|
|||||||
"hash/adler32"
|
"hash/adler32"
|
||||||
"io"
|
"io"
|
||||||
"strings"
|
"strings"
|
||||||
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/mitchellh/colorstring"
|
"github.com/mitchellh/colorstring"
|
||||||
@ -21,6 +22,7 @@ var colorize = colorstring.Colorize{
|
|||||||
type Console struct {
|
type Console struct {
|
||||||
Out io.Writer
|
Out io.Writer
|
||||||
maxLength int
|
maxLength int
|
||||||
|
l sync.Mutex
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *Console) Write(p []byte) (n int, err error) {
|
func (c *Console) Write(p []byte) (n int, err error) {
|
||||||
@ -31,9 +33,12 @@ func (c *Console) Write(p []byte) (n int, err error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
source := c.parseSource(event)
|
source := c.parseSource(event)
|
||||||
|
|
||||||
|
c.l.Lock()
|
||||||
if len(source) > c.maxLength {
|
if len(source) > c.maxLength {
|
||||||
c.maxLength = len(source)
|
c.maxLength = len(source)
|
||||||
}
|
}
|
||||||
|
c.l.Unlock()
|
||||||
|
|
||||||
return fmt.Fprintln(c.Out,
|
return fmt.Fprintln(c.Out,
|
||||||
colorize.Color(fmt.Sprintf("%s %s %s%s%s",
|
colorize.Color(fmt.Sprintf("%s %s %s%s%s",
|
||||||
|
@ -1,16 +1,12 @@
|
|||||||
package dagger
|
package dagger
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"archive/tar"
|
|
||||||
"context"
|
"context"
|
||||||
"errors"
|
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
|
||||||
"os"
|
"os"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"cuelang.org/go/cue"
|
|
||||||
"golang.org/x/sync/errgroup"
|
"golang.org/x/sync/errgroup"
|
||||||
|
|
||||||
"github.com/opentracing/opentracing-go"
|
"github.com/opentracing/opentracing-go"
|
||||||
@ -64,7 +60,7 @@ func NewClient(ctx context.Context, host string) (*Client, error) {
|
|||||||
type ClientDoFunc func(context.Context, *Deployment, Solver) error
|
type ClientDoFunc func(context.Context, *Deployment, Solver) error
|
||||||
|
|
||||||
// FIXME: return completed *Route, instead of *compiler.Value
|
// FIXME: return completed *Route, instead of *compiler.Value
|
||||||
func (c *Client) Do(ctx context.Context, state *DeploymentState, fn ClientDoFunc) (*compiler.Value, error) {
|
func (c *Client) Do(ctx context.Context, state *DeploymentState, fn ClientDoFunc) (*Deployment, error) {
|
||||||
lg := log.Ctx(ctx)
|
lg := log.Ctx(ctx)
|
||||||
eg, gctx := errgroup.WithContext(ctx)
|
eg, gctx := errgroup.WithContext(ctx)
|
||||||
|
|
||||||
@ -83,24 +79,14 @@ func (c *Client) Do(ctx context.Context, state *DeploymentState, fn ClientDoFunc
|
|||||||
})
|
})
|
||||||
|
|
||||||
// Spawn build function
|
// Spawn build function
|
||||||
outr, outw := io.Pipe()
|
|
||||||
eg.Go(func() error {
|
eg.Go(func() error {
|
||||||
defer outw.Close()
|
return c.buildfn(gctx, deployment, fn, events)
|
||||||
return c.buildfn(gctx, deployment, fn, events, outw)
|
|
||||||
})
|
})
|
||||||
|
|
||||||
// Spawn output retriever
|
return deployment, eg.Wait()
|
||||||
var out *compiler.Value
|
|
||||||
eg.Go(func() error {
|
|
||||||
defer outr.Close()
|
|
||||||
out, err = c.outputfn(gctx, outr)
|
|
||||||
return err
|
|
||||||
})
|
|
||||||
|
|
||||||
return out, eg.Wait()
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *Client) buildfn(ctx context.Context, deployment *Deployment, fn ClientDoFunc, ch chan *bk.SolveStatus, w io.WriteCloser) error {
|
func (c *Client) buildfn(ctx context.Context, deployment *Deployment, fn ClientDoFunc, ch chan *bk.SolveStatus) error {
|
||||||
lg := log.Ctx(ctx)
|
lg := log.Ctx(ctx)
|
||||||
|
|
||||||
// Scan local dirs to grant access
|
// Scan local dirs to grant access
|
||||||
@ -116,15 +102,6 @@ func (c *Client) buildfn(ctx context.Context, deployment *Deployment, fn ClientD
|
|||||||
// Setup solve options
|
// Setup solve options
|
||||||
opts := bk.SolveOpt{
|
opts := bk.SolveOpt{
|
||||||
LocalDirs: localdirs,
|
LocalDirs: localdirs,
|
||||||
// FIXME: catch output & return as cue value
|
|
||||||
Exports: []bk.ExportEntry{
|
|
||||||
{
|
|
||||||
Type: bk.ExporterTar,
|
|
||||||
Output: func(m map[string]string) (io.WriteCloser, error) {
|
|
||||||
return w, nil
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Call buildkit solver
|
// Call buildkit solver
|
||||||
@ -154,15 +131,14 @@ func (c *Client) buildfn(ctx context.Context, deployment *Deployment, fn ClientD
|
|||||||
span, _ := opentracing.StartSpanFromContext(ctx, "Deployment.Export")
|
span, _ := opentracing.StartSpanFromContext(ctx, "Deployment.Export")
|
||||||
defer span.Finish()
|
defer span.Finish()
|
||||||
|
|
||||||
stateSource, err := deployment.State().Source()
|
computed := deployment.Computed().JSON().PrettyString()
|
||||||
if err != nil {
|
st := llb.
|
||||||
return nil, compiler.Err(err)
|
Scratch().
|
||||||
}
|
File(
|
||||||
|
llb.Mkfile("computed.json", 0600, []byte(computed)),
|
||||||
st := llb.Scratch().File(
|
llb.WithCustomName("[internal] serializing computed values"),
|
||||||
llb.Mkfile("state.cue", 0600, stateSource),
|
|
||||||
llb.WithCustomName("[internal] serializing state to CUE"),
|
|
||||||
)
|
)
|
||||||
|
|
||||||
ref, err := s.Solve(ctx, st)
|
ref, err := s.Solve(ctx, st)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
@ -185,45 +161,6 @@ func (c *Client) buildfn(ctx context.Context, deployment *Deployment, fn ClientD
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Read tar export stream from buildkit Build(), and extract cue output
|
|
||||||
func (c *Client) outputfn(ctx context.Context, r io.Reader) (*compiler.Value, error) {
|
|
||||||
lg := log.Ctx(ctx)
|
|
||||||
|
|
||||||
// FIXME: merge this into deployment output.
|
|
||||||
out := compiler.EmptyStruct()
|
|
||||||
|
|
||||||
tr := tar.NewReader(r)
|
|
||||||
for {
|
|
||||||
h, err := tr.Next()
|
|
||||||
if errors.Is(err, io.EOF) {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("read tar stream: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
lg := lg.
|
|
||||||
With().
|
|
||||||
Str("file", h.Name).
|
|
||||||
Logger()
|
|
||||||
|
|
||||||
if !strings.HasSuffix(h.Name, ".cue") {
|
|
||||||
lg.Debug().Msg("skipping non-cue file from exporter tar stream")
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
lg.Debug().Msg("outputfn: compiling & merging")
|
|
||||||
|
|
||||||
v, err := compiler.Compile(h.Name, tr)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
if err := out.FillPath(cue.MakePath(), v); err != nil {
|
|
||||||
return nil, fmt.Errorf("%s: %w", h.Name, compiler.Err(err))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return out, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (c *Client) logSolveStatus(ctx context.Context, ch chan *bk.SolveStatus) error {
|
func (c *Client) logSolveStatus(ctx context.Context, ch chan *bk.SolveStatus) error {
|
||||||
parseName := func(v *bk.Vertex) (string, string) {
|
parseName := func(v *bk.Vertex) (string, string) {
|
||||||
// Pattern: `@name@ message`. Minimal length is len("@X@ ")
|
// Pattern: `@name@ message`. Minimal length is len("@X@ ")
|
||||||
|
@ -2,6 +2,7 @@ package compiler
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"errors"
|
"errors"
|
||||||
|
"fmt"
|
||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
"cuelang.org/go/cue"
|
"cuelang.org/go/cue"
|
||||||
@ -19,8 +20,8 @@ func Compile(name string, src interface{}) (*Value, error) {
|
|||||||
return DefaultCompiler.Compile(name, src)
|
return DefaultCompiler.Compile(name, src)
|
||||||
}
|
}
|
||||||
|
|
||||||
func EmptyStruct() *Value {
|
func NewValue() *Value {
|
||||||
return DefaultCompiler.EmptyStruct()
|
return DefaultCompiler.NewValue()
|
||||||
}
|
}
|
||||||
|
|
||||||
// FIXME can be refactored away now?
|
// FIXME can be refactored away now?
|
||||||
@ -28,6 +29,10 @@ func Wrap(v cue.Value, inst *cue.Instance) *Value {
|
|||||||
return DefaultCompiler.Wrap(v, inst)
|
return DefaultCompiler.Wrap(v, inst)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func InstanceMerge(src ...*Value) (*Value, error) {
|
||||||
|
return DefaultCompiler.InstanceMerge(src...)
|
||||||
|
}
|
||||||
|
|
||||||
func Cue() *cue.Runtime {
|
func Cue() *cue.Runtime {
|
||||||
return DefaultCompiler.Cue()
|
return DefaultCompiler.Cue()
|
||||||
}
|
}
|
||||||
@ -72,9 +77,9 @@ func (c *Compiler) Cue() *cue.Runtime {
|
|||||||
return &(c.Runtime)
|
return &(c.Runtime)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Compile an empty struct
|
// Compile an empty value
|
||||||
func (c *Compiler) EmptyStruct() *Value {
|
func (c *Compiler) NewValue() *Value {
|
||||||
empty, err := c.Compile("", "")
|
empty, err := c.Compile("", "_")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
panic(err)
|
panic(err)
|
||||||
}
|
}
|
||||||
@ -93,6 +98,34 @@ func (c *Compiler) Compile(name string, src interface{}) (*Value, error) {
|
|||||||
return c.Wrap(inst.Value(), inst), nil
|
return c.Wrap(inst.Value(), inst), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// InstanceMerge merges multiple values and mirrors the value in the cue.Instance.
|
||||||
|
// FIXME: AVOID THIS AT ALL COST
|
||||||
|
// Special case: we must return an instance with the same
|
||||||
|
// contents as v, for the purposes of cueflow.
|
||||||
|
func (c *Compiler) InstanceMerge(src ...*Value) (*Value, error) {
|
||||||
|
var (
|
||||||
|
v = c.NewValue()
|
||||||
|
inst = v.CueInst()
|
||||||
|
err error
|
||||||
|
)
|
||||||
|
|
||||||
|
c.lock()
|
||||||
|
defer c.unlock()
|
||||||
|
|
||||||
|
for _, s := range src {
|
||||||
|
inst, err = inst.Fill(s.val)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("merge failed: %w", err)
|
||||||
|
}
|
||||||
|
if err := inst.Value().Err(); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
v = c.Wrap(inst.Value(), inst)
|
||||||
|
return v, nil
|
||||||
|
}
|
||||||
|
|
||||||
func (c *Compiler) DecodeJSON(path string, data []byte) (*Value, error) {
|
func (c *Compiler) DecodeJSON(path string, data []byte) (*Value, error) {
|
||||||
inst, err := cuejson.Decode(c.Cue(), path, data)
|
inst, err := cuejson.Decode(c.Cue(), path, data)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -18,59 +18,8 @@ import (
|
|||||||
"github.com/rs/zerolog/log"
|
"github.com/rs/zerolog/log"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Contents of a deployment serialized to a file
|
|
||||||
type DeploymentState struct {
|
|
||||||
// Globally unique deployment ID
|
|
||||||
ID string `json:"id,omitempty"`
|
|
||||||
|
|
||||||
// Human-friendly deployment name.
|
|
||||||
// A deployment may have more than one name.
|
|
||||||
// FIXME: store multiple names?
|
|
||||||
Name string `json:"name,omitempty"`
|
|
||||||
|
|
||||||
// Cue module containing the deployment plan
|
|
||||||
// The input's top-level artifact is used as a module directory.
|
|
||||||
PlanSource Input `json:"plan,omitempty"`
|
|
||||||
|
|
||||||
Inputs []inputKV `json:"inputs,omitempty"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type inputKV struct {
|
|
||||||
Key string `json:"key,omitempty"`
|
|
||||||
Value Input `json:"value,omitempty"`
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *DeploymentState) SetInput(key string, value Input) error {
|
|
||||||
for i, inp := range s.Inputs {
|
|
||||||
if inp.Key != key {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
// Remove existing inputs with the same key
|
|
||||||
s.Inputs = append(s.Inputs[:i], s.Inputs[i+1:]...)
|
|
||||||
}
|
|
||||||
|
|
||||||
s.Inputs = append(s.Inputs, inputKV{Key: key, Value: value})
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Remove all inputs at the given key, including sub-keys.
|
|
||||||
// For example RemoveInputs("foo.bar") will remove all inputs
|
|
||||||
// at foo.bar, foo.bar.baz, etc.
|
|
||||||
func (s *DeploymentState) RemoveInputs(key string) error {
|
|
||||||
newInputs := make([]inputKV, 0, len(s.Inputs))
|
|
||||||
for _, i := range s.Inputs {
|
|
||||||
if i.Key == key {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
newInputs = append(newInputs, i)
|
|
||||||
}
|
|
||||||
s.Inputs = newInputs
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
type Deployment struct {
|
type Deployment struct {
|
||||||
st *DeploymentState
|
state *DeploymentState
|
||||||
|
|
||||||
// Layer 1: plan configuration
|
// Layer 1: plan configuration
|
||||||
plan *compiler.Value
|
plan *compiler.Value
|
||||||
@ -79,19 +28,16 @@ type Deployment struct {
|
|||||||
input *compiler.Value
|
input *compiler.Value
|
||||||
|
|
||||||
// Layer 3: computed values
|
// Layer 3: computed values
|
||||||
output *compiler.Value
|
computed *compiler.Value
|
||||||
|
|
||||||
// All layers merged together: plan + input + output
|
|
||||||
state *compiler.Value
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewDeployment(st *DeploymentState) (*Deployment, error) {
|
func NewDeployment(st *DeploymentState) (*Deployment, error) {
|
||||||
empty := compiler.EmptyStruct()
|
|
||||||
d := &Deployment{
|
d := &Deployment{
|
||||||
st: st,
|
state: st,
|
||||||
plan: empty,
|
|
||||||
input: empty,
|
plan: compiler.NewValue(),
|
||||||
output: empty,
|
input: compiler.NewValue(),
|
||||||
|
computed: compiler.NewValue(),
|
||||||
}
|
}
|
||||||
|
|
||||||
// Prepare inputs
|
// Prepare inputs
|
||||||
@ -109,23 +55,20 @@ func NewDeployment(st *DeploymentState) (*Deployment, error) {
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if err := d.mergeState(); err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
return d, nil
|
return d, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d *Deployment) ID() string {
|
func (d *Deployment) ID() string {
|
||||||
return d.st.ID
|
return d.state.ID
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d *Deployment) Name() string {
|
func (d *Deployment) Name() string {
|
||||||
return d.st.Name
|
return d.state.Name
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d *Deployment) PlanSource() Input {
|
func (d *Deployment) PlanSource() Input {
|
||||||
return d.st.PlanSource
|
return d.state.PlanSource
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d *Deployment) Plan() *compiler.Value {
|
func (d *Deployment) Plan() *compiler.Value {
|
||||||
@ -136,12 +79,8 @@ func (d *Deployment) Input() *compiler.Value {
|
|||||||
return d.input
|
return d.input
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d *Deployment) Output() *compiler.Value {
|
func (d *Deployment) Computed() *compiler.Value {
|
||||||
return d.output
|
return d.computed
|
||||||
}
|
|
||||||
|
|
||||||
func (d *Deployment) State() *compiler.Value {
|
|
||||||
return d.state
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// LoadPlan loads the plan
|
// LoadPlan loads the plan
|
||||||
@ -149,12 +88,12 @@ func (d *Deployment) LoadPlan(ctx context.Context, s Solver) error {
|
|||||||
span, ctx := opentracing.StartSpanFromContext(ctx, "deployment.LoadPlan")
|
span, ctx := opentracing.StartSpanFromContext(ctx, "deployment.LoadPlan")
|
||||||
defer span.Finish()
|
defer span.Finish()
|
||||||
|
|
||||||
planSource, err := d.st.PlanSource.Compile()
|
planSource, err := d.state.PlanSource.Compile()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
p := NewPipeline("[internal] source", s, nil)
|
p := NewPipeline("[internal] source", s)
|
||||||
// execute updater script
|
// execute updater script
|
||||||
if err := p.Do(ctx, planSource); err != nil {
|
if err := p.Do(ctx, planSource); err != nil {
|
||||||
return err
|
return err
|
||||||
@ -171,8 +110,7 @@ func (d *Deployment) LoadPlan(ctx context.Context, s Solver) error {
|
|||||||
}
|
}
|
||||||
d.plan = plan
|
d.plan = plan
|
||||||
|
|
||||||
// Commit
|
return nil
|
||||||
return d.mergeState()
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Scan all scripts in the deployment for references to local directories (do:"local"),
|
// Scan all scripts in the deployment for references to local directories (do:"local"),
|
||||||
@ -203,15 +141,22 @@ func (d *Deployment) LocalDirs() map[string]string {
|
|||||||
}
|
}
|
||||||
// 1. Scan the deployment state
|
// 1. Scan the deployment state
|
||||||
// FIXME: use a common `flow` instance to avoid rescanning the tree.
|
// FIXME: use a common `flow` instance to avoid rescanning the tree.
|
||||||
inst := d.state.CueInst()
|
src, err := compiler.InstanceMerge(d.plan, d.input)
|
||||||
flow := cueflow.New(&cueflow.Config{}, inst, newTaskFunc(inst, noOpRunner))
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
flow := cueflow.New(
|
||||||
|
&cueflow.Config{},
|
||||||
|
src.CueInst(),
|
||||||
|
newTaskFunc(src.CueInst(), noOpRunner),
|
||||||
|
)
|
||||||
for _, t := range flow.Tasks() {
|
for _, t := range flow.Tasks() {
|
||||||
v := compiler.Wrap(t.Value(), inst)
|
v := compiler.Wrap(t.Value(), src.CueInst())
|
||||||
localdirs(v.Lookup("#up"))
|
localdirs(v.Lookup("#up"))
|
||||||
}
|
}
|
||||||
|
|
||||||
// 2. Scan the plan
|
// 2. Scan the plan
|
||||||
plan, err := d.st.PlanSource.Compile()
|
plan, err := d.state.PlanSource.Compile()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
panic(err)
|
panic(err)
|
||||||
}
|
}
|
||||||
@ -219,95 +164,31 @@ func (d *Deployment) LocalDirs() map[string]string {
|
|||||||
return dirs
|
return dirs
|
||||||
}
|
}
|
||||||
|
|
||||||
// FIXME: this is just a 3-way merge. Add var args to compiler.Value.Merge.
|
|
||||||
func (d *Deployment) mergeState() error {
|
|
||||||
// FIXME: make this cleaner in *compiler.Value by keeping intermediary instances
|
|
||||||
// FIXME: state.CueInst() must return an instance with the same
|
|
||||||
// contents as state.v, for the purposes of cueflow.
|
|
||||||
// That is not currently how *compiler.Value works, so we prepare the cue
|
|
||||||
// instance manually.
|
|
||||||
// --> refactor the compiler.Value API to do this for us.
|
|
||||||
var (
|
|
||||||
state = compiler.EmptyStruct()
|
|
||||||
stateInst = state.CueInst()
|
|
||||||
err error
|
|
||||||
)
|
|
||||||
|
|
||||||
stateInst, err = stateInst.Fill(d.plan.Cue())
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("merge base & input: %w", err)
|
|
||||||
}
|
|
||||||
stateInst, err = stateInst.Fill(d.input.Cue())
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("merge base & input: %w", err)
|
|
||||||
}
|
|
||||||
stateInst, err = stateInst.Fill(d.output.Cue())
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("merge output with base & input: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
state = compiler.Wrap(stateInst.Value(), stateInst)
|
|
||||||
|
|
||||||
// commit
|
|
||||||
d.state = state
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
type UpOpts struct{}
|
|
||||||
|
|
||||||
// Up missing values in deployment configuration, and write them to state.
|
// Up missing values in deployment configuration, and write them to state.
|
||||||
func (d *Deployment) Up(ctx context.Context, s Solver, _ *UpOpts) error {
|
func (d *Deployment) Up(ctx context.Context, s Solver) error {
|
||||||
span, ctx := opentracing.StartSpanFromContext(ctx, "r.Compute")
|
span, ctx := opentracing.StartSpanFromContext(ctx, "deployment.Up")
|
||||||
defer span.Finish()
|
defer span.Finish()
|
||||||
|
|
||||||
lg := log.Ctx(ctx)
|
// Reset the computed values
|
||||||
|
d.computed = compiler.NewValue()
|
||||||
|
|
||||||
// Cueflow cue instance
|
// Cueflow cue instance
|
||||||
inst := d.state.CueInst()
|
src, err := compiler.InstanceMerge(d.plan, d.input)
|
||||||
|
|
||||||
// Reset the output
|
|
||||||
d.output = compiler.EmptyStruct()
|
|
||||||
|
|
||||||
// Cueflow config
|
|
||||||
flowCfg := &cueflow.Config{
|
|
||||||
UpdateFunc: func(c *cueflow.Controller, t *cueflow.Task) error {
|
|
||||||
if t == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
lg := lg.
|
|
||||||
With().
|
|
||||||
Str("component", t.Path().String()).
|
|
||||||
Str("state", t.State().String()).
|
|
||||||
Logger()
|
|
||||||
|
|
||||||
if t.State() != cueflow.Terminated {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
// Merge task value into output
|
|
||||||
err := d.output.FillPath(t.Path(), t.Value())
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
lg.
|
|
||||||
Error().
|
|
||||||
Err(err).
|
|
||||||
Msg("failed to fill task result")
|
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
return nil
|
|
||||||
},
|
|
||||||
}
|
|
||||||
// Orchestrate execution with cueflow
|
// Orchestrate execution with cueflow
|
||||||
flow := cueflow.New(flowCfg, inst, newTaskFunc(inst, newPipelineRunner(inst, s)))
|
flow := cueflow.New(
|
||||||
|
&cueflow.Config{},
|
||||||
|
src.CueInst(),
|
||||||
|
newTaskFunc(src.CueInst(), newPipelineRunner(src.CueInst(), d.computed, s)),
|
||||||
|
)
|
||||||
if err := flow.Run(ctx); err != nil {
|
if err := flow.Run(ctx); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
return nil
|
||||||
span, _ := opentracing.StartSpanFromContext(ctx, "merge state")
|
|
||||||
defer span.Finish()
|
|
||||||
|
|
||||||
return d.mergeState()
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type DownOpts struct{}
|
type DownOpts struct{}
|
||||||
@ -333,7 +214,7 @@ func noOpRunner(t *cueflow.Task) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func newPipelineRunner(inst *cue.Instance, s Solver) cueflow.RunnerFunc {
|
func newPipelineRunner(inst *cue.Instance, computed *compiler.Value, s Solver) cueflow.RunnerFunc {
|
||||||
return cueflow.RunnerFunc(func(t *cueflow.Task) error {
|
return cueflow.RunnerFunc(func(t *cueflow.Task) error {
|
||||||
ctx := t.Context()
|
ctx := t.Context()
|
||||||
lg := log.
|
lg := log.
|
||||||
@ -358,7 +239,7 @@ func newPipelineRunner(inst *cue.Instance, s Solver) cueflow.RunnerFunc {
|
|||||||
Msg("dependency detected")
|
Msg("dependency detected")
|
||||||
}
|
}
|
||||||
v := compiler.Wrap(t.Value(), inst)
|
v := compiler.Wrap(t.Value(), inst)
|
||||||
p := NewPipeline(t.Path().String(), s, NewFillable(t))
|
p := NewPipeline(t.Path().String(), s)
|
||||||
err := p.Do(ctx, v)
|
err := p.Do(ctx, v)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
span.LogFields(otlog.String("error", err.Error()))
|
span.LogFields(otlog.String("error", err.Error()))
|
||||||
@ -379,6 +260,29 @@ func newPipelineRunner(inst *cue.Instance, s Solver) cueflow.RunnerFunc {
|
|||||||
Msg("failed")
|
Msg("failed")
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Mirror the computed values in both `Task` and `Result`
|
||||||
|
if p.Computed().IsEmptyStruct() {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := t.Fill(p.Computed().Cue()); err != nil {
|
||||||
|
lg.
|
||||||
|
Error().
|
||||||
|
Err(err).
|
||||||
|
Msg("failed to fill task")
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Merge task value into output
|
||||||
|
if err := computed.FillPath(t.Path(), p.Computed()); err != nil {
|
||||||
|
lg.
|
||||||
|
Error().
|
||||||
|
Err(err).
|
||||||
|
Msg("failed to fill task result")
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
lg.
|
lg.
|
||||||
Info().
|
Info().
|
||||||
Dur("duration", time.Since(start)).
|
Dur("duration", time.Since(start)).
|
||||||
|
@ -10,6 +10,7 @@ import (
|
|||||||
"path"
|
"path"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
|
"cuelang.org/go/cue"
|
||||||
"github.com/docker/distribution/reference"
|
"github.com/docker/distribution/reference"
|
||||||
bk "github.com/moby/buildkit/client"
|
bk "github.com/moby/buildkit/client"
|
||||||
"github.com/moby/buildkit/client/llb"
|
"github.com/moby/buildkit/client/llb"
|
||||||
@ -33,15 +34,15 @@ type Pipeline struct {
|
|||||||
s Solver
|
s Solver
|
||||||
state llb.State
|
state llb.State
|
||||||
result bkgw.Reference
|
result bkgw.Reference
|
||||||
out *Fillable
|
computed *compiler.Value
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewPipeline(name string, s Solver, out *Fillable) *Pipeline {
|
func NewPipeline(name string, s Solver) *Pipeline {
|
||||||
return &Pipeline{
|
return &Pipeline{
|
||||||
name: name,
|
name: name,
|
||||||
s: s,
|
s: s,
|
||||||
state: llb.Scratch(),
|
state: llb.Scratch(),
|
||||||
out: out,
|
computed: compiler.NewValue(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -60,6 +61,10 @@ func (p *Pipeline) FS() fs.FS {
|
|||||||
return NewBuildkitFS(p.result)
|
return NewBuildkitFS(p.result)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (p *Pipeline) Computed() *compiler.Value {
|
||||||
|
return p.computed
|
||||||
|
}
|
||||||
|
|
||||||
func isComponent(v *compiler.Value) bool {
|
func isComponent(v *compiler.Value) bool {
|
||||||
return v.Lookup("#up").Exists()
|
return v.Lookup("#up").Exists()
|
||||||
}
|
}
|
||||||
@ -246,7 +251,7 @@ func (p *Pipeline) Copy(ctx context.Context, op *compiler.Value, st llb.State) (
|
|||||||
return st, err
|
return st, err
|
||||||
}
|
}
|
||||||
// Execute 'from' in a tmp pipeline, and use the resulting fs
|
// Execute 'from' in a tmp pipeline, and use the resulting fs
|
||||||
from := NewPipeline(op.Lookup("from").Path().String(), p.s, nil)
|
from := NewPipeline(op.Lookup("from").Path().String(), p.s)
|
||||||
if err := from.Do(ctx, op.Lookup("from")); err != nil {
|
if err := from.Do(ctx, op.Lookup("from")); err != nil {
|
||||||
return st, err
|
return st, err
|
||||||
}
|
}
|
||||||
@ -446,7 +451,7 @@ func (p *Pipeline) mount(ctx context.Context, dest string, mnt *compiler.Value)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
// eg. mount: "/foo": { from: www.source }
|
// eg. mount: "/foo": { from: www.source }
|
||||||
from := NewPipeline(mnt.Lookup("from").Path().String(), p.s, nil)
|
from := NewPipeline(mnt.Lookup("from").Path().String(), p.s)
|
||||||
if err := from.Do(ctx, mnt.Lookup("from")); err != nil {
|
if err := from.Do(ctx, mnt.Lookup("from")); err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
@ -488,7 +493,7 @@ func (p *Pipeline) Export(ctx context.Context, op *compiler.Value, st llb.State)
|
|||||||
Bytes("contents", contents).
|
Bytes("contents", contents).
|
||||||
Msg("exporting string")
|
Msg("exporting string")
|
||||||
|
|
||||||
if err := p.out.Fill(string(contents)); err != nil {
|
if err := p.computed.FillPath(cue.MakePath(), string(contents)); err != nil {
|
||||||
return st, err
|
return st, err
|
||||||
}
|
}
|
||||||
case "json":
|
case "json":
|
||||||
@ -504,7 +509,7 @@ func (p *Pipeline) Export(ctx context.Context, op *compiler.Value, st llb.State)
|
|||||||
Interface("contents", o).
|
Interface("contents", o).
|
||||||
Msg("exporting json")
|
Msg("exporting json")
|
||||||
|
|
||||||
if err := p.out.Fill(o); err != nil {
|
if err := p.computed.FillPath(cue.MakePath(), o); err != nil {
|
||||||
return st, err
|
return st, err
|
||||||
}
|
}
|
||||||
case "yaml":
|
case "yaml":
|
||||||
@ -520,7 +525,7 @@ func (p *Pipeline) Export(ctx context.Context, op *compiler.Value, st llb.State)
|
|||||||
Interface("contents", o).
|
Interface("contents", o).
|
||||||
Msg("exporting yaml")
|
Msg("exporting yaml")
|
||||||
|
|
||||||
if err := p.out.Fill(o); err != nil {
|
if err := p.computed.FillPath(cue.MakePath(), o); err != nil {
|
||||||
return st, err
|
return st, err
|
||||||
}
|
}
|
||||||
default:
|
default:
|
||||||
@ -550,7 +555,7 @@ func unmarshalAnything(data []byte, fn unmarshaller) (interface{}, error) {
|
|||||||
|
|
||||||
func (p *Pipeline) Load(ctx context.Context, op *compiler.Value, st llb.State) (llb.State, error) {
|
func (p *Pipeline) Load(ctx context.Context, op *compiler.Value, st llb.State) (llb.State, error) {
|
||||||
// Execute 'from' in a tmp pipeline, and use the resulting fs
|
// Execute 'from' in a tmp pipeline, and use the resulting fs
|
||||||
from := NewPipeline(op.Lookup("from").Path().String(), p.s, nil)
|
from := NewPipeline(op.Lookup("from").Path().String(), p.s)
|
||||||
if err := from.Do(ctx, op.Lookup("from")); err != nil {
|
if err := from.Do(ctx, op.Lookup("from")); err != nil {
|
||||||
return st, err
|
return st, err
|
||||||
}
|
}
|
||||||
@ -683,7 +688,7 @@ func (p *Pipeline) DockerBuild(ctx context.Context, op *compiler.Value, st llb.S
|
|||||||
// docker build context. This can come from another component, so we need to
|
// docker build context. This can come from another component, so we need to
|
||||||
// compute it first.
|
// compute it first.
|
||||||
if context.Exists() {
|
if context.Exists() {
|
||||||
from := NewPipeline(op.Lookup("context").Path().String(), p.s, nil)
|
from := NewPipeline(op.Lookup("context").Path().String(), p.s)
|
||||||
if err := from.Do(ctx, context); err != nil {
|
if err := from.Do(ctx, context); err != nil {
|
||||||
return st, err
|
return st, err
|
||||||
}
|
}
|
||||||
|
56
dagger/state.go
Normal file
56
dagger/state.go
Normal file
@ -0,0 +1,56 @@
|
|||||||
|
package dagger
|
||||||
|
|
||||||
|
// Contents of a deployment serialized to a file
|
||||||
|
type DeploymentState struct {
|
||||||
|
// Globally unique deployment ID
|
||||||
|
ID string `json:"id,omitempty"`
|
||||||
|
|
||||||
|
// Human-friendly deployment name.
|
||||||
|
// A deployment may have more than one name.
|
||||||
|
// FIXME: store multiple names?
|
||||||
|
Name string `json:"name,omitempty"`
|
||||||
|
|
||||||
|
// Cue module containing the deployment plan
|
||||||
|
// The input's top-level artifact is used as a module directory.
|
||||||
|
PlanSource Input `json:"plan,omitempty"`
|
||||||
|
|
||||||
|
// User Inputs
|
||||||
|
Inputs []inputKV `json:"inputs,omitempty"`
|
||||||
|
|
||||||
|
// Computed values
|
||||||
|
Computed string `json:"output,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type inputKV struct {
|
||||||
|
Key string `json:"key,omitempty"`
|
||||||
|
Value Input `json:"value,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *DeploymentState) SetInput(key string, value Input) error {
|
||||||
|
for i, inp := range s.Inputs {
|
||||||
|
if inp.Key != key {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
// Remove existing inputs with the same key
|
||||||
|
s.Inputs = append(s.Inputs[:i], s.Inputs[i+1:]...)
|
||||||
|
}
|
||||||
|
|
||||||
|
s.Inputs = append(s.Inputs, inputKV{Key: key, Value: value})
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Remove all inputs at the given key, including sub-keys.
|
||||||
|
// For example RemoveInputs("foo.bar") will remove all inputs
|
||||||
|
// at foo.bar, foo.bar.baz, etc.
|
||||||
|
func (s *DeploymentState) RemoveInputs(key string) error {
|
||||||
|
newInputs := make([]inputKV, 0, len(s.Inputs))
|
||||||
|
for _, i := range s.Inputs {
|
||||||
|
if i.Key == key {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
newInputs = append(newInputs, i)
|
||||||
|
}
|
||||||
|
s.Inputs = newInputs
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
@ -1,28 +0,0 @@
|
|||||||
package dagger
|
|
||||||
|
|
||||||
import (
|
|
||||||
"os"
|
|
||||||
|
|
||||||
cueflow "cuelang.org/go/tools/flow"
|
|
||||||
)
|
|
||||||
|
|
||||||
var ErrNotExist = os.ErrNotExist
|
|
||||||
|
|
||||||
// Something which can be filled in-place with a cue value
|
|
||||||
type Fillable struct {
|
|
||||||
t *cueflow.Task
|
|
||||||
}
|
|
||||||
|
|
||||||
func NewFillable(t *cueflow.Task) *Fillable {
|
|
||||||
return &Fillable{
|
|
||||||
t: t,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (f *Fillable) Fill(x interface{}) error {
|
|
||||||
// Use a nil pointer receiver to discard all values
|
|
||||||
if f == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
return f.t.Fill(x)
|
|
||||||
}
|
|
@ -15,7 +15,7 @@ repo: git.#Repository & {
|
|||||||
// Host the application with Netlify
|
// Host the application with Netlify
|
||||||
www: netlify.#Site & {
|
www: netlify.#Site & {
|
||||||
// Site name can be overridden
|
// Site name can be overridden
|
||||||
name: string | *"dagger-example-react"
|
name: string | *"dagger-examples-react"
|
||||||
|
|
||||||
// Deploy the output of yarn build
|
// Deploy the output of yarn build
|
||||||
// (Netlify build feature is not used, to avoid extra cost).
|
// (Netlify build feature is not used, to avoid extra cost).
|
||||||
|
@ -1,3 +0,0 @@
|
|||||||
package testing
|
|
||||||
|
|
||||||
nonConcrete: string
|
|
@ -4,8 +4,18 @@ import "dagger.io/dagger/op"
|
|||||||
|
|
||||||
foo: "value"
|
foo: "value"
|
||||||
bar: "another value"
|
bar: "another value"
|
||||||
|
computed: {
|
||||||
#up: [
|
string
|
||||||
|
#up: [
|
||||||
op.#FetchContainer & {ref: "busybox"},
|
op.#FetchContainer & {ref: "busybox"},
|
||||||
op.#Exec & {args: ["true"]},
|
op.#Exec & {
|
||||||
]
|
args: ["sh", "-c", """
|
||||||
|
printf test > /export
|
||||||
|
"""]
|
||||||
|
},
|
||||||
|
op.#Export & {
|
||||||
|
source: "/export"
|
||||||
|
format: "string"
|
||||||
|
},
|
||||||
|
]
|
||||||
|
}
|
||||||
|
@ -1,14 +1,14 @@
|
|||||||
todoApp:
|
www:
|
||||||
account:
|
account:
|
||||||
name: ENC[AES256_GCM,data:/T5E/8rU3YVxpA==,iv:dscf+kcVg0jYd5ICyJJL/f80BHyUDtfHWW9cRSP7aCU=,tag:xHZpdqNoYa58NbAX691vhg==,type:str]
|
name: ENC[AES256_GCM,data:EsPTWeiDCrVeUQ==,iv:9/tZQOrrjQejsK6NFcgQO9HaAnjIUv1Qc+S0slds+4o=,tag:n3NYautI94ilmEWG9UeFzA==,type:str]
|
||||||
token: ENC[AES256_GCM,data:6RBg3KafhrJYS8hbACxNRtnkpSNeablIiqHhUZTjz04X3cPyP4igPn7cnw==,iv:kj0vn83GaAtocUBdacO5D+6SJu6sVxfs58LYvUdpivA=,tag:CzZFkr9sXFHOOQihqYBdTA==,type:str]
|
token: ENC[AES256_GCM,data:Jx7oVJXcMX3hBmC6Kld7jxOOH/3CGSAzC7rRhHgs25iLFZG+F3iN5fYYVA==,iv:9SVNNv5CTM0AZns0x7x5bSI6jW93jSh8Xt21hXN1g28=,tag:QKBf7OsKs1TxvMA4gdA53Q==,type:str]
|
||||||
sops:
|
sops:
|
||||||
kms: []
|
kms: []
|
||||||
gcp_kms: []
|
gcp_kms: []
|
||||||
azure_kv: []
|
azure_kv: []
|
||||||
hc_vault: []
|
hc_vault: []
|
||||||
lastmodified: '2021-03-19T00:41:47Z'
|
lastmodified: '2021-04-09T01:47:48Z'
|
||||||
mac: ENC[AES256_GCM,data:eaDHLQHEH8wtCiDhcl9wWIC+WwitEBRr4K/p7chOKJ8sBc3wlH+dT3HVX5hTYX5rwU0ai2wgfcOg6mw4mDksIJvU4JpZV/AIo16ETjZDhRSc5hU1hUAiOQCYuJPqyGa12HzxzA81mcFm8kJETnoTYiTo7ApG9ShmOCp6XPezjPk=,iv:5Pd9eGTrVnr0ITNj/TrgJz9UE2SjVqqmOxxZOw0Spus=,tag:VGxPwp7hXHtz9Zvk+6WsqA==,type:str]
|
mac: ENC[AES256_GCM,data:UqOr8wGUwf6iVnurG/dvpiZVN0k9NrLTaB5CQVn+QTRQybgYuLZLTJuTFNlAqFGvNO07OoGkDx/Vmhw9F6nJS0qUcHC5iWg+Bxaa7anHwer7fkA/xTjKpqJnE6iveq+hzumDgeFbGL7+EJvbyxtJioF/LnWa5gnduBVacizjbKc=,iv:x4ulAfd2R0BdkGF4zkMn+wX+Y8wF3jEeFOu3+5t+wz4=,tag:e5uOip0iAt9xJa6RJGukCA==,type:str]
|
||||||
pgp:
|
pgp:
|
||||||
- created_at: '2021-03-18T22:02:51Z'
|
- created_at: '2021-03-18T22:02:51Z'
|
||||||
enc: |
|
enc: |
|
||||||
|
@ -51,10 +51,11 @@ test::cli::newdir() {
|
|||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "simple"
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "simple"
|
||||||
|
|
||||||
test::one "CLI: new: verify we have the right plan" --stdout='{
|
test::one "CLI: new: verify we have the right plan" --stdout='{
|
||||||
foo: "value"
|
"bar": "another value",
|
||||||
bar: "another value"
|
"computed": "test",
|
||||||
|
"foo": "value"
|
||||||
}' \
|
}' \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -f cue -d "simple" -c
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -f cue -d "simple" -c -f json
|
||||||
}
|
}
|
||||||
|
|
||||||
test::cli::newgit() {
|
test::cli::newgit() {
|
||||||
@ -65,13 +66,13 @@ test::cli::newgit() {
|
|||||||
DAGGER_STORE="$(mktemp -d -t dagger-store-XXXXXX)"
|
DAGGER_STORE="$(mktemp -d -t dagger-store-XXXXXX)"
|
||||||
export DAGGER_STORE
|
export DAGGER_STORE
|
||||||
|
|
||||||
disable test::one "CLI: new git: --plan-git" \
|
test::one "CLI: new git: --plan-git" \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-git https://github.com/samalba/dagger-test.git simple
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-git https://github.com/samalba/dagger-test.git simple
|
||||||
|
|
||||||
disable test::one "CLI: new git: verify plan can be upped" \
|
test::one "CLI: new git: verify plan can be upped" \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "simple"
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "simple"
|
||||||
|
|
||||||
disable test::one "CLI: new git: verify we have the right plan" --stdout='{
|
test::one "CLI: new git: verify we have the right plan" --stdout='{
|
||||||
foo: "value"
|
foo: "value"
|
||||||
bar: "another value"
|
bar: "another value"
|
||||||
}' \
|
}' \
|
||||||
@ -89,20 +90,45 @@ test::cli::query() {
|
|||||||
test::one "CLI: query: initialize simple" \
|
test::one "CLI: query: initialize simple" \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-dir "$d"/cli/simple simple
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-dir "$d"/cli/simple simple
|
||||||
|
|
||||||
test::one "CLI: query: concrete" --stdout='{
|
test::one "CLI: query: before up" --stdout='{
|
||||||
foo: "value"
|
"bar": "another value",
|
||||||
bar: "another value"
|
"foo": "value"
|
||||||
}' \
|
}' \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -f cue -d "simple" -c
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple"
|
||||||
|
|
||||||
|
test::one "CLI: query: concrete should fail" --exit=1 \
|
||||||
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple" -c
|
||||||
|
|
||||||
test::one "CLI: query: target" --stdout='"value"' \
|
test::one "CLI: query: target" --stdout='"value"' \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -f cue -d "simple" foo
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple" foo
|
||||||
|
|
||||||
test::one "CLI: query: initialize nonconcrete" \
|
test::one "CLI: query: compute missing values" \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-dir "$d"/cli/nonconcrete nonconcrete
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "simple"
|
||||||
|
|
||||||
test::one "CLI: query: non concrete" --exit=1 \
|
test::one "CLI: query: all values" --stdout='{
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -f cue -d "nonconcrete" -c
|
"bar": "another value",
|
||||||
|
"computed": "test",
|
||||||
|
"foo": "value"
|
||||||
|
}' \
|
||||||
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple"
|
||||||
|
|
||||||
|
test::one "CLI: query: concrete should work" --stdout='{
|
||||||
|
"bar": "another value",
|
||||||
|
"computed": "test",
|
||||||
|
"foo": "value"
|
||||||
|
}' \
|
||||||
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple" -c
|
||||||
|
|
||||||
|
test::one "CLI: query --no-computed" --stdout='{
|
||||||
|
"bar": "another value",
|
||||||
|
"foo": "value"
|
||||||
|
}' \
|
||||||
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple" --no-computed
|
||||||
|
|
||||||
|
test::one "CLI: query: --no-plan" --stdout='{
|
||||||
|
"computed": "test"
|
||||||
|
}' \
|
||||||
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple" -c --no-plan
|
||||||
}
|
}
|
||||||
|
|
||||||
test::cli::plan() {
|
test::cli::plan() {
|
||||||
@ -117,19 +143,22 @@ test::cli::plan() {
|
|||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-dir "$d"/cli/simple simple
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-dir "$d"/cli/simple simple
|
||||||
|
|
||||||
test::one "CLI: plan dir" \
|
test::one "CLI: plan dir" \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" -d "simple" plan dir "$d"/cli/nonconcrete
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" -d "simple" plan dir "$d"/cli/simple
|
||||||
|
|
||||||
test::one "CLI: plan dir: query non-concrete" --exit=1 \
|
test::one "CLI: plan dir: verify we have the right plan" --stdout='{
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" -d "simple" query -c
|
"bar": "another value",
|
||||||
|
"foo": "value"
|
||||||
|
}' \
|
||||||
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" -d "simple" query
|
||||||
|
|
||||||
disable test::one "CLI: plan git (Disabled: external dependency)" \
|
test::one "CLI: plan git" \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" -d "simple" plan git https://github.com/samalba/dagger-test.git
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" -d "simple" plan git https://github.com/samalba/dagger-test.git
|
||||||
|
|
||||||
disable test::one "CLI: plan git: verify we have the right plan" --stdout='{
|
test::one "CLI: plan git: verify we have the right plan" --stdout='{
|
||||||
foo: "value"
|
"bar": "another value",
|
||||||
bar: "another value"
|
"foo": "value"
|
||||||
}' \
|
}' \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -f cue -d "simple" -c
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "simple" -c
|
||||||
}
|
}
|
||||||
|
|
||||||
test::cli::input() {
|
test::cli::input() {
|
||||||
@ -140,21 +169,40 @@ test::cli::input() {
|
|||||||
DAGGER_STORE="$(mktemp -d -t dagger-store-XXXXXX)"
|
DAGGER_STORE="$(mktemp -d -t dagger-store-XXXXXX)"
|
||||||
export DAGGER_STORE
|
export DAGGER_STORE
|
||||||
|
|
||||||
test::one "CLI: new input" \
|
test::one "CLI: input: new" \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-dir "$d"/cli/input "input"
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" new --plan-dir "$d"/cli/input "input"
|
||||||
|
|
||||||
test::one "CLI: up: missing input" \
|
test::one "CLI: input: up missing input" \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "input" --stdout='{"foo":"bar"}'
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "input"
|
||||||
|
|
||||||
test::one "CLI: input dir" \
|
test::one "CLI: input: query missing input" \
|
||||||
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "input" --stdout='{
|
||||||
|
"foo": "bar"
|
||||||
|
}'
|
||||||
|
|
||||||
|
test::one "CLI: input: set dir" \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" input -d "input" dir "source" ./tests/cli/input/testdata
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" input -d "input" dir "source" ./tests/cli/input/testdata
|
||||||
|
|
||||||
test::one "CLI: up: input is set with input dir" \
|
test::one "CLI: input: up with input dir" \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "input" --stdout='{"bar":"thisisatest\n","foo":"bar","source":{}}'
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "input"
|
||||||
|
|
||||||
disable test::one "CLI: input git" \
|
test::one "CLI: input: query with input dir" \
|
||||||
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "input" --stdout='{
|
||||||
|
"bar": "thisisatest\n",
|
||||||
|
"foo": "bar",
|
||||||
|
"source": {}
|
||||||
|
}'
|
||||||
|
|
||||||
|
test::one "CLI: input: set git" \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" input -d "input" git "source" https://github.com/samalba/dagger-test-simple.git
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" input -d "input" git "source" https://github.com/samalba/dagger-test-simple.git
|
||||||
|
|
||||||
disable test::one "CLI: up: input is set with input git" \
|
test::one "CLI: input: up with input git" \
|
||||||
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "input" --stdout='{"bar":"testgit\n","foo":"bar","source":{}}'
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" up -d "input"
|
||||||
|
|
||||||
|
test::one "CLI: query with input git" \
|
||||||
|
"$dagger" "${DAGGER_BINARY_ARGS[@]}" query -d "input" --stdout='{
|
||||||
|
"bar": "testgit\n",
|
||||||
|
"foo": "bar",
|
||||||
|
"source": {}
|
||||||
|
}'
|
||||||
}
|
}
|
||||||
|
Reference in New Issue
Block a user