2018-06-24 16:40:48 +03:00
|
|
|
package migrate
|
|
|
|
|
|
|
|
import (
|
|
|
|
"fmt"
|
|
|
|
nurl "net/url"
|
2020-04-07 12:23:20 +03:00
|
|
|
"runtime"
|
2018-06-24 16:40:48 +03:00
|
|
|
"strings"
|
2020-04-07 12:23:20 +03:00
|
|
|
|
2020-06-16 12:08:36 +03:00
|
|
|
crontriggers "github.com/hasura/graphql-engine/cli/metadata/cron_triggers"
|
|
|
|
|
2020-04-07 12:23:20 +03:00
|
|
|
"github.com/hasura/graphql-engine/cli/metadata"
|
|
|
|
"github.com/hasura/graphql-engine/cli/metadata/actions"
|
|
|
|
"github.com/hasura/graphql-engine/cli/metadata/allowlist"
|
|
|
|
"github.com/hasura/graphql-engine/cli/metadata/functions"
|
|
|
|
"github.com/hasura/graphql-engine/cli/metadata/querycollections"
|
|
|
|
"github.com/hasura/graphql-engine/cli/metadata/remoteschemas"
|
|
|
|
"github.com/hasura/graphql-engine/cli/metadata/tables"
|
|
|
|
"github.com/hasura/graphql-engine/cli/metadata/types"
|
|
|
|
"github.com/hasura/graphql-engine/cli/metadata/version"
|
|
|
|
|
|
|
|
"github.com/hasura/graphql-engine/cli"
|
|
|
|
"github.com/pkg/errors"
|
2018-06-24 16:40:48 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
// MultiError holds multiple errors.
|
|
|
|
type MultiError struct {
|
|
|
|
Errs []error
|
|
|
|
}
|
|
|
|
|
|
|
|
// NewMultiError returns an error type holding multiple errors.
|
|
|
|
func NewMultiError(errs ...error) MultiError {
|
|
|
|
compactErrs := make([]error, 0)
|
|
|
|
for _, e := range errs {
|
|
|
|
if e != nil {
|
|
|
|
compactErrs = append(compactErrs, e)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return MultiError{compactErrs}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Error implements error. Mulitple errors are concatenated with 'and's.
|
|
|
|
func (m MultiError) Error() string {
|
|
|
|
var strs = make([]string, 0)
|
|
|
|
for _, e := range m.Errs {
|
|
|
|
if len(e.Error()) > 0 {
|
|
|
|
strs = append(strs, e.Error())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return strings.Join(strs, " and ")
|
|
|
|
}
|
|
|
|
|
|
|
|
// suint64 safely converts int to uint64
|
|
|
|
// see https://goo.gl/wEcqof
|
|
|
|
// see https://goo.gl/pai7Dr
|
|
|
|
func suint64(n int64) uint64 {
|
|
|
|
if n < 0 {
|
|
|
|
panic(fmt.Sprintf("suint(%v) expects input >= 0", n))
|
|
|
|
}
|
|
|
|
return uint64(n)
|
|
|
|
}
|
|
|
|
|
2020-10-29 10:08:12 +03:00
|
|
|
/*
|
2018-06-24 16:40:48 +03:00
|
|
|
// newSlowReader turns an io.ReadCloser into a slow io.ReadCloser.
|
|
|
|
// Use this to simulate a slow internet connection.
|
|
|
|
func newSlowReader(r io.ReadCloser) io.ReadCloser {
|
|
|
|
return &slowReader{
|
|
|
|
rx: r,
|
|
|
|
reader: bufio.NewReader(r),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
type slowReader struct {
|
|
|
|
rx io.ReadCloser
|
|
|
|
reader *bufio.Reader
|
|
|
|
}
|
|
|
|
|
|
|
|
func (b *slowReader) Read(p []byte) (n int, err error) {
|
|
|
|
time.Sleep(10 * time.Millisecond)
|
|
|
|
c, err := b.reader.ReadByte()
|
|
|
|
if err != nil {
|
|
|
|
return 0, err
|
|
|
|
} else {
|
|
|
|
copy(p, []byte{c})
|
|
|
|
return 1, nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (b *slowReader) Close() error {
|
|
|
|
return b.rx.Close()
|
2020-10-29 10:08:12 +03:00
|
|
|
} */
|
2018-06-24 16:40:48 +03:00
|
|
|
|
|
|
|
var errNoScheme = fmt.Errorf("no scheme")
|
|
|
|
|
|
|
|
// schemeFromUrl returns the scheme from a URL string
|
|
|
|
func schemeFromUrl(url string) (string, error) {
|
|
|
|
u, err := nurl.Parse(url)
|
|
|
|
if err != nil {
|
|
|
|
return "", err
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(u.Scheme) == 0 {
|
|
|
|
return "", errNoScheme
|
|
|
|
}
|
|
|
|
|
|
|
|
return u.Scheme, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// FilterCustomQuery filters all query values starting with `x-`
|
|
|
|
func FilterCustomQuery(u *nurl.URL) *nurl.URL {
|
|
|
|
ux := *u
|
|
|
|
vx := make(nurl.Values)
|
|
|
|
for k, v := range ux.Query() {
|
|
|
|
if len(k) <= 1 || (len(k) > 1 && k[0:2] != "x-") {
|
|
|
|
vx[k] = v
|
|
|
|
}
|
|
|
|
}
|
|
|
|
ux.RawQuery = vx.Encode()
|
|
|
|
return &ux
|
|
|
|
}
|
2020-04-07 12:23:20 +03:00
|
|
|
|
|
|
|
func NewMigrate(ec *cli.ExecutionContext, isCmd bool) (*Migrate, error) {
|
2020-04-08 13:59:21 +03:00
|
|
|
dbURL := GetDataPath(ec)
|
2020-04-07 12:23:20 +03:00
|
|
|
fileURL := GetFilePath(ec.MigrationDir)
|
2020-04-28 14:59:57 +03:00
|
|
|
t, err := New(fileURL.String(), dbURL.String(), isCmd, int(ec.Config.Version), ec.Config.ServerConfig.TLSConfig, ec.Logger)
|
2020-04-07 12:23:20 +03:00
|
|
|
if err != nil {
|
|
|
|
return nil, errors.Wrap(err, "cannot create migrate instance")
|
|
|
|
}
|
|
|
|
// Set Plugins
|
|
|
|
SetMetadataPluginsWithDir(ec, t)
|
2020-04-22 13:22:02 +03:00
|
|
|
if ec.Config.Version == cli.V2 {
|
|
|
|
t.EnableCheckMetadataConsistency(true)
|
|
|
|
}
|
2020-04-07 12:23:20 +03:00
|
|
|
return t, nil
|
|
|
|
}
|
|
|
|
|
2020-04-08 13:59:21 +03:00
|
|
|
func GetDataPath(ec *cli.ExecutionContext) *nurl.URL {
|
|
|
|
url := ec.Config.ServerConfig.ParsedEndpoint
|
2020-04-07 12:23:20 +03:00
|
|
|
host := &nurl.URL{
|
2020-04-08 13:59:21 +03:00
|
|
|
Scheme: "hasuradb",
|
|
|
|
Host: url.Host,
|
|
|
|
Path: url.Path,
|
|
|
|
RawQuery: ec.Config.ServerConfig.APIPaths.GetQueryParams().Encode(),
|
2020-04-07 12:23:20 +03:00
|
|
|
}
|
2020-04-08 13:59:21 +03:00
|
|
|
q := host.Query()
|
2020-04-07 12:23:20 +03:00
|
|
|
// Set sslmode in query
|
|
|
|
switch scheme := url.Scheme; scheme {
|
|
|
|
case "https":
|
|
|
|
q.Set("sslmode", "enable")
|
|
|
|
default:
|
|
|
|
q.Set("sslmode", "disable")
|
|
|
|
}
|
2020-04-09 12:30:47 +03:00
|
|
|
for k, v := range ec.HGEHeaders {
|
|
|
|
q.Add("headers", fmt.Sprintf("%s:%s", k, v))
|
2020-04-07 12:23:20 +03:00
|
|
|
}
|
|
|
|
host.RawQuery = q.Encode()
|
|
|
|
return host
|
|
|
|
}
|
|
|
|
|
|
|
|
func SetMetadataPluginsWithDir(ec *cli.ExecutionContext, drv *Migrate, dir ...string) {
|
|
|
|
var metadataDir string
|
|
|
|
if len(dir) == 0 {
|
|
|
|
metadataDir = ec.MetadataDir
|
|
|
|
} else {
|
|
|
|
metadataDir = dir[0]
|
|
|
|
}
|
|
|
|
plugins := make(types.MetadataPlugins, 0)
|
|
|
|
if ec.Config.Version == cli.V2 && metadataDir != "" {
|
|
|
|
plugins = append(plugins, version.New(ec, metadataDir))
|
|
|
|
plugins = append(plugins, tables.New(ec, metadataDir))
|
|
|
|
plugins = append(plugins, functions.New(ec, metadataDir))
|
|
|
|
plugins = append(plugins, querycollections.New(ec, metadataDir))
|
|
|
|
plugins = append(plugins, allowlist.New(ec, metadataDir))
|
|
|
|
plugins = append(plugins, remoteschemas.New(ec, metadataDir))
|
|
|
|
plugins = append(plugins, actions.New(ec, metadataDir))
|
2020-06-16 12:08:36 +03:00
|
|
|
plugins = append(plugins, crontriggers.New(ec, metadataDir))
|
2020-04-07 12:23:20 +03:00
|
|
|
} else {
|
|
|
|
plugins = append(plugins, metadata.New(ec, ec.MigrationDir))
|
|
|
|
}
|
|
|
|
drv.SetMetadataPlugins(plugins)
|
|
|
|
}
|
|
|
|
|
|
|
|
func GetFilePath(dir string) *nurl.URL {
|
|
|
|
host := &nurl.URL{
|
|
|
|
Scheme: "file",
|
|
|
|
Path: dir,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Add Prefix / to path if runtime.GOOS equals to windows
|
|
|
|
if runtime.GOOS == "windows" && !strings.HasPrefix(host.Path, "/") {
|
|
|
|
host.Path = "/" + host.Path
|
|
|
|
}
|
|
|
|
return host
|
|
|
|
}
|