package driver import ( "context" "database/sql" "strings" "sync" "time" "github.com/neilotoole/sq/libsq/core/lg/lgm" "github.com/neilotoole/sq/libsq/core/options" "github.com/neilotoole/sq/libsq/driver/dialect" "github.com/neilotoole/sq/libsq/core/lg/lga" "github.com/neilotoole/sq/libsq/core/lg" "golang.org/x/exp/slog" "github.com/neilotoole/sq/libsq/core/cleanup" "github.com/neilotoole/sq/libsq/core/kind" "github.com/neilotoole/sq/libsq/core/errz" "github.com/neilotoole/sq/libsq/ast/render" "github.com/neilotoole/sq/libsq/core/sqlmodel" "github.com/neilotoole/sq/libsq/core/sqlz" "github.com/neilotoole/sq/libsq/source" ) // ConfigureDB configures DB using o. It is no-op if o is nil. func ConfigureDB(ctx context.Context, db *sql.DB, o options.Options) { o2 := options.Effective(o, OptConnMaxOpen, OptConnMaxIdle, OptConnMaxIdleTime, OptConnMaxLifetime) lg.FromContext(ctx).Debug("Setting config on DB conn", "config", o2) db.SetMaxOpenConns(OptConnMaxOpen.Get(o2)) db.SetMaxIdleConns(OptConnMaxIdle.Get(o2)) db.SetConnMaxIdleTime(OptConnMaxIdleTime.Get(o2)) db.SetConnMaxLifetime(OptConnMaxLifetime.Get(o2)) } var ( // OptConnMaxOpen controls sql.DB.SetMaxOpenConn. OptConnMaxOpen = options.NewInt( "conn.max-open", 0, `Maximum number of open connections to the database. A value of zero indicates no limit.`, "source", "sql", ) // OptConnMaxIdle controls sql.DB.SetMaxIdleConns. OptConnMaxIdle = options.NewInt( "conn.max-idle", 2, `Set the maximum number of connections in the idle connection pool. If conn.max-open is greater than 0 but less than the new conn.max-idle, then the new conn.max-idle will be reduced to match the conn.max-open limit. If n <= 0, no idle connections are retained.`, "source", ) // OptConnMaxIdleTime controls sql.DB.SetConnMaxIdleTime. OptConnMaxIdleTime = options.NewDuration( "conn.max-idle-time", time.Second*2, `Sets the maximum amount of time a connection may be idle. Expired connections may be closed lazily before reuse. If n <= 0, connections are not closed due to a connection's idle time.`, "source", ) // OptConnMaxLifetime controls sql.DB.SetConnMaxLifetime. OptConnMaxLifetime = options.NewDuration( "conn.max-lifetime", time.Minute*10, `Set the maximum amount of time a connection may be reused. Expired connections may be closed lazily before reuse. If n <= 0, connections are not closed due to a connection's age.`, "source", ) // OptMaxRetryInterval is the maximum interval to wait // between retries. OptMaxRetryInterval = options.NewDuration( "retry.max-interval", time.Second*3, `The maximum interval to wait between retries. If an operation is retryable (for example, if the DB has too many clients), repeated retry operations back off, typically using a Fibonacci backoff.`, "source", ) // OptTuningErrgroupLimit controls the maximum number of goroutines that can be spawned // by an errgroup. OptTuningErrgroupLimit = options.NewInt("tuning.errgroup-limit", 16, `Controls the maximum number of goroutines that can be spawned by an errgroup. Note that this is the limit for any one errgroup, but not a ceiling on the total number of goroutines spawned, as some errgroups may themselves start an errgroup. This knob is primarily for internal use. Ultimately it should go away in favor of dynamic errgroup limit setting based on availability of additional DB conns, etc.`, "tuning") // OptTuningRecChanSize is the size of the buffer chan for record // insertion/writing. OptTuningRecChanSize = options.NewInt("tuning.record-buffer", 1024, `Controls the size of the buffer channel for record insertion/writing.`, "tuning", ) ) // Provider is a factory that returns Driver instances. type Provider interface { // DriverFor returns a driver instance for the given type. DriverFor(typ source.DriverType) (Driver, error) } // DatabaseOpener opens a Database. type DatabaseOpener interface { // Open returns a Database instance for src. This operation can // take a long time if opening the DB requires an import of data. // For example, with file-based sources such as CSV, invoking Open // will ultimately read and import all CSV rows from the file. // Thus, set a timeout on ctx as appropriate for the source. Open(ctx context.Context, src *source.Source) (Database, error) } // JoinDatabaseOpener can open a join database. type JoinDatabaseOpener interface { // OpenJoin opens an appropriate Database for use as // a work DB for joining across sources. OpenJoin(ctx context.Context, src1, src2 *source.Source, srcN ...*source.Source) (Database, error) } // ScratchDatabaseOpener opens a scratch database. A scratch database is // typically a short-lived database used as a target for loading // non-SQL data (such as CSV). type ScratchDatabaseOpener interface { // OpenScratch returns a database for scratch use. OpenScratch(ctx context.Context, name string) (Database, error) } // Driver is the core interface that must be implemented for each type // of data source. type Driver interface { DatabaseOpener // DriverMetadata returns driver metadata. DriverMetadata() Metadata // ValidateSource verifies that the source is valid for this driver. It // may transform the source into a canonical form, which is returned in // the "src" return value (the original source is not changed). An error // is returned if the source is invalid. ValidateSource(source *source.Source) (src *source.Source, err error) // Ping verifies that the source is reachable, or returns an error if not. // The exact behavior of Ping() is driver-dependent. Ping(ctx context.Context, src *source.Source) error // Truncate truncates tbl in src. If arg reset is true, the // identity counter for tbl should be reset, if supported // by the driver. Some DB impls may reset the identity // counter regardless of the val of reset. Truncate(ctx context.Context, src *source.Source, tbl string, reset bool) (affected int64, err error) } // SQLDriver is implemented by Driver instances for SQL databases. type SQLDriver interface { Driver // Dialect returns the SQL dialect. Dialect() dialect.Dialect // Renderer returns the SQL renderer for this driver. Renderer() *render.Renderer // CurrentSchema returns the current schema name. CurrentSchema(ctx context.Context, db sqlz.DB) (string, error) // TableColumnTypes returns the column type info from // the SQL driver. If len(colNames) is 0, info is returned // for all columns in the table. TableColumnTypes(ctx context.Context, db sqlz.DB, tblName string, colNames []string) ([]*sql.ColumnType, error) // RecordMeta returns the result metadata (the metadata for // each col) from colTypes. RecordMeta is preferred over // sql.Rows.ColumnTypes because of the inconsistent behavior // of various SQL driver implementations wrt reporting // "nullable" information and other quirks. The returned // metadata may differ from the original metadata returned // by rows.ColumnTypes. // // The caller should typically should invoke rows.Next before // this method is invoked, as some implementations do not return // complete column type info until after the first call to rows.Next. // // RecordMeta also returns a NewRecordFunc which can be // applied to the scan row from sql.Rows. RecordMeta(colTypes []*sql.ColumnType) (sqlz.RecordMeta, NewRecordFunc, error) // PrepareInsertStmt prepares a statement for inserting // values to destColNames in destTbl. numRows specifies // how many rows of values are inserted by each execution of // the insert statement (1 row being the prototypical usage). // It is the caller's responsibility to close the execer. // // Note that db must guarantee a single connection: that is, db // must be a sql.Conn or sql.Tx. PrepareInsertStmt(ctx context.Context, db sqlz.DB, destTbl string, destColNames []string, numRows int) (*StmtExecer, error) // PrepareUpdateStmt prepares a statement for updating destColNames in // destTbl, using the supplied where clause (which may be empty). // The where arg should use question mark "?" as the placeholder: it will // be translated to the appropriate driver-specific placeholder. For example, // the where arg could be: // // "actor_id = ? AND first_name = ?". // // Use the returned StmtExecer per its documentation. It is the caller's // responsibility to close the execer. // // Note that db must guarantee a single connection: that is, db // must be a sql.Conn or sql.Tx. PrepareUpdateStmt(ctx context.Context, db sqlz.DB, destTbl string, destColNames []string, where string) (*StmtExecer, error) // CreateTable creates the table defined by tblDef. Some implementations // may not honor all of the fields of tblDef, e.g. an impl might not // build the foreign key constraints. At a minimum the implementation // must honor the table name and column names and kinds from tblDef. CreateTable(ctx context.Context, db sqlz.DB, tblDef *sqlmodel.TableDef) error // TableExists returns true if there's an existing table tbl in db. TableExists(ctx context.Context, db sqlz.DB, tbl string) (bool, error) // CopyTable copies fromTable into a new table toTable. // If copyData is true, fromTable's data is also copied. // Constraints (keys, defaults etc.) may not be copied. The // number of copied rows is returned in copied. CopyTable(ctx context.Context, db sqlz.DB, fromTable, toTable string, copyData bool) (copied int64, err error) // DropTable drops tbl from db. If ifExists is true, an "IF EXISTS" // or equivalent clause is added, if supported. DropTable(ctx context.Context, db sqlz.DB, tbl string, ifExists bool) error // AlterTableRename renames a table. AlterTableRename(ctx context.Context, db sqlz.DB, tbl, newName string) error // AlterTableAddColumn adds column col to tbl. The column is appended // to the list of columns (that is, the column position cannot be // specified). AlterTableAddColumn(ctx context.Context, db sqlz.DB, tbl, col string, knd kind.Kind) error // AlterTableRenameColumn renames a column. AlterTableRenameColumn(ctx context.Context, db sqlz.DB, tbl, col, newName string) error } // Database models a database handle. It is conceptually equivalent to // stdlib sql.DB, and in fact encapsulates a sql.DB instance. The // realized sql.DB instance can be accessed via the DB method. type Database interface { // DB returns the sql.DB object for this Database. DB() *sql.DB // SQLDriver returns the underlying database driver. This // may be different from the type reported by the // Database source. SQLDriver() SQLDriver // Source returns the data source for which this connection was opened. Source() *source.Source // SourceMetadata returns metadata about the data source. SourceMetadata(ctx context.Context) (*source.Metadata, error) // TableMetadata returns metadata for the specified table in the data source. TableMetadata(ctx context.Context, tblName string) (*source.TableMetadata, error) // Close is invoked to close and release any underlying resources. Close() error } // Metadata holds driver metadata. type Metadata struct { // Type is the driver type, e.g. "mysql" or "csv", etc. Type source.DriverType `json:"type"` // Description is typically the long name of the driver, e.g. // "MySQL" or "Microsoft Excel XLSX". Description string `json:"description"` // Doc is optional documentation, typically a URL. Doc string `json:"doc,omitempty"` // UserDefined is true if this driver is the product of a // user driver definition, and false if built-in. UserDefined bool `json:"user_defined"` // IsSQL is true if this driver is a SQL driver. IsSQL bool `json:"is_sql"` // Monotable is true if this is a non-SQL document type that // effectively has a single table, such as CSV. Monotable bool `json:"monotable"` } // Databases provides a mechanism for getting Database instances. // Note that at this time instances returned by Open are cached // and then closed by Close. This may be a bad approach. type Databases struct { log *slog.Logger drvrs Provider mu sync.Mutex scratchSrcFn ScratchSrcFunc dbases map[string]Database clnup *cleanup.Cleanup } // NewDatabases returns a Databases instances. func NewDatabases(log *slog.Logger, drvrs Provider, scratchSrcFn ScratchSrcFunc) *Databases { return &Databases{ log: log, drvrs: drvrs, mu: sync.Mutex{}, scratchSrcFn: scratchSrcFn, dbases: map[string]Database{}, clnup: cleanup.New(), } } // Open returns an opened Database for src. The returned Database // may be cached and returned on future invocations for the // same handle. Thus, the caller should typically not close // the Database: it will be closed via d.Close. // // NOTE: This entire logic re caching/not-closing is a bit sketchy, // and needs to be revisited. // // Open implements DatabaseOpener. func (d *Databases) Open(ctx context.Context, src *source.Source) (Database, error) { lg.FromContext(ctx).Debug(lgm.OpenSrc, lga.Src, src) d.mu.Lock() defer d.mu.Unlock() dbase, ok := d.dbases[src.Handle] if ok { return dbase, nil } drvr, err := d.drvrs.DriverFor(src.Type) if err != nil { return nil, err } dbase, err = drvr.Open(ctx, src) if err != nil { return nil, err } d.clnup.AddC(dbase) d.dbases[src.Handle] = dbase return dbase, nil } // OpenScratch returns a scratch database instance. It is not // necessary for the caller to close the returned Database as // its Close method will be invoked by d.Close. // // OpenScratch implements ScratchDatabaseOpener. func (d *Databases) OpenScratch(ctx context.Context, name string) (Database, error) { const msgCloseScratch = "close scratch db" scratchSrc, cleanFn, err := d.scratchSrcFn(ctx, name) if err != nil { // if err is non-nil, cleanup is guaranteed to be nil return nil, err } d.log.Debug("Opening scratch src", lga.Src, scratchSrc) drvr, err := d.drvrs.DriverFor(scratchSrc.Type) if err != nil { lg.WarnIfFuncError(d.log, msgCloseScratch, cleanFn) return nil, err } sqlDrvr, ok := drvr.(SQLDriver) if !ok { lg.WarnIfFuncError(d.log, msgCloseScratch, cleanFn) return nil, errz.Errorf("driver for scratch source %s is not a SQLDriver but is %T", scratchSrc.Handle, drvr) } var backingDB Database backingDB, err = sqlDrvr.Open(ctx, scratchSrc) if err != nil { lg.WarnIfFuncError(d.log, msgCloseScratch, cleanFn) return nil, err } d.clnup.AddE(cleanFn) return backingDB, nil } // OpenJoin opens an appropriate database for use as // a work DB for joining across sources. // // Note: There is much work to be done on this method. At this time, only // two sources are supported. Ultimately OpenJoin should be able to // inspect the join srcs and use heuristics to determine the best // location for the join to occur (to minimize copying of data for // the join etc.). Currently the implementation simply delegates // to OpenScratch. // // OpenJoin implements JoinDatabaseOpener. func (d *Databases) OpenJoin(ctx context.Context, src1, src2 *source.Source, srcN ...*source.Source) (Database, error) { if len(srcN) > 0 { return nil, errz.Errorf("Currently only two-source join is supported") } names := []string{src1.Handle, src2.Handle} for _, src := range srcN { names = append(names, src.Handle) } d.log.Debug("OpenJoin: [%s]", strings.Join(names, ",")) return d.OpenScratch(ctx, "joindb__"+strings.Join(names, "_")) } // Close closes d, invoking Close on any instances opened via d.Open. func (d *Databases) Close() error { d.log.Debug("Closing databases(s)...", lga.Count, d.clnup.Len()) return d.clnup.Run() } // requireSingleConn returns nil if db is a type that guarantees a // single database connection. That is, requireSingleConn returns an // error if db does not have type *sql.Conn or *sql.Tx. func requireSingleConn(db sqlz.DB) error { switch db.(type) { case *sql.Conn, *sql.Tx: default: return errz.Errorf("db must be guaranteed single-connection (sql.Conn or sql.Tx) but was %T", db) } return nil }