mirror of
https://github.com/neilotoole/sq.git
synced 2024-12-19 14:11:45 +03:00
221 lines
6.3 KiB
Go
221 lines
6.3 KiB
Go
package sqlz
|
|
|
|
import (
|
|
"database/sql"
|
|
"fmt"
|
|
"reflect"
|
|
"strconv"
|
|
"time"
|
|
|
|
"github.com/neilotoole/sq/libsq/errz"
|
|
)
|
|
|
|
// Record is a []interface{} row of field values returned from a query.
|
|
//
|
|
// In the codebase, we distinguish between a "Record" and
|
|
// a "ScanRow", although both are []interface{} and are closely related.
|
|
//
|
|
// An instance of ScanRow is passed to the sql rows.Scan method, and
|
|
// its elements may include implementations of the sql.Scanner interface
|
|
// such as sql.NullString, sql.NullInt64 or even driver-specific types.
|
|
//
|
|
// A Record is typically built from a ScanRow, unwrapping and
|
|
// munging elements such that the Record only contains standard types:
|
|
//
|
|
// nil, *int64, *float64, *bool, *string, *[]byte, *time.Time
|
|
//
|
|
// It is an error for a Record to contain elements of any other type.
|
|
type Record []interface{}
|
|
|
|
// ValidRecord checks that each element of the record vals is
|
|
// of an acceptable type. On the first unacceptable element,
|
|
// the index of that element and an error are returned. On
|
|
// success (-1, nil) is returned.
|
|
//
|
|
// These acceptable types, per the stdlib sql pkg, are:
|
|
// nil, *int64, *float64, *bool, *string, *[]byte, *time.Time
|
|
func ValidRecord(recMeta RecordMeta, rec Record) (i int, err error) {
|
|
// FIXME: ValidRecord should check the values of rec to see if they match recMeta's kinds
|
|
|
|
var val interface{}
|
|
for i, val = range rec {
|
|
switch val := val.(type) {
|
|
case nil, *int64, *float64, *bool, *string, *[]byte, *time.Time:
|
|
continue
|
|
default:
|
|
return i, errz.Errorf("field [%d] has unacceptable record value type %T", i, val)
|
|
}
|
|
}
|
|
|
|
return -1, nil
|
|
}
|
|
|
|
// FieldMeta is a bit of a strange entity, and in an ideal
|
|
// world, it wouldn't exist. It's here because:
|
|
//
|
|
// - The DB driver impls that sq utilizes (postgres, sqlite, etc)
|
|
// often have individual quirks (not reporting nullability etc)
|
|
// that necessitate modifying sql.ColumnType so that there's
|
|
// consistent behavior across the drivers.
|
|
//
|
|
// - We wanted to retain (and supplement) the method set of
|
|
// sql.ColumnType (basically, use the same "interface", even
|
|
// though sql.ColumnType is a struct, not interface) so that
|
|
// devs don't need to learn a whole new thing.
|
|
//
|
|
// - For that reason, stdlib sql.ColumnType needs to be
|
|
// supplemented with sqlz.Kind, and there needs to
|
|
// be a mechanism for modifying sql.ColumnType's fields.
|
|
//
|
|
// - But sql.ColumnType is sealed (its fields cannot be changed
|
|
// from outside its package).
|
|
//
|
|
// - Hence this construct where we have FieldMeta (which
|
|
// abstractly is an adapter around sql.ColumnType) and also
|
|
// a data holder struct (ColumnDataType), which permits the
|
|
// mutation of the column type fields.
|
|
//
|
|
// Likely there's a better design available than this one,
|
|
// but it suffices.
|
|
type FieldMeta struct {
|
|
data *ColumnTypeData
|
|
}
|
|
|
|
// NewFieldMeta returns a new instance backed by the data arg.
|
|
func NewFieldMeta(data *ColumnTypeData) *FieldMeta {
|
|
return &FieldMeta{data: data}
|
|
}
|
|
|
|
func (fm *FieldMeta) String() string {
|
|
nullMsg := "?"
|
|
if fm.data.HasNullable {
|
|
nullMsg = strconv.FormatBool(fm.data.Nullable)
|
|
}
|
|
|
|
return fmt.Sprintf("{name:%s, kind:%s, dbtype:%s, scan:%s, nullable:%s}",
|
|
fm.data.Name, fm.data.Kind.String(), fm.data.DatabaseTypeName, fm.ScanType().String(), nullMsg)
|
|
}
|
|
|
|
// Name is documented by sql.ColumnType.Name.
|
|
func (fm *FieldMeta) Name() string {
|
|
return fm.data.Name
|
|
}
|
|
|
|
// Length is documented by sql.ColumnType.Length.
|
|
func (fm *FieldMeta) Length() (length int64, ok bool) {
|
|
return fm.data.Length, fm.data.HasLength
|
|
}
|
|
|
|
// DecimalSize is documented by sql.ColumnType.DecimalSize.
|
|
func (fm *FieldMeta) DecimalSize() (precision, scale int64, ok bool) {
|
|
return fm.data.Precision, fm.data.Scale, fm.data.HasPrecisionScale
|
|
}
|
|
|
|
// ScanType is documented by sql.ColumnType.ScanType.
|
|
func (fm *FieldMeta) ScanType() reflect.Type {
|
|
return fm.data.ScanType
|
|
}
|
|
|
|
// Nullable is documented by sql.ColumnType.Nullable.
|
|
func (fm *FieldMeta) Nullable() (nullable, ok bool) {
|
|
return fm.data.Nullable, fm.data.HasNullable
|
|
}
|
|
|
|
// DatabaseTypeName is documented by sql.ColumnType.DatabaseTypeName.
|
|
func (fm *FieldMeta) DatabaseTypeName() string {
|
|
return fm.data.DatabaseTypeName
|
|
}
|
|
|
|
// Kind returns the data kind for the column.
|
|
func (fm *FieldMeta) Kind() Kind {
|
|
return fm.data.Kind
|
|
}
|
|
|
|
// RecordMeta is a slice of *FieldMeta, encapsulating the metadata
|
|
// for a record.
|
|
type RecordMeta []*FieldMeta
|
|
|
|
// Names returns the column names.
|
|
func (rm RecordMeta) Names() []string {
|
|
names := make([]string, len(rm))
|
|
for i, col := range rm {
|
|
names[i] = col.Name()
|
|
}
|
|
|
|
return names
|
|
}
|
|
|
|
// NewScanRow returns a new []interface{} that can be scanned
|
|
// into by sql.Rows.Scan.
|
|
func (rm RecordMeta) NewScanRow() []interface{} {
|
|
dests := make([]interface{}, len(rm))
|
|
|
|
for i, col := range rm {
|
|
dests[i] = reflect.New(col.data.ScanType).Interface()
|
|
}
|
|
|
|
return dests
|
|
}
|
|
|
|
// Kinds returns the data kinds for the record.
|
|
func (rm RecordMeta) Kinds() []Kind {
|
|
kinds := make([]Kind, len(rm))
|
|
for i, col := range rm {
|
|
kinds[i] = col.Kind()
|
|
}
|
|
|
|
return kinds
|
|
}
|
|
|
|
// ScanTypes returns the scan types for the record.
|
|
func (rm RecordMeta) ScanTypes() []reflect.Type {
|
|
scanTypes := make([]reflect.Type, len(rm))
|
|
for i, col := range rm {
|
|
scanTypes[i] = col.ScanType()
|
|
}
|
|
|
|
return scanTypes
|
|
}
|
|
|
|
// ColumnTypeData contains the same data as sql.ColumnType
|
|
// as well SQ's derived data kind. This type exists with
|
|
// exported fields instead of methods (as on sql.ColumnType)
|
|
// due to the need to work with the fields for testing, and
|
|
// also because for some drivers it's useful to twiddle with
|
|
// the scan type.
|
|
//
|
|
// This is all a bit ugly.
|
|
type ColumnTypeData struct {
|
|
Name string
|
|
|
|
HasNullable bool
|
|
HasLength bool
|
|
HasPrecisionScale bool
|
|
|
|
Nullable bool
|
|
Length int64
|
|
DatabaseTypeName string
|
|
Precision int64
|
|
Scale int64
|
|
ScanType reflect.Type
|
|
|
|
Kind Kind
|
|
}
|
|
|
|
// NewColumnTypeData returns a new instance with field values
|
|
// taken from col, supplemented with the kind param.
|
|
func NewColumnTypeData(col *sql.ColumnType, kind Kind) *ColumnTypeData {
|
|
ct := &ColumnTypeData{
|
|
Name: col.Name(),
|
|
DatabaseTypeName: col.DatabaseTypeName(),
|
|
ScanType: col.ScanType(),
|
|
Kind: kind,
|
|
}
|
|
|
|
ct.Nullable, ct.HasNullable = col.Nullable()
|
|
ct.Length, ct.HasLength = col.Length()
|
|
ct.Precision, ct.Scale, ct.HasPrecisionScale = col.DecimalSize()
|
|
|
|
return ct
|
|
}
|