2022-02-08 12:24:34 +03:00
|
|
|
|
-- | Postgres DDL RunSQL
|
|
|
|
|
--
|
|
|
|
|
-- Escape hatch for running raw SQL against a postgres database.
|
|
|
|
|
--
|
|
|
|
|
-- 'runRunSQL' executes the provided raw SQL.
|
|
|
|
|
--
|
|
|
|
|
-- 'isSchemaCacheBuildRequiredRunSQL' checks for known schema-mutating keywords
|
|
|
|
|
-- in the raw SQL text.
|
|
|
|
|
--
|
|
|
|
|
-- See 'Hasura.Server.API.V2Query' and 'Hasura.Server.API.Query'.
|
2021-02-14 09:07:52 +03:00
|
|
|
|
module Hasura.Backends.Postgres.DDL.RunSQL
|
2021-09-24 01:56:37 +03:00
|
|
|
|
( runRunSQL,
|
|
|
|
|
RunSQL (..),
|
|
|
|
|
isSchemaCacheBuildRequiredRunSQL,
|
|
|
|
|
)
|
|
|
|
|
where
|
|
|
|
|
|
|
|
|
|
import Control.Monad.Trans.Control (MonadBaseControl)
|
|
|
|
|
import Data.Aeson
|
|
|
|
|
import Data.HashMap.Strict qualified as M
|
|
|
|
|
import Data.HashSet qualified as HS
|
|
|
|
|
import Data.Text.Extended
|
|
|
|
|
import Database.PG.Query qualified as Q
|
|
|
|
|
import Hasura.Backends.Postgres.DDL.EventTrigger
|
|
|
|
|
import Hasura.Backends.Postgres.DDL.Source
|
|
|
|
|
( ToMetadataFetchQuery,
|
|
|
|
|
fetchFunctionMetadata,
|
|
|
|
|
fetchTableMetadata,
|
|
|
|
|
)
|
|
|
|
|
import Hasura.Backends.Postgres.SQL.Types
|
|
|
|
|
import Hasura.Base.Error
|
|
|
|
|
import Hasura.EncJSON
|
|
|
|
|
import Hasura.Prelude
|
|
|
|
|
import Hasura.RQL.DDL.Schema
|
|
|
|
|
import Hasura.RQL.DDL.Schema.Diff
|
|
|
|
|
import Hasura.RQL.Types hiding
|
|
|
|
|
( ConstraintName,
|
|
|
|
|
fmFunction,
|
|
|
|
|
tmComputedFields,
|
|
|
|
|
tmTable,
|
|
|
|
|
)
|
|
|
|
|
import Hasura.SQL.AnyBackend qualified as AB
|
|
|
|
|
import Hasura.Server.Utils (quoteRegex)
|
|
|
|
|
import Hasura.Session
|
|
|
|
|
import Hasura.Tracing qualified as Tracing
|
|
|
|
|
import Text.Regex.TDFA qualified as TDFA
|
|
|
|
|
|
|
|
|
|
data RunSQL = RunSQL
|
|
|
|
|
{ rSql :: Text,
|
|
|
|
|
rSource :: !SourceName,
|
|
|
|
|
rCascade :: !Bool,
|
|
|
|
|
rCheckMetadataConsistency :: !(Maybe Bool),
|
|
|
|
|
rTxAccessMode :: !Q.TxAccess
|
|
|
|
|
}
|
|
|
|
|
deriving (Show, Eq)
|
2021-02-14 09:07:52 +03:00
|
|
|
|
|
2021-05-27 18:06:13 +03:00
|
|
|
|
instance FromJSON RunSQL where
|
|
|
|
|
parseJSON = withObject "RunSQL" $ \o -> do
|
|
|
|
|
rSql <- o .: "sql"
|
|
|
|
|
rSource <- o .:? "source" .!= defaultSource
|
|
|
|
|
rCascade <- o .:? "cascade" .!= False
|
|
|
|
|
rCheckMetadataConsistency <- o .:? "check_metadata_consistency"
|
|
|
|
|
isReadOnly <- o .:? "read_only" .!= False
|
|
|
|
|
let rTxAccessMode = if isReadOnly then Q.ReadOnly else Q.ReadWrite
|
2021-09-24 01:56:37 +03:00
|
|
|
|
pure RunSQL {..}
|
2021-05-27 18:06:13 +03:00
|
|
|
|
|
|
|
|
|
instance ToJSON RunSQL where
|
|
|
|
|
toJSON RunSQL {..} =
|
|
|
|
|
object
|
2021-09-24 01:56:37 +03:00
|
|
|
|
[ "sql" .= rSql,
|
|
|
|
|
"source" .= rSource,
|
|
|
|
|
"cascade" .= rCascade,
|
|
|
|
|
"check_metadata_consistency" .= rCheckMetadataConsistency,
|
|
|
|
|
"read_only"
|
|
|
|
|
.= case rTxAccessMode of
|
|
|
|
|
Q.ReadOnly -> True
|
|
|
|
|
Q.ReadWrite -> False
|
2021-05-27 18:06:13 +03:00
|
|
|
|
]
|
|
|
|
|
|
2022-02-08 12:24:34 +03:00
|
|
|
|
-- | Check for known schema-mutating keywords in the raw SQL text.
|
|
|
|
|
--
|
|
|
|
|
-- See Note [Checking metadata consistency in run_sql].
|
2021-05-27 18:06:13 +03:00
|
|
|
|
isSchemaCacheBuildRequiredRunSQL :: RunSQL -> Bool
|
|
|
|
|
isSchemaCacheBuildRequiredRunSQL RunSQL {..} =
|
|
|
|
|
case rTxAccessMode of
|
2021-09-24 01:56:37 +03:00
|
|
|
|
Q.ReadOnly -> False
|
2021-05-27 18:06:13 +03:00
|
|
|
|
Q.ReadWrite -> fromMaybe (containsDDLKeyword rSql) rCheckMetadataConsistency
|
2021-09-24 01:56:37 +03:00
|
|
|
|
where
|
|
|
|
|
containsDDLKeyword =
|
|
|
|
|
TDFA.match
|
|
|
|
|
$$( quoteRegex
|
|
|
|
|
TDFA.defaultCompOpt
|
|
|
|
|
{ TDFA.caseSensitive = False,
|
|
|
|
|
TDFA.multiline = True,
|
|
|
|
|
TDFA.lastStarGreedy = True
|
|
|
|
|
}
|
|
|
|
|
TDFA.defaultExecOpt
|
|
|
|
|
{ TDFA.captureGroups = False
|
|
|
|
|
}
|
|
|
|
|
"\\balter\\b|\\bdrop\\b|\\breplace\\b|\\bcreate function\\b|\\bcomment on\\b"
|
|
|
|
|
)
|
2021-05-27 18:06:13 +03:00
|
|
|
|
|
|
|
|
|
{- Note [Checking metadata consistency in run_sql]
|
|
|
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
|
SQL queries executed by run_sql may change the Postgres schema in arbitrary
|
|
|
|
|
ways. We attempt to automatically update the metadata to reflect those changes
|
|
|
|
|
as much as possible---for example, if a table is renamed, we want to update the
|
|
|
|
|
metadata to track the table under its new name instead of its old one. This
|
|
|
|
|
schema diffing (plus some integrity checking) is handled by withMetadataCheck.
|
|
|
|
|
|
|
|
|
|
But this process has overhead---it involves reloading the metadata, diffing it,
|
|
|
|
|
and rebuilding the schema cache---so we don’t want to do it if it isn’t
|
|
|
|
|
necessary. The user can explicitly disable the check via the
|
|
|
|
|
check_metadata_consistency option, and we also skip it if the current
|
|
|
|
|
transaction is in READ ONLY mode, since the schema can’t be modified in that
|
|
|
|
|
case, anyway.
|
|
|
|
|
|
|
|
|
|
However, even if neither read_only or check_metadata_consistency is passed, lots
|
|
|
|
|
of queries may not modify the schema at all. As a (fairly stupid) heuristic, we
|
|
|
|
|
check if the query contains any keywords for DDL operations, and if not, we skip
|
|
|
|
|
the metadata check as well. -}
|
2021-02-14 09:07:52 +03:00
|
|
|
|
|
2021-09-24 01:56:37 +03:00
|
|
|
|
fetchMeta ::
|
|
|
|
|
(ToMetadataFetchQuery pgKind, BackendMetadata ('Postgres pgKind), MonadTx m) =>
|
|
|
|
|
TableCache ('Postgres pgKind) ->
|
|
|
|
|
FunctionCache ('Postgres pgKind) ->
|
|
|
|
|
m ([TableMeta ('Postgres pgKind)], [FunctionMeta ('Postgres pgKind)])
|
2021-02-14 09:07:52 +03:00
|
|
|
|
fetchMeta tables functions = do
|
|
|
|
|
tableMetaInfos <- fetchTableMetadata
|
|
|
|
|
functionMetaInfos <- fetchFunctionMetadata
|
|
|
|
|
|
|
|
|
|
let getFunctionMetas function =
|
|
|
|
|
let mkFunctionMeta rawInfo =
|
|
|
|
|
FunctionMeta (rfiOid rawInfo) function (rfiFunctionType rawInfo)
|
2021-09-24 01:56:37 +03:00
|
|
|
|
in maybe [] (map mkFunctionMeta) $ M.lookup function functionMetaInfos
|
2021-02-14 09:07:52 +03:00
|
|
|
|
|
|
|
|
|
mkComputedFieldMeta computedField =
|
|
|
|
|
let function = _cffName $ _cfiFunction computedField
|
2021-09-24 01:56:37 +03:00
|
|
|
|
in map (ComputedFieldMeta (_cfiName computedField)) $ getFunctionMetas function
|
2021-02-14 09:07:52 +03:00
|
|
|
|
|
|
|
|
|
tableMetas = flip map (M.toList tableMetaInfos) $ \(table, tableMetaInfo) ->
|
2021-09-24 01:56:37 +03:00
|
|
|
|
TableMeta table tableMetaInfo $
|
|
|
|
|
fromMaybe [] $
|
|
|
|
|
M.lookup table tables <&> \tableInfo ->
|
|
|
|
|
let tableCoreInfo = _tiCoreInfo tableInfo
|
|
|
|
|
computedFields = getComputedFieldInfos $ _tciFieldInfoMap tableCoreInfo
|
|
|
|
|
in concatMap mkComputedFieldMeta computedFields
|
2021-02-14 09:07:52 +03:00
|
|
|
|
|
|
|
|
|
functionMetas = concatMap getFunctionMetas $ M.keys functions
|
|
|
|
|
|
|
|
|
|
pure (tableMetas, functionMetas)
|
|
|
|
|
|
2022-02-08 12:24:34 +03:00
|
|
|
|
-- | Used as an escape hatch to run raw SQL against a database.
|
2021-09-24 01:56:37 +03:00
|
|
|
|
runRunSQL ::
|
|
|
|
|
forall (pgKind :: PostgresKind) m.
|
|
|
|
|
( BackendMetadata ('Postgres pgKind),
|
|
|
|
|
ToMetadataFetchQuery pgKind,
|
|
|
|
|
CacheRWM m,
|
|
|
|
|
HasServerConfigCtx m,
|
|
|
|
|
MetadataM m,
|
|
|
|
|
MonadBaseControl IO m,
|
|
|
|
|
MonadError QErr m,
|
|
|
|
|
MonadIO m,
|
|
|
|
|
Tracing.MonadTrace m,
|
|
|
|
|
UserInfoM m
|
|
|
|
|
) =>
|
|
|
|
|
RunSQL ->
|
|
|
|
|
m EncJSON
|
|
|
|
|
runRunSQL q@RunSQL {..} = do
|
2021-09-01 20:56:46 +03:00
|
|
|
|
sourceConfig <- askSourceConfig @('Postgres pgKind) rSource
|
|
|
|
|
traceCtx <- Tracing.currentContext
|
|
|
|
|
userInfo <- askUserInfo
|
|
|
|
|
let pgExecCtx = _pscExecCtx sourceConfig
|
|
|
|
|
if (isSchemaCacheBuildRequiredRunSQL q)
|
|
|
|
|
then do
|
|
|
|
|
-- see Note [Checking metadata consistency in run_sql]
|
2021-09-24 01:56:37 +03:00
|
|
|
|
withMetadataCheck @pgKind rSource rCascade rTxAccessMode $
|
|
|
|
|
withTraceContext traceCtx $
|
|
|
|
|
withUserInfo userInfo $
|
|
|
|
|
execRawSQL rSql
|
2021-09-01 20:56:46 +03:00
|
|
|
|
else do
|
2021-09-15 23:45:49 +03:00
|
|
|
|
runTxWithCtx pgExecCtx rTxAccessMode $ execRawSQL rSql
|
2021-02-14 09:07:52 +03:00
|
|
|
|
where
|
2021-05-27 18:06:13 +03:00
|
|
|
|
execRawSQL :: (MonadTx n) => Text -> n EncJSON
|
|
|
|
|
execRawSQL =
|
|
|
|
|
fmap (encJFromJValue @RunSQLRes) . liftTx . Q.multiQE rawSqlErrHandler . Q.fromText
|
|
|
|
|
where
|
|
|
|
|
rawSqlErrHandler txe =
|
2021-09-24 01:56:37 +03:00
|
|
|
|
(err400 PostgresError "query execution failed") {qeInternal = Just $ ExtraInternal $ toJSON txe}
|
2021-02-14 09:07:52 +03:00
|
|
|
|
|
|
|
|
|
-- | @'withMetadataCheck' cascade action@ runs @action@ and checks if the schema changed as a
|
|
|
|
|
-- result. If it did, it checks to ensure the changes do not violate any integrity constraints, and
|
|
|
|
|
-- if not, incorporates them into the schema cache.
|
2021-07-23 02:06:10 +03:00
|
|
|
|
-- TODO(antoine): shouldn't this be generalized?
|
2021-09-24 01:56:37 +03:00
|
|
|
|
withMetadataCheck ::
|
|
|
|
|
forall (pgKind :: PostgresKind) a m.
|
|
|
|
|
( BackendMetadata ('Postgres pgKind),
|
|
|
|
|
ToMetadataFetchQuery pgKind,
|
|
|
|
|
CacheRWM m,
|
|
|
|
|
HasServerConfigCtx m,
|
|
|
|
|
MetadataM m,
|
|
|
|
|
MonadBaseControl IO m,
|
|
|
|
|
MonadError QErr m,
|
|
|
|
|
MonadIO m
|
|
|
|
|
) =>
|
|
|
|
|
SourceName ->
|
|
|
|
|
Bool ->
|
|
|
|
|
Q.TxAccess ->
|
|
|
|
|
Q.TxET QErr m a ->
|
|
|
|
|
m a
|
2021-02-14 09:07:52 +03:00
|
|
|
|
withMetadataCheck source cascade txAccess action = do
|
2021-10-29 17:42:07 +03:00
|
|
|
|
SourceInfo _ preActionTables preActionFunctions sourceConfig _ _ <- askSourceInfo @('Postgres pgKind) source
|
2021-02-14 09:07:52 +03:00
|
|
|
|
|
|
|
|
|
(actionResult, metadataUpdater) <-
|
2021-09-24 01:56:37 +03:00
|
|
|
|
liftEitherM $
|
|
|
|
|
runExceptT $
|
|
|
|
|
runTx (_pscExecCtx sourceConfig) txAccess $ do
|
|
|
|
|
-- Drop event triggers so no interference is caused to the sql query
|
|
|
|
|
forM_ (M.elems preActionTables) $ \tableInfo -> do
|
|
|
|
|
let eventTriggers = _tiEventTriggerInfoMap tableInfo
|
|
|
|
|
forM_ (M.keys eventTriggers) (liftTx . dropTriggerQ)
|
|
|
|
|
|
|
|
|
|
-- Get the metadata before the sql query, everything, need to filter this
|
|
|
|
|
(preActionTableMeta, preActionFunctionMeta) <- fetchMeta preActionTables preActionFunctions
|
|
|
|
|
|
|
|
|
|
-- Run the action
|
|
|
|
|
actionResult <- action
|
|
|
|
|
|
|
|
|
|
-- Get the metadata after the sql query
|
|
|
|
|
(postActionTableMeta, postActionFunctionMeta) <- fetchMeta preActionTables preActionFunctions
|
|
|
|
|
|
|
|
|
|
let preActionTableMeta' = filter (flip M.member preActionTables . tmTable) preActionTableMeta
|
2021-10-22 17:49:15 +03:00
|
|
|
|
tablesDiff = getTablesDiff preActionTableMeta' postActionTableMeta
|
|
|
|
|
FunctionsDiff droppedFuncs alteredFuncs = getFunctionsDiff preActionFunctionMeta postActionFunctionMeta
|
|
|
|
|
overloadedFuncs = getOverloadedFunctions (M.keys preActionFunctions) postActionFunctionMeta
|
2021-09-24 01:56:37 +03:00
|
|
|
|
|
|
|
|
|
-- Do not allow overloading functions
|
|
|
|
|
unless (null overloadedFuncs) $
|
2021-02-14 09:07:52 +03:00
|
|
|
|
throw400 NotSupported $
|
2021-09-24 01:56:37 +03:00
|
|
|
|
"the following tracked function(s) cannot be overloaded: "
|
|
|
|
|
<> commaSeparated overloadedFuncs
|
|
|
|
|
|
|
|
|
|
-- Report back with an error if cascade is not set
|
2021-10-22 17:49:15 +03:00
|
|
|
|
indirectDeps <- getIndirectDependencies source tablesDiff
|
|
|
|
|
when (indirectDeps /= [] && not cascade) $ reportDependentObjectsExist indirectDeps
|
2021-09-24 01:56:37 +03:00
|
|
|
|
|
|
|
|
|
metadataUpdater <- execWriterT $ do
|
|
|
|
|
-- Purge all the indirect dependents from state
|
|
|
|
|
for_ indirectDeps \case
|
|
|
|
|
SOSourceObj sourceName objectID -> do
|
|
|
|
|
AB.dispatchAnyBackend @BackendMetadata objectID $ purgeDependentObject sourceName >=> tell
|
|
|
|
|
_ ->
|
|
|
|
|
pure ()
|
|
|
|
|
|
|
|
|
|
-- Purge all dropped functions
|
|
|
|
|
let purgedFuncs = flip mapMaybe indirectDeps \case
|
|
|
|
|
SOSourceObj _ objectID
|
|
|
|
|
| Just (SOIFunction qf) <- AB.unpackAnyBackend @('Postgres pgKind) objectID ->
|
|
|
|
|
Just qf
|
|
|
|
|
_ -> Nothing
|
|
|
|
|
for_ (droppedFuncs \\ purgedFuncs) $
|
|
|
|
|
tell . dropFunctionInMetadata @('Postgres pgKind) source
|
|
|
|
|
|
|
|
|
|
-- Process altered functions
|
|
|
|
|
forM_ alteredFuncs $ \(qf, newTy) -> do
|
|
|
|
|
when (newTy == FTVOLATILE) $
|
|
|
|
|
throw400 NotSupported $
|
|
|
|
|
"type of function " <> qf <<> " is altered to \"VOLATILE\" which is not supported now"
|
|
|
|
|
|
|
|
|
|
-- update the metadata with the changes
|
2021-10-22 17:49:15 +03:00
|
|
|
|
processTablesDiff source preActionTables tablesDiff
|
2021-09-24 01:56:37 +03:00
|
|
|
|
|
|
|
|
|
pure (actionResult, metadataUpdater)
|
2021-02-14 09:07:52 +03:00
|
|
|
|
|
|
|
|
|
-- Build schema cache with updated metadata
|
|
|
|
|
withNewInconsistentObjsCheck $
|
2021-09-24 01:56:37 +03:00
|
|
|
|
buildSchemaCacheWithInvalidations mempty {ciSources = HS.singleton source} metadataUpdater
|
2021-02-14 09:07:52 +03:00
|
|
|
|
|
|
|
|
|
postActionSchemaCache <- askSchemaCache
|
|
|
|
|
|
|
|
|
|
-- Recreate event triggers in hdb_catalog
|
2021-05-21 05:46:58 +03:00
|
|
|
|
let postActionTables = fromMaybe mempty $ unsafeTableCache @('Postgres pgKind) source $ scSources postActionSchemaCache
|
2021-02-14 09:07:52 +03:00
|
|
|
|
serverConfigCtx <- askServerConfigCtx
|
2021-09-24 01:56:37 +03:00
|
|
|
|
liftEitherM $
|
|
|
|
|
runPgSourceWriteTx sourceConfig $
|
Role-invariant schema constructors
We build the GraphQL schema by combining building blocks such as `tableSelectionSet` and `columnParser`. These building blocks individually build `{InputFields,Field,}Parser` objects. Those object specify the valid GraphQL schema.
Since the GraphQL schema is role-dependent, at some point we need to know what fragment of the GraphQL schema a specific role is allowed to access, and this is stored in `{Sel,Upd,Ins,Del}PermInfo` objects.
We have passed around these permission objects as function arguments to the schema building blocks since we first started dealing with permissions during the PDV refactor - see hasura/graphql-engine@5168b99e463199b1934d8645bd6cd37eddb64ae1 in hasura/graphql-engine#4111. This means that, for instance, `tableSelectionSet` has as its type:
```haskell
tableSelectionSet ::
forall b r m n.
MonadBuildSchema b r m n =>
SourceName ->
TableInfo b ->
SelPermInfo b ->
m (Parser 'Output n (AnnotatedFields b))
```
There are three reasons to change this.
1. We often pass a `Maybe (xPermInfo b)` instead of a proper `xPermInfo b`, and it's not clear what the intended semantics of this is. Some potential improvements on the data types involved are discussed in issue hasura/graphql-engine-mono#3125.
2. In most cases we also already pass a `TableInfo b`, and together with the `MonadRole` that is usually also in scope, this means that we could look up the required permissions regardless: so passing the permissions explicitly undermines the "single source of truth" principle. Breaking this principle also makes the code more difficult to read.
3. We are working towards role-based parsers (see hasura/graphql-engine-mono#2711), where the `{InputFields,Field,}Parser` objects are constructed in a role-invariant way, so that we have a single object that can be used for all roles. In particular, this means that the schema building blocks _need_ to be constructed in a role-invariant way. While this PR doesn't accomplish that, it does reduce the amount of role-specific arguments being passed, thus fixing hasura/graphql-engine-mono#3068.
Concretely, this PR simply drops the `xPermInfo b` argument from almost all schema building blocks. Instead these objects are looked up from the `TableInfo b` as-needed. The resulting code is considerably simpler and shorter.
One way to interpret this change is as follows. Before this PR, we figured out permissions at the top-level in `Hasura.GraphQL.Schema`, passing down the obtained `xPermInfo` objects as required. After this PR, we have a bottom-up approach where the schema building blocks themselves decide whether they want to be included for a particular role.
So this moves some permission logic out of `Hasura.GraphQL.Schema`, which is very complex.
PR-URL: https://github.com/hasura/graphql-engine-mono/pull/3608
GitOrigin-RevId: 51a744f34ec7d57bc8077667ae7f9cb9c4f6c962
2022-02-17 11:16:20 +03:00
|
|
|
|
forM_ (M.elems postActionTables) $ \(TableInfo coreInfo _ eventTriggers _) -> do
|
2021-09-24 01:56:37 +03:00
|
|
|
|
let table = _tciName coreInfo
|
|
|
|
|
columns = getCols $ _tciFieldInfoMap coreInfo
|
|
|
|
|
forM_ (M.toList eventTriggers) $ \(triggerName, eti) -> do
|
|
|
|
|
let opsDefinition = etiOpsDef eti
|
|
|
|
|
flip runReaderT serverConfigCtx $ mkAllTriggersQ triggerName table columns opsDefinition
|
2021-02-14 09:07:52 +03:00
|
|
|
|
|
|
|
|
|
pure actionResult
|