graphql-engine/server/src-lib/Hasura/GraphQL/Parser/Monad.hs
Auke Booij caf9957aca Remove Unique from Definition
GraphQL types can refer to each other in a circular way. The PDV framework used to use values of type `Unique` to recognize two fragments of GraphQL schema as being the same instance. Internally, this is based on `Data.Unique` from the `base` package, which simply increases a counter on every creation of a `Unique` object.

**NB**: The `Unique` values are _not_ used for knot tying the schema combinators themselves (i.e. `Parser`s). The knot tying for `Parser`s is purely based on keys provided to `memoizeOn`. The `Unique` values are _only_ used to recognize two pieces of GraphQL _schema_ as being identical. Originally, the idea was that this would help us with a perfectly correct identification of GraphQL types. But this fully correct equality checking of GraphQL types was never implemented, and does not seem to be necessary to prevent bugs.

Specifically, these `Unique` values are stored as part of `data Definition a`, which specifies a part of our internal abstract syntax tree for the GraphQL types that we expose. The `Unique` values get initialized by the `SchemaT` effect.

In #2894 and #2895, we are experimenting with how (parts of) the GraphQL types can be hidden behind certain permission predicates. This would allow a single GraphQL schema in memory to serve all roles, implementing #2711. The permission predicates get evaluated at query parsing time when we know what role is doing a certain request, thus outputting the correct GraphQL types for that role.

If the approach of #2895 is followed, then the `Definition` objects, and thus the `Unique` values, would be hidden behind the permission predicates. Since the permission predicates are evaluated only after the schema is already supposed to be built, this means that the permission predicates would prevent us from initializing the `Unique` values, rendering them useless.

The simplest remedy to this is to remove our usage of `Unique` altogether from the GraphQL schema and schema combinators. It doesn't serve a functional purpose, doesn't prevent bugs, and requires extra bookkeeping.

PR-URL: https://github.com/hasura/graphql-engine-mono/pull/2980
GitOrigin-RevId: 50d3f9e0b9fbf578ac49c8fc773ba64a94b1f43d
2021-12-01 16:21:35 +00:00

210 lines
8.0 KiB
Haskell
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

-- | Monad transformers for GraphQL schema construction and query parsing.
module Hasura.GraphQL.Parser.Monad
( SchemaT,
runSchemaT,
ParseT,
runParseT,
ParseError (..),
reportParseErrors,
)
where
import Control.Monad.Validate
import Data.Dependent.Map (DMap)
import Data.Dependent.Map qualified as DM
import Data.GADT.Compare.Extended
import Data.IORef
import Data.Kind qualified as K
import Data.Parser.JSONPath
import Data.Proxy (Proxy (..))
import Data.Sequence.NonEmpty qualified as NE
import Hasura.Base.Error
import Hasura.GraphQL.Parser.Class
import Hasura.Prelude
import Language.Haskell.TH qualified as TH
import System.IO.Unsafe (unsafeInterleaveIO)
import Type.Reflection (Typeable, typeRep, (:~:) (..))
-- -------------------------------------------------------------------------------------------------
-- schema construction
newtype SchemaT n m a = SchemaT
{ unSchemaT :: StateT (DMap ParserId (ParserById n)) m a
}
deriving (Functor, Applicative, Monad, MonadError e)
runSchemaT :: forall m n a. Monad m => SchemaT n m a -> m a
runSchemaT = flip evalStateT mempty . unSchemaT
-- | see Note [SchemaT requires MonadIO]
instance
(MonadIO m, MonadParse n) =>
MonadSchema n (SchemaT n m)
where
memoizeOn name key buildParser = SchemaT do
let parserId = ParserId name key
parsersById <- get
case DM.lookup parserId parsersById of
Just (ParserById parser) -> pure parser
Nothing -> do
-- We manually do eager blackholing here using a MutVar rather than
-- relying on MonadFix and ordinary thunk blackholing. Why? A few
-- reasons:
--
-- 1. We have more control. We arent at the whims of whatever
-- MonadFix instance happens to get used.
--
-- 2. We can be more precise. GHCs lazy blackholing doesnt always
-- kick in when youd expect.
--
-- 3. We can provide more useful error reporting if things go wrong.
-- Most usefully, we can include a HasCallStack source location.
cell <- liftIO $ newIORef Nothing
-- We use unsafeInterleaveIO here, which sounds scary, but
-- unsafeInterleaveIO is actually far more safe than unsafePerformIO.
-- unsafeInterleaveIO just defers the execution of the action until its
-- result is needed, adding some laziness.
--
-- That laziness can be dangerous if the action has side-effects, since
-- the point at which the effect is performed can be unpredictable. But
-- this action just reads, never writes, so that isnt a concern.
parserById <-
liftIO $
unsafeInterleaveIO $
readIORef cell >>= \case
Just parser -> pure $ ParserById parser
Nothing ->
error $
unlines
[ "memoize: parser was forced before being fully constructed",
" parser constructor: " ++ TH.pprint name
]
put $! DM.insert parserId parserById parsersById
parser <- unSchemaT buildParser
liftIO $ writeIORef cell (Just parser)
pure parser
-- We can add a reader in two places. I'm not sure which one is the correct
-- one. But since we don't seem to change the values that are being read, I
-- don't think it matters.
deriving instance Monad m => MonadReader a (SchemaT n (ReaderT a m))
instance
(MonadIO m, MonadParse n) =>
MonadSchema n (ReaderT a (SchemaT n m))
where
memoizeOn name key = mapReaderT (memoizeOn name key)
{- Note [SchemaT requires MonadIO]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
The MonadSchema instance for SchemaT requires MonadIO, which is unsatisfying.
The only reason the constraint is needed is to implement knot-tying via IORefs
(see Note [Tying the knot] in Hasura.GraphQL.Parser.Class), which really only
requires the power of ST. Using ST would be much nicer, since we could discharge
the burden locally, but unfortunately we also want to use MonadUnique, which
is handled by IO in the end.
This means that we need IO at the base of our monad, so to use STRefs, wed need
a hypothetical STT transformer (i.e. a monad transformer version of ST). But
such a thing isnt safe in general, since reentrant monads like ListT or ContT
would incorrectly share state between the different threads of execution.
In theory, this can be resolved by using something like Vault (from the vault
package) to create “splittable” sets of variable references. That would allow
you to create a transformer with an STRef-like interface that works over any
arbitrary monad. However, while the interface would be safe, the implementation
of such an abstraction requires unsafe primitives, and to the best of my
knowledge no such transformer exists in any existing libraries.
So we decide it isnt worth the trouble and just use MonadIO. If `eff` ever pans
out, it should be able to support this more naturally, so we can fix it then. -}
-- | A key used to distinguish calls to 'memoize'd functions. The 'TH.Name'
-- distinguishes calls to completely different parsers, and the @a@ value
-- records the arguments.
data ParserId (t :: ((K.Type -> K.Type) -> K.Type -> K.Type, K.Type)) where
ParserId :: (Ord a, Typeable p, Typeable a, Typeable b) => TH.Name -> a -> ParserId '(p, b)
instance GEq ParserId where
geq
(ParserId name1 (arg1 :: a1) :: ParserId t1)
(ParserId name2 (arg2 :: a2) :: ParserId t2)
| _ :: Proxy '(p1, b1) <- Proxy @t1,
_ :: Proxy '(p2, b2) <- Proxy @t2,
name1 == name2,
Just Refl <- typeRep @a1 `geq` typeRep @a2,
arg1 == arg2,
Just Refl <- typeRep @p1 `geq` typeRep @p2,
Just Refl <- typeRep @b1 `geq` typeRep @b2 =
Just Refl
| otherwise = Nothing
instance GCompare ParserId where
gcompare
(ParserId name1 (arg1 :: a1) :: ParserId t1)
(ParserId name2 (arg2 :: a2) :: ParserId t2)
| _ :: Proxy '(p1, b1) <- Proxy @t1,
_ :: Proxy '(p2, b2) <- Proxy @t2 =
strengthenOrdering (compare name1 name2)
`extendGOrdering` gcompare (typeRep @a1) (typeRep @a2)
`extendGOrdering` strengthenOrdering (compare arg1 arg2)
`extendGOrdering` gcompare (typeRep @p1) (typeRep @p2)
`extendGOrdering` gcompare (typeRep @b1) (typeRep @b2)
`extendGOrdering` GEQ
-- | A newtype wrapper around a 'Parser' that rearranges the type parameters
-- so that it can be indexed by a 'ParserId' in a 'DMap'.
--
-- This is really just a single newtype, but its implemented as a data family
-- because GHC doesnt allow ordinary datatype declarations to pattern-match on
-- type parameters, and we want to match on the tuple.
data family ParserById (m :: K.Type -> K.Type) (a :: ((K.Type -> K.Type) -> K.Type -> K.Type, K.Type))
newtype instance ParserById m '(p, a) = ParserById (p m a)
-- -------------------------------------------------------------------------------------------------
-- query parsing
newtype ParseT m a = ParseT
{ unParseT :: ReaderT JSONPath (ValidateT (NESeq ParseError) m) a
}
deriving (Functor, Applicative, Monad)
runParseT ::
Functor m =>
ParseT m a ->
m (Either (NESeq ParseError) a)
runParseT =
unParseT
>>> flip runReaderT []
>>> runValidateT
instance MonadTrans ParseT where
lift = ParseT . lift . lift
instance Monad m => MonadParse (ParseT m) where
withPath f x = ParseT $ withReaderT f $ unParseT x
parseErrorWith code text = ParseT $ do
path <- ask
lift $ refute $ NE.singleton ParseError {peCode = code, pePath = path, peMessage = text}
data ParseError = ParseError
{ pePath :: JSONPath,
peMessage :: Text,
peCode :: Code
}
reportParseErrors ::
MonadError QErr m =>
NESeq ParseError ->
m a
reportParseErrors errs = case NE.head errs of
-- TODO: Our error reporting machinery doesnt currently support reporting
-- multiple errors at once, so were throwing away all but the first one
-- here. It would be nice to report all of them!
ParseError {pePath, peMessage, peCode} ->
throwError (err400 peCode peMessage) {qePath = pePath}