A while back I released the unification-fd library, which gives a generic implementation of first-order unification of non-cyclic terms. I've given a few talks on how the library is implemented and what optimizations it performs, but that's not the topic for today. Today, I'm going to talk about how to use it.
Unification is a widely useful operation and, consequently, comes in many different flavors. The version currently supported by the library is the sort used by logic programming languages like Prolog, Curry, Dyna, and MiniKanren; which is the same sort that's used for unification-based type inference algorithms like Hindley–Damas–Milner. Of these two examples, the logic programming example is the simpler one to discuss— at least for folks who've used a language like Prolog before. So let's start from there.
Caveat Emptor: This post is something of a stream of consciousness. I've gotten a few requests for tutorials on how to use the library, but the requests weren't terribly specific about what problems people've had or what's been difficult to figure out. So I'm shooting in the dark as far as what folks need and how much background they have. I'm going to assume you're familiar with Prolog and the basics of what unification is and does.
Preemptive apology: I started writing this post months and months (and months) ago, but unintentionally dropped it after running into a certain issue and then getting distracted and moving onto other things. Actually, this happened at least twice. I'm terribly sorry about that. So, apologies for not tackling the disjunction issue in this post. I'll come back to it later, but figured this post really needs to get out the door already.
A term, in Prolog, is just a fancy name for a value of some algebraic data type. In most variants of Prolog there's no explicit definition of the ADT, no restriction on what the constructors are, and no type checking to ensure that subterms have a particular shape. That is, Prolog is what's called a single-sorted logic; in other words, Prolog is an untyped/unityped language. With unification-fd we can implement multi-sorted (aka typed) logics, but for this tutorial we're going to stick with Prolog's single-sorted approach.
Control.Unification we'll see a handful of types and type classes, followed by a bunch of operators. The
UTerm data type captures the recursive structure of logic terms. (
UTerm is the free monad, if you're familiar with that terminology.) That is, given some functor
t which describes the constructors of our logic terms, and some type
v which describes our logic variables, the type
UTerm t v is the type of logic terms: trees with multiple layers of
t structure and leaves of type
v. For our single-sorted logic, here's an implementation of
data T a = T String [a]
String gives the name of the term constructor, and the list gives the ordered sequence of subterms. Thus, the Prolog term
foo(bar,baz(X)) would be implemented as
UTerm$T "foo" [UTerm$T "bar" , UTerm$T "baz" [UVar x]]. If we're going to be building these terms directly, then we probably want to define some smart constructors to reduce the syntactic noise:
foo x y = UTerm$T "foo" [x,y] bar = UTerm$T "bar"  baz x = UTerm$T "baz" [x]
Now, we can implement the Prolog term as
foo bar (baz x). If you prefer a more Prolog-like syntax, you can use uncurried definitions for smart constructors that take more than one argument.
In order to use our
T data type with the rest of the API, we'll need to give a
Unifiable instance for it. Before we do that we'll have to give
Traversable instances. These are straightforward and can be automatically derived with the appropriate language pragmas.
Unifiable class gives one step of the unification process. Just as we only need to specify one level of the ADT (i.e.,
T) and then we can use the library's
UTerm to generate the recursive ADT, so too we only need to specify one level of the unification (i.e.,
zipMatch) and then we can use the library's operators to perform the recursive unification, subsumption, etc.
zipMatch function takes two arguments of type
t a. The abstract
t will be our concrete
T type. The abstract
a is polymorphic, which ensures that we can't mess around with more than one level of the term at once. If we abandon that guarantee, then you can think of it as if
UTerm T v. Thus,
t a means
T (UTerm T v); and
T (UTerm T v) is essentially the type
UTerm T v with the added guarantee that the values aren't in fact variables. Thus, the arguments to
zipMatch are non-variable terms.
zipMatch method has the rather complicated return type:
Maybe (t (Either a (a,a))). Let's unpack this a bit by thinking about how unification works. When we try to unify two terms, first we look at their head constructors. If the constructors are different, then the terms aren't unifiable, so we return
Nothing to indicate that unification has failed. Otherwise, the constructors match, so we have to recursively unify their subterms. Since the
T structures of the two terms match, we can return
Just t0 where
t0 has the same
T structure as both input terms. Where we still have to recursively unify subterms, we fill
Right(l,r) values where
l is a subterm of the left argument to
r is the corresponding subterm of the right argument. Thus,
zipMatch is a generalized zipping function for combining the shared structure and pairing up substructures. And now, the implementation:
instance Unifiable T where zipMatch (T m ls) (T n rs) | m /= n = Nothing | otherwise = T n <$> pairWith (\l r -> Right(l,r)) ls rs
list-extras:Data.List.Extras.Pair. is a version of
zip which returns
Nothing if the lists have different lengths. So, if the names
n match, and if the two arguments have the same number of subterms, then we pair those subterms off in order; otherwise, either the names or the lengths don't match, so we return
T example, we don't need to worry about the
Left option. The reason it's there is to support feature structures and other sparse representations of terms. That is, consider the following type:
newtype FS k a = FS (Map k a)
Using this type, our logic terms are sets of key–subterm pairs. When unifying maps like these, what do we do if one argument has a binding for a particular key but the other argument does not? In the
T example we assumed that subterms which couldn't be paired together (because the lists were different lengths) meant the unification must fail. But for
FS it makes more sense to assume that terms which can't be paired up automatically succeed! That is, we'd like to assume that all the keys which are not explicitly present in the
Map k a are implicitly present and each one is bound to a unique logic variable. Since the unique logic variables are implicit, there's no need to actually keep track of them, we'll just implicitly unify them with the subterm that can't be paired off.
This may make more sense if you see the
instance (Ord k) => Unifiable (FS k) where zipMatch (FS ls) (FS rs) = Just . FS $ unionWith (\(Left l) (Left r) -> Right(l,r)) (fmap Left ls) (fmap Left rs)
We start off by mapping
Left over both the
ls and the
rs. We then call
unionWith to pair things up. For any given key, if both
rs specify a subterm, then these subterms will be paired up as
Right(l,r). If we have extra subterms from either
rs, however, then we keep them around as
Left l or
Left r. Thus, the
Unifiable instance for
FS performs a union of the
FS structure, whereas the instance for
T performs an intersection of
Left option can be used in any situation where you can immediately resolve the unification of subterms, whereas the
Right option says you still have work to do.1
The library ships with two implementations of logic variables. The
IntVar implementation uses
Int as the names of variables, and uses an
IntMap to keep track of the environment. The
STVar implementation uses
STRefs, so we can use actual mutation for binding logic variables, rather than keeping an explicit environment around. Of course, mutation complicates things, so the two implementations have different pros and cons.
Performing unification has the side effect of binding logic variables to terms. Thus, we'll want to use a monad in order to keep track of these effects. The
BindingMonad type class provides the definition of what we need from our ambient monad. In particular, we need to be able to generate fresh logic variables, to bind logic variables, and to lookup what our logic variables are bound to. The library provides the necessary instances for both
You can, of course, provide your own implementations of
BindingMonad. However, doing so is beyond the scope of the current tutorial. For simplicity, we'll use the
IntVar implementation below.
When embedding Prolog programs into Haskell, the main operators we want to consider are those in the section titled "Operations on two terms". These are structural equality (i.e., equality modulo substitution), structural equivalence (i.e., structural equality modulo alpha-variance), unification, and subsumption.
Consider the following Horn clause in Prolog:
example1(X,Y,Z) :- X = Y, Y = Z.
To implement this in Haskell we want a function which takes in three arguments, unifies the first two, and then unifies the second two. Thus,2
example1 x y z = do x =:= y y =:= z
To run this program we'd use one of the functions
execIntBindingT, depending on whether we care about the binding state, the resulting logic term, or both. Of course, since the unifications may fail, we also need to run the underlying error monad, using something like
runErrorT3,4. And since these are both monad transformers, we'll need to use
runIdentity or the like in order to run the base monad. Thus, the functions to execute the entire monad stack will look like:
-- Aliases to simplify our type signatures. N.B., the -- signatures are not actually required to get things -- to typecheck. type PrologTerm = UTerm T IntVar type PrologFailure = UnificationFailure T IntVar type PrologBindingState = IntBindingState T -- N.B., the @FallibleBindingMonad@ isn't yet a monad -- for Prolog because it doesn't support backtracking. type FallibleBindingMonad = ErrorT PrologFailure (IntBindingT T Identity) -- N.B., this definition runs into certain issues. type PrologMonad = ErrorT PrologFailure (IntBindingT T Logic) runFBM :: FallibleBindingMonad a -> (Either PrologFailure a, PrologBindingState) runFBM = runIdentity . runIntBindingT . runErrorT
Here are some more examples:
-- A helper function to reduce boilerplate. First we get -- a free variable, then we embed it into @PrologTerm@, -- and then we embed it into some error monad (for -- capturing unification failures). getFreeVar = lift (UVar <$> freeVar) -- example2(X,Z) :- X = Y, Y = Z. example2 x z = do y <- getFreeVar x =:= y y =:= z -- example3(X,Z) :- example1(X,Y,Z). example3 x z = do y <- getFreeVar example1 x y z -- example4(X) :- X = bar; X = backtrack. example4 x = (x =:= bar) <|> (x =:= atom "backtrack")
The complete code for this post can be found here online, or at
./test/tutorial/tutorial1.hs in the Darcs repo. Notably, there are some complications about the semantics of
example4; it doesn't mean what you think it should mean. We'll tackle that problem and fix it later on in the tutorial series (in part 4 or thereabouts).
Term Factoring and Clause Resolution Automata (CRAs)
Note that for the above examples, the Haskell functions only execute the right-hand side of the Horn clause. In Prolog itself, there's also a process of searching through all the Horn clauses in a program and deciding which one to execute next. A naive way to implement that search process would be to have a list of all the Horn clauses and walk through it, trying to unify the goal with the left-hand side of each clause and executing the right-hand side if it matches. A more efficient way would be to compile all the right-hand sides into a single automaton, allowing us to match the goal against all the right-hand sides at once. (The idea here is similar to compiling a bunch of strings together into a trie or regex.)
Constructing optimal CRAs is NP-complete in general, though it's feasible if we have an arbitrary ordering of clauses (e.g., Prolog's top–down order for trying each clause). The unification-fd library does not implement any support for CRAs at present, though it's something I'd like to add in the future. For more information on this topic, see Dawson et al. (1995) Optimizing Clause Resolution: Beyond Unification Factoring and Dawson et al. (1996) Principles and Practice of Unification Factoring.
In addition to unification itself, it's often helpful to have various other operators on hand.
One such operator is the subsumption operator. Whereas unification looks for a most-general substitution which when applied to both arguments yields terms which are structurally equal (i.e.,
l =:= r computes the most general
s such that
s l === s r), subsumption applies the substitution to only one side. That is,
r just in case
r is a substitution instance of
l (i.e., there exists a substitution
s such that
s l === r). The symbolic name
(<:=) comes from the fact that when
r we also say that
l is less defined5 than
r. Subsumption shows up in cases where we have to hold
r fixed for some reason, such as when implementing polymorphism or subtyping.
Other operators work on just one term, such as determining the free variables of a term, explicitly applying the ambient substitution to obtain a pure term, or cloning a term to make a copy where all the free variables have been renamed to fresh variables. These sorts of operators aren't used very often in logic programming itself, but are crucial for implementing logic programming languages.
Hopefully that gives a quick idea of how the library's API is set up. Next time I'll walk through an implementation of Hindley–Damas–Milner type inference, and then higher-ranked polymorphism à la Peyton Jones et al. (2011) Practical type inference for arbitrary-rank types. After that, I'll discuss the complications about backtracking choice I noticed when writing this post, and walk through how to fix them. If there's still interest after that, I can get into some of the guts of the library's implementation— like ranked path compression, maximal structure sharing, and so on.
If you have any particular questions you'd like me to address, drop me a line.
 Older versions of the library used the type
zipMatch :: forall a b. t a -> t b -> Maybe (t (a,b)) in order to ensure that we did in fact properly pair up subterms from the two arguments. Unfortunately I had to relax that guarantee in order to add support for feature structures. ↩
 N.B., a more efficient implementation is:
example1' x y z = do y' <- x =:= y y' =:= z
The unification operator returns a new term which guarantees maximal structure sharing with both of its arguments. The implementation of unification makes use of observable structure sharing, so by capturing
y' and using it in lieu of
y, the subsequent unifications can avoid redundant work. ↩
ErrorT transformer was deprecated by transformers-0.4.1.0, though it still works for this tutorial. Unfortunately, the preferred
ExceptT does not work since
UnificationFailure doesn't have a
Monoid instance as of unification-fd-0.9.0. The definition of
UnificationFailure already contains a hack to get it to work with
ErrorT, but future versions of the library will remove that hack and will require users to specify their own monoid for combining errors. The
First monoid captures the current behavior, though one may prefer to use other monoids such as a monoid that gives a trace of the full execution path, or witnesses for all the backtracks, etc. ↩
 To be honest, I don't entirely recall why I had the error monad explicitly separated out as a monad transformer over the binding monad, rather than allowing these two layers to be combined. Since it's so awkward, I'm sure there was some reason behind it, I just failed to make note of why. If there turns out to be no decent reason for it, future versions of the library may remove this fine-grain distinction. ↩
 The symbolic name for subsumption is chosen to reflect the meaning of more/less defined (rather than more/less grounded) so that the subsumption ordering coincides with the domain ordering (think of logic variables as being bottom). This is the standard direction for looking at subsumption; though, of course, we could always consider the dual ordering instead. ↩