Functions

Next, we’ll build diamondback which adds support for

  • User-Defined Functions

In the process of doing so, we will learn abou t

  • Static Checking
  • Calling Conventions
  • Tail Recursion

Plan

  1. Defining Functions
  2. Checking Functions
  3. Compiling Functions
  4. Compiling Tail Calls

1. Defining Functions

First, lets add functions to our language.

As always, lets look at some examples.

Example: Increment

For example, a function that increments its input:

We have a function definition followed by a single “main” expression, which is evaluated to yield the program’s result, which, in this case, is 11.

Example: Factorial

Here’s a somewhat more interesting example:

This program should produce the result

5
4
3
2
1
0
120

Suppose we modify the above to produce intermediate results:

we should now get:

5
4
3
2
1
0
1
1
2
6
24
120
120

Example: Mutually Recursive Functions

For this language, the function definitions are global * any function can call any other function.

This lets us write mutually recursive functions like:

QUIZ What should be the result of executing the above?

  1. false true false true 0
  2. true false true false 0
  3. false false false false 0
  4. true true true true 0

Types

Lets add some new types to represent programs.

Bindings

Lets create a special type that represents places where variables are bound,

A Bind is basically just an Id decorated with an a which will let us save extra metadata like tags or source positions

  • The metadata will make it easy to report errors.

We will use Bind at two places:

  1. Let-bindings,
  2. Function parameters.

It will be helpful to have a function to extract the Id corresponding to a Bind

Programs

A program is a list of declarations and main expression.

Declarations

Each function lives is its own declaration,

Expressions

Finally, lets add function application (calls) to the source expressions:

An application or call comprises

  • an Id, the name of the function being called,
  • a list of expressions corresponding to the parameters, and
  • a metadata/tag value of type a.

(Note: that we are now using Bind instead of plain Id at a Let.)

Examples Revisited

Lets see how the examples above are represented:

ghci> parseFile "tests/input/incr.diamond"
Prog {pDecls = [Decl { fName = Bind "incr" ()
                     , fArgs = [Bind "n" ()]
                     , fBody = Prim2 Plus (Id "n" ()) (Number 1 ()) ()
                     , fLabel = ()}
               ]
     , pBody = App "incr" [Number 5 ()] ()
     }

ghci> parseFile "tests/input/fac.diamond"
Prog { pDecls = [ Decl {fName = Bind "fac" ()
                , fArgs = [Bind "n" ()]
                , fBody = Let (Bind "t" ()) (Prim1 Print (Id "n" ()) ())
                          (If (Prim2 Less (Id "n" ()) (Number 1 ()) ())
                             (Number 1 ())
                             (Prim2 Times (Id "n" ())
                                (App "fac" [Prim2 Minus (Id "n" ()) (Number 1 ()) ()] ())
                                ()) ()) ()
                , fLabel = ()}
                ]
     , pBody  = App "fac" [Number 5 ()] ()
     }

2. Static Checking

Next, we will look at an increasingly important aspect of compilation, pointing out bugs in the code at compile time

Called Static Checking because we do this without (i.e. before) compiling and running (“dynamicking”) the code.

There is a huge spectrum of checks possible:

  • Code Linting jslint, hlint
  • Static Typing
  • Static Analysis
  • Contract Checking
  • Dependent or Refinement Typing

Increasingly, this is the most important phase of a compiler, and modern compiler engineering is built around making these checks lightning fast. For more, see this interview of Anders Hejlsberg the architect of the C# and TypeScript compilers.

Static Well-formedness Checking

We will look at code linting and, later in the quarter, type systems in 131.

For the former, suppose you tried to compile:

We would like compilation to fail, not silently, but with useful messages:

$ make tests/output/err-fac.result

Errors found!

tests/input/err-fac.diamond:6:13-14: Unbound variable 'm'

         6|      n * fac(m - 1)
                         ^

tests/input/err-fac.diamond:8:1-9: Function 'fact' is not defined

         8|  fact(5) + fac(3, 4)      
             ^^^^^^^^

tests/input/err-fac.diamond:(8:11)-(9:1): Wrong arity of arguments at call of fac

         8|  fact(5) + fac(3, 4)
                       ^^^^^^^^^

We get multiple errors:

  1. The variable m is not defined,
  2. The function fact is not defined,
  3. The call fac has the wrong number of arguments.

Next, lets see how to update the architecture of our compiler to support these and other kinds of errors.

Types

An error message type:

We make it an exception (that can be thrown):

We can create errors with:

We can throw errors with:

We display errors with:

which takes something like:

and produce a pretty message (that requires reading the source file),

tests/input/err-fac.diamond:6:13-14: Unbound variable 'm'

         6|      n * fac(m - 1)
                         ^

We can put it all together by

Which runs the compiler and if any UserError are thrown, catch-es and renders the result.

Transforms

Next, lets insert a checker phase into our pipeline:

Compiler Pipeline with Checking Phase
Compiler Pipeline with Checking Phase

In the above, we have defined the types:

Catching Multiple Errors

To make using a language and compiler pleasant, we should return as many errors as possible in each run.

  • Its rather irritating to get errors one-by-one.

We will implement this by writing the functions

which will recursively walk over the entire program, declaration and expression and return the list of all errors.

  • If this list is empty, we just return the source unchanged,
  • Otherwise, we throw the list of found errors (and exit.)

Thus, our check function looks like this:

Well-formed Programs, Declarations and Expressions

The bulk of the work is done by:

This function,

  1. creates fEnv, a map from function-names to the function-arity (number of params),
  2. computes the errors for each declaration (given functions in fEnv),
  3. concatenates the resulting lists of errors.

Traversals

Lets look at how we might find three types of errors:

  1. “unbound variables”
  2. “undefined functions”

(In your assignment, you will look for many more.)

The helper function wellFormedD creates an initial variable environment vEnv containing the functions parameters, and uses that (and fEnv) to walk over the body-expressions.

The helper function wellFormedE starts with the input vEnv0 (which has just) the function parameters, and fEnv that has the defined functions, and traverses the expression:

  • At each definition Let x e1 e2, the variable x is added to the environment used to check e2,
  • At each use Id x we check if x is in vEnv and if not, create a suitable UserError
  • At each call App f es we check if f is in fEnv and if not, create a suitable UserError.

You should understand the above and be able to easily add extra error checks.

QUIZ Which function(s) would we have to modify to add large number errors (i.e. errors for numeric literals that may cause overflow)?

  1. wellFormed :: BareProgram -> [UserError]
  2. wellFormedD :: FunEnv -> BareDecl -> [UserError]
  3. wellFormedE :: FunEnv -> Env -> Bare -> [UserError]
  4. 1 and 2
  5. 2 and 3

QUIZ Which function(s) would we have to modify to add variable shadowing errors ?

  1. wellFormed :: BareProgram -> [UserError]
  2. wellFormedD :: FunEnv -> BareDecl -> [UserError]
  3. wellFormedE :: FunEnv -> Env -> Bare -> [UserError]
  4. 1 and 2
  5. 2 and 3

QUIZ Which function(s) would we have to modify to add duplicate parameter errors ?

  1. wellFormed :: BareProgram -> [UserError]
  2. wellFormedD :: FunEnv -> BareDecl -> [UserError]
  3. wellFormedE :: FunEnv -> Env -> Bare -> [UserError]
  4. 1 and 2
  5. 2 and 3

QUIZ Which function(s) would we have to modify to add duplicate function errors ?

  1. wellFormed :: BareProgram -> [UserError]
  2. wellFormedD :: FunEnv -> BareDecl -> [UserError]
  3. wellFormedE :: FunEnv -> Env -> Bare -> [UserError]
  4. 1 and 2
  5. 2 and 3

3. Compiling Functions

Compiler Pipeline for Functions
Compiler Pipeline for Functions

In the above, we have defined the types:

Tagging

Compiler Pipeline ANF
Compiler Pipeline ANF

The tag phase simply recursively tags each function body and the main expression

ANF Conversion

Compiler Pipeline ANF
Compiler Pipeline ANF
  • The normalize phase (i.e. anf) is recursively applied to each function body.

  • In addition to Prim2 operands, each call’s arguments should be transformed into an immediate expression

Generalize the strategy for binary operators

  • from (2 arguments) to n-arguments.

Strategy

Now, lets look at compiling function definitions and calls.

Compiler Pipeline with Checking Phase
Compiler Pipeline with Checking Phase

We need a co-ordinated strategy for definitions and calls.

Definitions * Each definition is compiled into a labeled block of Asm * That implements the body of the definitions. * (But what about the parameters)?

Calls * Each call of f(args) will execute the block labeled f * (But what about the parameters)?

Strategy: The Stack

Stack Frames
Stack Frames

We will use our old friend, the stack to

  • pass parameters
  • have local variables for called functions.

Calling Convention

Recall that we are using the C calling convention that ensures the following stack layout:

Stack Layout
Stack Layout

Strategy: Definitions

When the function body starts executing, the parameters x1, x2, … xn are at [ebp + 4*2], [ebp + 4*3], … [ebp + 4*(n+1)].

  1. Ensure that enough stack space is allocated i.e. that esp and ebp are properly managed

  2. Compile body with initial Env mapping parameters to -2, -3,…,-(n+1).

Strategy: Calls

As before we must ensure that the parameters actually live at the above address.

  1. Before the call, push the parameter values onto the stack in reverse order,
  2. Call the appropriate function (using its label),
  3. After the call, clear the stack by incrementing esp appropriately.

NOTE:

At both definition and call, if you are compiling on MacOS, you need to also respect the 16-Byte Stack Alignment Invariant

Types

We already have most of the machinery needed to compile calls.

Lets just add a new kind of Label for each user-defined function:

We will also extend the Arg type to include information about size directives

We will often need to specify that an Arg is a double word
(the other possibilities are – single word and byte) which we needn’t worry about.

Implementation

Lets can refactor our compile functions into:

that respectively compile Program, Decl and Expr.

In order to simplify stack managment as in Cobra lets have a helper function that compiles the body of each function:

compileBody env e will wrap the Asm generated by compileExpr env e with the code that manages esp and ebp.

Compiling Programs

To compile a Program we compile each Decl and the main body expression

QUIZ:

Does it matter whether we put the code for e before ds?

  1. Yes
  2. No

QUIZ:

Does it matter what order we compile the ds ?

  1. Yes
  2. No

Compiling Declarations

To compile a single Decl we

  1. Create a block starting with a label for the function’s name (so we know where to call),
  2. Invoke compileBody to fill in the assembly code for the body, using the initial Env obtained from the function’s formal parameters.

The initial Env is created by paramsEnv which returns an Env mapping each parameter to its stack position

(Recall that bindId extracts the Id from each Bind)

Finally, as in cobra, compileBody env e wraps the assmbly for e with the code that manages esp and ebp.

Compiling Calls

Finally, lets extend code generation to account for calls:

The function param converts an immediate expressions (corresponding to function arguments)

The Sized DWordPtr specifies that each argument will occupy a double word (i.e. 4 bytes) on the stack.

EXERCISE The hard work in compiling calls is done by:

which implements the strategy for calls. Fill in the implementation of call yourself. As an example, of its behavior, consider the (source) program:

The call add2(12, 7) is represented as:

The code for the above call is generated by

where arg converts source values into assembly Arg which should generate the equivalent of the assembly:

4. Compiling Tail Calls

Our language doesn’t have loops. While recursion is more general, it is more expensive because it uses up stack space (and requires all the attendant management overhead). For example (the python program):

  • Requires a single stack frame
  • Can be implemented with 2 registers

But, the “equivalent” diamond program

  • Requires 10000 stack frames …
  • One for fac(10000), one for fac(9999) etc.

Tail Recursion

Fortunately, we can do much better.

A tail recursive function is one where the recursive call is the last operation done by the function, i.e. where the value returned by the function is the same as the value returned by the recursive call.

We can rewrite sumTo using a tail-recursive loop function:

Visualizing Tail Calls

Lets compare the execution of the two versions of sumTo

Plain Recursion

  • Each call pushes a frame onto the call-stack;
  • The results are popped off and added to the parameter at that frame.

Tail Recursion

  • Accumulation happens in the parameter (not with the output),
  • Each call returns its result without further computation

No need to use call-stack, can make recursive call in place. * Tail recursive calls can be compiled into loops!

Tail Recursion Strategy

Instead of using call to make the call, simply:

  1. Move the call’s arguments to the (same) stack position (as current args),
  2. Free current stack space by resetting esp and ebp (as just prior to ret c.f. exitCode),
  3. Jump to the start of the function.

That is, here’s what a naive implementation would look like:

but a tail-recursive call can instead be compiled as:

which has the effect of executing loop literally as if it were a while-loop!

Requirements

To implement the above strategy, we need a way to:

  1. Identify tail calls in the source Expr (AST),
  2. Compile the tail calls following the above strategy.

Types

We can do the above in a single step, i.e., we could identify the tail calls during the code generation, but its cleaner to separate the steps into:

Labeling Expr with Tail Calls
Labeling Expr with Tail Calls

In the above, we have defined the types:

Transforms

Thus, to implement tail-call optimization, we need to write two transforms:

1. To Label each call with True (if it is a tail call) or False otherwise:

2. To Compile tail calls, by extending compileExpr

Labeling Tail Calls

Which Calls are Tail Calls?
Which Calls are Tail Calls?

The Expr in non tail positions

  • Prim1
  • Prim2
  • Let (“bound expression”)
  • If (“condition”)

cannot contain tail calls; all those values have some further computation performed on them.

However, the Expr in tail positions

  • If (“then” and “else” branch)
  • Let (“body”)

can contain tail calls (unless they appear under the first case)

Algorithm: Traverse Expr using a Bool

  • Initially True but
  • Toggled to False under non-tail positions,
  • Used as “tail-label” at each call.

NOTE: All non-calls get a default tail-label of False.

EXERCISE: How could we modify the above to only mark tail-recursive calls, i.e. to the same function (whose declaration is being compiled?)

Compiling Tail Calls

Finally, to generate code, we need only add a special case to compileExpr

That is, if the call is not labeled as a tail call, generate code as before. Otherwise, use tailcall which implements our tail recursion strategy

EXERCISE

Does the above strategy work always? Can you think of situations where it may go horribly wrong?

Recap

We just saw how to add support for first-class function

  • Definitions, and
  • Calls

and a way in which an important class of

  • Tail Recursive functions can be compiled as loops.

Later, we’ll see how to represent functions as values using closures.