123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885 |
- :orphan:
- =======================================================
- Kaleidoscope: Extending the Language: Mutable Variables
- =======================================================
- .. contents::
- :local:
- Chapter 7 Introduction
- ======================
- Welcome to Chapter 7 of the "`Implementing a language with
- LLVM <index.html>`_" tutorial. In chapters 1 through 6, we've built a
- very respectable, albeit simple, `functional programming
- language <http://en.wikipedia.org/wiki/Functional_programming>`_. In our
- journey, we learned some parsing techniques, how to build and represent
- an AST, how to build LLVM IR, and how to optimize the resultant code as
- well as JIT compile it.
- While Kaleidoscope is interesting as a functional language, the fact
- that it is functional makes it "too easy" to generate LLVM IR for it. In
- particular, a functional language makes it very easy to build LLVM IR
- directly in `SSA
- form <http://en.wikipedia.org/wiki/Static_single_assignment_form>`_.
- Since LLVM requires that the input code be in SSA form, this is a very
- nice property and it is often unclear to newcomers how to generate code
- for an imperative language with mutable variables.
- The short (and happy) summary of this chapter is that there is no need
- for your front-end to build SSA form: LLVM provides highly tuned and
- well tested support for this, though the way it works is a bit
- unexpected for some.
- Why is this a hard problem?
- ===========================
- To understand why mutable variables cause complexities in SSA
- construction, consider this extremely simple C example:
- .. code-block:: c
- int G, H;
- int test(_Bool Condition) {
- int X;
- if (Condition)
- X = G;
- else
- X = H;
- return X;
- }
- In this case, we have the variable "X", whose value depends on the path
- executed in the program. Because there are two different possible values
- for X before the return instruction, a PHI node is inserted to merge the
- two values. The LLVM IR that we want for this example looks like this:
- .. code-block:: llvm
- @G = weak global i32 0 ; type of @G is i32*
- @H = weak global i32 0 ; type of @H is i32*
- define i32 @test(i1 %Condition) {
- entry:
- br i1 %Condition, label %cond_true, label %cond_false
- cond_true:
- %X.0 = load i32* @G
- br label %cond_next
- cond_false:
- %X.1 = load i32* @H
- br label %cond_next
- cond_next:
- %X.2 = phi i32 [ %X.1, %cond_false ], [ %X.0, %cond_true ]
- ret i32 %X.2
- }
- In this example, the loads from the G and H global variables are
- explicit in the LLVM IR, and they live in the then/else branches of the
- if statement (cond\_true/cond\_false). In order to merge the incoming
- values, the X.2 phi node in the cond\_next block selects the right value
- to use based on where control flow is coming from: if control flow comes
- from the cond\_false block, X.2 gets the value of X.1. Alternatively, if
- control flow comes from cond\_true, it gets the value of X.0. The intent
- of this chapter is not to explain the details of SSA form. For more
- information, see one of the many `online
- references <http://en.wikipedia.org/wiki/Static_single_assignment_form>`_.
- The question for this article is "who places the phi nodes when lowering
- assignments to mutable variables?". The issue here is that LLVM
- *requires* that its IR be in SSA form: there is no "non-ssa" mode for
- it. However, SSA construction requires non-trivial algorithms and data
- structures, so it is inconvenient and wasteful for every front-end to
- have to reproduce this logic.
- Memory in LLVM
- ==============
- The 'trick' here is that while LLVM does require all register values to
- be in SSA form, it does not require (or permit) memory objects to be in
- SSA form. In the example above, note that the loads from G and H are
- direct accesses to G and H: they are not renamed or versioned. This
- differs from some other compiler systems, which do try to version memory
- objects. In LLVM, instead of encoding dataflow analysis of memory into
- the LLVM IR, it is handled with `Analysis
- Passes <../WritingAnLLVMPass.html>`_ which are computed on demand.
- With this in mind, the high-level idea is that we want to make a stack
- variable (which lives in memory, because it is on the stack) for each
- mutable object in a function. To take advantage of this trick, we need
- to talk about how LLVM represents stack variables.
- In LLVM, all memory accesses are explicit with load/store instructions,
- and it is carefully designed not to have (or need) an "address-of"
- operator. Notice how the type of the @G/@H global variables is actually
- "i32\*" even though the variable is defined as "i32". What this means is
- that @G defines *space* for an i32 in the global data area, but its
- *name* actually refers to the address for that space. Stack variables
- work the same way, except that instead of being declared with global
- variable definitions, they are declared with the `LLVM alloca
- instruction <../LangRef.html#alloca-instruction>`_:
- .. code-block:: llvm
- define i32 @example() {
- entry:
- %X = alloca i32 ; type of %X is i32*.
- ...
- %tmp = load i32* %X ; load the stack value %X from the stack.
- %tmp2 = add i32 %tmp, 1 ; increment it
- store i32 %tmp2, i32* %X ; store it back
- ...
- This code shows an example of how you can declare and manipulate a stack
- variable in the LLVM IR. Stack memory allocated with the alloca
- instruction is fully general: you can pass the address of the stack slot
- to functions, you can store it in other variables, etc. In our example
- above, we could rewrite the example to use the alloca technique to avoid
- using a PHI node:
- .. code-block:: llvm
- @G = weak global i32 0 ; type of @G is i32*
- @H = weak global i32 0 ; type of @H is i32*
- define i32 @test(i1 %Condition) {
- entry:
- %X = alloca i32 ; type of %X is i32*.
- br i1 %Condition, label %cond_true, label %cond_false
- cond_true:
- %X.0 = load i32* @G
- store i32 %X.0, i32* %X ; Update X
- br label %cond_next
- cond_false:
- %X.1 = load i32* @H
- store i32 %X.1, i32* %X ; Update X
- br label %cond_next
- cond_next:
- %X.2 = load i32* %X ; Read X
- ret i32 %X.2
- }
- With this, we have discovered a way to handle arbitrary mutable
- variables without the need to create Phi nodes at all:
- #. Each mutable variable becomes a stack allocation.
- #. Each read of the variable becomes a load from the stack.
- #. Each update of the variable becomes a store to the stack.
- #. Taking the address of a variable just uses the stack address
- directly.
- While this solution has solved our immediate problem, it introduced
- another one: we have now apparently introduced a lot of stack traffic
- for very simple and common operations, a major performance problem.
- Fortunately for us, the LLVM optimizer has a highly-tuned optimization
- pass named "mem2reg" that handles this case, promoting allocas like this
- into SSA registers, inserting Phi nodes as appropriate. If you run this
- example through the pass, for example, you'll get:
- .. code-block:: bash
- $ llvm-as < example.ll | opt -mem2reg | llvm-dis
- @G = weak global i32 0
- @H = weak global i32 0
- define i32 @test(i1 %Condition) {
- entry:
- br i1 %Condition, label %cond_true, label %cond_false
- cond_true:
- %X.0 = load i32* @G
- br label %cond_next
- cond_false:
- %X.1 = load i32* @H
- br label %cond_next
- cond_next:
- %X.01 = phi i32 [ %X.1, %cond_false ], [ %X.0, %cond_true ]
- ret i32 %X.01
- }
- The mem2reg pass implements the standard "iterated dominance frontier"
- algorithm for constructing SSA form and has a number of optimizations
- that speed up (very common) degenerate cases. The mem2reg optimization
- pass is the answer to dealing with mutable variables, and we highly
- recommend that you depend on it. Note that mem2reg only works on
- variables in certain circumstances:
- #. mem2reg is alloca-driven: it looks for allocas and if it can handle
- them, it promotes them. It does not apply to global variables or heap
- allocations.
- #. mem2reg only looks for alloca instructions in the entry block of the
- function. Being in the entry block guarantees that the alloca is only
- executed once, which makes analysis simpler.
- #. mem2reg only promotes allocas whose uses are direct loads and stores.
- If the address of the stack object is passed to a function, or if any
- funny pointer arithmetic is involved, the alloca will not be
- promoted.
- #. mem2reg only works on allocas of `first
- class <../LangRef.html#first-class-types>`_ values (such as pointers,
- scalars and vectors), and only if the array size of the allocation is
- 1 (or missing in the .ll file). mem2reg is not capable of promoting
- structs or arrays to registers. Note that the "sroa" pass is
- more powerful and can promote structs, "unions", and arrays in many
- cases.
- All of these properties are easy to satisfy for most imperative
- languages, and we'll illustrate it below with Kaleidoscope. The final
- question you may be asking is: should I bother with this nonsense for my
- front-end? Wouldn't it be better if I just did SSA construction
- directly, avoiding use of the mem2reg optimization pass? In short, we
- strongly recommend that you use this technique for building SSA form,
- unless there is an extremely good reason not to. Using this technique
- is:
- - Proven and well tested: clang uses this technique
- for local mutable variables. As such, the most common clients of LLVM
- are using this to handle a bulk of their variables. You can be sure
- that bugs are found fast and fixed early.
- - Extremely Fast: mem2reg has a number of special cases that make it
- fast in common cases as well as fully general. For example, it has
- fast-paths for variables that are only used in a single block,
- variables that only have one assignment point, good heuristics to
- avoid insertion of unneeded phi nodes, etc.
- - Needed for debug info generation: `Debug information in
- LLVM <../SourceLevelDebugging.html>`_ relies on having the address of
- the variable exposed so that debug info can be attached to it. This
- technique dovetails very naturally with this style of debug info.
- If nothing else, this makes it much easier to get your front-end up and
- running, and is very simple to implement. Let's extend Kaleidoscope with
- mutable variables now!
- Mutable Variables in Kaleidoscope
- =================================
- Now that we know the sort of problem we want to tackle, let's see what
- this looks like in the context of our little Kaleidoscope language.
- We're going to add two features:
- #. The ability to mutate variables with the '=' operator.
- #. The ability to define new variables.
- While the first item is really what this is about, we only have
- variables for incoming arguments as well as for induction variables, and
- redefining those only goes so far :). Also, the ability to define new
- variables is a useful thing regardless of whether you will be mutating
- them. Here's a motivating example that shows how we could use these:
- ::
- # Define ':' for sequencing: as a low-precedence operator that ignores operands
- # and just returns the RHS.
- def binary : 1 (x y) y;
- # Recursive fib, we could do this before.
- def fib(x)
- if (x < 3) then
- 1
- else
- fib(x-1)+fib(x-2);
- # Iterative fib.
- def fibi(x)
- var a = 1, b = 1, c in
- (for i = 3, i < x in
- c = a + b :
- a = b :
- b = c) :
- b;
- # Call it.
- fibi(10);
- In order to mutate variables, we have to change our existing variables
- to use the "alloca trick". Once we have that, we'll add our new
- operator, then extend Kaleidoscope to support new variable definitions.
- Adjusting Existing Variables for Mutation
- =========================================
- The symbol table in Kaleidoscope is managed at code generation time by
- the '``NamedValues``' map. This map currently keeps track of the LLVM
- "Value\*" that holds the double value for the named variable. In order
- to support mutation, we need to change this slightly, so that
- ``NamedValues`` holds the *memory location* of the variable in question.
- Note that this change is a refactoring: it changes the structure of the
- code, but does not (by itself) change the behavior of the compiler. All
- of these changes are isolated in the Kaleidoscope code generator.
- At this point in Kaleidoscope's development, it only supports variables
- for two things: incoming arguments to functions and the induction
- variable of 'for' loops. For consistency, we'll allow mutation of these
- variables in addition to other user-defined variables. This means that
- these will both need memory locations.
- To start our transformation of Kaleidoscope, we'll change the
- NamedValues map so that it maps to AllocaInst\* instead of Value\*. Once
- we do this, the C++ compiler will tell us what parts of the code we need
- to update:
- .. code-block:: c++
- static std::map<std::string, AllocaInst*> NamedValues;
- Also, since we will need to create these allocas, we'll use a helper
- function that ensures that the allocas are created in the entry block of
- the function:
- .. code-block:: c++
- /// CreateEntryBlockAlloca - Create an alloca instruction in the entry block of
- /// the function. This is used for mutable variables etc.
- static AllocaInst *CreateEntryBlockAlloca(Function *TheFunction,
- const std::string &VarName) {
- IRBuilder<> TmpB(&TheFunction->getEntryBlock(),
- TheFunction->getEntryBlock().begin());
- return TmpB.CreateAlloca(Type::getDoubleTy(TheContext), 0,
- VarName.c_str());
- }
- This funny looking code creates an IRBuilder object that is pointing at
- the first instruction (.begin()) of the entry block. It then creates an
- alloca with the expected name and returns it. Because all values in
- Kaleidoscope are doubles, there is no need to pass in a type to use.
- With this in place, the first functionality change we want to make belongs to
- variable references. In our new scheme, variables live on the stack, so
- code generating a reference to them actually needs to produce a load
- from the stack slot:
- .. code-block:: c++
- Value *VariableExprAST::codegen() {
- // Look this variable up in the function.
- Value *V = NamedValues[Name];
- if (!V)
- return LogErrorV("Unknown variable name");
- // Load the value.
- return Builder.CreateLoad(V, Name.c_str());
- }
- As you can see, this is pretty straightforward. Now we need to update
- the things that define the variables to set up the alloca. We'll start
- with ``ForExprAST::codegen()`` (see the `full code listing <#id1>`_ for
- the unabridged code):
- .. code-block:: c++
- Function *TheFunction = Builder.GetInsertBlock()->getParent();
- // Create an alloca for the variable in the entry block.
- AllocaInst *Alloca = CreateEntryBlockAlloca(TheFunction, VarName);
- // Emit the start code first, without 'variable' in scope.
- Value *StartVal = Start->codegen();
- if (!StartVal)
- return nullptr;
- // Store the value into the alloca.
- Builder.CreateStore(StartVal, Alloca);
- ...
- // Compute the end condition.
- Value *EndCond = End->codegen();
- if (!EndCond)
- return nullptr;
- // Reload, increment, and restore the alloca. This handles the case where
- // the body of the loop mutates the variable.
- Value *CurVar = Builder.CreateLoad(Alloca);
- Value *NextVar = Builder.CreateFAdd(CurVar, StepVal, "nextvar");
- Builder.CreateStore(NextVar, Alloca);
- ...
- This code is virtually identical to the code `before we allowed mutable
- variables <LangImpl5.html#code-generation-for-the-for-loop>`_. The big difference is that we
- no longer have to construct a PHI node, and we use load/store to access
- the variable as needed.
- To support mutable argument variables, we need to also make allocas for
- them. The code for this is also pretty simple:
- .. code-block:: c++
- Function *FunctionAST::codegen() {
- ...
- Builder.SetInsertPoint(BB);
- // Record the function arguments in the NamedValues map.
- NamedValues.clear();
- for (auto &Arg : TheFunction->args()) {
- // Create an alloca for this variable.
- AllocaInst *Alloca = CreateEntryBlockAlloca(TheFunction, Arg.getName());
- // Store the initial value into the alloca.
- Builder.CreateStore(&Arg, Alloca);
- // Add arguments to variable symbol table.
- NamedValues[Arg.getName()] = Alloca;
- }
- if (Value *RetVal = Body->codegen()) {
- ...
- For each argument, we make an alloca, store the input value to the
- function into the alloca, and register the alloca as the memory location
- for the argument. This method gets invoked by ``FunctionAST::codegen()``
- right after it sets up the entry block for the function.
- The final missing piece is adding the mem2reg pass, which allows us to
- get good codegen once again:
- .. code-block:: c++
- // Promote allocas to registers.
- TheFPM->add(createPromoteMemoryToRegisterPass());
- // Do simple "peephole" optimizations and bit-twiddling optzns.
- TheFPM->add(createInstructionCombiningPass());
- // Reassociate expressions.
- TheFPM->add(createReassociatePass());
- ...
- It is interesting to see what the code looks like before and after the
- mem2reg optimization runs. For example, this is the before/after code
- for our recursive fib function. Before the optimization:
- .. code-block:: llvm
- define double @fib(double %x) {
- entry:
- %x1 = alloca double
- store double %x, double* %x1
- %x2 = load double, double* %x1
- %cmptmp = fcmp ult double %x2, 3.000000e+00
- %booltmp = uitofp i1 %cmptmp to double
- %ifcond = fcmp one double %booltmp, 0.000000e+00
- br i1 %ifcond, label %then, label %else
- then: ; preds = %entry
- br label %ifcont
- else: ; preds = %entry
- %x3 = load double, double* %x1
- %subtmp = fsub double %x3, 1.000000e+00
- %calltmp = call double @fib(double %subtmp)
- %x4 = load double, double* %x1
- %subtmp5 = fsub double %x4, 2.000000e+00
- %calltmp6 = call double @fib(double %subtmp5)
- %addtmp = fadd double %calltmp, %calltmp6
- br label %ifcont
- ifcont: ; preds = %else, %then
- %iftmp = phi double [ 1.000000e+00, %then ], [ %addtmp, %else ]
- ret double %iftmp
- }
- Here there is only one variable (x, the input argument) but you can
- still see the extremely simple-minded code generation strategy we are
- using. In the entry block, an alloca is created, and the initial input
- value is stored into it. Each reference to the variable does a reload
- from the stack. Also, note that we didn't modify the if/then/else
- expression, so it still inserts a PHI node. While we could make an
- alloca for it, it is actually easier to create a PHI node for it, so we
- still just make the PHI.
- Here is the code after the mem2reg pass runs:
- .. code-block:: llvm
- define double @fib(double %x) {
- entry:
- %cmptmp = fcmp ult double %x, 3.000000e+00
- %booltmp = uitofp i1 %cmptmp to double
- %ifcond = fcmp one double %booltmp, 0.000000e+00
- br i1 %ifcond, label %then, label %else
- then:
- br label %ifcont
- else:
- %subtmp = fsub double %x, 1.000000e+00
- %calltmp = call double @fib(double %subtmp)
- %subtmp5 = fsub double %x, 2.000000e+00
- %calltmp6 = call double @fib(double %subtmp5)
- %addtmp = fadd double %calltmp, %calltmp6
- br label %ifcont
- ifcont: ; preds = %else, %then
- %iftmp = phi double [ 1.000000e+00, %then ], [ %addtmp, %else ]
- ret double %iftmp
- }
- This is a trivial case for mem2reg, since there are no redefinitions of
- the variable. The point of showing this is to calm your tension about
- inserting such blatant inefficiencies :).
- After the rest of the optimizers run, we get:
- .. code-block:: llvm
- define double @fib(double %x) {
- entry:
- %cmptmp = fcmp ult double %x, 3.000000e+00
- %booltmp = uitofp i1 %cmptmp to double
- %ifcond = fcmp ueq double %booltmp, 0.000000e+00
- br i1 %ifcond, label %else, label %ifcont
- else:
- %subtmp = fsub double %x, 1.000000e+00
- %calltmp = call double @fib(double %subtmp)
- %subtmp5 = fsub double %x, 2.000000e+00
- %calltmp6 = call double @fib(double %subtmp5)
- %addtmp = fadd double %calltmp, %calltmp6
- ret double %addtmp
- ifcont:
- ret double 1.000000e+00
- }
- Here we see that the simplifycfg pass decided to clone the return
- instruction into the end of the 'else' block. This allowed it to
- eliminate some branches and the PHI node.
- Now that all symbol table references are updated to use stack variables,
- we'll add the assignment operator.
- New Assignment Operator
- =======================
- With our current framework, adding a new assignment operator is really
- simple. We will parse it just like any other binary operator, but handle
- it internally (instead of allowing the user to define it). The first
- step is to set a precedence:
- .. code-block:: c++
- int main() {
- // Install standard binary operators.
- // 1 is lowest precedence.
- BinopPrecedence['='] = 2;
- BinopPrecedence['<'] = 10;
- BinopPrecedence['+'] = 20;
- BinopPrecedence['-'] = 20;
- Now that the parser knows the precedence of the binary operator, it
- takes care of all the parsing and AST generation. We just need to
- implement codegen for the assignment operator. This looks like:
- .. code-block:: c++
- Value *BinaryExprAST::codegen() {
- // Special case '=' because we don't want to emit the LHS as an expression.
- if (Op == '=') {
- // Assignment requires the LHS to be an identifier.
- VariableExprAST *LHSE = dynamic_cast<VariableExprAST*>(LHS.get());
- if (!LHSE)
- return LogErrorV("destination of '=' must be a variable");
- Unlike the rest of the binary operators, our assignment operator doesn't
- follow the "emit LHS, emit RHS, do computation" model. As such, it is
- handled as a special case before the other binary operators are handled.
- The other strange thing is that it requires the LHS to be a variable. It
- is invalid to have "(x+1) = expr" - only things like "x = expr" are
- allowed.
- .. code-block:: c++
- // Codegen the RHS.
- Value *Val = RHS->codegen();
- if (!Val)
- return nullptr;
- // Look up the name.
- Value *Variable = NamedValues[LHSE->getName()];
- if (!Variable)
- return LogErrorV("Unknown variable name");
- Builder.CreateStore(Val, Variable);
- return Val;
- }
- ...
- Once we have the variable, codegen'ing the assignment is
- straightforward: we emit the RHS of the assignment, create a store, and
- return the computed value. Returning a value allows for chained
- assignments like "X = (Y = Z)".
- Now that we have an assignment operator, we can mutate loop variables
- and arguments. For example, we can now run code like this:
- ::
- # Function to print a double.
- extern printd(x);
- # Define ':' for sequencing: as a low-precedence operator that ignores operands
- # and just returns the RHS.
- def binary : 1 (x y) y;
- def test(x)
- printd(x) :
- x = 4 :
- printd(x);
- test(123);
- When run, this example prints "123" and then "4", showing that we did
- actually mutate the value! Okay, we have now officially implemented our
- goal: getting this to work requires SSA construction in the general
- case. However, to be really useful, we want the ability to define our
- own local variables, let's add this next!
- User-defined Local Variables
- ============================
- Adding var/in is just like any other extension we made to
- Kaleidoscope: we extend the lexer, the parser, the AST and the code
- generator. The first step for adding our new 'var/in' construct is to
- extend the lexer. As before, this is pretty trivial, the code looks like
- this:
- .. code-block:: c++
- enum Token {
- ...
- // var definition
- tok_var = -13
- ...
- }
- ...
- static int gettok() {
- ...
- if (IdentifierStr == "in")
- return tok_in;
- if (IdentifierStr == "binary")
- return tok_binary;
- if (IdentifierStr == "unary")
- return tok_unary;
- if (IdentifierStr == "var")
- return tok_var;
- return tok_identifier;
- ...
- The next step is to define the AST node that we will construct. For
- var/in, it looks like this:
- .. code-block:: c++
- /// VarExprAST - Expression class for var/in
- class VarExprAST : public ExprAST {
- std::vector<std::pair<std::string, std::unique_ptr<ExprAST>>> VarNames;
- std::unique_ptr<ExprAST> Body;
- public:
- VarExprAST(std::vector<std::pair<std::string, std::unique_ptr<ExprAST>>> VarNames,
- std::unique_ptr<ExprAST> Body)
- : VarNames(std::move(VarNames)), Body(std::move(Body)) {}
- Value *codegen() override;
- };
- var/in allows a list of names to be defined all at once, and each name
- can optionally have an initializer value. As such, we capture this
- information in the VarNames vector. Also, var/in has a body, this body
- is allowed to access the variables defined by the var/in.
- With this in place, we can define the parser pieces. The first thing we
- do is add it as a primary expression:
- .. code-block:: c++
- /// primary
- /// ::= identifierexpr
- /// ::= numberexpr
- /// ::= parenexpr
- /// ::= ifexpr
- /// ::= forexpr
- /// ::= varexpr
- static std::unique_ptr<ExprAST> ParsePrimary() {
- switch (CurTok) {
- default:
- return LogError("unknown token when expecting an expression");
- case tok_identifier:
- return ParseIdentifierExpr();
- case tok_number:
- return ParseNumberExpr();
- case '(':
- return ParseParenExpr();
- case tok_if:
- return ParseIfExpr();
- case tok_for:
- return ParseForExpr();
- case tok_var:
- return ParseVarExpr();
- }
- }
- Next we define ParseVarExpr:
- .. code-block:: c++
- /// varexpr ::= 'var' identifier ('=' expression)?
- // (',' identifier ('=' expression)?)* 'in' expression
- static std::unique_ptr<ExprAST> ParseVarExpr() {
- getNextToken(); // eat the var.
- std::vector<std::pair<std::string, std::unique_ptr<ExprAST>>> VarNames;
- // At least one variable name is required.
- if (CurTok != tok_identifier)
- return LogError("expected identifier after var");
- The first part of this code parses the list of identifier/expr pairs
- into the local ``VarNames`` vector.
- .. code-block:: c++
- while (1) {
- std::string Name = IdentifierStr;
- getNextToken(); // eat identifier.
- // Read the optional initializer.
- std::unique_ptr<ExprAST> Init;
- if (CurTok == '=') {
- getNextToken(); // eat the '='.
- Init = ParseExpression();
- if (!Init) return nullptr;
- }
- VarNames.push_back(std::make_pair(Name, std::move(Init)));
- // End of var list, exit loop.
- if (CurTok != ',') break;
- getNextToken(); // eat the ','.
- if (CurTok != tok_identifier)
- return LogError("expected identifier list after var");
- }
- Once all the variables are parsed, we then parse the body and create the
- AST node:
- .. code-block:: c++
- // At this point, we have to have 'in'.
- if (CurTok != tok_in)
- return LogError("expected 'in' keyword after 'var'");
- getNextToken(); // eat 'in'.
- auto Body = ParseExpression();
- if (!Body)
- return nullptr;
- return std::make_unique<VarExprAST>(std::move(VarNames),
- std::move(Body));
- }
- Now that we can parse and represent the code, we need to support
- emission of LLVM IR for it. This code starts out with:
- .. code-block:: c++
- Value *VarExprAST::codegen() {
- std::vector<AllocaInst *> OldBindings;
- Function *TheFunction = Builder.GetInsertBlock()->getParent();
- // Register all variables and emit their initializer.
- for (unsigned i = 0, e = VarNames.size(); i != e; ++i) {
- const std::string &VarName = VarNames[i].first;
- ExprAST *Init = VarNames[i].second.get();
- Basically it loops over all the variables, installing them one at a
- time. For each variable we put into the symbol table, we remember the
- previous value that we replace in OldBindings.
- .. code-block:: c++
- // Emit the initializer before adding the variable to scope, this prevents
- // the initializer from referencing the variable itself, and permits stuff
- // like this:
- // var a = 1 in
- // var a = a in ... # refers to outer 'a'.
- Value *InitVal;
- if (Init) {
- InitVal = Init->codegen();
- if (!InitVal)
- return nullptr;
- } else { // If not specified, use 0.0.
- InitVal = ConstantFP::get(TheContext, APFloat(0.0));
- }
- AllocaInst *Alloca = CreateEntryBlockAlloca(TheFunction, VarName);
- Builder.CreateStore(InitVal, Alloca);
- // Remember the old variable binding so that we can restore the binding when
- // we unrecurse.
- OldBindings.push_back(NamedValues[VarName]);
- // Remember this binding.
- NamedValues[VarName] = Alloca;
- }
- There are more comments here than code. The basic idea is that we emit
- the initializer, create the alloca, then update the symbol table to
- point to it. Once all the variables are installed in the symbol table,
- we evaluate the body of the var/in expression:
- .. code-block:: c++
- // Codegen the body, now that all vars are in scope.
- Value *BodyVal = Body->codegen();
- if (!BodyVal)
- return nullptr;
- Finally, before returning, we restore the previous variable bindings:
- .. code-block:: c++
- // Pop all our variables from scope.
- for (unsigned i = 0, e = VarNames.size(); i != e; ++i)
- NamedValues[VarNames[i].first] = OldBindings[i];
- // Return the body computation.
- return BodyVal;
- }
- The end result of all of this is that we get properly scoped variable
- definitions, and we even (trivially) allow mutation of them :).
- With this, we completed what we set out to do. Our nice iterative fib
- example from the intro compiles and runs just fine. The mem2reg pass
- optimizes all of our stack variables into SSA registers, inserting PHI
- nodes where needed, and our front-end remains simple: no "iterated
- dominance frontier" computation anywhere in sight.
- Full Code Listing
- =================
- Here is the complete code listing for our running example, enhanced with
- mutable variables and var/in support. To build this example, use:
- .. code-block:: bash
- # Compile
- clang++ -g toy.cpp `llvm-config --cxxflags --ldflags --system-libs --libs core mcjit native` -O3 -o toy
- # Run
- ./toy
- Here is the code:
- .. literalinclude:: ../../../examples/Kaleidoscope/Chapter7/toy.cpp
- :language: c++
- `Next: Compiling to Object Code <LangImpl08.html>`_
|