Deep Dive To (Tail) Recursion - 1 Intro Recursion as one (if no the most) of important concepts.

Posted on February 12, 2017

From time to time someone raises concern about performance of recursive algorithms in some of many conversations I have with my colleagues, friends and folks from local meetup groups. Even though this might seem to be clear to people with certain level of functional programming experience it’s certainly not generally well understood topic. I’ve found that there is a lot of confusion going around and some folks seems to be pretty scared when you pull all that weird terminology right of the box. This is why I still think it’s worth spending some time exploring what recursion and tail recursion really means and how you can use both in practice.

Recursion

Recursion can be found in many areas of mathematics and computer science. There are some examples:

Shell

rm -r folder_name

where -r means apply command rm recursively to all sub files and sub folders.

Algorithms

function length(arr) {
    if (!arr[0]) {
        return 0;
    }
    arr.pop();

    return 1 + length(arr);
}

where function contains call to itself.

length() implementation is just used for demonstration purposes. Please don’t try to use it in real world! There are many problems with the code above. Performance will be really bad and also function will cause mutations in original array which becomes empty after being passed to the function due to pop() call! If you thinking about using something similar in your code base fuckUpLength would be probably better name for such function.

Algebraic Data Types

data RoseTree a = Node a [RoseTree a]

where type definition might be recursive (type containing itself)

Sequences

For instance our friend Fibonacci


$$ \begin{align*} F_0 &= 0, \\ F_1 &= 1, \\ F_n &= F_{n-1} + F_{n-2} \end{align*} $$

where fibonnaci n is define in terms of fibbonaci n-1 and fibbonaci n-2.

Yo man I heard you like…

Even though these three might seem completely unrelated it’s super important to understand that underlaying concept is same.

Let’s use rm example for instance. This command executes itself for every sub-folder of folder it’s executed. This simple principle allows rm to work with any depth of directories in file system.

More generally recursion is property of self-reference. Acronym “GNU” Gnu’s Not Unix is refering to itself (G stays for GNU itself). Therefore we call such acronym recursive acronym.

You can have a look at wikipedia where you can find more in case you’re interested.

Loops

Loops are fairly popular technique used in place of recursion in many programs. It’s quite important to understand that loops are workaround for underlaying recursive nature of program or algorithm. The idea is simple - instead of dealing with function calling itself we loop sequence of “instructions” over data. In some sense loops are just restrictive abstraction over GOTO but core idea stays same. Program checks some condition (like i < arr.length) proceed some instructions, apply some mutation to data it started with (eg. i++) and GOTO condition again to either executes instructions with new state or exists subroutine if condition is not satisfied.

This solves problem with stack overflow - function can’t return (exit) until all its sub-calls (recursive calls) return) and is generally much closer to what we want hardware to do than recursion itself. Let’s rewrite length() implementation from previous examples using loop.

function length(arr) {
    var len = 0;
    while(arr[0]) {
      len++;
      arr.pop()
    }

    return len;
}

Even this implementation of length is bad. Function still performs mutation on array using pop(). Please do not use such code for anything but learning.

Loops are so widely used technique that many programmers kind of defaults to it. This is probably because most widely used languages derives more or less from C (think about C++, Java, C#, JavaScript…). Anyway I would say this is huge tradeoff. Translating algorithm for machine is hard and when people writing programs starts to think as machine they quite often miss general logic in program (Tend to solves adhock cases here and there instead abstracting concepts to composable pieces). Even it might seem easier to use loops over recursion to most developers it’s definitely not easier when it comes to more complex algorithms or design of large systems.

There are so many things that can go wrong in simple for loop that it’s even not worth trying to list them. Concept of loops also heavily relies on mutations which are hard to deal with especially when concurrency (in multithreaded environment) is involved. This is nothing new. Even imperative languages introduced many concepts to address issues like this like for example iterators which I’m not really going to cover in this article but you can find tons of material about them on the internet.

Nevertheless I still think that recursion over power any other concept and should be the thing we defaults to instinctively when thinking about problem.

Even most dynamic programming practices often starts with recursive definition and translates that implementation to loop just to gain better performance characteristics later.

In case you’re interested learning more about dynamic programming I recommend to look at MIT’s public Introduction to Algorithms.

There is still one issue using technique like dynamic programming in my opinion. Once you optimize code to loops it’s stay that way. This means every time you or anyone else will need to tweak something in you’re code youl’ll need to walk through code you was no able to put down from head in first place. You can imagine this won’t be pleasing experience. Wouldn’t it be nice to have recursive implementation and hand optimization to compiler instead?

Good news folks! There are compilers capable of such optimization out there! The only requirement on your side is to keep your implementation tail recursive!

We will look what this mean and how you can use it in next part. Till then let force be with you my friends.