Lecture 23: High Performance in Dynamic Languages

Flash and JavaScript are required for this feature.

Download the video from Internet Archive.

Description: Professor Steven Johnson talks about a new dynamic language called Julia as an alternative to the two-language approach for interactive math. Julia is then compared to Python and C. 

Instructor: Steven Johnson

The following content is provided under a Creative Commons license. Your support will help MIT Open CourseWare continue to offer high quality educational resources for free. To make a donation or to view additional materials from hundreds of MIT courses, visit MIT Open CourseWare at ocw.mit.edu.

CHARLES E. LEISERSON: Hey, everybody. Let's get going. Who here has heard of the FFT? That's most of you. So I first met Steve Johnson when he worked with one of my graduate students, now former graduate student, Matteo Frigo. And they came up with a really spectacular piece of performance engineering for the FFT, a system they call FFTW for the Fastest Fourier Transform in the West. And it has, for over years and years been a staple of anybody doing signal processing will know FFTW.

So anyway, it's a great pleasure to welcome Steve Johnson, who is going to talk about some of the work that he's been doing on dynamic languages, such as Julia and Python.

STEVEN JOHNSON: Yeah. Thanks.

CHARLES E. LEISERSON: Is that pretty actuate?

STEVEN JOHNSON: Yeah. Yeah, so I'm going to talk, as I said, about high level dynamic languages and how you get performance in these. And so most of you have probably used Python, or R, and Matlab. And so these are really popular for people doing in technical computing, statistics, and anything where you want kind of interactive exploration. You'd like to have a dynamically typed language where you can just type x equals 3. And then three lines later, you said, oh, x is an array. Because you're doing things interactively. You don't have to be stuck with a particular set of types. And there's a lot of choices for these.

But they usually hit a wall when it comes to writing performance critical code in these languages. And so traditionally, people doing some serious computing in these languages have a two-language solution. So they do high level exploration, and productivity and so forth in Python or whatever. But when they need to write performance critical code, then you drop down to a lower level language, Fortran, or C, or Cython, or one of these things. And you use Python as the glue for these low level kernels.

And the problem-- and this is workable. I've done this myself. Many of you have probably done this. But when you drop down from Python to C, or even to Cython, there there's a huge discontinuous jump in the complexity of the coding. And there's usually a lot of generality. When you write code in C or something like that, it's specific to a very small set of types, whereas the nice thing about high level languages is you can write generic code that works for a lot of different types.

So at this point, there's often someone who pops up and says, oh, well, I did performance programming in Python. And everyone knows you just need to vectorize your code, right? So basically, what they mean is you rely on mature external libraries that you pass on a big block of data. It does a huge amount of computation and comes back. And so you never write your own loops.

And this is great. If there's someone who's already written the code that you need, you should try and leverage that as much as possible. But somebody has to write those. And eventually, that person will be you. And because eventually if you do scientific computing, you run into a problem inevitably that you just can't express in terms of existing libraries very easily or at all.

So this was the state of affairs for a long time. And a few years ago, starting in Alan Edelman's group at MIT, there was a proposal for a new language called Julia, which tries to be as high level and interactive as-- it's a dynamically typed language, you know, as Matlab, or Python, and so forth. But general purpose language like Python, very productive for technical work, so really oriented towards scientific numerical computing.

But you can write a loop, and you write low level code in that that's as fast as C. So that was the goal. The first release was in 2013. So it's a pretty young language. The 1.0 release was in August of this year. So before that point every year there was a new release, 0.1, 0.2, Point3. And every year, it would break all your old code, and you'd have to update everything to keep it working.

So now they said, OK, it's stable. We'll add new features. We'll make it faster. But from this point onwards, for least until 2.0, many years in the future it will be backwards compatible. So there's lots of-- in my experience, this pretty much holds up. I haven't found any problem where there was a nice highly optimized C or Fortran code where I couldn't write equivalent code or equivalent performance, equivalently performing code in Julia given enough time, right?

Obviously, if something is-- there's a library with 100,000 lines of code. It takes quite a long time to rewrite that in any other language. So there are lots of benchmarks that illustrate this. The goal of Julia is usually to stay within a factor of 2 of C. In my experience, it's usually within a factor of a few percent if you know what you're doing.

So there's a very simple example that I like to use, which is generating a Vandermonde matrix. So giving a vector a value as alpha 1 alpha 2 to alpha n. And you want to make an n by m matrix whose columns are just those entries to 0 with power, first power squared, cubed, and so forth element-wise. All right, so this kind of matrix shows up in a lot of problems. So most matrix and vector libraries have a built in function to do this and Python.

In NumPy, there is a function called numpy.vander to do this. And if you look at-- it's generating a big matrix. It could be performance critical. So they can implement it in Python. So if you look at the NumPy implementation, it's a little Python shim that calls immediately to C. And then if you look at the C code-- I won't scroll through it-- but it's several hundred lines of code. It's quite long and complicated.

And all that several hundred lines of code is doing is just figuring out what types to work with, like what kernels to dispatch to. And at the end of that, it dispatches to a kernel that does the actual work. And that kernel is also C code, but that C code was generated by a special purpose code generation. So it's quite involved to get good performance for this while still being somewhat type generic.

So their goal is to have something that works for basically any NumPy array and any NumPy type, which there's a handful, like maybe a dozen scalar types that it should work with, all right? So if you're implementing this in C, it's really trivial to write 20 lines of code that implements this but only for double precision, a point or two double position array, all right? So the difficulty is getting type generic in C.

So in Julia. Here is the implementation in Julia. It looks at first glance much like what roughly what a C or Fourier implementation would look like. It's just implemented the most simple way. It's just two nested loops. So just basically, you loop across. And as you go across, you accumulate powers by multiplying repeatedly by x. That's all it is. And it just fills in the array.

The performance of that graph here is the time for the NumPy implementation divided by the time for the Julie implementation as a function of n for an n by n matrix. The first data point, I think there's something funny going on that's not 10,000 times slower. But for a 10 by 10, 20 by 20 array, the NumPy version is actually 10 times slower because it's basically the overhead that's imposed by all going through all those layers.

Once you get to 100 by 100 matrix, the overhead doesn't matter. And then it's all this optimized C code, generation and so forth is pretty much the same speed as the Julia code. Except the Julia code there, as I said, it looks much like C code would, except there's no types. It's Vander x. There's no type declaration. x can be anything.

And, in fact, this works with any container type as long as it has an indexing operation. And any numeric type-- it could be real numbers. It could be complex numbers. It could be quarternians, anything that supports the times operation. And there's also a call to 1. So 1 returns the multiplicative identity for whatever, so whatever group you're in you need to have a 1, right? That's the first column.

That might be a different type of 1 for a different object, right? It might be an array of matrices, for example. And then the 1 is the identity matrix. So, in fact. There are even cases where you can do. Get significantly faster than optimize C and Fortran codes. So I found this when I was implementing special functions, so things like the error function, or polygamma function, or the inverse of the error function.

I've consistently found that I can get often two to three times faster than optimized C and Fortran libraries out there, partly because I'm smarter than people who wrote those libraries, but-- no-- mainly because in Julia, I'm using basically the same expansions, the same series, rational functions that everyone else is using. The difference is then in Julia, it has built-in techniques for what's called metaprogramming or co-generation.

So usually, the special functions involved lots of polynomial evaluations. That's what they boil down to. And you can basically write co-generation that generates very optimized inline evaluation of the specific polynomials for these functions that would be really awkward to write in Fortran. You'd either have to write it all by hand or write a separate routine, a separate program that wrote Fortran code for you. So you can do this. It's a high level languages allow you to do tricks for performance that it would be really hard to do in low level languages.

So mainly what I wanted to talk about is give some idea of why Julia can be fast. And to understand this, you also need to understand why is Python slow. And in general, what's going on in determining the performance in a language like this? What do you need in the language to enable you to compile it to fast code while still, still being completely generic like this Vander function, which works on any type. Even user-defined, numeric type, user-defined container type will be just as fast. There's no privileged-- in fact, if you look at Julia, almost all of Julia is implemented in Julia.

Integer operations and things like that, the really basic types, most of that is implemented in Julia, right? Obviously, if you're multiplying two 32-bit integers. At some point, it's calling an assembly language instruction. But even that, calling out to the assembly is actually in Julia. So at this point, I want to switch over to sort of a live calculation.

So this is from a notebook that I developed as part of a short course with Alan Edelman, who's sitting over there, [INAUDIBLE] on performance optimization high level languages. And so I want to go through just a very simple calculation. Of course, you would never-- in any language, usually you would have a built-- often have a built-in function for this. But it's just a sum function just written up there.

So we need to have a list, an array, of n numbers. We're just going to add them up. And if we can't make this fast, then we have real problems. And we're not going to be able to do anything in this list. So this is the simple sort of thing where if someone doesn't provide this for you, you're going to have to write a loop to do this.

So I'm going to look at it not just in Julia but also in Python, in C, and Python with NumPy and so forth. So this document that I'm showing you here is a Jupyter Notebook. Some of you may have seen this kind of thing. So Jupyter is this really nice-- they provide this really nice browser-based front end when I can put in equations, and text, and code, and results, and graphs all in one Mathematical notebook document.

And you can plug in different languages. So initially, it was for Python. But we plugged in Julia. And now there's R, and there's 30 different-- like 100 different languages that you can plug in to the same front end. OK, so I'll start with the C implementation of this.

So this is a Julia notebook, but I can easily compile and call out to C. So I just made a string that has just-- you know there's 10 lines C implementation. It's just the most obvious function that just takes in a pointer to an array of doubles and it's length. And it just loops over them and sums them up, just what you would do.

And then I'll compile it with GCC dash 03 and link it to a shared library, and load that shared library in Julia and just call it. So there's a function called C call in Julia where I can just call out to a C library with a 0 overhead, basically. So it's nice because you have lots of existing C libraries out there. You don't want to lose them. So I just say C call, and we call this c_sum function in my library. It returns a flow 64. It takes two parameters, a size t and a flow 64.

And I'm going to pass the length of my array. And the array-- and it'll automatically-- a Julia array, of course, is just a bunch of numbers. And it'll pass a pointer to that under the hood. So do that. And I wrote a little function to call relerr that computes the relative error between the fractional difference between x and y.

And I'll just check it. I'll just generate 10 to the 7 random numbers in 01 and compare that to the Julia because Julia has a built-in function called sum, that sums and array. And it's giving the same answer to 13 decimal places, so not quite machine precision, but there's 10 to the 7 numbers. So the error is kind of accumulative when you add it across.

OK so, as I'm calling it, it's giving the right answer. And now I want to just benchmark the C implementation, use that as kind of a baseline for this. This should be pretty fast for an array of floating point values. So there's a Julia package called benchmark tools. As you probably know from this class, benchmarking is a little bit tricky. So this will take something, run it lots of times, collect some statistics, return the minimum time, or you can also get the variance and other things.

So I'm going to get that number. B time is something called a macro in Julia. So it takes an expression, rewrites it into something that basically has a loop, and times it, and does all that stuff. OK, so it takes 11 milliseconds to sum 10 to the 7 numbers with a straight C, C loop compiled with jcc-03, no special tricks, OK? And so that's 1 gigaflop, basically, billion operations per second. So it's not hitting the peak rate of the CPU.

But, of course, there's additional calculations. This array doesn't fit in cache, and so forth. OK. So now let's-- before I do anything in Julia, let's do some Python. But I'll do a trick. I can call Python from Julia, so that way I can just do everything from one notebook using a package I wrote called PyCall. And PyCall just calls directly out to lib Python.

So with no virtually no overhead, so it's just like calling Python from within Python. I'm calling directly out to lib Python functions to call. And I can pass any type I want, and call any function, and do conversions back and forth. OK, so I'm going to take that array. I'll convert it to a Python list object. So I don't want to time the overhead of converting my array to a Python array. So I'll just convert ahead of time.

And just start with a built-- Python has a built-in function called sum. So I'll use the built-in sum function. And I'll get this Py object for it. I'll call it PySum on this list and make sure it's giving the right answer OK is the difference is 10 to the minus 13 again. And now let's benchmark it. Oops. There we go.

So it takes a few seconds because it has to run it a few times and catch up with statistics. OK, so it takes 40 milliseconds. That's not bad. It's actually it's four or five times slower than C, but it's pretty good, OK? So and why is it five times slower than C? Is it is it because-- the glib answer is, oh, well, Python is slow because it's interpreted.

But the sum function is actually written in C. Here's the C implementation of the sum function that I'm calling. And I'm just linking to the GitHub code. There's a whole bunch of boilerplate that just checks with the type of the object, and then has some loops and so forth. And so if you look carefully, it turns out it's actually doing really well. And the reason it does really well is it has a fast path. If you have a list where everything is a number type, so then it has an optimized implementation for that case. But it's still five times slower than C.

And they've spent a lot of work on it. It used to be 10 times slower than C a couple of years ago. So they do a lot of work on optimizing this. And so why aren't they able to get C speed? Since they have a C implementation of a sum, are they just dumb, you know? No. It's because the semantics of the data type prevent them from getting anything faster than that.

And this is one of the things you learn when you do high level performance in high level languages. You have to think about data types, and you have to think about what the semantics are. And that that greatly constrains what any conceivable compiler can do. And if the language doesn't provide you with the ability to express the semantics you want, then you're dead. And that's one of the basic things that Julia does. So what does a Python list?

Right, so you have-- you can have three, four, right? A Python list is a bunch of objects, Python objects. But the Python numbers can be anything. They can be any type. So it's completely heterogeneous types. So, of course, a particular list like in this case can be homogeneous. But the data structure has to be heterogeneous because, in fact, I can take that homogeneous thing. At any point, I can assign the third element to a string, right? And it has to support that.

So think about what that means for how it has to be implemented in memory. So what this has to do-- so this is a list of-- in this case, three items. But what are those items? So if they can be an item of any type, they could be things that-- they could be another array, right? It could be of different sizes and so forth. You don't want to have an array where everything is a different size, first of all. So it has to be this an array of pointers where the first pointer is 3, turned out to be 3. The next one is four. The next one is 2, all right?

But it can't just be that. It can't just be pointer to-- if this is a you 64-bit number, it's can't just be pointer to one 64-bit value in memory, because it has to know. It has to somehow store what type the subject is. So there has to be a type tag this says this is an integer. And this one has to have a type tag that says it's a string. So this is sometimes called the box.

So you have a value, you have a type tag plus a value. And so imagine what even the most optimized C imitation has to do given this kind of data structure. OK, here's the first element. It has to chase the pointer and then ask what type of object is it, OK? Then depending on what type of object is it, so I initialize my sum to that. Then I read the next object. I have to chase the second pointer, read the type tag, figure out the type. This is all done at run time, right?

And then, oh, this is another integer that tells me I want to use the plus function for two integers, OK? And then I read the next value, which maybe-- which plus function it's using depends upon the type of the object. It's an to object-oriented language. I can define my own type. If it has its own plus function, it should work with sum.

So it's looking up the types of the objects at runtime. It's looking at the plus function at runtime. And not only that, but each time it does a loop iteration it has to add two things and allocate a result. That result in general might be another-- it has to be a box because the type might change as you're summing through. If you start with integers, and then you get a floating point value, and then you get an array, the type will change. So each time you do a loop iteration, it allocates another box.

So what happens is the C implementation is a fast path. If they're all integer types, I think it doesn't reallocate that box for the sum it's accumulating all the time. And it caches the value of the plus function it's using. So it's a little bit faster. But still, it has to inspect every type tag and chase all these pointers for every element of the array, whereas the C implementation of sum, if you imagine what this is this compiles down to, for each loop iteration, what does it do?

It increments a pointer to the next element. At compile time, the types are all flow 64. So it flushes flow 64 value into a register. And then it has a running sum in another register, calls one machine instruction to add that running sum, and then it checks to see if we're done, an if statement there, and then goes on, all right? So just a few instructions and in a very tight loop here, whereas each loop iteration here has to be lots and lots of instructions to chase all these pointers to get the type tag-- and that's in the fast case where they're all the same type and it's optimized for that.

So where was I? So wrong thing. So that's the Python sum function. Now most many of you, you've used Python know that there is another type of array and there's a whole library called NumPy for working with numerics. So what problem is that addressing? So the basic problem is this data structure. This data structure, as soon as you have a list of items-- it can be any type-- you're dead, right? There's no way to make this as fast as a C loop over a double pointer.

So to make it fast, what you need to have is a way to say, oh, every element of this array is the same type. So I don't need to store type tags for every element. I can store a type tag once for the whole array. So there, there is a tag. There is a type, which is, say, float 64, OK? There is maybe a length of the array. And then there's just a bunch of values one after the other.

So this is just 1.0, 3.7, 8.9. And each of these are just 8 bytes, an 8-byte double in C notation, right? So it's just one after the other. So it reads this once, reads the length, and then it dispatches to code that says, OK, now-- basically dispatches to the equivalent of my C code, all right? So now once it knows the type and the length, then OK, it says it runs this, OK?

And that can be quite fast. And the only problem is you cannot write, implement this in Python. So Python doesn't provide you a way to have that semantics to have a list of objects where you say they all have to be the same type. There is no way to enforce that, or to inform the language of that in Python. And then to tell it-- oh, for this, since these are all the same type, you can throw away the boxes.

Every Python object looks like this. So there's no way to tell Python. Oh, well, these are all the same types. You don't need to store the type tags. You don't need to have pointers. You don't need to have reference counting. You can just slam the values into memory one after the other. It doesn't provide you with that facility, and so there's no way to-- you can make a fast Python compiler that will do this.

So NumPy is implemented in C. Even with-- some of you are familiar with PyPy, which is an attempt to make a fast-- like a tracing jit for Python. So when they poured into NumPy to PyPy, or they attempted to, even then they could implement more of it in Python. But they had to implement the core in C.

OK, but given that, I can do this. I can import the NumPy module into Julia, get its sum function, and benchmark the NumPy sum function, and-- OK, again, it takes a few seconds to run. OK, and it takes 3.8 milliseconds. So the C was 10 milliseconds. So it's actually doing faster than the C code, almost a little over twice as fast, actually.

And what's going on is their C code is better than my C code. Their C code is using SIMD instructions. So and at this point, I'm sure that you guys all know about these things where you can read in two numbers or four numbers into one giant register and one instruction add all four numbers at once.

OK, so what about if we go in the other direction? We write our own Python sum function. So we don't use the Python sum implemented in C. I write our own Python. So here is a little my sum function in Python. Only works for floating point. Oh, I initialize S to 0.0. So really, it only accumulates floating point values. But that's OK.

And then I just loop for x and a, s equals s plus x, return s is the most obvious thing you would write in Python. OK, and checked that it works. Yeah, errors 10 the minus 13th. It's giving the right answer. And now let's time it.

So remember that C was 10 milliseconds. NumPy was 5 milliseconds, and then built-in Python was like, the sum was 50 milliseconds operating on this list. So now we have C code operating on this list with 50 milliseconds. And now we have Python code operating on this list. And that is 230 milliseconds, all right? So it's quite a bit slower.

And it's because, basically, in Python, there's-- in the pure Python code, there's no way to implement this fast path that checks-- oh they're all the same type, so I can cash the plus function and so forth. I don't think it's feasible to implement that. And so basically then on every loop iteration has to look up the plus function dynamically and allocate a new box for the result and do that 10 to 7 times.

Now, so there's a built-in sum function in Julia. So [INAUDIBLE] benchmark that as a-- it's actually implemented in Julia. It's not implemented in C. I won't show you the code for the built-in one because it's a little messy because it's actually computing the sum more accurately than the loop that have done. So that's 3.9 milliseconds. So it's comparable to the NumPy code, OK? So it's also using SIMD, but so this is also fast.

So now so why can Julie do that? So it has to be that the array type, first of all, has the type attached to it. So you can see the type of the array is an array of low 64. So there's a type tag attached to the array itself. So somehow, that's involved. So it looks more like an NumPy ray in memory than a Python list.

You can make the equivalent of a Python list that's called an array of any. So if I convert this to an array of any, so an array of any is something where the elements types can be any Julia type. And so then it has to be stored as something like this as an array of pointers to boxes. And when I do that-- let's see-- [INAUDIBLE] there it is-- then it's 355 milliseconds. So it's actually even worse than Python.

So the Julia-- the Python, they spent a lot of time optimizing their code past for things that had to allocate lots of boxes all the time. So in Julia, it's usually understood that if you're writing optimized code you're going to do it not on arrays of pointers to boxes. You're going to write on homogeneous arrays or things where the types are known at compile time. OK, so let's write our own Julia sum function. So this is a a Julia sum function. There's no type declaration. It works on any container type.

I initialize s for this function called 0 for the element type of a. So it initializes into the additive identity. So it will work on any container of anything that supports a plus function that has an additive identity. So it's completely generic. It looks a lot like the Python code except for there's no 0 function in Python. And let's make sure it gives the right answer. It does.

And let's benchmark it. So this is the code you'd like to be able to right. You'd like to be able to write high level code that's a straight level straight loop. Unlike the C code, it's completely generic, right? It works on any container, anything you can loop over, and anything that has a plus function, so an array of quarternians or whatever. And a benchmark, it's 11 milliseconds. It's the same as the C code I wrote in the beginning.

It's not using SIMD. So the instructions are-- that's where the additional factors of 2 come. But it's the same as the non SIMD C code. And, in fact, if I want to use SIMD, there is a little tag you can put on a loop to tell-- it says compile this with llvm. Tell llvm to try and vectorize the loop. Sometimes it can. Sometimes it can't. But something like this, it should be able to vectorize it simple enough. You don't need to hand code SIMD instructions for a loop this simple. Yeah?

AUDIENCE: Why don't you always put the [INAUDIBLE]?

STEVEN JOHNSON: So a yeah, why isn't it the default? Because most code, the compiler cannot autovectorize. So it increases the completion time and often blows to the code size for no benefit. So it's only really-- it's really only relatively simple loops on doing simple operations and arrays that benefit from SIMD. So you don't want it to be there.

Yeah, so now it's 4.3 milliseconds. So it's about the same as the NumPy and so forth. It's a little slower than the NumPy. It's interesting. A year ago when I tried this, it was almost exactly the same speed as the NumPy. And then since then, both the NumPy and the Julia have gotten better. But the NumPy got better more. So there's something going on with basically how well the compiler can use AVX instructions by-- it seems like we're still investigating what that is. But it's an llvm limitation looks like.

So as it's still completely type generic. So it I make a random array of complex numbers, and then I sum them-- which one am I calling now? Am I calling the vector? My sum is the vectorized one, right? So complex numbers-- each complex number is two floating point numbers. So it should take about twice the time to naively think, right? So twice the number of operations to add, the same number of complex numbers, the same number of real numbers.

Yeah, and it does. It takes about 11 milliseconds. So 10 to the 7, which is about twice the 5 milliseconds it took for the same number of real numbers. And the code works for everything. So why? OK, so what's going on here? So-- OK, so we saw this my sum function. I'll just take out the SIMD for now. And we did all that. OK. OK, and it works for any type. It doesn't even have to be an array.

So, for example, there is another container type called a set in Julia, which is just an unordered collection of unique elements. But you can also loop over it. If it's a set of integers, you can also summit. And I'm waiting for the benchmarks to complete. So let me allocate a set. It says there's no type declarations here. Mysum a-- there was no a-- has to be an array of particular type or it doesn't even have to be an array, right? So set is a different data structure. And so it's a set of integers.

The sets are unique. If I add something to already the set-- it's in there-- it won't add it twice. And it supports the fast checking. Is 2 in the set? Is 3 in the set? It doesn't have to look through all the elements. It's a hash internally. But I can call my mysum function on it, and it sums up 2 plus 17 plus 6.24, which is hopefully 49, right? So OK, so what's going on here?

So suppose you define-- the key, one of the keys, there's several things that are going on in to make Julia fast. One key thing is that when you have a function like this mysum, or even here's a simpler function-- f of x equals x plus 1-- when I call it with a particular type of argument, like an integer, or an array of integers, or whatever, then it compiles a specialized version of that for that type.

So here's f of x equals x plus 1. It works on any type supporting plus. So if I call f of 3-- so here I'm passing a 64-bit integer. When it did that, it says, OK, x is a 64-bit integer. I'm going to compile a specialized version of f with that knowledge. And then when I call with a different type, 3.1, now x is a floating point number. It will compile a specialized version with that value.

If I call it with another integer, it says, oh, that version was already compiled. [INAUDIBLE] I'll re-use it. So it only compiles it the first time you call it with a particular type. If I call it with a string, it'll give an error because it doesn't know how to add plus to a string. So it's a particular-- OK, so what is going on?

So we can actually look at the compiled code. So the function is called code, these macros called code llvm and code native. They say, OK, when I call f of 1, what's the-- do people know what llvm is? You guys-- OK, so llvm compiles to byte code first and then it goes to machine codes. So you can see the llvm bit code or byte code, or whatever it's called. And you can see the native machine code. So here's the llvm byte code that it compiles to [INAUDIBLE].

So it's a bit called add i640 basically one llvm instruction, which turns into one machine instruction, load effective address is actually a 64-bit edition function instruction. So let's think about what had to happen there. So you have f of x equals x plus 1.

Now you want to compile for x is a int 64, so 64-bit integer, or in Julia, we'd say x colon colon int 64. So colon colon means that this is of this type, OK? So this is 64-bit integer type. So what does it have to do? It first has to figure out what plus function to call. So plus, it has lots of-- there is a plus for two matrices, a plus for lots of different things. So depending on the types of the arguments, it decides on which plus function to call.

So it first realizes, oh, this is an integer. Oh, this is an integer. This is also a 64-bit integer. So that means I'm going to call the plus function for two integers. So I'm going to look into that function. And then, oh, that one returns an int 64. So that's a return value for my function. And oh, by the way, this function is so simple that I'm going to inline it. So it's type specializing.

And this process of going from x is an integer to that, to figure out the type of the output, is called type inference, OK? So. in general, for type inference, it is given the types of the inputs, it tries to infer the types of the outputs and, in fact, all intermediate values as well.

Now what makes it a dynamic language is this can fail. So in some languages like ml or some other languages, you don't really declare types. But they're designed so they could give the types the inputs. So you can figure out everything. And if it can't figure out everything, it gives an error, basically. It has to infer everything. So Julia is a dynamic language. This can fail and have a fallback. If it doesn't know the type, it can stick things in a box.

But obviously, the fast path is when it succeeds. And one of the key things is you have to try and make this kind of thing work in a language. You have to design the language so that at least for all the built-in constructs, the standard library, in general, in the culture for people designing packages and so forth, to design things so that this type inference can succeed.

And I'll give a counter example to that in a minute, right? So and this works recursively. So it's not suppose I define a function g of x equals f of x times 2, OK? And then I called g of 1. So it's going to say, OK, x here is an integer, OK? I'm going to call f with an integer argument. Oh, I should compile f for an integer argument, figure out its return type, use its returned type to figure out what time's function to call and do all of this at compile time, right? Not at runtime, ideally.

So we can look at the llvm code for this. And, in fact, so remember, f of x adds 1 to x. And then we're multiplying by 2. So the result computes 2x plus 2. And llvm is smart enough. So f is so simple that it inlines it. And then llvm is smart enough that I don't have to-- well, I know it does it by one shift instruction to multiply x by 2. And then it adds 2. So it actually combines the times 2 and the plus 1. So it does constant folding, OK?

And it can continue on. If you look at h of x equals g of x times 2, then that compiles to 1 shift instruction to multiply x by 4 and then adding 4. So you want the-- so this process cascades. So you can even do it for recursive function. So here's a stupid implementation of the Fibonacci number and calculation of recursive limitation, right? So this is given n. It's an integer.

OK, if n is less than 3, returns 1. Otherwise, it adds the previous two numbers. I can compute the first call, listen to the first 10 integers. Here's the first 10 Fibonacci numbers. There's also an acute notation in Julia. You can say fib dot. So if you do f dot arguments, it calls the function element Y's on a collection and returns a collection. So F dot 1 to 10 returns the first 10 Fibonacci numbers.

And I can call-- there's a function called code 1 type that'll tell me what-- it'll tell me the output of type inference. N is a 1. And it goes through-- this is kind of a hard to read format. But this is like the-- after one of the compiler passes called lowering, but yeah. It's figure out the types of every intermediate calls.

So here it's invoking main dot fib. It's recursively. And it's figured out that the return type is also int 64. So it knows everything, OK? So you'll notice that here I declared a type. I've said that this is an integer, OK? I don't have to do that for type inference. This doesn't help the compiler at all because it does type inference depending on what I pass. So what this is is more like a filter.

It says that if I pass-- this function only accepts integers. If you pass something else, you should throw an error. Because if I don't want this function, because if I pass 3.7, if I fill out any number, if you look at the 3.7, I can check whether it's less than three. You can call it recursively. I mean, the function would run. It would just give nonsense, right? So I want to prevent someone from passing nonsense for this.

So that's one reason to do a type declaration. But another reason is to do something called dispatch. So what we can do is we can define different versions of the function for different arguments. So. for example. another nicer version of that is a factorial function. So here is a stupid recursive implementation of a factorial function that takes an integer argument and just recursively calls itself an n minus 1. You can call it a 10 factorial, OK?

If I want 100 factorial, I need to use a different type, not 64-bit integers. I need some arbitrary precision integer. And since I said it was an integer, if I call in 3.7, it'll given an error. So that's good. But now I can find a different version of this. So actually, there is a generalization of factorial to arbitrary real, in fact, even complex numbers called a gamma function. And so I can define a fallback that works for any type of number that calls a gamma function from someplace else.

And then if I can pass it to floating point value, I can-- if you take the factorial minus 1/2, it turns out that's actually a square root of pi. So if I square it, it gives pi, all right? So now I have one function and I have two methods, all right? So these types here, so there's a hierarchy of types. So this is what's called an abstract type, which most of you have probably seen.

So there's a type called number. And underneath, there's a class of subtypes called integer. And underneath, there is, for example, int 64 or int 8 for 64-bit integers. And underneath number there's actually another subtype called real. And underneath that there's a couple of subtypes. And then there's say flow 64 or float 32 for a single precision 32-bit floating point number and so forth. So there's a hierarchy of these things.

When I specify something can take integer I'm just saying so this type is not help the compiler. It's to provide a filter. So this method only works for these types. And this other method only works-- my second method works for any number type. So I have one thing that works for any number. Whoops. Here it is-- One that works for any number type, and one method that only works for integers. So when I call it for 3, which ones does it call, because it actually called both methods.

And what it does is it calls the most specific one. It calls the one that sort of farthest down the tree. So if I have a method defined for number and one defined for integer, if I pass an integer, it'll do this. If I have one that's defined for number, one that's defined by integer, and one that's defined specifically for int 8, and I call it a --pass an 8 bit integer, it'll call that version, all right? So it gives you a filter.

But, in general, you can do this on multiple arguments. So this is like the key abstraction in Julia, something called multiple dispatch. So this was not invented by Julia. I guess it was present in Small Talk, and Dylan. It's been in a bunch of languages. It's been floating around for a while. But it's not been in a lot of mainstream languages, not in a high performance way. And you can think of it as a generalization of advertising and programming.

So I'm sure all of you have done object oriented programming in Python or C++ or something like this. So in object-oriented programming typically the way you think of it is this is you save an object. It's usually spelled object dot method xy for example, right? And what it does, is this type, the object type determines the method, right?

So you can have a method called plus. But it would actually call a different class function for a complex number or a real number, something like that, or a method called length, which for a Python list would call a different function than for an NumPy array, OK? In Julia, the way you would spell the same thing would you'd say method. And you wouldn't say object dot method.

So you don't think of the-- here, you think of the object as sort of owning the method. all right? And Julia-- the object would just be maybe the first argument. In fact, under the hood, if you looking in Python, for example, the object is passed as an implicit first argument called self, all right? So it actually is doing this. It's just different spelling of the same thing. But as soon as you read it this way, you realized what Python and what op languages are doing is they're looking at the type of the first argument to determine the method. But why just the first argument?

In a multiple dispatch language, you look at all the types. So this is sometimes-- in Julia, this will sometimes be called single dispatch because determining the method is called dispatch, figuring out which function is spelled length, which function actually are you calling this dispatching to the right function. So this is called multiple dispatch.

And it's clearest if you look at something like a plus function. So a plus function, if you do a plus b, which plus you do really should depend on both a and b, right? It shouldn't depend on just a or just b. And so it's actually quite awkward in languages, in o op languages like Python or especially C++ to overload a plus operation that operates on sort of mixed types. As a consequence, for example, in C++ there's a built-in complex number type.

So you can have a complex float, or complex double, complex and complex with different real types. But you can't add a complex float to a complex double. You can't add a single-precision complex number to a double-precision complex number or any mixed operation because-- any mixed complex operation because it can't figure out who owns the method. It doesn't have a way of doing that kind of promotion, all right?

So in Julia, so now you can have a method for adding a float 32 to a float 32, but also a method for adding a-- I don't know-- let's see, adding a complex number to a real number, for example. You want to specialize-- or a real number to a complex number. You want to specialize things. In fact, we can click on the link here and see the code.

So the complex number to a real number in Julia looks like this. It's the most obvious thing. It's implemented in Julia. Plus complex real creates new complex number. But you only have to add the real part. You can leave the imaginary part alone. And this works on any complex type. OK, so there's too many methods for-- OK, I can shrink that. Let's see.

So but there's another type inference thing called-- I'll just mention it briefly. So one of the things you've to do to make this type inference process work is given the types of the arguments you have to figure out the type of the return value, OK? So that means when you assign a function, it has to be what's called type stable. The type of the result should depend on the types of the arguments and not on the values of the arguments because the types are known at compile time. The values are only known at runtime.

And it turns out if you don't have this in mind, in C, you have no choice but to obey this. But in something like Python and dynamic language, like Python and Matlab, if you're not thinking about this, it's really easy to design things so that it doesn't work, so that it's not true. So a classic example is a square root function. All right, so suppose I pass an integer to it, OK? So the square root of-- let's do square root of 5, all right? The result has to be floating point number, right? It's 2.23-- whatever.

So if I do square root of 4, of course, that square root is an integer. But if I return an integer for that type, then it wouldn't be type stable anymore. Than the return value type would depend on the value of the input, whether it's a perfect square or not, all right? So it was returned to floating point value, even if the input is an integer. Yes?

AUDIENCE: If you have enough methods to find for a bunch of different types of lookup function? Can that become really slow?

STEVEN JOHNSON: Well, so the lookup comes-- it comes at compile time. So it's really kind of irrelevant. At least if type inference succeeds, if type inference fails, then it's runtime, it's slower. But it's not like-- it's like a tree surge. So it's not-- it's not as slow as you might think. But most of the time you don't worry about that because if you care about performance you want to arrange your code so that type inference succeeds.

So you prototype maybe to-- this is something that you do in performance optimization. Like when you're prototyping, you don't care about types, you say x equals 3. And the next line you say x equals an array-- whatever. But when you're optimizing your code, then, OK, you tweak it a little bit to make sure that things don't change types willy nilly and that the types of function depend on the types of the arguments not on the values.

So, as I mentioned, square root is what really confuses people at first, is if you take square root of minus 1, you might think you should get a complex value. And instead, it gives you an error, right? And basically, what are the choices here? It could give you an error. It could give you a complex value. But if it gave you a complex value, then the return type of square root would depend upon the value of the input, not just the type.

So Matlab, for example, if you take square root of minus 1, it will happily give you a complex number. But as a result, if you have Matlab, Matlab has a compiler, right? But it has many, many challenges/ but one simple thing to understand is if the Matlab compiler suite sees a square root function, anywhere in your function, even if it knows the inputs that are real, it doesn't know if the outputs are complex or real unless it can prove that the inputs were positive or non-negative, right?

And that means it could then compile two code pass for the output, all right? But then suppose it calls square root again or some other function, right? You quickly get a combinatorial explosion of possible code paths because of possible types. And so at some point, you just give up and put things in a box. But as soon as you put things in a box, and you're looking up types at runtime, you're dead from a performance perspective.

So Python, actually-- s if you want a complex result from square root, you have to give it a complex input. So in Julia, a complex number, the I is actually m. They decide I is too useful for loop variables. So I and J. So m is the complex unit. And if you take square root of a complex input, it gives you a complex output.

So Python actually does the same thing. So if in Python if it takes square root of a negative value, it gives an error unless you give it a complex input. But Python made other mistakes. So, for example, in Python, an integer is guaranteed never to overflow. If you add 11 plus 1 plus 1 over and over again in Python, eventually overflow the size of a 64-bit integer, and Python will just switch under the hood to an arbitrary position in an integer, which seem like a nice idea probably at the time.

And the rest in Python is so slow that the performance cost of this test makes no difference in a typical Python code. But it makes it very difficult to compile. Because that means if you have integer inputs and you see x plus 1 in Python, the compiler cannot just compile that to one instruction, because unless you can somehow prove that x is sufficiently small.

So in Julia, integer arithmetic will overflow. But the default integer arithmetic it's 64 bits. So in practice that never overflows unless you're doing number theory. And you usually know if you're doing number theory, and then use arbitrary precision integers. It was much worse in the days-- this is something people worried a lot before you were born when there were 16-bit machines, right? And integers, it's really, really easy to overflow 16 bits because the biggest sine value is then 32,767, right?

So it's really easy to overflow it. So you're constantly worrying about overflow. And even 32 bits, the biggest sign value is 2 billion. It's really easy to overflow that, even just counting bytes, right? You can have files that are bigger than 2 gigabytes easily nowadays. So some people worried about this all the time. There were 64-bit integers.

Basically, 64-bit integer will never overflow if it's counting objects that exist in the real universe, like bytes, or loop iterations, or something like that. So you just-- then you just say, OK, it's either 64 bits, or you're doing number theory. You should big ints. So OK. So let me talk about-- the final thing I want to talk about-- let's see, how much [INAUDIBLE] good-- is defining our own types.

So this is the real test of the language, right? So it's easy to make a language where there is a certain built-in set of functions and built-in types, and those things are fast. So, for example, for Python, there actually is a compiler called numba that does exactly what Julia does. It looks at the arguments, type specializes things, and then calls llvm and compiles it to fast code.

But it only works if you're only container type as an NumPy array and you're only scalar type is one of the 12 scalar types that NumPy supports. If you have your own user defined number type or your own user defined container type, then it doesn't work. And user-defined container types, it's probably easy to understand why that's useful. User defined number types are extremely useful as well.

So, for example, there's a package in Julia that provides the number type called dual numbers. And those have the property that if you pass them into the function they compute the function in its derivative. And just a slightly different. It basically carries around function an derivative values and has a slightly different plus and times and so forth that just do the product rule and so forth. And it just propagates derivatives. And then if you have Julia code, like that Vandermonde function, it will just compute its derivative as well.

OK, so I want to be able to find my own type. So a very simple type that I might want to add would be points 2D vectors in two space, right? So, of course, I could have an array of two values. But an array is a really heavyweight object for just two values, right? If I know at compile time there's two values that I don't need to have a pointer to-- I can actually store them in registers. I I can unroll the loop over these and everything should be faster.

You can get an order of magnitude and speed by specializing on the number of elements for small arrays compared to just a general ray data structure. So let's make a point, OK? So this is-- and I'm going to go through several iterations, starting with a slow iteration. I'm going to define a mutable struct. OK, so this will be a mutable object where I can add a [INAUDIBLE] point that has two values x and y. It can be of any type.

I'll define a plus function that can add them. And it does the most obvious thing. It adds the x components, adds the y components. I'll define a 0 function that's the additive identity that just returns the point 0, 0. And then I can construct an object Point34. I can say 034 plus 0.56. It works. It can hold actually-- right now it's very generic, and probably too generic.

So they act like the real part can be a floating point number and the imaginary. The x can be a floating point number here and the y is a complex number of two integers, or even I can make a string and an array. It doesn't even make sense. So I probably should have restricted the types of x and y a little bit just to prevent the user from putting in something that makes no sense at all, OK?

So these things, they can be anything. So this type is not ideal in several ways. So let's think about how this has to be stored in memory. So this is a 0.1. 0.11, 3.7, right? So in memory it's-- there is an x and there is a y. But x and y can be of any type. So that means they have to be pointers to boxes. There's pointer to int 1 and there's a pointer to a float 64, in this case, 3.7.

So oh, this already, we know this is not going to be good news for performance. And it's mutable. So that mutable struct means if I take p equals a point, I can then say p dot x equals 7. I can change the value, which seems like a harmless thing to do, but actually is a big problem because, for example, if I make an array of a three piece, and then I say p dot y equals 8, and I look at that array, it has to change the y component, OK?

So if I have a p, or in general, if I have a p that's looking at that object, if this is an object you can mutate, it means that if I have another element, a q, it's also pointing the same object. And I say p, p dot x equals 4, then q dot x had better also before at that point. So to have mutable semantics, to have the semantics of something you can change, and other references can see that change, that means that this object has to be stored in memory on the heap as a pointer to two objects so that you have other pointer to the same object, and I mutate it, and the other references see it.

It can't just be stuck in a register or something like that. It has to be something that other references can see. So this is bad. So if I have, so I can call 0.1 dot aa calls the constructor element-wise. A is this array of 10 to the 7 random numbers. I was benchmarking them before. That was taking 10 milliseconds, OK?

And I can sum it. I can call the built-in some function on this. I can even call my sum function on this because it supports a 0 function. And so it supports a plus. So here, I have an array. If i just go back up, I have an array here of 10 to the 7 values of type 0.1. So the type of 0.1 is attached to the array.

So the array and memory, so I have an array of 0.1, the one here means it's a one-dimensional array. There's also 2D, 3D, and so forth. That looks like a 0.1 value, a 0.1 value, a 0.1 value. But each one of those now-- sorry-- has to be a pointer to an x and a y, which themselves are pointers to boxes. All right, so summing is going to be really slow because there's a lot of pointer chasing. It has to run time, check what's the type of x, what's the type of y.

And, in fact, it was. It took instead of 10 milliseconds, it took 500 or 600 milliseconds. So to do better, we need to do two things. So, first of all, x and y, we have to be able to see what type they are, OK? It can't be just any arbitrary old thing that has to be a pointer to a box, OK? And the point object has to be mutable. It has to be something where if I have p equals something, q equals something, I can't change p and expect q to see it. Otherwise, if it's mutable, those semantics have to be implemented as some pointer to an object someplace. Then you're dead.

So I can just say struct. So struct now is not mutable. It doesn't have the mutable keyword. And I can give the arguments types. I can say they're both flow 64. And x and y are both the same type. They're both 64. But floating point numbers, I'll define plus the same way, 0 the same way, and now I can add them and so forth.

But now if I make an array of these things, and if I say p dot x equals 6, it will give an error. It says you can't actually mutate. Don't even try to mutate it because we can't support those semantics on this type. But that means so that type is actually-- if you look at look at that in memory, what the compiler is allowed to do and what it does do for this is if you have an array of points 0.21, then it looks like just the x value, the y value, The value, the y value, and so forth. But each of these are exactly one 8 byte flow 64.

And all the types are known at compile time. And so if I sum them, it should take about 0.20 milliseconds, right? Because summing real numbers was 10 milliseconds. And this is twice as many because you have to sum the x's, sum the y's. And let's benchmark it. And let's see.

Oh, actually, some of the real numbers took 5 milliseconds. So something should take about 10. Let's see if that's still true. Yeah, it took about 10. So actually, the compiler is smart enough. So, first of all, it stores this in line as one big block, consecutive block of memory.

And then when you sum them, remember our sum function. Well, this is the built-in sum. But our sum function will work in the same way. The compiler-- llvm will be smart enough to say, oh, I can load this into a register, load y into a register, call, have a tight loop where I basically call one instruction to sum the x's, one instruction to sum the y's, and then repeat. And so it's about as good as you could get.

But you paid a big price. We've lost all generality, right? These can only be two 64-bit floating point numbers. I can't have two single-precision numbers or-- this is like a struct of two doubles in C. So if I have to do this to get performance in Julia, then it would suck. Basically, I'm basically writing C code in a slightly higher level syntax. I'm losing that benefit of using a high level language.

So the way you get around this is you define-- what you want is to define something like this 0.2 type. But you don't want to define just one type. You want to define a whole family of types. You want to define this for two float 64s, two float 32s. In fact, you want to define an infinite family of types, at two things of any type you want as long as they're two real numbers, two real types.

And so the way you do that in Julia is a parametrized type. This is called parametric polymorphism. It's similar to what you see in C++ templates. So now I have a struct. It's not mutable-- Point3. But the curly braces t. It says it's parametrized by type t. So x and y-- I've restricted it slightly here. I've said x and y had to be the same type. I didn't have to do that, but I could have had two parameters, one for the type of x, one to the type of y.

But most the time you'd be doing something like this, you'd want them both to be the same type. But they could be both 64s, both float 32s, both integers, whatever. So t is any type that less than colon means is a subtype of. So t is any subtype of real. It could be float 64. It could be int 64. It could be int 8. It could be big float. It could be a user defined type. It doesn't care.

So this is really not-- it's a Point3 here. It is a whole hierarchy. So I'm not defining one type. I'm defining a whole set of types. So Point3 is a set of types. There's a point Point3 int 64. There is a Point3 float 32, a float 64 and so on. Infinitely, many types, as many as you want, and basically, it'll create more types on the fly just by instantiating.

So, for example, otherwise it looks the same. The plus function is basically the same. I add the x components, the y components. The 0 function is the same. Except now I make sure there's zeros of type t, whatever that type is. And now if I say Point34, now I'm instantiating a particular instance of this. And now that particular instance of Point3 we'll have-- this is an abstract type. We'll have one of these concrete types. And the concrete type it has in this case is a Point3 of two int 64s, two 64-bit integers, OK?

And I can add them. And actually, adding mixed types will already work because the plus, the addition function here, it works for any 2.3s. I didn't say there had to Point3s of the same type. Any two of these, they don't have to be two of these. They could be one of these and one of these. And then it determines the type of the result by the type of the [INAUDIBLE] it does type inference.

So if you have a point Point3 of two int 64s and Point3 of two float 32s, it says, oh, p dot x is an int 64. Q dot x is a full 64. Oh, which plus function do I call? There is a plus function from that mixing. And it promotes the result of flow 64. So that means that that sum is flow 64. The other sum is flow 64. Oh, then I'm creating a Point3 of flow 64s. So this kind of mixed promotion is done automatically. You can actually define your own promotion rules in Julia as well. And I can make an array.

And so now if I have an array of Point3 float 64, so this type is attached to the whole array. And this is not an arbitrary Point3. It's a Point3 of two float 64s. So it gets stored again as just 10 to the 7 elements of xy, xy where each one is 8 bytes 8 byes, 8 bytes, one after the other. The compiler knows all the types. And when you submit, it knows everything at compile time. And it will sum to these things. But I loaded this into a register, load this into a register called one instruction-- add them.

And so the sum function should be fast. So we can call the built-in sum function. We can call our own sum function. Our own some function, I didn't put SIMD here, so it's going to be twice as slow. But Yeah. Yeah?

AUDIENCE: Will this work with SIMD?

STEVEN JOHNSON: Yeah. Yeah. In fact, if you look, the built-in sum function, the built-in sum function is implemented in Julia. It just hasn't [INAUDIBLE] SIMD on the sum. So yeah. llvm is smart enough that if you give it a struct of two values and load them-- and if you tell it that you're adding these two values to these two values these two values to these two values, it will actually use SIMD instructions, I think. Oh, maybe not.

No, wait. Did my sum use SIMD? I'm confused. I thought it did.

AUDIENCE: [INAUDIBLE] removed it.

STEVEN JOHNSON: I thought I removed it. Yeah, so maybe it's not smart enough to use SIMD. I've seen in some cases where it's smart enough to use-- huh, yeah. OK, so they're the same speed. OK, no. I take it back. So maybe llvm is not smart enough in this case to use SIMD automatically. We could try putting the SIMD annotation there and try it again. But I thought it was, but maybe not.

Let's see. Let's put SIMD. So redefine that. And then just rerun this. So it'll notice that I've changed the definition. It'll recompile it. But the B time, since it times at multiple times, the first time it calls it, it's slow because it's compiling it. But it takes the minimum over several times. So let's see.

Yeah, this is the problem in general with vectorizing compilers if they're not that smart if you're using anything other than just an array of an elementary data type. Yeah, no. It didn't make any difference. So I took it back. So for more complicated data structures, you often have to use SIMD structure explicitly. And there is a way to do that in Julia. And there is a higher level library on top of that. You can basically credit a tuple and then add things and it will do SIMD acceleration. But yeah.

So anyway, so that's the upside here. There's a whole bunch of-- like the story of why Julia can be compiled with fast code, it's a combination of lots of little things. But there are a few big things. One is that its specialized thing is compile times. But, of course, you could do that in any language. So that relies on designing the language so that you can do type inference.

It relies on having these kind of parametrized types and giving you a way to talk about types and attach types to other types. So the array you notice probably-- let's see-- and now that you understand what these little braces mean, you can see that the array is defined in Julia as another parametrized type. It's parametrized by the type of the element and also by the dimensionality. So it uses the same mechanism to attach types to an array. And you can have your own-- the array type in Julia is implemented mostly in Julia. And there are other packages that implement their own types of arrays that have the same performance.

One of the goals of Julia is to build in as little as possible so that there's not some set of privileged types that the compiler knows about and everything else is second class. It's like user code is just as good as the built-in code. And, in fact, the built-in code is mostly just implemented in Julia. There's a small core that's implemented in C for bootstrapping, basically. Yeah. So having parametrized types, having another technicalities, having all concrete types are final in Julia.

A concrete type is something you can actually store in memory. So Point3864 is something you can actually have, right? An object of two integers is that type. So it is concrete, as opposed to this thing. This is an abstract type. You can't actually have one of these. You can only have one of the instances of the concrete types. So but there are no-- this is final.

It's not possible to have a subtype of this because if you could, then you'd be dead because this is an array of these things. If the compiler has to know it's actually these things and not some subtype of this, all right, whereas in other languages, like Python, you can have subtypes of concrete types. And so then even if you said this is an array of a particular Python type, it wouldn't really know it's that type, or it might be some subtype of that. That's one of the reasons why you can't implement NumPy in Python. This is-- there's no way to say this is really that type and nothing else in the language level. Yeah?

AUDIENCE: Will this compilation in Julia work?

STEVEN JOHNSON: Oh, yeah. So and it's calling llvm. So basically, the stage is you call-- so there's a few passes. OK, so and one of the fun things is you can actually inspect all the passes and almost intercept all of them practically. So, of course, typing code like this, first, it gets parsed, OK? And you can macro those things [INAUDIBLE] actually are functions that are called right after parsing. They can just take the parse code and rewrite it arbitrarily.

So they can extend the language that way. But then it parsed, maybe rewritten by a macro. And then you get an abstract syntax tree. And then when you call it, let's say f of 3, then says, oh, x is an integer. Int 64, it runs a type inference pass. It tries to figure out what's the type of everything, which version of plus to call and so forth.

Then it decides whether to inline some things. And then once it's done all that, it spits out llvm byte code, then calls llvm, and compiles it to machine code. And then it caches that some place for the next time you call, you call f of 4, f with another integer. It doesn't repeat the same processes. Notice it's cached. So that's-- so yeah. At the lowest level, it's just the llvm.

So then there's tons of things I haven't showed you. So I haven't showed you-- I mentioned metaprogramming. So it has this macro facility. So you can basically write syntax that rewrites other syntax, which is really cool for code generation. You can also intercept it after the type inference phase. You can write something called the generated function that basically takes-- because at parse time, it knows how things are spelled. And you can rewrite how they're spelled. But it doesn't know what anything actually means. It does knows x is a symbol. It doesn't know x as an integer-- whatever. It just knows the spelling.

So when you actually compile f of x, at that point, it knows x is an integer. And so you can write something called a generator or a stage function that basically runs at that time and says, oh, you told me x is an integer. Now I'll rewrite the code based on that. And so this is really useful for-- there's some cool facilities for multidimensional arrays. Because the dimensionality of the array is actually part of the type.

So you can say, oh, this is a three-dimensional array. I'll write three loops. Oh, you have a four-dimensional array. I'll write four loops. And it can rewrite the code depending on the dimensionality with code generation. So you can have code that basically generates any number of nested loops depending on the types of the arguments. And all the generation is done in compiled time after type inference. So it knows the dimensionality of the array.

And yeah, so there's lots of fun things like that. Of course, it has parallel facilities. They're not quite as advanced as Cilk at this point, but that's the direction there they're heading. There's no global interpreter lock like in Python. There's no interpreter. So there's a threading facility. And there's a pool of workers. And you can thread a loop. And the garbage collection is threading aware. So that's safe.

And they're gradually having more and more powerful run times, hopefully, eventually hooking into some of Professor Leiserson's advanced threading compiler, taper compiler, or whatever it is. And there's also-- most of what I do in my research is more coarse grained distributed memory parallelism, so running on supercomputers and stuff like that.

And there's MPI. There is a remote procedure call library. There's different flavors of that. But yeah. So any other questions? Yeah?

AUDIENCE: How do you implement the big number type?

STEVEN JOHNSON: The big num type in Julia is actually calling GIMP. So that's one of those things. Let me just-- let me make a new notebook. So if I say I know big int 3, 3,000, and then I'd say that to the, say, factorial. I think there's a built-in factorial of that. All right, so this is called the big num type, right? It's something where the number of digits changes at run time.

So, of course, these are orders of magnitude slower than hardware things. Basically, it has to implement it as a loop of digits in some base. And when you add or multiply, you have to loop over those at runtime. These big num libraries, they are quite large and heavily optimized. And so nobody has reimplemented one in Julia. They're just calling out to a C library called the GNU multi-precision library.

And for floating point values, there is something called big float. So big float of pi is that I can actually-- let's set precision. Big float to 1000. That's 1,000 binary digits. And then say big float of pi. And [INAUDIBLE] more. By the way, you might have-- so I can have a variable alpha-- oops-- alpha hat sub 2 equals 17. That's allowed.

All that's happening here is that Julia allows almost arbitrary unicode things for identifiers. So I can have-- make it bigger so we can have an identifier Koala, right? So there's two issues here. So one is just you have a language that allows those things as identifiers. So Python 3 also allows Unicode identifiers, although I think Julia out of all the existing-- the common languages-- it's probably the widest unicode support.

Most languages only allow a very narrow range of unicode characters for identifiers. So Python would allow the koala, but Python 3 would not allow with alpha hat sub 2 because the numeric subscript unicode characters it doesn't allow. The other thing is how do you type these things. And that's more of an editor thing.

And so in Julia, we implemented initially in the repl and in Jupiter. And now all the editors support, you can just to tab completion of latex. So I can type in gamma, tab, and the tab completes to the unicode character. I can say dot. And it puts a dot over it and backslash superscript 4. And it puts a 4. And that's allowed. So it's quite nice.

So when I'm typing emails, and I put equations in emails, I go to the Julia rappel and tab complete all my LaTeX characters so that I can put equations in emails because It's the easiest way to type these Unicode math characters. But yeah. So IPython borrowed this. So now do the same thing in the IPython notebooks as well.

So it's really fun. Because if you read old math codes, especially old Fortran codes or things, you see lots of variables that are named alpha hat or something like that, alpha hat underscore 3. It's so much nicer to have a variable that's actually the alpha hat sub 3. So that's cute.

CHARLES E. LEISERSON: Steve, thanks very much. Thanks. This was great.

[APPLAUSE]

We are, as Steve mentioned, looking actually at a project to merge the Julia technology with the Cilk technology. And so we're right now in the process of putting together the grant proposal. And if that gets funded, there may be some UROPS.