Playing with Actors ------------------- As I hinted in an earlier post, I have some vague plans for a distributed public MOO-like system based on Carl Hewitt's actor model of computation [1]. Under this model computations proceed by a large number of "actors" (atomic units of computation) asynchronously passing messages to one another. Besides being a useful way to think about concurrency formally, the model has many practical applications. (E.g. Erlang, although that was developed independently.) Today I wanted to write about some of my recent attempts to get to grips with the basics of the actor model by implementing some of its ideas in (can you guess?) Scheme. Before we get started though, I need to include the following warning: I AM NOT A REAL COMPUTER SCIENTIST. I.e. don't believe anything you read here! Okay, now that's out of the way... The basic capabilities of actors are very simple. As stated on the Wikipedia page [2], an actor is a computational entity which, upon receipt of a message, may do any of the following: (1) send messages to other actors whose address it has, (2) create new actors, (3) update its "behaviour", i.e. how the actor will respond to subsequent messages. That's essentially it. There are some interesting side effects to this list. For instance, (1) implies that actor addresses must be "unforgeable" (otherwise actors could "guess" another's address and send it a message). Also, (3) suggests that the processing of a message must be somehow atomic on the scale of the actor. Okay, so how can we go about modeling this in Scheme? (Actually, Scheme and the actor model have some interesting shared history, in that Sussman and Steele actually designed Scheme with the goal of experimenting with this model. [3]) To begin, we need a way of representing an actor. But what is an actor? From the description above, it is simply some kind of "address" associated with a "behaviour", where a "behaviour" describes a procedure for responding to messages. Aha! Scheme has a wonderful way of representing procedures: the lambda expression! So here's our first code snippet: (define actor-table (make-hash-table)) (define (make-actor-with-address address behaviour) (hash-table-set! actor-table address behaviour) address) This defines a global hash table to store the mapping from addresses to behaviours, as well as a procedure which adds a address->behaviour pair to that table. The value returned is the address. Of course, this method for making actors is a bit clumsy - we want each actor to be given some unique address, not one that is predetermined by the creator. (Using predetermined addresses would no doubt lead to violations of rule (1).) Thus we also define (define next-actor-address 1) (define (make-actor behaviour) (make-actor-with-address next-actor-address behaviour) (let ((address next-actor-address)) (set! next-actor-address (+ next-actor-address 1)) address)) This is a fairly silly approach, but with some discipline on the part of the programmer it does ensure that each actor is given a unique address. Okay, so far so good. Now how do we go about sending messages to actors? To start thinking about this, lets define a message dispatch procedure: (define (dispatch-message address message) (let ((behaviour (hash-table-ref/default actor-table address '()))) (if (null? behaviour) (print "Warning: discarded message " message " to unknown actor " address) (let ((value (apply behaviour (cons address message)))) (case value ((sleep) 'do-nothing) ((done) (hash-table-delete! actor-table address)) (else (hash-table-set! actor-table address value))))))) All this does is take an address and a message (some arbitrary scheme object), look the address up in the hash table to find the corresponding behaviour, then apply this behaviour to the both the actor address and the message object. (Sneaking in the actor address allows the actor to send messages to itself.) The only non-obvious part is what is done with the return value of the behaviour procedure. It's non-obvious precisely because this is not indicated at all by the model itself since there's no requirement that actors return a response to a received message, nor even that actors implicitly know from whom a given message comes. Because of this, we're free to use the return value of the behaviour procedure in any way we like, and in the case above I've chosen to interpret the return value as the means of signifying actor behaviour updates. Symbols 'sleep and 'done are used to respectively keep the same behaviour and delete the actor from the hash map. Otherwise the response is assumed to be a new procedure to use as the actor's behaviour. Our implementation now has a fairly concrete actor representation. The only thing that's really missing at this point is a way to "send" messages, i.e. schedule them to be delivered asynchronously to actors using the above dispatch mechanism. This is probably a good point to emphasize the difference between concurrency and parallelism. Parallelism involves multiple tasks literally running at the same time, which requires multiple cores, CPUs, computers or whatever. Concurrency is a more general concept that describes tasks whose overall execution spans the same period of time, but doesn't necessarily require that the two tasks are ever simultaneously being executed: interleaving the steps in the tasks is enough. Why is this important? Because it means that for our purposes of exploring the model, it's enough to forget about threads, mutexes etc and instead implement the core of our actor processing machine using a simple FIFO queue. (define message-queue (make-fifo)) Here make-fifo returns a FIFO "f" to which a object "x" can be added using (fifo-push f x) and from which objects can be retrieved using (fifo-pop f). With this queue in place, the send-message function is easy: (define (send-message actor . message) (fifo-push message-queue (cons actor message))) I.e. sending a message amounts to simply adding it to the FIFO queue. "That's great," I hear you say, "but how on earth does anything get run?" Easy! We just set up a loop to iteratively dispatch messages at the end of the queue to their intended destination, like so: (define (next-addressed-msg) (if (fifo-empty? message-queue) '() (fifo-pop message-queue))) (define (process-next-message) (let ((addressed-msg (next-addressed-msg))) (if (null? addressed-msg) '() (let ((address (car addressed-msg)) (message (cdr addressed-msg))) (dispatch-message address message))))) (define (run) (unless (null? (process-next-message)) (run))) Done! Okay, so let's try it out! To do this I'm going to steal the first and simplest factorial example from a wonderful blog post [5] by Dale Schumacher. (I found his posts really fascinating in general, take a look!) This involves defining two actors as follows: (define println (make-actor (lambda (self . message) (apply print message) 'sleep))) (define factorial (make-actor (lambda (self customer . message) (match message ((n) (send-message self customer n 1) 'sleep) ((0 acc) (send-message customer acc) 'sleep) ((n acc) (send-message self customer (- n 1) (* acc n)) 'sleep))))) (I'm cheating a little by using the pattern matching macro implemented in Chicken's "matchable" egg, but it clarifies things hugely.) The first actor "println" simply prints any message it gets. This may seem like a peculiar inclusion in our first example, but remember that actors don't return values - they can only send more messages. We use the println actor as the destination for the result of the computation. The second actor "factorial" does all of the work. Firstly, note that its message structure is actually broken down into two parts: the "customer" argument and the "message" argument. The "customer" part tells factorial where the result of the computation should eventually be sent, while the "message" part is either a simple positive integer n, or a pair (n acc). In the former case, factorial sends _itself_ a message of the form (n 1). In the latter case, it again sends itself a message but this time (n-1 acc*n). This continues until factorial receives a message of the form (0 acc) at which point acc holds the result which factorial passes on to whatever actor has the address contained in "customer". To actually try this out we can kick the process off by evaluating the following two forms: (send-message factorial println 5) (run) The first adds a message (asking factorial to send the factorial of 5 to the actor whose address is bound to println) to the message queue. The second simply dispatches messages from that queue until none remain. The result is, unsurprisingly, 120 What's actually going on here? To get a better picture we can use the time-honoured trick of sprinkling print statements everywhere to trace the execution. Lets make our program tell us whenever a message is queued and dispatched. Furthermore, lets give the actors human-readable address names to make the trace even easier to read. The result of the above evaluation is then Queued message (println 5) to factorial Dispatching message (println 5) to factorial Queued message (println 5 1) to factorial Dispatching message (println 5 1) to factorial Queued message (println 4 5) to factorial Dispatching message (println 4 5) to factorial Queued message (println 3 20) to factorial Dispatching message (println 3 20) to factorial Queued message (println 2 60) to factorial Dispatching message (println 2 60) to factorial Queued message (println 1 120) to factorial Dispatching message (println 1 120) to factorial Queued message (println 0 120) to factorial Dispatching message (println 0 120) to factorial Queued message (120) to println Dispatching message (120) to println 120 We can now clearly see the sequence of messages that produced the final result, which was dutifully sent to the println actor to be displayed on the screen. Now here's an interesting question: what happens when we schedule two separate factorial evaluations simultaneously? Let's try computing 5! and 7!: (send-message factorial println 5) (send-message factorial println 7) (run) The result is as we'd expect, although note that the ordering here is a product of the scheduler implementation and is not defined by the abstract model: 120 5040 Okay, that seems to work. What does the message trace look like for this? Queued message (println 5) to factorial Queued message (println 7) to factorial Dispatching message (println 5) to factorial Queued message (println 5 1) to factorial Dispatching message (println 7) to factorial Queued message (println 7 1) to factorial Dispatching message (println 5 1) to factorial Queued message (println 4 5) to factorial Dispatching message (println 7 1) to factorial Queued message (println 6 7) to factorial Dispatching message (println 4 5) to factorial Queued message (println 3 20) to factorial Dispatching message (println 6 7) to factorial Queued message (println 5 42) to factorial Dispatching message (println 3 20) to factorial Queued message (println 2 60) to factorial Dispatching message (println 5 42) to factorial Queued message (println 4 210) to factorial Dispatching message (println 2 60) to factorial Queued message (println 1 120) to factorial Dispatching message (println 4 210) to factorial Queued message (println 3 840) to factorial Dispatching message (println 1 120) to factorial Queued message (println 0 120) to factorial Dispatching message (println 3 840) to factorial Queued message (println 2 2520) to factorial Dispatching message (println 0 120) to factorial Queued message (120) to println Dispatching message (println 2 2520) to factorial Queued message (println 1 5040) to factorial Dispatching message (120) to println 120 Dispatching message (println 1 5040) to factorial Queued message (println 0 5040) to factorial Dispatching message (println 0 5040) to factorial Queued message (5040) to println Dispatching message (5040) to println 5040 Wow! Notice that the computations of both factorials are interleaved, but the two computations do not interfere with one another at all. More than that, the factorial actor doesn't ever know or care to which computation the message being processed actually belongs. This is possible because the state of the computation is entirely specified by the content of the messages flying around. Is it possible to store state in the actor behaviours themselves? Of course! Again stealing (ahem borrowing) from [4] we can create an actor that works as a counter: (define ((make-counter-behaviour value) self customer . args) (match args (('get) (send-message customer value) 'sleep) (('inc delta) (make-counter-behaviour (+ value delta))))) Here we've actually defined a higher order procedure to construct behaviour procedures for us. A given counter behaviour responds to 'get and 'inc messages: 'get sends the a value (defined by the behaviour) to a customer actor, while 'inc replaces the existing behaviour with a new one in which the value is increased by some amount. Here's what it looks like in practice: (define counter (make-actor (make-counter-behaviour 0))) (send-and-run counter println 'get) (send-message counter println 'inc 3) (send-message counter println 'get) (send-message counter println 'inc 2) (send-message counter println 'get) (run) The result is exactly what you'd expect: 0 3 5 and the trace is as follows: Queued message (println get) to counter Dispatching message (println get) to counter Queued message (0) to println Dispatching message (0) to println 0 Queued message (println inc 3) to counter Queued message (println get) to counter Queued message (println inc 2) to counter Queued message (println get) to counter Dispatching message (println inc 3) to counter Updating behaviour of counter Dispatching message (println get) to counter Queued message (3) to println Dispatching message (println inc 2) to counter Updating behaviour of counter Dispatching message (println get) to counter Queued message (5) to println Dispatching message (3) to println 3 Dispatching message (5) to println 5 Again, remember that the precise _order_ in which the messages arrive is not defined by the model and thus is entirely implementation dependent. This of course means that the result is in some sense stochastic: not well defined. However, what's I think is important is that - provided the model is implemented in line with the above three rules - none of the synchronization headaches often associated with concurrent programming can occur. None of this is magic of course, its' just that the model enforces a discipline which avoids these difficulties. (State is local to actors, never shared between actors, and message responses are atomic.) Anyway, I hope you found that as fun as I did! The code for these exercises can be cloned from git://thelambdalab.xyz/actors.git or you can browse it via gopher [5]. Until next time, fellow gophers. -- [1]: http://worrydream.com/refs/Hewitt-ActorModel.pdf [2]: https://en.wikipedia.org/wiki/Actor_model [3]: doi:10.1023/A:1010079421970 [4]: http://www.dalnefre.com/wp/2010/05/deconstructing-the-actor-model/ [5]: gopher://thelambdalab.xyz/1/scripts/browse-git.scm|actors.git