A practical introduction to kill-safe, concurrent programming in Racket
1 The catter API
<make-catter>
<catter-stop>
<catter-read>
<catter-read-evt>
<example>
2 Design
3 Creating the catter
<logger>
<catter-struct>
<make-catter-body>
<make-manager>
3.1 Creating the subprocess
<subproc-creation>
3.2 Reader setup
<reader-setup>
4 The Reader Thread
<make-reader>
<reader-loop>
<error-handling>
5 The Manager
5.1 The Manager Loop
<useful-structs>
<manager-loop>
5.2 Handling a new request
<new-request>
5.3 Handling responses from the Reader
<reader-response>
5.4 Events that stop the catter
<stop>
5.5 Safely handling callers that have gone away
<pending-requests>
5.6 Sending responses to the caller
<response-attempts>
5.7 Submitting requests to the subprocess
<submit-inputs>
6 Sending requests and receiving responses
<catter-stop-body>
<catter-read-body>
6.1 Using negative acknowledgements to good effect
<catter-read-evt-body>
<handle-read>
7 Complete Code
<*>
8 Closing thoughts
9 References and acknowledgements
8.9

A practical introduction to kill-safe, concurrent programming in Racket

Nikhil Marathe <me at nikhilism dot com>

Racket’s default concurrency model is based on green threads. It uses a model based on ConcurrentML (CML), which is very similar to Go, but allows further composition of concurrency primitives. In one of my posts I had given a quick summary of this and pointers to other resources. I thought it would be good to walk through building a somewhat useful concurrent program to show how to use it. As an experiment, this is written as a literate program, using scribble/lp2/manual.

The use case is to provide a Racket wrapper for git-cat-file. This command lets you retrieve the contents of a Git object (say a commit+path pair). Imagine a task that requires getting the contents of a specific file in a Git repository at a specific commit. A common uses case includes generating some code from the contents at a specific commit. We would like to allow concurrent requests for file contents from multiple green threads and our program should correctly serve responses to the calling threads.

To avoid spawning a subprocess for every request, the program will use cat-file’s batch mode. This batch execution will be wrapped in a nice API that will be usable by multiple threads.

Beyond the use of ConcurrentML primitives, which allow dynamically building up sets of events to wait on, and associating actions to run when those events are selected, this program also demonstrates kill-safety. That is, the program should exhibit liveness and safety in the presence of various faults. A client thread that exits badly should not cause the cat-file service to get into an inconsistent state, and the service should be able to respond to clients going away, instead of leaking memory or blocking on trying to respond.

This tutorial assumes some familiarity with Racket, Go-style concurrency (including channels), green threads, Git and running subprocesses.

Racket threads are a unit of concurrency and isolation, not parallelism. See Places for a parallelism mechanism that could be easily integrated into this code.

1 The catter API

A user creates a catter using make-catter, passing it the Git repository path. The returned catter can be safely used from multiple threads.

(define (make-catter repo-path)
  <make-catter-body>)

The catter can be stopped via custodian shutdown, or using catter-stop!.

(define (catter-stop! a-catter)
  <catter-stop-body>)

There are 2 ways to submit read requests. The blocking catter-read, which will return the contents of the file, as a bytes (or raise an error),

(define (catter-read cat commit file-path)
  <catter-read-body>)

and catter-read-evt

(define (catter-read-evt cat commit file-path)
  <catter-read-evt-body>)

which offers a synchronizable event that can be waited on with other events. This is a non-blocking API. This will become ready for synchronization when the file content is available, and the synchronization result is the file contents. For example, the consumer could sync on this with a alarm-evt to do something else if a response is not received within a given time.

Here is a complete example of using catter. It hard-codes a few different files, and gets their content at a specific commit, but in different threads.

(define (do-stuff-with-file file-path contents)
  (printf "~a contents ~a~n" file-path contents))
 
(define files
  (list
   "info.rkt"
   "main.rkt"
   "does-not-exist"))
 
(define commit "master")
 
; Uses the repository of this tutorial.
(define cat (make-catter (current-directory)))
(define tasks
  (for/list ([file files])
    (thread
     (lambda ()
       (do-stuff-with-file file (catter-read cat commit file))))))
(map sync tasks)
(catter-stop! cat)

2 Design

The catter manages a git-cat-file subprocess. Requests for file contents need to be written to the stdin of the subprocess, and responses are read from stdout. Responses have a specific pattern: a single line with response metadata, a new line, then the contents, then a new line. To avoid having to juggle this state of where we are, we will have a dedicated reader thread to read stdout. Just like goroutines, Racket threads are very cheap to create and are pre-emptible.

Besides managing the subprocess, the catter has to handle incoming requests, and deliver responses to the correct calling thread. The implementation will make liberal use of channels for this, as these are also cheap to create. Channels are always synchronous (like CML). They don’t contain values. Rather, they act as synchronization points where threads can rendezvous and exchange a single value.

Bounded, buffered channels are present in the standard library. See make-async-channel.

To perform all this management, the catter has a manager thread that spins forever. The manager thread is the heart of catter, and exhibits the kill safety characteristics we want.

3 Creating the catter

To help debug the code, lets define a helper logger

(define-logger git-cat)

This defines a logger named git-cat, and several helper functions like:

(log-git-cat-debug "Debug log")
(log-git-cat-info "Info log")

Both the Dr. Racket Log window and the terminal will show log output based on the log level. This can be set to debug by setting the environment variable PLTSTDERR="debug@git-cat", or making a similar change in the Dr. Racket Log window.

catter is a struct. req-ch is a channel for incoming requests. stop-ch is a channel to tell the catter to stop (only used by catter-stop!) and mgr is the manager thread.

(struct catter (req-ch stop-ch mgr))

The catter constructor is really simple. It creates the two channels and the manager thread, and returns a catter instance with its fields set.

(define req-ch (make-channel))
(define stop-ch (make-channel))
(catter req-ch stop-ch (make-manager repo-path req-ch stop-ch))

make-manager creates a new instance of the manager thread, which is the most complex part of the program. We are going to break it down into somewhat bite-size pieces.

(define (make-manager repo-path req-ch stop-ch)
  (thread/suspend-to-kill
   (lambda ()
     <subproc-creation>
 
     <reader-setup>
 
     <manager-loop>)))

The use of thread/suspend-to-kill is crucial. Unlike a thread created with thread, this suspend-to-kill thread is not actually terminated by kill-thread or custodian-shutdown-all. It remains suspended and can be resumed by thread-resume. This prevents rogue users of catter from denying service to other threads. It also prevents deadlock from mistaken use. In this admittedly contrived situation

(define cat (make-catter ...))
 
(thread
 (lambda ()
   (sleep 5)
   (catter-read ...)))
 
(catter-stop! cat)

a naive implementation would lead to the second thread never receiving a response, as catter-read would submit a request, but the manager thread would be dead. A suspended thread will only truly terminate when there are no references to it any longer.

3.1 Creating the subprocess

As soon as the catter is created, the manager thread creates the git-cat-file subprocess.

(define-values (cat-proc proc-out proc-in _)
  (parameterize ([current-subprocess-custodian-mode 'kill])
    ; current-error-port is not a valid file-stream-port? when run in DrRacket.
    (let ([err-port (if (file-stream-port? (current-error-port))
                        (current-error-port)
                        (let ([fn (make-temporary-file "rkt_git_cat_stderr_~a")])
                          (log-git-cat-debug "stderr is not a file-stream. Using ~a instead." fn)
                          (open-output-file fn #:exists 'truncate)))])
      (subprocess #f #f err-port
                  (find-executable-path "git")
                  "-C" repo-path
                  "cat-file"
                  "--batch=%(objectname) %(objecttype) %(objectsize) %(rest)"))))
 
(file-stream-buffer-mode proc-in 'line)

This part is fairly straightforward, beyond the workaround for Dr. Racket. We use subprocess to spawn git cat-file, passing a few parameters. The crucial one is the --batch= format string. The %(rest) bit request’s that anything in the input request following the object name is reflected back in the response. This allows delivering file contents to the correct read request. The parameterization of current-subprocess-custodian-mode ensures the process is forcibly killed on custodian shutdown. This does mean that if any thread creates the catter within a custodian and shuts down the custodian, any threads spawned outside the custodian, that use the catter, will receive a non-functional catter. However our kill-safe design will still make sure those threads don’t block. Finally, we set the stdin mode to line buffered. This way the port implementation will automatically flush when we write a line, saving us some effort.

The closest analogue to parameters in mainstream languages are thread-locals, or Python’s contextvars.

3.2 Reader setup

This is pretty simple, with the actual details in the The Reader Thread.

(define reader-resp-ch (make-channel))
(define reader (make-reader proc-out reader-resp-ch))

The reader-resp-ch channel is used to read responses from the reader thread. It will be ready for synchronization whenever the reader has file contents or errors.

Let’s digress to the reader thread for a bit, before diving into the meat of the implementation.

4 The Reader Thread

The reader is a regular thread, since kill-safety doesn’t require it to stay suspended. It will read git-cat-file batch output and put a response on the resp-ch. We parameterize the input port so that all read operations can default to the current-input-port instead of requiring proc-out to be passed to each one.

(define (make-reader proc-out resp-ch)
  (thread
   (lambda ()
     (parameterize ([current-input-port proc-out])
       <reader-loop>))))

The reader loop is fairly straightforward.

(let loop ()
  (define line (read-line))
  (when (not (eof-object? line))
    (match (string-split line)
      [(list object-name "blob" object-len-s key)
       (define object-len (string->number object-len-s))
       (define contents (read-bytes object-len))
       (if (= (bytes-length contents) object-len)
           (begin
             ; consume the newline
             (read-char)
             (channel-put resp-ch (cons key contents))
             (loop))
           ; The subprocess must have exited
           (error 'unexpected-eof))]
      <error-handling>)))

It tries to read one line, which is assumed to be the metadata line. This is split by whitespace and matched to the "blob" output. This indicates we got successful file contents, in which case we can use match’s powerful ability to introduce bindings to access the various metadata. We do make sure to read exactly the file contents. If we are unable to do this, something is wrong (the subprocess exited), and we error. Otherwise the key and contents are written out to the channel as a cons pair. The reader enters the next iteration of the loop to read additional output.

[(list object-name (and msg (or "ambiguous" "missing")))
 (channel-put resp-ch
              (cons object-name
                    (exn:fail
                     (format "~a object ~v" msg object-name)
                     (current-continuation-marks))))
 (loop)]

We must also handle some error cases in the batch output. These can happen if the input commit/file-path combination was invalid. Unlike end-of-file, these are user errors and should be relayed to the caller. So we match those forms, and then create an exn:fail object with some extra information. That is written out to the channel, with other parts of the API knowing how to handle exceptions.

The (and ...) matcher has a neat trick where an identifier can be used as the first argument to bind the matched pattern to it.

All other exceptions will cause the reader to fail and exit, and the manager will observe that.

5 The Manager

This is the crux of the program. We are going to continue from where we left off in <make-manager>. We will observe several useful features of the CML concurrency model here.

5.1 The Manager Loop

The manager is an infinite loop that maintains 3 lists and a boolean.

  1. pending is a list of Pending instances, tracking outstanding cat-file requests that have been accepted and for whom we are awaiting a response from the subprocess.

  2. write-requests is a list of bytes containing UTF-8 encoded lines we want to write to the subprocesses’ standard input. I’ll explain why we maintain this list instead of directly writing to the subprocess.

  3. response-attempts is a list of Response-Attempts. On each loop iteration we will try to send a response back to the thread that issued the request.

  4. closed? a boolean that tracks whether this catter is no longer operational, due to clean shutdown or error.

The structs in these lists are:

(struct Request (commit path ch nack-evt))
(struct Pending (key ch nack-evt))
(struct Response-Attempt (ch nack-evt resp))
 
(define (pending->response p resp)
  (Response-Attempt (Pending-ch p)
                    (Pending-nack-evt p)
                    resp))

Requests are passed to the manager thread by <catter-read-evt>. They contain the channel to respond on, the commit+path for the request, and a negative acknowledgement (NACK) event, used for cancellation, that will be explained later.

Pending tracks the key associated with the request. Each Request is immediately converted to a Pending. The channel and NACK event are from the Request struct.

Once the subprocess returns a response, a Pending request becomes a Response-Attempt, using the pending->response helper. resp is either a bytes containing the file contents or an error.

The loop itself is a single call to sync. However each event in the call has some complexity, so the overall code is long. Let’s look at it one at at time. The one invariant we must preserve is that loop always calls itself regardless of which event gets selected.

We use apply plus append to assemble the various combinations of single events and lists that we need to pass to sync. This means sync may be waiting on tens of events at once. Each event is going to have extra behavior associated with it using handle-evt, all of which promise to call the loop again.

For Pythonistas, the closest analogue to sync is asyncio.wait. For Go folks, it is select. For JavaScript it seems to be Promise.race(). However, IMO, sync is better in several ways that this margin note is too small to explain.

(let loop ([pending null]
           [write-requests null]
           [response-attempts null]
           [closed? #f])
  (apply
   sync
 
   <new-request>
   <reader-response>
   <stop>
   (append
    <pending-requests>
    <response-attempts>
    <submit-inputs>)))

To restate, each iteration of the loop, sync will pick one event out of any that are ready for synchronization. It will return the synchronization result of that event. Our events are actually wrapper events returned by handle-evt. This means, when the wrapped event is ready for synchronization, the wrapper is also ready. However, the synchronization result of the event from handle-evt is the return value of the function provided as the second argument of handle-evt. Due to Racket’s tail call optimizations, it is safe and fast for these functions to call loop as long as it is done in tail position.

5.2 Handling a new request

(handle-evt req-ch
            (match-lambda
              [(Request commit path resp-ch nack-evt)
               (if closed?
                   (loop pending
                         write-requests
                         (cons
                          (Response-Attempt resp-ch nack-evt (exn:fail "catter stopped" (current-continuation-marks)))
                          response-attempts)
                         closed?)
                   (let* ([key (format "~a:~a" commit path)]
                          [p (Pending key resp-ch nack-evt)]
                          [write-req
                           (string->bytes/utf-8 (format "~a ~a~n" key key))])
                     (loop (cons p pending)
                           (cons write-req write-requests)
                           response-attempts
                           closed?)))]))

When we receive a new request from other threads, it will come in on the req-ch channel as a Request. The synchronization result of a channel is the value received on the channel. We can use the convenient match-lambda to directly create a function that will bind to the members of the struct.

If the catter is already closed, we don’t create a pending request. Instead we queue up a response attempt to the calling thread and continue looping. This handles the kill-safety aspect, where callers can continue to rely on the catter responding to them, as long as they have a reference to it.

In the common case where the catter is running, we create a key using the commit and path. We do this because the object name of the response will not be the commit+file path, so we need a way to associate the request and response. The key is also the input request object name. We rely on cat-file’s input syntax treating everything after the object name as something that will be echoed back in the %(rest) output format. We loop with this request line ready to be written.

5.3 Handling responses from the Reader

The next possible event is the reader responding with a pair of the key and contents. Handling it is straightforward. This handler tries to find an entry in pending that matches the key. It creates a Response-Attempt from that, and then removes it from the pending list (by returning #f to the filter). It loops again with the two modified lists.

If anybody knows how to express the filter in a nicer way, please tell me!

(handle-evt reader-resp-ch
            (match-lambda
              [(cons got-key contents)
               (define found #f)
               (define new-pending
                 (filter
                  (lambda (p)
                    (if (and (not found) (equal? got-key (Pending-key p)))
                        (begin
                          (set! found (pending->response p contents))
                          #f)
                        #t))
                  pending))
               (loop new-pending
                     write-requests
                     (if found
                         (cons found response-attempts)
                         response-attempts)
                     closed?)]))

5.4 Events that stop the catter

There are 3 situations in which the catter can no longer process new requests.

  1. The subprocess exits. This could happen for a multitude of reasons, including the user killing the process or it encountering an error. The value returned by subprocess is can act as an event that becomes ready on process exit.

  2. The reader thread exits, due to an 'eof or due to an error. thread is ready for synchronization in that case.

  3. An explicit stop via catter-stop!.

In all 3 cases, we want to do the same thing, so we compose choice-evt and handle-evt. This is pretty neat!

<stop> ::=
(if closed?
    never-evt
    (handle-evt (choice-evt cat-proc reader stop-ch)
                (lambda (_)
                  (close-output-port proc-in)
                  (subprocess-wait cat-proc)
                  (define status (subprocess-status cat-proc))
                  (when (not (zero? status))
                    (log-git-cat-error "subprocess exited with non-zero exit code: ~a" status))
                  (define new-puts
                    (map
                     (curryr pending->response (exn:fail "terminated" (current-continuation-marks)))
                     pending))
                  (loop null
                        null
                        (append new-puts response-attempts)
                        #t))))

Closing the subprocess input port is essential to cause the subprocess to quit. Then we wait for the process to quit and log an error if it did not exit successfully. At this point, any pending requests cannot be fulfilled. They get converted to Response-Attempts with an exception as the result. The next loop also discards all write-requests as those will no longer succeed. The most important action is to set closed? to #t.

In the iterations where the catter is already closed, we use never-evt instead, which is never ready in a sync. This will keep the loop running to process other events.

5.5 Safely handling callers that have gone away

This is the first case where we are going to do something that will seem weird, but is surprisingly powerful. I’m also not sure if a facility like this truly exists in other concurrency implementations.

We are going to handle the case where a caller submitted a request, but then disappeared before receiving the response. It either is no longer interested, or experienced some error and terminated. Regardless, we don’t want to keep the request attempt pending on the manager end. We are going to leverage negative acknowledgement events. the Using negative acknowledgements to good effect section explains this in more detail, and I recommend jumping to that section and then coming back here. On this end, we have an existing nack event. This nack event will be ready when it was not selected in a sync, or the calling thread went away. We can use it to safely remove the pending request. Note that the catter will still submit the request to the subprocess. We are only using this to discard responses. We will see another appearance from nack events in the response-attempts handler, where it will be more important.

Removing any requests that are not yet sent to the subprocess from write-requests is left as an exercise for the reader.

(for/list ([req (in-list pending)])
  (handle-evt (Pending-nack-evt req)
              (lambda (_)
                (loop (remq req pending)
                      write-requests
                      response-attempts
                      closed?))))

5.6 Sending responses to the caller

(for/list ([res (in-list response-attempts)])
  (match-let ([(Response-Attempt ch nack-evt response) res])
    (handle-evt (choice-evt (channel-put-evt ch response) nack-evt)
                (lambda (_)
                  (loop pending
                        write-requests
                        (remq res response-attempts)
                        closed?)))))

So far, we have been queuing reader responses into response-attempts, instead of resolving the caller’s request. A channel has no buffer, which means if the caller thread has gone away, a channel-put will deadlock. To avoid this, we use the same nack event from above, along with channel-put-evt. channel-put-evt will become ready when the channel has a receiver waiting on the other side and the value is successfully sent. By performing a choice on the nack event and the put event, we can handle success and error. Due to how the public API will set this up, we are guaranteed that one of the two must succeed. Either the thread is still waiting on the other end, or the nack event will be ready. The action in both cases is the same – remove the response attempt from future iterations. Pretty neat!

5.7 Submitting requests to the subprocess

(for/list ([req (in-list write-requests)])
  (handle-evt (write-bytes-avail-evt req proc-in)
              (lambda (_)
                (loop pending (remq req write-requests) response-attempts closed?))))

One of the things to keep in mind when writing concurrent code is that we have to avoid any action that could cause the thread to block. Racket is communicating with the subprocess over pipes, and pipes have a limited capacity. The process on either end can block once this capacity is reached. If the subprocess is waiting for the reader thread to read from stdout, then it can block on stdin. This would be bad for the manager thread because it would not be able to service new incoming requests until that was resolved. Any and all actions in our loop should execute without blocking and re-enter the sync. This is why, in Handling a new request, we didn’t try to write to the subprocess in the event handler. In this chunk of code, we use write-bytes-avail-evt instead. This will only become ready if it is possible to write to the pipe immediately, without blocking. If so, it will perform the write. It offers the guarantee that if the event is not selected by sync, no bytes have been written. This ensures we don’t write the same request multiple times. When this write succeeds, the handler removes it from future iterations.

The use of events to write to the process stdin means that requests may be submitted to the subprocess out of order from how they were received by the manager thread. This makes the key crucial to match callers. However it is safe to have multiple requests for the same key and resolve them in any order.

6 Sending requests and receiving responses

Phew! Good job on making it past the manager loop. There are still a couple of confusing things in there. Hopefully, seeing the public API used by client threads will clear those up. See The catter API for the declarations.

(thread-resume (catter-mgr a-catter) (current-thread))
(channel-put (catter-stop-ch a-catter) 'stop)

Stopping the catter is pretty simple. Just issue a write (doesn’t matter what we write) to the channel. However, the call to thread-resume is key! Remember how the manager thread was created with thread/suspend-to-kill so it doesn’t terminate? Well, it may be suspended right now if all custodians that were managing it have gone away. thread-resume associates the manager thread with the current thread, which means it acquires all the custodians of the current thread. With a non-empty custodian set, the thread can start running again. This is crucial because we are about to write to a channel that better have a reader on the other end. I encourage reading the kill-safety paper for the details.

The other public APIs will similarly need to be careful to resume the thread!

(sync (catter-read-evt cat commit file-path))

The synchronous request version is easy. We take the async form and block on it.

6.1 Using negative acknowledgements to good effect

(define evt (nack-guard-evt
             (lambda (nack-evt)
               (define resp-ch (make-channel))
               (thread-resume (catter-mgr cat) (current-thread))
               (channel-put (catter-req-ch cat) (Request commit file-path resp-ch nack-evt))
               resp-ch)))
(handle-evt evt
            <handle-read>)

This function provides the nack events the manager loop has been monitoring. nack events are a core part of the Racket/CML concurrency model, and are manufactured using nack-guard-evt. I encourage reading the documentation for guard-evt and nack-guard-evt carefully, and playing with them a bit, as it isn’t immediately obvious what is happening. To get a response back from the manager, we are using a channel. We want to guard against the API caller no longer waiting on the response. To do this, we wrap the channel with a nack event. The nack-guard-evt function returns an event. Every time this event is used with sync, it calls the function passed to nack-guard-evt with a new nack event. There are now two events in play, and they are linked.

The first event is the return value of nack-guard-evt. This event becomes ready when the return value of the function passed to nack-guard-evt becomes ready. In this case, that is the channel on which the manager thread will deliver the response.

The second event is the nack-evt passed to the function. This event is sent to the manager. When (and only when!) the first event is used in a sync, but is not chosen as the result of the sync, the nack event will become ready. It will also become ready if the thread unexpectedly exits the sync. If the first event is chosen, then the nack event will never become ready. This duality plays very nicely with the manager thread in Sending responses to the caller.

Finally, in true CML form, we can compose this existing composition with handle-evt to raise an exception or deliver the result based on the response from the manager.

(lambda (resp)
  (if (exn:fail? resp)
      (raise resp)
      resp))

Keep in mind that these functions, and handle-evt are not immediately running any code. They are associating code to be run when events are passed to sync.

With that, we have completed the entire public API. I hope this gave you a taste of concurrency using Racket’s CML primitives.

7 Complete Code

Some imports round off the program. Also see the full listing.

<*> ::=

8 Closing thoughts

There are a lot of things I like about this concurrency model. Composition is a powerful way to perform subsequent handling on concurrent operations. The kill-safety and liveness properties are pretty nice.

I don’t have enough experience to say how negative acknowledgement events compare to other cancellation mechanisms like cancellation tokens, Go’s context.WithCancel() or Python’s cancellation exception. One of the advantages I realize is that with a specific value to watch for, it is harder to mess up cancellation handling for others. In Python, a task has to handle CancelledError, and if it swallows the exception, it can mess up calling code as well. I am also not aware of any mechanism in Python to know if another thread exits in error.

The async/await model also has no guarantees on only one event out of several getting chosen. For example, a primitive like read-line-evt guarantees that data has been read from the buffer only if the event was selected. On the contrary, using asyncio.wait() to drive several tasks in Python makes no such guarantees.

At first I thought CML/Racket suffer from the function coloring problem. However, it doesn’t. If you were to use a blocking I/O function in a Racket thread, you would not block other threads, the way using a blocking I/O function in any task in Python can block other tasks. Green threads performing sequential operations can continue to use regular I/O functions, and the runtime will transparently pre-empt and reschedule them. There is no alternate syntax for "red" functions. The multitude of *-evt functions is only required wherever your thread has to actually deal with concurrency.

There are a few remaining holes in this code. It does not handle Breaks (User interrupts). It should be fairly straight-forward to support that by using the various /enable-break forms. The manager thread is also susceptible to bugs that would raise an exception and lead to the thread exiting loop, at which point catter-read would still deadlock. The solution probably is to have a catch-all exception handler at the top-level that transitions to closed.

The manager thread loop maintaining local state requires cramming the entire implementation into one long block of code. It would be nice to refactor each event’s handling into a named function. However the only reasonable way to do that while not having to pass 4 arguments (and receive changed values) to every function is to define these functions within the (let loop ...), so the functions close over the arguments. At that point, I’m not sure if there is a readability win.

Finally, in this demo I’ve chosen to focus on concurrency, but an application with sufficient scale could run multiple cat-file subprocesses.

As I have said in other places, Racket as a language tends to be ignored or maligned. However it has so much to teach other languages. This concurrency paradigm is definitely one of them. As far as I know, ML, Guile and F# (Hopac) are the only other languages that have something similar. I hope this tutorial can convince some of you that there are better ways to think about concurrency out there beyond the async/await local minima the industry is currently in.

9 References and acknowledgements

The Racket Docs are an excellent source of reference information for events, threads, and all the other possible standard library items that can act as events. Racket also provides ways for user structs to act as events, and the unsafe FFI APIs allow arbitrary file descriptors and other polling mechanisms to plug into this system.

Flatt and Findler’s Kill-Safe Synchronization Abstractions, while 20 years old at this point, is a really good reference to the implementation of CML in Racket, and the problems they set out to solve. It is also an excellent example of Racket internalizing extra-linguistic mechanisms.

Andy Wingo’s A New Concurrent ML is a really good introduction to the nitty gritty implementation details of CML, as are his related posts comparing it to Go.

Bogdan Popa’s resource-pool-lib is another good application of CML patterns in Racket that I consulted heavily while wrapping my head around this model.

John Reppy is the creator of Concurrent ML. He has several papers laying out its design and specifics. I read CML: A Higher-order Concurrent Language and thought it was a good primer on CML, although the APIs are very different compared to their Racket forms.

Finally, thanks to EmEf and benknoble on the Racket forums for helping me with a few things.