pHluid: The Design of a Parallel Functional Language
Transcription
pHluid: The Design of a Parallel Functional Language
pHluid : The Design of a Parallel Functional Language Implementation on Workstations Cormac Flanagan Rice University Department of Computer Science Houston, Texas 77251-1892, USA [email protected] Rishiyur S. Nikhil Digital Equipment Corp. Cambridge Research Laboratory One Kendall Square, Bldg. 700 Cambridge, Massachusetts 02139, USA [email protected] Abstract This paper describes the distributed memory implementation of a shared memory parallel functional language. The language is Id, an implicitly parallel, mostly functional language that is currently evolving into a dialect of Haskell. The target is a distributed memory machine, because we expect these to be the most widely available parallel platforms in the future. The dicult problem is to bridge the gap between the shared memory language model and the distributed memory machine model. The language model assumes that all data is uniformly accessible, whereas the machine has a severe memory hierarchy: a processor's access to remote memory (using explicit communication) is orders of magnitude slower than its access to local memory. Thus, avoiding communication is crucial for good performance. The Id language, and its general dataow-inspired compilation to multithreaded code are described elsewhere. In this paper, we focus on our new parallel runtime system and its features for avoiding communication and for tolerating its latency when necessary: multithreading, scheduling and load balancing the distributed heap model and distributed coherent cacheing, and parallel garbage collection. We have completed the rst implementation, and we present some preliminary performance measurements. Keywords: parallel and distributed implementations garbage collection and run-time systems data ow. 1 Introduction This paper describes the distributed memory implementation of a shared memory parallel functional language. The language is Id 19], an implicitly parallel, mostly functional language designed in the dataow group at MIT. Id is semantically similar to Haskell 14] and is in fact currently evolving into pH 1], a dialect of Haskell. Economics seems to dictate that most scalable parallel platforms in the next ve to ten years will be clusters of SMPs (symmetric/shared memory multiprocessors), i.e., machines consisting of a number of nodes that communicate using message passing over a switched interconnection network, where each node may be a small SMP (bus-based, 2-4 processors). Larger shared memory machines are of course possible, as demonstrated by the Stanford DASH multiprocessor 17], and the KSR machines 15], but they are likely to be high-end machines and not widely available. Further, scalable shared memory machines (like the DASH) are also built with physically distributed memories for scalability, and face some similar problems with their memory hierarchies. Even at small numbers of processors (such as 4 to 8), many more people are likely to have access to clusters of uniprocessors of that size than SMPs. Thus, our target for Id is a distributed memory, message passing machine. In our initial implementation, each node is a conventional uniprocessor workstation, not an SMP. We know how to extend this to exploit SMP nodes and believe it will be easy (handling distributed memory is the major hurdle). The dicult problem is to bridge the gap between the shared memory model of the language and the distributed memory model of the machine. In Id, as in Haskell, SML and Scheme, the language model assumes that all data is uniformly accessible, whereas the machine has a severe memory hierarchy: a processor's access to remote memory (using explicit message passing) is typically orders of magnitude slower than its access to local memory. Thus, minimizing communication, avoiding communication if possible, and tolerating the latency of remote operations, are all crucial for good performance. The pHluid system is a compiler and runtime system for the Id language that we have been building at Digital's Cambridge Research Laboratory for some years. Based on ideas originating in dataow architectures1 11] the compiler produces multithreaded code for conventional machines. In this paper, we focus on novel aspects of a new parallel runtime system for pHluid, in particular features that avoid communication and tolerate its latency when necessary: multithreading, scheduling and load balancing a distributed heap model and distributed coher1 The name pHluid is a play on Id, pH and data ow. ent cacheing, and parallel garbage collection. We also present some preliminary performance data. 2 Background on Id and its compilation to multithreaded code The important messages of this section are summarized in its last paragraph. Readers familiar with topics such as Id, Id compilation, dataow, message driven execution, ne grain multithreading, etc. may wish to skip this section and just read the last paragraph.] Id 19] is an implicitly parallel, mostly functional, language designed in the dataow group at MIT. It has many features common to other modern functional languages like Haskell and SML{ higher-order functions, a Hindley-Milner polymorphic type system with userde ned algebraic types, pattern-matching notation, array and list comprehensions, etc. The main novelty of Id is its implicitly parallel evaluation model: everything is evaluated eagerly, except for expressions inside conditionals and inside lambdas. This is described in more detail in 19], but a key behavior relevant to this paper is that most data structures have I-structure semantics : given an expression of the form: cons e1 code for latency tolerance{ we do not start with a compiler for sequential machines and add parallelism as an afterthought. Figure 1 shows an example P-RISC assembler translation of the following function that counts the nodes in a binary tree: def leaves Empty = 1 | leaves (Node x l r) = leaves l + leaves r It is displayed here in graphical form as a control-ow graph, but it is trivial to linearize with labels and explicit control transfers. The following features are worth e2 we allocate the data structure and evaluate e1 and e2 in parallel. The reference to the cons cell is immediately available as the result of the expression. Any consumer of the data structure that attempts to read the head (or the tail) of the cell will automatically block, if necessary, until e1 (or e2) has completed evaluation and is available in the data structure2 The major phases of our pHluid compiler for Id are: Parsing, typechecking, simpli cation, lambda lifting, optimization, etc., eventually producing PRISC assembler, a ne grain multithreaded abstract machine code (\parallel RISC"). Translation and peephole optimization, converting P-RISC assembler to Gnu C. Gnu C compilation, and linking with our new parallel runtime system, written in Gnu C. We use various C extensions provided by Gnu C, such as rst class C labels, and mapping C variables that contain frequently accessed data such as the heap allocation pointer into speci c machine registers. First class C labels allow us to represent ne grain, dynamically scheduled threads conveniently. These Gnu C facilities have also been used by other researchers for the same purposes. An important point is that the entire compiler is geared towards producing ne grain multithreaded 2 This non-strict behavior in fact makes Id semantically closer to Haskell than to SML, despite its eager evaluation. Recognizing this semantic similarity, and because of various other syntactic similarities, Id is currently evolving into pH 1], a dialect of Haskell. Figure 1. P-RISC assembler: An example noting: The function's code can be viewed as a collection of threads , each of which is activated by the arrival of a message (in practice, we optimize away many messages, both statically and dynamically). Messages always arrive at entry instructions, whose arguments correspond to the message payload. At the top, a call message arrives, which allocates a frame for this function, and delivers three arguments to the top entry instruction: a continuation frame pointer, cfp a continuation label (instruction pointer) cip, and the tree itself, T. The if tests if the tree is empty if so, the result is 1, and the return instruction sends a message to the continuation (cfp, cip) carrying the result. If the test fails, we initiate two \heap I-structure loads": the hiloads conceptually send messages to the heap locations T1] and T2] requesting their contents, passing the current frame pointer (implicitly) and the labels a: and b: respectively (explicitly) as their respective continuations. The current function invocation then goes dormant, with no threads active. The heap location T1] eventually responds with a message, kicking o the thread at a:, placing the value (the left sub-tree) into local variable L. The last action of this thread is to initiate a recursive call by sending a call message containing the continuation (current frame pointer and label c:) and argument subtree L. This recursive call eventually returns a message that starts the thread at c:, loading the count of the left subtree into local variable t1. Similar actions take place at b: and and d:. Thus, two threads (from c: and d:) arrive at the join instruction. This instruction counts up the local variable j (initialized to 0). Only the last thread proceeds, nding that the count j has reached the terminal count (2) earlier threads die at the join. Thus, t1 and t2 are guaranteed ready when the sum is computed. This nal thread computes the sum and returns the result. It is undetermined whether thread a: executes before b: and whether c: executes before d:. They are scheduled asynchronously, as and when messages arrive. Thus, although these threads execute in some sequential order (because they all execute on the same processor), the two overall load actions (communicating to the heap and back) occur in parallel, and the two recursive function invocations execute in parallel. This kind of fundamental attention to latency is quite unique to Id compilers. The round-trip time for the request and response messages of a function call of course depend on how much work the function does, and whether the function call is done locally or remotely. The round-trip time for an hiload depends on a number of factors: { Whether the heap location is local or remote. { The current load on the processor that owns the heap location, which aects how soon it can handle the request. { Whether the heap location is \empty" or not when the request message arrives. Due to the non-strict, I-structure semantics, the producer may not yet have delivered the value in this case, the request message has to be queued on that location until the value arrives. However, note that all these sources of delay are handled uniformly by couching the code as multiple, ne grain, message-driven threads, each of which never suspends . All this was by way of background, and is discussed in more detail in 18]. The important points to remember for purposes of this paper are : Parallelism in pHluid is at two levels: the function call is the unit of work distribution across processors. This is real parallelism, in that these processors actually execute simultaneously. Within each function are a number of ne grain threads that are scheduled asynchronously based on message arrival. This multithreading is pseudo-parallelism, in that all threads in a function invocation are multiplexed on the same processor, but it is vital because it permits overlapping communication, synchronization and congestion latencies with useful work, and is a clean model for adaptively responding to the dynamic parallelism of the program. Unlike many other parallel languages, threads in pHluid are unrelated to function boundaries. Each function invocation allocates a frame that contains, among other things, all the local variables for all the threads of the function. However, because thread invocation does not involve any frame or stack resource allocation, threads are extremely lightweight. All functions share the same view of the heap, i.e., the heap is a shared memory . Threads in pHluid never suspend. All function calls and heap accesses are couched as split-phase transactions: one thread issues a request message, and a response message later initiates a separate continuation thread. For heap accesses, the request message may be queued at the heap location if the value is not yet available. Thus, data access synchronization occurs at heap locations, never in the accessing functions. Note that checking the full/empty state of a heap location is not an issue that is unique to I-structures in Id. Almost exactly the same issue is faced in lazy language implementations where we have to check whether a location has been evaluated yet or still contains a closure. This relationship is not surprising, because they are two dierent ways of implementing non-strictness (and they are both equally dicult to optimize away). 3 The new parallel runtime system This section describes the novel aspects of pHluid's new parallel runtime system. Recall that our target platform is a distributed memory machine: each PE (Processing Element) is a conventional uniprocessor workstation. These PEs communicate by explicit messagepassing using active messages 21], i.e., each message contains a code pointer (a \handler"), and the message is consumed by simply jumping to the handler with the message itself as an argument. The handler is responsible for extracting items from the message and executing arbitrary code that may free or reuse the message. 3.1 Scheduling and Work Distribution 3.1.1 Multithreading When a thread is executed, it may enable other threads. For example, consider the following code fragment describing two threads, starting at labels a: and b:, respectively: a: ... hiload b T1] ... foo ... b: ... A fast entry point which assumes the call arguments are already in registers. A xed queue entry point which assumes that the arguments need to be retrieved from a call record on the xed queue. A stealable queue entry point which assumes that the arguments need to be retrieved from a call record on the stealable queue. the thread at a: executes an hiload instruction to initiate access to a heap location, and continues (at foo ). Conceptually, it sends a message to the heap location, which eventually produces a response message that enables the other thread at b:. Each processor maintains a scheduling stack containing threads that are ready to run. Each entry in the stack consists of a code pointer, a pointer to the frame of the enclosing function, and a variable number of arguments. A thread is pushed on this stack when a response to a split-phase operation is received. When the current thread terminates, the system pops and executes the next thread on this stack. We avoid a stack empty check on pops from the scheduling stack by including a stack empty handler as the bottom entry in this stack. This handler is invoked whenever the system runs out of executable threads. The handler pushes itself back onto the stack, and then calls the scheduler (described below) to create new threads. The scheduler invokes a call record by jumping either to the xed or to the steabable queue entry point. The fast entry point is used for an optimization described below. The code at the entry point then extracts the arguments from the call record, allocates a frame and starts evaluating the function's body. The function later terminates when one of its threads sends a return message to the function's continuation, deallocates the function's frame, and dies. 3.1.2 Function calls 3.1.3 Work Scheduling In pHluid, the function call is the unit of work distribution across PEs. By default, the system chooses how to do this distribution. However, by means of a source code annotation at a function call site, the programmer can direct that the call must run on a speci c PE. Primitive functions are available to discover the PE on which the current function is executing, and the total number of PEs in the current run (this number is decided when starting the executable Id code is not compiled with a knowledge of the number of PEs, expect for the special case where we know we wish to compile for a uniprocessor). Each function call is encoded as a split-phase action. The function call itself involves creating a call record containing the function pointer (or codeblock , described below) and its arguments, including the continuation argument. The programmer can annotate any function call to specify which processor it should run on. When there is no annotation (the usual case), the runtime system chooses where to execute the function. Each processor maintains two scheduling queues containing call records: The xed queue contains call records that must be executed on this processor. The stealable queue contains call records that can be executed on any processor. For function calls that have a processor annotation, the call record is dispatched to the xed queue of that processor. For calls without a processor annotation, the call record is placed on the local stealable queue. Call records may migrate between the stealable queues of various processors according to the work stealing algorithm described below. However, a function call does not migrate once it has begun executing. Each function is described by a codeblock , which consists of three function entry points: Naive parallel work scheduling algorithms can result in an exponential growth of the memory requirements of an application as compared to a sequential execution by creating a very large number of parallel threads, each of which simultaneously requires some local storage. The pHluid work scheduling algorithm is designed to avoid this problem by reducing the number of simultaneously active functions, while still exploiting the available parallelism to keep all the PEs busy. The scheduler gives work on the scheduling stack a priority higher than either the xed or the stealable queue, in order to complete existing function activations (if possible) before creating new ones. The scheduler also invokes call records on the xed queue in preference to those on the stealable queue, since call records on the stealable queue may later be stolen to keep a dierent PE active. The scheduler tries to approximate the depth rst traversal of the function call tree performed by a sequential execution by treating the xed and stealable queues as LIFO queues, or stacks. When we execute a function call with no processor annotation, or with an explicit processor annotation specifying the current PE, we push the call record on the appropriate local queue. Each PE's scheduler always pops work from these queues. The net eect of our scheduler is that once all PEs have work, each PE tends to settle into a depth- rst traversal of the call tree similar to that performed by a sequential implementation. 3.1.4 Work Stealing When a PE becomes idle, it sends a steal message to a randomly-chosen victim PE. The steal message attempts to steal a call record from the victim PE's stealable queue. Of course, the victim PE may not have any work to steal, in which case the idle PE randomly chooses a new victim PE. In order to avoid repeatedly bothering active PEs with steal requests they cannot satisfy, we use a linear backo scheme to decide how long the idle PE must wait before asking that victim PE again, i.e., the wait time increases linearly with the number of times we fail. This scheme successfully adapts to both lowgranularity and high-granularity computations.3 If the victim PE's stealable queue is nonempty, then the steal message returns the call record from the bottom of that PE's stealable queue (i.e., away from the stack-like end) to the idle PE. The reason for this approach is that items deeper in the stack are likely to represent fatter chunks of work, being higher in the call tree 12, 16]. The net eect of our work-stealing algorithm is that we do not gratuitously fork functions to other PEs. Instead, work request messages are only sent when some PE is idle, and call records only migrate from busy PEs to idle PEs. 3.1.5 Optimizations The pHluid system is designed using general messagesending mechanisms that work in all cases. For example, the split-phase instruction: hiload b T0] conceptually involves sending a hiload message to the appropriate PE, which accesses the memory location T0], and then sends a split-phase return message to the original PE. For cases where the original PE contains the memory at location T0], we avoid the overhead of sending these messages by immediately performing the hiload operation and pushing the enabled thread at b: onto the scheduling stack. We also avoid the message sending overhead in other cases by inlining other split-phase operations where possible. We also optimize split-phase operations that are immediately followed by a halt instruction. For example, consider the following code fragment: a: hiload b T0] halt b: ... If the hiload operation can be satis ed on the current processor, then the pHluid system performs the load and immediately jumps to the thread at b:, without the overhead of pushing that thread on the scheduling stack and invoking the scheduler. Similarly, for a function call with no processor annotation followed immediately by a halt instruction, we avoid the overhead of manipulating the stealable queue by simply placing the arguments into appropriate registers and invoking the fast entry point (described above) of that function's codeblock. 3 Our work-stealing algorithm is a variant of that used in the Cilk system developed at MIT8]. Our algorithm was developed jointly with Martin Carlisle of Princeton University. 3.2 Distributed memory model and distributed coherent cacheing The memory of each PE is divided into ve regions: Compiled code and static data: Static data includes data structures that describe codeblocks and the other memory areas, etc. These are simply replicated at the same address on all PEs. The heap : this contains actual Id data structures (constructed data, arrays and closures), and is described in greater detail below. The store : this area contains objects that are explicitly allocated and deallocated by the runtime system, and which are never accessed remotely. These include frames for function invocations, the xed scheduling queue, heap access requests that are waiting on empty locations, etc. These objects are managed in a freelist organized by object size. Each PE maps its store at a dierent address, so that we can determine where a frame is located by looking at its address (frames are never accessed remotely, but we do need to know a frame's PE so that we can send a message to it, e.g., for sending a heap access response or for sending a result to a function's continuation). The stealable queue: in principle this could be allocated in the store, but because it is manipulated so heavily (since most function calls go through the stealable queue), we allocate a special region for it and treat it as a deque with contiguous entries, allowing cheap pushes and pops from both ends. The scheduling stack: this stack is also allocated in a contiguous memory area for performance reasons. In a previous sequential implementation, and in an initial parallel implementation, we allocated everything in a single heap (as in the SML/NJ implementation 4, 3]). However, we chose to separate out those objects that can be explicitly deallocated and which are not accessed remotely, in order to reduce garbage collection pressure. Although explicit deallocation of system objects incurs a small overhead, it does not require any communication, and signi cantly reduces the frequency of communication-intensive global garbage collection. 3.2.1 The heap Heap objects are represented quite conventionally, as contiguous chunks of memory. An object consists of a header word followed by one word for each eld (arrays have extra words for index bounds and pre-computed index calculation coecients). We use the lowest-order bit to tag immediates. Thus, for example, integers lose 1 bit of precision, but this is not a problem on our current platform, 64-bit Alpha workstations. If a eld of a heap object does not contain an immediate value, it always contains a pointer, either to another heap object or to a deferred list of continuations waiting for the eld to transition from empty to full . Since the eld contains a pointer in either case, and since our pointers are always aligned to word boundaries, this frees up another low-order bit to use as a full/empty bit and distinguish between these two cases. This works even if the eld is of immediate type (e.g., integer), because for the duration that it remains empty, it contains only a pointer and not an immediate, and the full/empty bit is available. The heap is the only area that requires global access. We use the operating system's virtual memory mapping facilities (mmap) so that the heap occupies the same address range on all PEs. The heap consists of a number of xed-size \pages", each of which is currently 1 KB in size. We partition \ownership" of these pages across the PEs. This, although each PE sees the whole address range, it only owns pages representing only a chunk that is 1=P of the total heap size (where P is the number of PEs), and it can only allocate objects into these pages. The remaining pages are treated as a cache for the data owned by other PEs. The advantage of this approach is that heap addresses do not have to be translated or indirected in going from one PE to another. Further, by examining an object's heap address, a PE can cheaply determine whether it owns the object or only a cached copy. The downside of this approach is that large amounts of address space are required as we increase the number of processors. Although this may somewhat limit the scalability of our system on 32-bit machines, it is not a problem on nextgeneration 64-bit architectures. We conjecture that the amount of physical memory required per PE will scale in a reasonable fashion, since only a small portion of the cache pages may actually be used between garbage collection cycles. If this conjecture does not hold in practice, we intend to implement a scheme that limits the number of cache pages in use at any time by explicitly deallocating cache pages. 3.2.2 Heap cacheing Because Id is a mostly functional language, the vast majority of objects in the heap are I-structure objects. These objects are accessed via the hiload operation, therefore we design our heap caching protocol to optimize this operation. Every heap location is tagged to specify whether it is full or empty in order to implement I-structure semantics. When a heap object is allocated, all its slots are initially empty, and these slots become full when they are initialized. Once an I-structures is initialized, it can never be changed. This functional nature of I-structures allows us to implement a very simple cache-coherence strategy. The invariant maintained by our strategy is that each cache page is always consistent with the corresponding \real" page, in the sense that the only allowable dierence is that a cache page may be empty at a location that is full (or de ned) in the corresponding real page. We associate a valid bit with every cache page. If a cache page is marked invalid, it implies that each Istructure in the page should be interpreted as being empty. Initially, every cache page is marked invalid, thus ensuring that the cache invariant holds at the start of program execution. On a local hiload, i.e., an hiload to a heap address that is owned by the current PE, we check the full/empty tag if full, we return the value (more accurately, as described before, push the value and enabled thread on the local scheduling stack). If empty, we queue the hiload's continuation on the heap location, allocating the queue entries themselves in the store, not the heap. On an hiload to a location owned by some other PE, we check our cache at that location. If we're lucky, the location is full, and the value can be read immediately. Otherwise, we need to retrieve the location's value from the owner PE of that page. Since we need to communicate with the owner PE anyway, it makes sense to request an up-to-date copy of that page at the same time, in the expectation that future hiloads to the same page can be satis ed immediately using the new copy of the page. Provided there is no outstanding request for that page, we simply send a hiload message to the owner of the page, and also request an up-todate copy of the page. However, if there is already an outstanding request for that page, then we queue the hiload locally, and reprocess it as soon as the new copy is received, since the new copy may allow us to satisfy the hiload locally, without performing any additional communication. An histore (\heap I-store") operation always writes through the cache to the owner PE (there, it may cause waiting continuations to be released). However, most histores are to locally-owned locations. Consider a typical Id constructor expression of the form (e1,e2): the heap allocation for the pair, and the two histores initializing the pair are part of the same function, and so the histores are guaranteed to be local, since memory allocation is always performed using the local memory of each PE. Because the histore operation only initializes a previously empty location, it does not aect the consistency of existing cache copies of the page containing that location. Thus, the PE that owns a page never has to send any \invalidation" messages to other PEs that have previously obtained copies of that page. In fact, the owner does not have to keep track of which PEs obtained cached copies, and it is perfectly safe for the owner to send multiple copies of the same page at dierent times. A new copy of the page is guaranteed to be consistent with a previous copy| the only dierences will be that some previously empty locations may now be full. A general observation here is that our distributed cachecoherence protocol relies heavily on the mostly-functional nature of the source language the protocol is trivial compared to those required for imperative languages 2]. We mentioned that Id is not a pure functional language { it has side-eecting constructs that operate on \Mstructures". However, as in SML and unlike Scheme, these side-eects are constrained by the type system to speci c objects and operations, and this allows the compiler to isolate them into special P-RISC assembler instructions: hmload (for \heap M-structure load") and hmstore (for \heap M-structure store"). These never interfere with the \functional" hiload and histore operations, i.e., the same heap location cannot be accessed by both I-structure and M-structure operations. Thus, the consistency of I-structure cacheing is not compromised. Our preliminary implementation of these sideeecting instructions does not involve caching { they always read and write through to the owner. We expect the overhead to be acceptable on many programs because, as in SML, side eecting operations have a much lower frequency compared to normal functional operations. However, programs that use M-structures extensively will require a more sophisticated implementation of these operations. To avoid deadlock, the pHluid system uses separate queues for regular computation messages and GC messages. Each processor treats its incoming computation message queue as part of its root set, and updates all messages (via the function move object) to refer only to to-space objects. We ensure that all outstanding computation messages are received, and hence updated, by their destination PE during GC by sending a flush message along the FIFO pipe connecting each pair of PEs. Detecting termination of a global GC cycle is dicult because each PE only has direct knowledge of its own status, and messages characterizing the status of other PEs may be out-of-date by the time they are received. We characterize the status of a PE as inactive once that PE has: 3.3 Parallel distributed garbage collection 1. Copied all known live objects into its to-space, 2. Updated all entries in the computation message queue, 3. Received all incoming flush messages, and 4. Received a response to each move object splitphase operation. A PE that is inactive may become active again on receipt of a move object message. The global garbage collection cycle can only terminate when all PEs are inactive simultaneously. We detect this situation using the following protocol: Conceptually, the PEs are organized as a ring. At any time a particular PE (the \terminator-PE") is responsible for detecting GC termination. When this PE becomes inactive, it tries to send a PEs-inactive? message around the ring of PEs. If this message encounters an active PE, then that PE becomes the terminator-PE. Otherwise, the message is successfully forwarded around the ring and returned to the terminator-PE. At this stage, the terminator-PE needs to ensure that all PEs are still inactive. It does this by sending a PEs-still-inactive? message around the ring of PEs. If this message encounters a PE which has been active since receiving the PEs-inactive? message, then that PE becomes the terminator-PE. Otherwise, the PEs-still-inactive? message is successfully returned to the terminator PE. At this stage we can guarantee that all PEs are inactive. The terminator-PE then broadcasts a message announcing the termination of the GC cycle. On receipt of this message, each PE returns an acknowledgement and resumes computation. To ensure that distinct GC cycles do not overlap, a PE that exhausts its heap space sends a request for a new GC cycle to the terminator-PE of the previous cycle. That terminator-PE waits, if necessary, until all PEs acknowledge termination of the last GC cycle, before broadcasting a message initiating the new cycle. The pHluid system uses a \stop-and-copy" approach for performing parallel garbage collection. Whenever any PE runs out of heap space, all the processors suspend their current activities to participate in a global garbage collection cycle. The garbage collector is a parallel extension of the conventional Cheney two-space copying collector 9]. The rst action taken by each PE during garbage collection is to deallocate the cache pages (using munmap), since these pages will become inconsistent during garbage collection. Special care must be taken with hiload requests queued on a cache page. These hiload requests are dispatched to the owner PE, where they are queued on the speci c locations that they refer to. This communication can amortized by bundling it with other communication that is necessary during GC anyway. Next, each PE allocates (with mmap) a to-space into which local reachable objects will be copied. We expect that physical memory pages previously used for the cache will be reused for the to-space. Each PE maintains two pointers into its to-space in the usual manner: a scan pointer and a free pointer. Objects in to-space below the scan pointer only contain pointers to other objects in to-space, whereas objects above the scan pointer may refer to objects in fromspace. During GC, each PE then calls the function move object on each of its local roots. This function copies an object from the from-space into the to-space. For local objects, this function behaves exactly as in a conventional Cheney collector { it copies the object to the free pointer, incrementes the free pointer and returns the new address of that object. For non-local objects, the function sends a move object message to the owner PE of that object. That PE then copies the object into to-space, and returns an active message that, when invoked, updates the appropriate location in the memory of the original PE with the new address in to-space of the object. Since this operation is implemented in a split-phase manner, the original PE can continue copying other objects while the operation is in progress. 4 Preliminary Performance Measurements We have just completed a rst implementation of the system as described above. We have taken some preliminary performance measurements, but we have yet to perform a detailed analysis of these timings in order to understand exactly where the time goes. Our implementation (both the compiler and the compiled code) runs on Digital Alpha workstations under Digital Unix, using gcc to compile the C \object code" and the runtime system. The parallel system runs on any network of Alpha workstations. Currently, it uses UDP sockets as the basic packet delivery mechanism. UDP sockets are unreliable (packet delivery and packet order are not guaranteed), so we implemented a thin reliability layer above UDP sockets called RPP, for Reliable Packet Protocol this is still advantageous over TCP, which is a reliable protocol, for several reasons: it is packet based, not stream based, and so does not need \packet parsing" by the receiver, and it automatically multiplexes incoming packets from all peers into a single receive queue, instead of having to wait on N receive queues, one per peer. With RPP, It will be trivial for us in the future to exploit faster packet delivery mechanisms, even if they are unreliable. Our parallelism measurements were taken on a particular con guration with eight 225 MHz Alpha workstations with an ATM interconnection network, running Digital Unix. Communication is still rather expensive relative to computation speed. The following table shows the round trip latency from a UDP socket send to a UDP socket receive, for various packet sizes: Packet size Round trip latency (Bytes) (secs) 64 550 128 580 256 680 1024 908 4096 1820 The maximum bandwidth is about 12 MBytes/sec (close to the ATM maximum). Thus, it is not an easy platform on which to get any parallel speedup. For a sequential, uniprocessor baseline, we compared an Id program compiled for a single processor and executed on a single processor, with a conventional (sequential) C program that implements the same algorithm but uses conventional C idioms. The pHluid compiler was given a ag to produce code for a uniprocessor, so that it could optimize away all message-passing code. For the C compilation phase of the Id program, and for the C program, we used the same gcc compiler with the same -O2 optimization level. The program we used does a lot of heap allocation: paraffins(18) enumerates all paran molecules containing up to 18 carbon atoms, unique up to certain symmetry conditions{ the Id program is described in detail in 6]. The C version did not have any garbage collection (it simply had a large enough heap). One version of the C code used malloc() for each heap allocation, and another version did its own allocation out of a large chunk of memory allocated once using sbrk() at the start of the program. paraffins(18) pHluid C (malloc) C (sbrk) Time (secs) 0.53 0.40 0.20 Relative speed 1x 1.3x 2.6x We believe that these indicate that pHluid's uniprocessor performance for Id programs is approaching competitive performance for functional language implementations (the study in 13] measures the performance of another program, the pseudoknot benchmark. It compares a functional version compiled with over 25 functional language compilers, and a C version. The best functional versions were about 1.4x, 1.5x, 1.9x and 2.0x slower than C). The following table shows the elapsed time to call and return from a trivial function that takes a void argument and simply returns a void result (in Id, void is the trivial type with just one value of type void). We measured this for two cases: a local call, where the function executes on the same PE as the caller, and for a directed remote call, where the function is directed, by means of an annotation, to execute on a dierent PE from the caller (this does not involve any work stealing): Local function call 9 secs Remote function call 1987 secs Recall that the remote function call involves allocating a call record containing the function pointer, normal arguments and continuation arguments sending it to the remote PE at the remote PE, reading the message and executing its handler which, in this case, places it on the xed queue executing the scheduler which takes the call record o this queue, allocates a frame, and invokes the function at the xed queue entry point, which unmarshalls the normal and continuation arguments and, follows a similar process for the return. The messages in each direction are of the order of 100 Bytes, whose raw round-trip latency is about 580 secs (from the previous table) thus, the overhead is quite high, about 1400 secs, and is clearly an area with much room for improvement. The following table shows the speed of the hiload and histore heap access operations. All accesses are to full I-structure locations (but they still perform the tests for emptiness, etc.). 1 2 3 4 5 6 Operation (secs) Local hiload 7 Remote hiload, uncached, unmapped 1466 Remote hiload, uncached, mapped 1196 Remote hiload, cached 4 Local histore 5 Remote histore 1047 The rst line describes an hiload to a heap object that is local, i.e., on the same PE. The next three lines describe an hiload to a eld of a remote object, i.e., allocated on a dierent PE. Line 2 is for the case when the remote object has not yet been locally cached, and the page containing that object has not yet been mapped into the local address space. In line 3, the object is not yet locally cached, but it is on a page that has been locally mapped (due to an earlier remote access to some other location on that page). Line 5 is for the case when it has been locally cached. For lines 2 and 3, we send a relatively small request message, and we get back an entire heap page (1 KB in size). The raw UDP communication latencies for this are about 225 + 454 secs. Thus, there is still about 517 secs overhead beyond the raw communication cost to fetch and install a cache copy of a heap page, plus about 270 secs if we have to map the page in. Line 4 just shows that hiloads to cached data proceed much faster (we do not know why it is faster than the the local hiload in line 1). Similarly, lines 5 and 6 are for a local and remote histore, respectively. A remote histore involves a round-trip communication (the return message is treated as an acknowledgement that the histore has completed). We have been able to measure parallel speedups only for some small programs because of a current problem in an early phase of the pHluid compiler (which was implemented before the parallel runtime system) wherein list comprehensions are translated into sequential loops, thus eliminating parallelism (for example, our paraffins program uses list comprehensions heavily, and so we have not yet been able to measure its parallel performance). Figure 2 shows speedups for three small programs. The trees program is a synthetic benchmark that creates a balanced binary tree to a given depth, copies the tree several times, counts the leaves and returns the count. nqueens(12) is the standard N-queens problem with a board size of 12. matrix-multiply(500) creates two 500x500 oating point matrices, multiplies them, and sums the resulting matrix. The last program uses explicit distribution of work, directing a sub-matrix computation to be performed on a speci c PE. These gures were measured on a testbed system with \only" eight processors. Eight processors is certainly well within the range of SMPs, and one may argue that there is no need to go to a distributed memory implementation. However, we believe that it is important to look at distributed memory systems even at this size, for several reasons. First, eight-processor workstation clusters are a lot more common, and cheaper, than eightprocessor SMPs. Further, there is a growing trend away from massively parallel multicomputers towards achieving the same power by assembling small-degree clusters of small-degree SMPs. Although these gures are for relatively small programs and do not supply any diagnostic detail about how and where time is spent, we nd them encouraging considering that we are using such an expensive communication layer. It would not have been surprising to see very little speedup at all, or even a slowdown. It appears that our runtime system mechanisms for avoiding, eliminating and tolerating communication are eective. We hope to do much more detailed performance studies in the near future. 5 Related Work We are not aware of any other implementations of conventional shared memory functional languages for distributed memory machines, that address the cost of remote communications as aggressively as pHluid does. There have of course been several implementations on shared memory machines (such as 16] and 7]), and implementations of message-passing functional languages on distributed memory machines (such as Erlang 5]). However, implementing a shared memory language on a distributed memory machine appears to require a substantially dierent approach, with latency-tolerance a high priority throughout the design. The Berkeley Id/TAM compiler 10] shares the same dataow heritage as our pHluid compiler and, not surprisingly, has many similarities (although they share no code). TAM itself (Threaded Abstract Machine) is somewhat similar to our P-RISC Assembler, but our scheduling discipline for threads is quite dierent (these two scheduling disciplines have been compared in detail by Ellen Spertus at MIT, and is reported in 20]). The Berkeley Id-on-TAM system has been implemented on distributed memory machines with relatively fast communication facilities, such as the Connection Machine CM-5 and the MIT J-Machine, but not to our knowledge on workstation farms. The Berkeley Id/TAM system does not use distributed cacheing, nor does it have a garbage collector. We do not know what mechanism is used for work and data distribution, and if it has any automatic load balancing mechanism. We have become aware of other implementations of functional languages on distributed memory machines| on Transputer networks some years ago, a parallel Haskell at Glasgow over PVM more recently, and a distributed memory implementation of Sisal currently in progress| but we do not know what, if any, latency-tolerance features they use. All these systems start with a compiler optimized for sequential execution, and incrementally modify them for parallel execution. In contrast, the pHluid and Berkeley TAM compilers have latencytolerance at the forefront throughout, and consequently have a very dierent abstract machine model based on ne grain, message driven multithreading. 6 Conclusion In this paper we have described our implementation of a modern shared memory parallel functional language the novelty is in the way we target distributed memory machines, including workstations clusters with relatively poor communication facilities. Here, aggressive latency-tolerance is a central preoccupation, and is achieved by a combination of the compilation strategy (producing ne grain threads), a work-stealing mechanism that avoids distributing work unnecessarily, a distributed coherent cacheing mechanism for the heap that exploits the functional nature of the language, a memory organization that reduces pressure for global garbage collection by managing several dynamically allocated objects purely locally, and a parallel garbage collector. We also presented some preliminary performance measurements. Our pHluid implementation is new and still immature, and there is a lot of work ahead: making it sturdy enough to handle more substantial programs implementing it on distributed memory platforms with faster interconnects (such as Digital's Memory Channel, which achieves communication latencies of less than 5 secs), 9 9 nqueens(12) linear speedup 8 7 6 6 4 Speedup 7 6 5 5 4 5 4 3 3 3 2 2 2 1 1 1 0 0 0 1 2 3 4 5 6 Number of Processors 7 8 matrix multiply linear speedup 8 7 Speedup Speedup 9 trees linear speedup 8 0 0 1 2 3 4 5 6 Number of Processors 7 8 0 1 2 3 4 5 6 Number of Processors 7 8 Figure 2. Parallel speedups of some small Id programs taking detailed performance measurements and characterizing the system to understand the bottlenecks improving the heap cacheing model to be more economical in memory use (to avoid each PE having to map the full heap) exploiting SMPs as cluster nodes (i.e., using shared memory when available) adding support for large distributed arrays (currently no object, including arrays, may be larger than one PE's heap) etc. Acknowledgements: The \P-RISC to Gnu C" phase of the compiler was originally designed and implemented on Sparcstations by Derek Chiou of MIT's Lab for Computer Science. We ported this implementation to our Digital Alpha workstations. The work-stealing algorithm with linear backo was jointly developed with Martin Carlisle of Princeton University he implemented it in Cid, another parallel language implementation at Digital CRL. References 1] Aditya, S., Arvind, Augustsson, L., Maessen, J.-W., and Nikhil, R. S. Semantics of pH: A Parallel Dialect of Haskell. In in Proc. Haskell Workshop (at FPCA 95), La Jolla, CA (June 1995). 2] Agarwal, A., Simoni, R., Hennessy, J., and Horowitz, M. An Evaluation of Directory Schemes for Cache Coherence. In Proc. 15th. Ann. Intl. Symp. on Computer Architecture, Hawaii (May 1988). 3] Appel, A. Garbage Collection can be Faster than Stack Allocation. Information Processing Letters 25, 4 (1987), 275{279. 4] Appel, A., and MacQueen, D. B. A Standard ML Compiler. In Proc. Conf. on Functional Programming and Computer Architecture, Portland, Oregon (September 1987). Springer-Verlag LNCS 274. 5] Armstrong, J., Virding, R., and Williams, M. Concurrent Programming in Erlang. Prentice Hall, 1993. ISBN: 0-13-285792-8. 6] Arvind, Heller, S., and Nikhil, R. S. Programming Generality and Parallel Computers. ESCOM Science Publishers, P.O.Box 214, 2300 AE Leiden, The Netherlands, 1988, pp. 255{286. Proc. 4th Intl. Symp. on Biological and Arti cial Intelligence Systems, Trento, Italy, September 1988. 7] Augustsson, L., and Johnsson, T. Parallel Graph Reduction with the <nu,G>-machine. In Proc. Fourth Intl. Conf. on Functional Programming Languages and Computer Architecture, London (September 1989), pp. 202{213. 8] Blumofe, R. D., Joerg, C. F., Kuszmaul, B. C., Leiserson, C. E., Randall, K. H., and Zhou, Y. Cilk: An Ecient Multithreaded Runtime System. In Proc. 5th. ACM Symp. on Principles and Practice of Parallel Programming (PPoPP), Santa Barbara, CA (July 19-21 1995), pp. 207{216. 9] Cheney, C. J. A Nonrecursive List Compacting Algorithm. Communications of the ACM 13, 11 (November 1970), 677{678. 10] Culler, D. E., Goldstein, S. C., Schauser, K. E., and von Eicken, T. v. TAM { A Compiler Controlled Threaded Abstract Machine. J. Parallel and Distributed Computing, Special Issue on Dataow 18 (June 1993). 11] Gaudiot, J.-L., and Bic (editors), L. Advanced Topics in Data-ow Computing. Prentice Hall, 1991. 12] Halstead, R. Multilisp: A language for concurrent symbolic computataion. ACM Trans. Program. Lang. Syst. 7, 4 (1985), 501{538. 13] Hartel, Pieter, H., Feeley, M., Alt, M., Augustsson, L., Baumann, P., Beemster, M., Chailloux, E., Flood, C. H., Grieskamp, W., van Groningen, J. H. G., Hammond, K., Hausman, B., Ivory, M. Y., Jones, R. E., Lee, P., Leroy, X., Lins, R. D., Loosemore, S., Rojemo, N., Serrano, M., Talpin, J.-P., Thackray, J., Thomas, S., Weis, P., and Wentworth, E. P. Benchmarking Implementations of Functional Languages with \Pseudoknot", a FloatIntensive Benchmark. In Workshop on Implementation of Functional Languages, J. R. W. Glauert (editor), School of Information Systems, Univ. of East Anglia (September 1994). 14] Hudak, P., Peyton Jones, S., Wadler, P., Boutel, B., Fairbairn, J., Fasel, J., Guzman, M. M., Hammond, K., Hughes, J., Johnsson, T., Kieburtz, R., Nikhil, R., Partain, W., and Peterson, J. Report on the Programming Language Haskell, A Non-strict, Purely Functional Language, Version 1.2. ACM SIGPLAN Notices 27, 5 (May 1992). 15] Kendall Square Research. Kendall Square Research Technical Summary, 1992. 16] Kranz, D. A., Halstead Jr., R. H., and Mohr, E. Mul-T: A High Performance Parallel Lisp. In Proc. ACM Symp. on Programming Language Design and Implementation, Portland, Oregon (June 1989). 17] Lenoski, D., Laudon, J., Gharachorloo, K., Weber, W.-D., Gupta, A., Hennessy, J., Horowitz, M., and Lam, M. S. a. The Stanford DASH Multiprocessor. IEEE Computer (March 1992), 63{79. 18] Nikhil, R. S. A Multithreaded Implementation of Id using P-RISC Graphs. In Proc. 6th. Ann. Wkshp. on Languages and Compilers for Parallel Computing, Portland, Oregon, Springer-Verlag LNCS 768 (August 12-14 1993), pp. 390{405. 19] Nikhil, R. S. An Overview of the Parallel Language Id (a foundation for pH, a parallel dialect of Haskell). Tech. Rep. Draft, Digital Equipment Corp., Cambridge Research Laboratory, September 23 1993. 20] Spertus, E., and Dally, W. J. Evaluating the Locality Bene ts of Active Messages. In Proc. 5th. ACM SIGPLAN Symp. on Principles and Practice of Parallel Programming (PPoPP), Santa Barbara, CA (July 19-21 1995), pp. 189{198. 21] von Eicken, T., Culler, D. E., Goldstein, S. C., and Schauser, K. E. Active Messages: a Mechanism for Integrated Communication and Computation. In Proc. 19th. Ann. Intl. Symp. on Computer Architecture, Gold Coast, Australia (May 1992), pp. 256{266.