Skip to content

Instantly share code, notes, and snippets.

@Quodss
Last active March 5, 2024 20:00
Show Gist options
  • Save Quodss/a1aaa81941e61707843a75d45d901ea0 to your computer and use it in GitHub Desktop.
Save Quodss/a1aaa81941e61707843a75d45d901ea0 to your computer and use it in GitHub Desktop.

urwasm: WebAssembly interpreter suite on Urbit


WebAssembly is a low-level language for a portable virtual machine. Wasm is designed to be a compilation target for a variety of programming languages and its design is hardware independent and relatively simple, making its support ubiquitous in modern browsers. Its simple design made it a perfect first candidate for a first emulator of an conventional computational system on a novel functional computer: Urbit. In this paper I discuss the current state of urwasm project and some technical details, as well as describe the strategy to jet the interpreter of a state machine in a functional environment.


Introduction

... Martian code is so perfect that it cannot be contaminated, even in this sticky environment. The general structure of cross-planet computation is that Earth always calls Mars; Mars never calls Earth. The latter would be quite impossible, since Earth code is non-Maxwellian. There is only one way for a Maxwellian computer to run non-Maxwellian code: in a Maxwellian emulator. Any direct invocation is an implementation error by definition. Thus, Mars remains completely pure and Martian, even while glued firmly to Earth.

— moronlab.blogspot.com, Urbit: functional programming from scratch [1]

Urbit computer is specified as a Nock interpreter and a storage of an event log, represented as a list of nouns [2]. Nock is a Lisp-like typeless functional language with a tiny specification [3], which defines a function nock(sub, fol), where subject sub and formula fol are both nouns: binary trees of unsigned integers with an arbitrary length, or "S-expressions without an S". The state of the Urbit computer is then defined as nock(epic, [2 [0 3] [0 2]]), where epic is the event log.

This environment might seem very limiting at the first glance, but, by feeding proper events to this computer (Hoon-to-Nock compiler, Arvo OS written in Hoon etc.) we can create a personal server used by thousands of people today, which can host Web applications and communicate with other Urbit computers.

If we desire to run conventional programs on Urbit, we would have to build an interpreter in Nock that would evaluate the imported code within Urbit. A great candidate language for having an interpreter in Nock is WebAssembly (Wasm):

  • It is a low-level language, supported by many languages as a compilation target;
  • Its specification is small and hardware-independent when compared to other alternatives;
  • It is widely used in browsers as well as in other contexts, creating an incentive for developers to add Wasm support in their projects.

For this reason, I set out to build a Wasm interpreter on Urbit: urwasm [4]. At this moment, the project contains:

  • A parser from .wasm binary format to a noun;
  • An interpreter, which fully satisfies Wasm Core Specification from the latest published version [5];
  • A parser from .wat text format to a noun, implemented by calling a Wasm module with a wat2wasm parser and then feeding the result to the binary parser. This parser serves as a testing method and is very slow with the interpreter being unjetted.

In the first part of this paper I will briefly describe the implementation of the interpreter in urwasm. In the second part I will describe the jetting strategy for the interpreter, discuss the need for an interpreter of Language for Invocation of Assembly (Lia) that would encapsulate the Wasm interpreter, and sketch out the specification of Lia language and ++lia interpreter.

Wasm interpreter implementation in Hoon

The strategy for interpreting Wasm expressions was to model each instruction as a function $$local\textnormal{-}state \rightarrow local\textnormal{-}state,$$ where local-state is a noun which describes the state of the interpreter at any point during the computation: stack and local values, linear memory etc. An expression is a list of instructions, and can be modeled as a composition of all instructions in the list, which also makes it a function from local-state to local-state. Most of the instructions were implemented as gates with that type, while five instructions, call, call_indirect, block, loop and eval were treated directly in the evaluation loop.

The evaluation loop consists mainly of two functions: ++eval and ++apply. The first one takes an expression and local-state, producing local-state. It pops an instruction from the expression and applies it to the local state with ++apply, stopping if it reaches the end of the list or if an instruction triggered execution flow navigation. For example, return instruction would cause ++eval to stop applying the rest of the instructions and return the local state as is, with the branching coordinate contained in the local state.

image
++eval data flow

++apply gate takes an instruction and local-state, returning local-state. While for most of the instructions it would find an appropriate gate $-(local-state local-state) and apply it to the given local state, in the case of five instructions listed above it would call ++eval on the body of those instructions. Thus stack frames are divided not with a label value, like in the formal specification of Wasm, but with the depth of mutual recursion of ++eval and ++apply.

image
++apply data flow. Some details like entering/exiting frames are elided

The gates above are internal, and the outside caller would usually interact with two other gates, ++prep and ++invoke. The first one instantiates the module: loads the data segments into the linear memory, instantiates global values and runs start function if specified. The product of this gate is either global-state, which describes the state of the Wasm module in between function invocations, or a block on an unresolved import, or a crash or "trap" in WebAssembly terms.

++invoke takes a name of an exported function, input values with type annotation and global-state. It performs a typecheck on the supplied values an calls an appropriate function. The result, similarly to ++prep, is either a success with output values and updated state, a block or a trap.

Determinism of WebAssembly VM

The state of an Urbit computer, as already mentioned above, is a pure function of its stream of events, with the lifecycle function defined in Nock. No information other than the event log may impact the computation of this state, including the hardware on which the interpreter runs, or the implementation details of the interpreter. For a Wasm interpreter in Nock to be practical, it needs to be paired with some fast implementation of Wasm interpreter in another language, and the equivalence of both algorithms needs to be validated. That way, the Urbit computer can be a practical personal server without violating its simple definition: for an Urbit programmer the code of functions in Nock must be enough to reason about the system.

The demand of equivalence, however, poses some difficulties. Some instructions of Wasm are defined non-deterministically. For example, memory.grow instruction can return -1 and not increase the length of the linear memory even if the maximum size of the memory described in the module file is not exceeded. The choice is left to the embedder, which can choose to keep the size of the memory buffer based on its resources.

In addition, some numerical operators return a set of values, that is, a Wasm VM could return any value from that set. This typically happens with floating-point operators when they are given a NaN value, with the set being defined by the value being a canonical NaN or not.

Some numerical operators return an empty set of values. In other words, their behaviour is set to be undefined for certain inputs, and those operators are described as partial. This can happen, for example, with idiv_N operator when the second operand is equal to zero. The behaviour of Wasm VM is undefined in this case.

Therefore, Wasm interpreter in urwasm describes a deterministic Wasm machine, limited to a subset of behaviors:

  • If an empty set of values must be returned, the interpreter "traps": returns a deterministic error as a result, similar to ++mink;
  • If a set of values must be returned, then a single result from that set is returned. The choice was particular to each operator, but typically, if the set was a union of sets that depend on multiple input parameters, then the choice was made from the set given by the first parameter. Refer to /lib/op-def.hoon for more information.
  • memory.grow and similar instructions always attempt to grow the buffer when the limits in the module file allow to do that.

Examples of programs evaluated with urwasm

Some simple programs are included in the test suite. Most of them involve only numerical operations, while one of them calls a Wasm program compiled from Rust source which flips the order of characters in the string. Evaluation of this function requires multiple Wasm function invocations in order to e.g. allocate memory for the input string, and interaction with the global-state to read from and write to linear memory.

When Rust source is compiled, the compiler produces Javascript code which takes care of low-level handling of the state of Wasm runtime. In the case of Wasm interpreter in Hoon, JS code was manually translated to Hoon.

In addition, some other programs were run successfully. WebAssembly text format parsing in urwasm is implemented by composing wat2wasm parser from wat Rust crate with a binary parser in Hoon. Another example was presented at the 2023 Urbit Hackathon: an algorithm to simplify debts between a group of people, using FlowGraph Dinic algorithm for maximum flow computation from contest_algorithms Rust crate [6].

All examples above highlight incredible inefficiency of the interpreter: it takes about a second to flip a string with 27 characters, and about a minute to parse a small Wasm binary with compiled wat2wasm. While there is most likely room for improvement, the Hoon code serves first and foremost as a formal specification of a deterministic Wasm VM, translated from the mix of pseudocode and mathematical formulas to a tightly-specified language. This specification would then serve for the verification of jet correctness.

Jetting

To jet the interpreter, the gate written in Hoon must be paired with a code in C that must be extensionally equivalent to the Hoon code. In that sense function definition in Hoon would act as a formal mathematical specification of what the interpreter returns, while C code would act as the actual implementation of Wasm runtime by arriving to the same conclusion as Hoon code but faster. The purpose of the Hoon specification is then to provide a test bed for verifying correctness of the interpreter, either by testing or formal analysis of both programs via e.g. K verification framework [7].

Since Wasm is a portable language for a state machine, each invocation of a function from a Wasm module would either return a successful result with returned values and an updated state, or some flavor of failure (trap or blocking on unresolved external request, e.g. function call of an imported function). This gives us several possible strategies for jetting:

  • Have a bespoke Wasm interpreter in C that operates on nouns, and jet invoke gate;
  • Use an established Wasm runtime in C, and add serializer/deserializer to the jet, to convert Hoon representation of the module state to a representation in C and vice versa, and jet invoke gate;
  • Use an established Wasm runtime in C, and add serializer/deserializer to the Hoon specification, and use a representation of state close to the one in C as input and output in Hoon invoke gate, and jet invoke gate;
  • Don't jet the invoke gate. Instead, have a higher level function that executes a series of operations on a module and doesn't return the entirety of Wasm module state. Hoon specification of this function would use the Wasm interpreter in Hoon, and the jetting code in C would use Wasm runtime in C.

But first, why do we have to have access to the state in the first place?

Even for a simple source code the generated Wasm code might require multiple function invocations to get a desired result. Consider this Rust code of a function that flips the characters in a given string:

use wasm_bindgen::prelude::*;

#[wasm_bindgen]
pub fn process(input: String) -> String {
     let output_string: String = input.chars().rev().collect();
     output_string
}

After compiling this function to Wasm, you would get a module with five exported functions: process itself, add_to_stack_pointer, malloc, realloc and free. The call of the compiled function in JS would look like:

export function process(input) {
    let deferred2_0;
    let deferred2_1;
    try {
        const retptr = wasm.__wbindgen_add_to_stack_pointer(-16);
        const ptr0 = passStringToWasm0(input, wasm.__wbindgen_malloc, wasm.__wbindgen_realloc);
        const len0 = WASM_VECTOR_LEN;
        wasm.process(retptr, ptr0, len0);
        var r0 = getInt32Memory0()[retptr / 4 + 0];
        var r1 = getInt32Memory0()[retptr / 4 + 1];
        deferred2_0 = r0;
        deferred2_1 = r1;
        return getStringFromWasm0(r0, r1);
    } finally {
        wasm.__wbindgen_add_to_stack_pointer(16);
        wasm.__wbindgen_free(deferred2_0, deferred2_1, 1);
    }
}

Multiple invocations are necessary, some with the arguments received from outputs of other invoked functions. Having the access to the state of the module is thus crucial for any practical interpreter. In addition, we may have to read data from the state and save it as a noun, in order to perform I/O on that piece of information, since Arvo may only send effects when the computation of a given event has finalized. While jets could in theory perform system calls during their evaluation, doing so is considered a gross violation of jetting mechanism: jets as functions are supposed to be as pure as possible, to imitate their Nock definitions.

Now, let's quickly go through our options:

Bespoke Wasm interpreter that operates on nouns

Maybe it is the perfect option in long term, but this option is unfeasible in short/mid term due to it being a huge sink of dev time. In addition, it would require different kinds of optimizations added to a Nock interpreter that do not exist yet, like "data jets", when a noun is represented in a way that allows to perform certain operations faster. For example, Nock runtime could have a list of bytes (list @D) represented as a byte array in its memory, with some jetted functions configured to operate on that kind of nouns more efficiently.

Serializer/deserializer in the jet

The first solution that comes to mind is to add a serializer and deserializer functions to the jet of ++invoke arm, which would convert the noun representation of the state of the module to a representation legible to the jetting interpreter before evaluating the invoked function, and convert it back into a noun when the computation is finished. However, having to translate the state between two models twice for each function invocation would impose a lot of computational overhead.

Serializer/deserializer in the Hoon spec

A less obvious variation of the method above is to put the de/serializer in the Hoon code of ++invoke, and have the jetted gate take and return module state in the same representation as in the Wasm interpreter in C. The jet would then operate on the given state directly, without having to translate the state of the interpreter. Here the problem is that the implementation strategy leaks into the formal specification, making Hoon code jet-dependent. Replacement of the jetting Wasm runtime, including in the case of switching from Vere to Ares, would make us have to change the Hoon code, which I find to be antithetical to the Urbit project itself.

In addition, current model of jetting in both Vere and Ares requires that the core produced by the jetted arm is left read-only, without modification, so the output of the jet function cannot overwrite the core's sample. That would require the interpreter to copy the entire state of the module and operate on that, which creates prohibitive overhead for memory-heavy computations, like running a Linux VM.

Higher level interpreting function

This strategy would involve writing a function that takes a Wasm module, a list of operations to be performed and some other parameters, and then jetting this function. No intermediate state would be returned, saving us from having to convert it between different representations. However, a practical jet implementation would have to cache the intermediate state of the interpreter between Arvo events, otherwise it would have to reevaluate the operations each time the jetted function encounters a block on an unresolved import, or if the caller appended new operations to the list to interact with the module.

Lia interpreter

Let's return to our example with string flipping in Rust. How would our hypothetical higher-level function run the program?

Ignoring imports for now, consider this gate:

++  lia
  |=  [module=octs actions=(list action)]
  ^-  $%  [%0 out=(list value)]
          [%2 ~]
      ==
  ::  (...)

This gate would take a binary file of a Wasm module and a list of actions to be performed, and return either a list of values in the event of successful resolution, or an error. A value here is either a WebAssembly numerical value or a slice of the linear memory called octs in homage to Hoon. To flip a string, we would pass it the module file obtained from the Rust compiler and a list of actions obtained from parsing this code (ignoring realloc for simplicity):

# Lia (Language for Invocation of Assembly) scripting language
#
# "add_to_stack_pointer", "malloc" and "process"
# are functions exported by the wasm module from Rust example above
#
string0 = "Hello world".to_octs
retptr = __wbindgen_add_to_stack_pointer(-16)
ptr0 = __wbindgen_malloc(string0.len, 1)
memory.write(string0, ptr0)
process(retptr, ptr0, len0)
i32 r0 = memory.read(retptr, 4)
i32 r1 = memory.read(retptr+4, 4)
return memory.read(r0, r1)

This code is written in Language for Invocation of Assembly, Lia for short. I imagine it being a language with a tiny specification, whose only purpose it to describe sequences of actions to be performed with a Wasm module: function invocation, memory read/writes, as well as variable declarations, for loops and conditionals for expressivity.

The code above is essentially identical to the JS example. Ideally, it would be generated by the compiler itself, just like the JS code was also generated, with placeholders for input values, where the caller could insert their parameters.

Jet of ++lia would have to perform the same operations but with an interpreter in C. But considering the tiny spec of Lia, implementing it in both Hoon and C is not going to be a challenge.

Until now I described a non-interactive version of Lia interpreter: it takes a module and a list of actions and returns a result. But the WebAssembly module can and should be interactive: we want to invoke a function, perform some I/O based on the result, then invoke a function again etc. In addition, the module might contain imported functions, which we want to resolve to provide I/O to the Wasm program.

Interacting with an instantiated module in the case of ++invoke jetting is straightforward: the state of the module is a noun, and to interact with the state we would call various jetted gates that invoke functions and perform I/O on the store. Handling import function calls is also straightforward: ++invoke might return a blocked result with the name of the imported function, which is then resolved in the embedding context, modifying state of the module if necessary. How would the same be achieved in a stateless fashion?

Let's return to the Lia interpreter, this time with some additions:

++  lia
  |=  $:  module=octs
          actions=(list action)
          shop=(list (list value))
          ext-func=(map (pair cord cord) (list action))
          diff=(each (list action) (list value))
      ==
  ^-  $%  [%0 out=(list value)]
          [%1 name=path args=(list value)]
          [%2 ~]
      ==
  =>
    ?:  ?=(%.y -.diff)
      .(actions (weld actions p.diff))
    .(shop (snoc shop p.diff))
  ::  (...)

Now the interpreter can return a block with %1, asking for an external reference to be resolved and the result of the resolution to be appended to shop. The interpreter now takes a list of list of values shop, which represents the resolved Lia imports, ext-func, which provides definitions for imported functions for a given Wasm module, and diff, whose contents are appended to an appropriate field.

Notice the difference between imports in Lia and in Wasm: the namespace of the latter is defined with a pair of strings in compliance with the formal specification, and the input parameters of ++lia contain definitions of the imported functions for the Wasm module. Lia imports, on the other hand, are bound to path namespace and are external to ++lia.

In a trivial case the list of actions of a function definition in ext-func would contain a single call to a function which is external to Lia and is named with path, not (pair cord cord) like Wasm import functions. In a nontrivial case the list of actions could contain multiple actions to be performed, e.g. memory read followed by calling a Lia import function which takes octs. The purpose here is for the import calls to be able to surface not only Wasm values but Lia values like octs, giving us the richness of import calls that we could have with ++invoke jetting model without exposing the entirety of the module's state.

Caching of the store

Since Wasm execution is deterministic, then the state of a module, whether in Nock or in the jetting environment, is referentially transparent with regards to the input parameters of Lia. Here diff appears to be semantically useless, since instead of placing changes there we could've placed them directly into a proper field of the sample.

But diff is necessary for efficient computations: each time ++lia jet computes something with a Wasm runtime, it will save a cache of Lia interpreter in the jetting environment tagged with a (hash of a) noun [module actions shop ext-func]. If ++lia is computed later with the same first four arguments, then the jet would first look for a cache, and if it finds one, only compute the diff, either injecting results of Lia import resolution into a suspended jetting Lia interpreter, or performing a list of appended actions. Failing to find a cache, jet of ++lia would do the same thing as ++lia in Hoon: append the diff to an appropriate field and run the whole thing.

At the price of having to bother with cache reclamation and perhaps slightly longer event log replays we get the possibility of running Wasm runtime almost full speed, without having to de/serialize nouns to structs and vice versa or to copy the state of the Wasm module for every interaction with it. In Nock the state of the Wasm module is represented lazily with the first four parameters of ++lia, while the actual state on which we perform computations exists as a cache.

A typical interaction with ++lia would look like this:

  1. Call ++lia for the first time, with an empty diff,
  2. Either Lia returns a success %0 or a failure %2, or
    1. It returns a block %1, which is resolved outside of Lia.
    2. The resolution result is placed into diff, and ++lia is called again. Next time that same result must be placed directly to shop to get the right cache and avoid recalculating from scratch.
  3. If succeeded, new actions can be put in diff and ++lia can be called again to continue interacting with the module. After that these new actions must be placed directly to actions on subsequent calls of ++lia to get the right cache.

As for the cache reclamation, ++lia could include another input parameter hint ignored in Hoon code, that would serve as a hint to the jet on how to handle the cache, e.g. for how long to keep it in the memory. Dojo generators and Gall agents, for example, would likely benefit from different cache reclamation strategies.

Conclusion

The roadmap for urwasm project:

  • First priority:
    • Add unit tests and module validation in Hoon;
    • Create a Lia interpreter in Hoon and jet with Lia interpreter in C/Rust;
  • Next:
    • Enable caching in the jet of ++lia;
    • Add Lia parser to turn text to (list action)
    • Create default Gall agent interface to interact with Lia to enable building Gall apps in conventional languages;
  • Finally:
    • Add Wasi support;
    • Add Lia generation to common Wasm compilers in the same way they generate JS wrappers for the modules.

This paper presented a novel method of modelling of state machines in Nock and jetting of the modelling algorithm. In this method the state is represented implicitly through a list of actions performed on the initialized state of the machine, and the state of the jetting emulator of the state machine is kept as a cache identified with the list of actions. This gives the interpreter full computational speed due to lack of overhead caused by translating the state from noun representation to the jetting representation and back and by the fact that the jetting function is free to edit the state in place, as long as it updates the naming tag of the cache.

References

  1. https://moronlab.blogspot.com/2010/01/urbit-functional-programming-from.html
  2. https://media.urbit.org/whitepaper.pdf
  3. https://docs.urbit.org/language/nock/reference/definition
  4. https://github.com/Quodss/urwasm
  5. https://www.w3.org/TR/2022/WD-wasm-core-2-20220419/
  6. https://www.youtube.com/watch?v=h5V_gX33RiM
  7. https://github.com/runtimeverification/knock
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment