package bin_prot

  1. Overview
  2. Docs
A binary protocol generator


Dune Dependency






Part of Jane Street's Core library The Core suite of libraries is an industrial strength alternative to OCaml's standard library that was developed by Jane Street, the largest industrial user of OCaml.

Published: 23 May 2024


Bin_prot - OCaml Type Converter for Binary Protocols

What is Bin_prot?

This library contains functionality for reading and writing OCaml-values in a type-safe binary protocol. It is extremely efficient, typically supporting type-safe marshalling and unmarshalling of even highly structured values at speeds sufficient to saturate a gigabit connection. The protocol is also heavily optimized for size, making it ideal for long-term storage of large amounts of data.

The library is highly dependable and safe to use: a rigorous test suite has to date guaranteed that this library has never exhibited a bug in production systems in several years of use. Bin_prot has been successfully deployed in mission-critical financial applications, storing many terabytes of structured data derived from thousands of type definitions and typically processing millions of messages a day in realtime for low-latency applications that must not crash.

Since version two this library should work with all CPU architectures currently supported by OCaml, no matter the word size (32 or 64 bit), alignment requirements, or endianness. Endianness defines the byte order in which machine representations of integers (machine words) are stored in main memory.

Bin_prot provides users with a convenient and safe way of performing I/O on any extensionally defined OCaml type (see later sections for details). Functions, objects, first-class modules, as well as values whose type is bound through a polymorphic record field are hence not supported. This is hardly ever a limitation in practice.

As of now, there is no support for cyclic or shared values. Cyclic values will lead to non-termination whereas shared values, besides requiring more space when encoded, may lead to a substantial increase in memory footprint when they are read back. It would not be trivial to support these kinds of values in a type-safe way without noticeably sacrificing performance. If these kinds of values are needed, the user may want to use the as of today still unsafe marshalling functions provided by OCaml.

This library uses the machine stack for efficiency reasons. This can potentially lead to a crash if the stack limit is reached. Note that this is also a limitation of the (unsafe) standard marshalling functions shipped with OCaml. This problem can happen for large amounts of data if recursion in the type definition of the data structure is not limited to the last element. Only in the latter case will tail-recursion allow for (practically) unlimited amounts of data. If this exceedingly rare limitation ever turned out to be an issue in a user application, it can often be solved by redefining the data type to allow for tail-recursion. The limitation cannot be eliminated in this library without significant performance impact and increased complexity.


The API (.mli-files) in the bin_prot library directory (lib) is fully documented, and HTML-documentation can be built from it on installation. The documentation for the latest release can also be found online.

Module Common defines some globally used types, functions, exceptions, and values. Nat0 implements natural numbers including zero.

Modules Read_ml and Write_ml contain read and write functions respectively for all basic types.

The module Size allows you to compute the size of basic OCaml-values in the binary representation before writing them to a buffer. The code generator will also provide you with functions for your user-defined types.

Module Std predefines converters for most standard OCaml types. If you use the preprocessor macros to generate code from type definitions, make sure that the contents of this module is visible by e.g. adding the following at the top of files using this library:

open Bin_prot.Std

Note that you can shadow the definitions in the above module in the unlikely event that the predefined ways of converting data are unsatisfactory to you.

The module Type_class contains some extra definitions for type classes of basic values. These definitions can be passed to the function bin_dump in module Utils to marshal values into buffers of exact size using the binary protocol. However, if bounds on the size of values are known, it is usually more efficient to write them directly into preallocated buffers and just catch exceptions if the buffer limits are unexpectedly violated. Doing so should never cause a crash. That way one does not have to compute the size of the value, which can sometimes be almost as expensive as writing the value in the first place.

In module Utils the function bin_read_stream can be used to efficiently read size-prefixed values as written by bin_dump with the header flag set to true. This module also offers several useful functors. The ones for Binable types help users create readers and writers if a type needs to be converted to or from some intermediate representation before marshalling or after unmarshalling respectively. The functors for Iterable types are helpful if some (usually abstract) data type offers iteration over elements and if the series of iterated-over elements alone is sufficient to reconstruct the original value. This allows for a more compact protocol and for robustness against changes to the internal representation of the data type (e.g. sets, maps, etc.).


Consider the following type definition:

type t = A | B [@@deriving bin_io]

This will generate the functions bin_size_t, bin_write_t, and bin_read_t, as well as the type class values bin_writer_t, bin_reader_t, and bin_t. If you use the annotation bin_write instead of bin_io, then only the write and size functions and their type class will be generated. Specifying bin_read will generate the read functions and associated type class only. The annotation bin_type_class will generate the combined type class only, thus allowing the user to easily define their own reader and writer type classes. The code generator may also generate low-level entry points used for efficiency or backtracking.

The preprocessor can also generate signatures for conversion functions. Just add the wanted annotation to the type in a module signature for that purpose.

Specification of the Binary Protocol

The binary protocol does not contain any data other than the minimum needed to decode values. This means that the user is responsible for e.g. writing out the size of messages themselves if they want to be able to preallocate sufficiently sized buffers before reading. The Utils module provides some simple functions for that matter, though users may obtain optimum efficiency by managing buffers themselves.

Basic OCaml-values are written out character-wise as described below. The specification uses hex codes to define the character encoding. Some of these values require size/length information to be written out before the value (e.g. for lists, hash tables, strings, etc.). Size information is always encoded as natural numbers (Nat0.t). The little-endian format is used in the protocol for the contents of integers on all platforms.

The following definitions will be used in the encoding specifications below:

CODE_NEG_INT8  ->  0xff
CODE_INT16     ->  0xfe
CODE_INT32     ->  0xfd
CODE_INT64     ->  0xfc


This type encodes natural numbers including zero. It is frequently used by Bin_prot itself for encoding size information for e.g. lists, arrays, etc., and hence defined first here. Developers can reuse this type in their code, too, of course.

If the value of the underlying integer is lower than a certain range, this implies a certain encoding as provided on the right hand side of the following definitions:

<  0x000000080  ->  lower 8 bits of the integer                     (1 byte)
<  0x000010000  ->  CODE_INT16 followed by lower 16 bits of integer (3 bytes)
<  0x100000000  ->  CODE_INT32 followed by lower 32 bits of integer (5 bytes)
>= 0x100000000  ->  CODE_INT64 followed by all 64 bits of integer   (9 bytes)

The last line in the definitions above is only supported on 64 bit platforms due to word size limitations.

Appropriate exceptions will be raised if there is an overflow, for example if a 64 bit encoding is read on a 32 bit platform, or if the 32 bit or 64 bit encoding overflowed the 30 bit or 62 bit capacity of natural numbers on their respective platforms. The last kind of overflow is due to OCaml reserving one bit for GC-tagging and the sign bit being lost.

Unit values

()  ->  0x00


false  ->  0x00
true   ->  0x01


First the length of the string is written out as a Nat0.t. Then the contents of the string is copied verbatim.


Characters are written out verbatim.


This includes all integer types: int, int32, int64, nativeint. If the value is positive (including zero) and if it is:

<  0x00000080  ->  lower 8 bits of the integer                     (1 byte)
<  0x00008000  ->  CODE_INT16 followed by lower 16 bits of integer (3 bytes)
<  0x80000000  ->  CODE_INT32 followed by lower 32 bits of integer (5 bytes)
>= 0x80000000  ->  CODE_INT64 followed by all 64 bits of integer   (9 bytes)

If the value is negative and if it is:

>= -0x00000080  ->  CODE_NEG_INT8 followed by lower 8 bits of integer (2 bytes)
>= -0x00008000  ->  CODE_INT16 followed by lower 16 bits of integer   (3 bytes)
>= -0x80000000  ->  CODE_INT32 followed by lower 32 bits of integer   (5 bytes)
<  -0x80000000  ->  CODE_INT64 followed by all 64 bits of integer     (9 bytes)

All of the above branches will be considered when converting values of type int64. The case for CODE_INT64 will only be considered with types int and nativeint if the architecture supports it. int32 will never be encoded as a CODE_INT64. Appropriate exceptions will be raised if the architecture of or the type requested by the reader does not support some encoding or if there is an overflow. An overflow can only happen with values of type int, because one bit is reserved by OCaml for the GC-tag again.

The reason for this peculiar encoding is of statistical nature. It was assumed that small or positive numbers are much more frequent in practice than big or negative ones. The code is biased accordingly to achieve good compression and decoding performance. For example, a positive integer in the range from 0 to 127 requires only a single byte on the wire and only a single branch to identify it.


Floats are written out according to the 64 bit IEEE 754 floating point standard, i.e. their memory representation is copied verbatim.

References and lazy values

Same as the binary encoding of the value in the reference or of the lazily calculated value.

Option values

If the value is:

None    ->  0x00
Some v  ->  0x01 followed by the encoding of v

Tuples and records

Values in tuples and records are written out one after the other in the order specified in the type definition. Polymorphic record fields are supported unless a value of the type bound by the field were accessed, which would lead to an exception.

Sum types

Each of the n tags in a sum type is assigned an integer from 0 to n - 1 in exactly the same order as they occur in the type. If a value of this type needs to be written out, then if:

n <= 256    ->  write out lower 8 bits of n  (1 byte)
n <= 65536  ->  write out lower 16 bits of n (2 bytes)

Sum types with more tags are currently not supported and highly unlikely to occur in practice. Arguments to the tag are written out in the order of occurrence.

Polymorphic variants

The tags of these values are written out as four characters, more precisely as the 32 bit hash value computed by OCaml for the given tag in little-endian format. Any arguments associated with the tag are written out afterwards in the order of occurrence.

When polymorphic variants are being read, they will be matched in order of occurrence (left-to-right) in the type and depth-first in the case of included polymorphic types. The first type containing a match for the variant will be used for reading. E.g.:

type ab = [ `A | `B ] [@@deriving bin_io]
type cda = [ `C | `D | `A ] [@@deriving bin_io]
type abcda = [ ab | cda ] [@@deriving bin_io]

When reading type abcda, the reader associated with type ab rather than cda will be invoked if a value of type ```A`` can be read. This may not make a difference in this example, but is important to know if the user manually overrides converters. It is strongly recommended to not merge polymorphic variants if their readers might disagree about how to interpret a certain tag. This is inconsistent, confusing, and hard to debug.

Lists and arrays

For lists and arrays the length is written out as a Nat0.t first, followed by all values in the same order as in the data structure.

Hash tables

First the size of the hash table is written out as a Nat0.t. Then the writer iterates over each binding in the hash table and writes out the key followed by the value.

Note that this makes reading somewhat slower than if we used the internal (extensional) representation of the hash table, because all values have to be rehashed. On the other hand, the format becomes more robust in case the hash table implementation changes. This has in fact already happened in practice with the release of OCaml 4.00. Users should take note of this and make sure that all of their serialization routines remain future-proof by defining wire formats that are independent of the implementation of abstract data types.

Bigarrays of doubles (type vec) and characters (type bigstring)

First the dimension(s) are written out as Nat0.t. Then the contents is copied verbatim.

Polymorphic values

There is nothing special about polymorphic values as long as there are conversion functions for the type parameters. E.g.:

type 'a t = A | B of 'a [@@deriving bin_io]
type foo = int t [@@deriving bin_io]

In the above case the conversion functions will behave as if foo had been defined as a monomorphic version of t with int substituted for 'a on the right hand side.

Abstract data types

If you want to convert an abstract data type that may impose constraints on the well-formedness of values, you will have to roll your own conversion functions. Use the functions in module Read_c and Write_c to map between low-level and high-level representations, or implement those manually for maximum efficiency. The Utils module may also come in handy as described in earlier sections, e.g. if the value can be converted to and from an intermediate representation that does not impose constraints, or if some sort of iteration is supported by the data type.

Dependencies (10)

  1. dune >= "3.11.0"
  2. ppx_variants_conv >= "v0.17" & < "v0.18"
  3. ppx_stable_witness >= "v0.17" & < "v0.18"
  4. ppx_sexp_conv >= "v0.17" & < "v0.18"
  5. ppx_optcomp >= "v0.17" & < "v0.18"
  6. ppx_fields_conv >= "v0.17" & < "v0.18"
  7. ppx_custom_printf >= "v0.17" & < "v0.18"
  8. ppx_compare >= "v0.17" & < "v0.18"
  9. base >= "v0.17" & < "v0.18"
  10. ocaml >= "5.1.0"

Dev Dependencies


Used by (9)

  1. async_core < "108.00.02"
  2. bap-std >= "2.2.0" & < "2.4.0"
  3. bitvec-binprot
  4. bonsai >= "v0.17.0"
  5. core >= "v0.17.0"
  6. git < "1.1.0"
  7. hardcaml >= "v0.17.0"
  8. mirage-tc >= "0.2.0"
  9. ppx_bin_prot >= "v0.17.0"




Innovation. Community. Security.