Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/ahamez/protox
A fast, easy to use and 100% conformant Elixir library for Google Protocol Buffers (aka protobuf)
https://github.com/ahamez/protox
elixir json protobuf protobuf-message protobuf-runtime protoc protocol-buffers
Last synced: 5 days ago
JSON representation
A fast, easy to use and 100% conformant Elixir library for Google Protocol Buffers (aka protobuf)
- Host: GitHub
- URL: https://github.com/ahamez/protox
- Owner: ahamez
- License: mit
- Created: 2017-02-16T13:55:34.000Z (over 7 years ago)
- Default Branch: master
- Last Pushed: 2024-04-22T07:34:42.000Z (7 months ago)
- Last Synced: 2024-05-02T00:56:48.511Z (6 months ago)
- Topics: elixir, json, protobuf, protobuf-message, protobuf-runtime, protoc, protocol-buffers
- Language: Elixir
- Homepage:
- Size: 1.4 MB
- Stars: 261
- Watchers: 5
- Forks: 17
- Open Issues: 2
-
Metadata Files:
- Readme: README.md
- Changelog: CHANGELOG
- Funding: .github/FUNDING.yml
- License: LICENSE
Awesome Lists containing this project
- freaking_awesome_elixir - Elixir - Elixir implementation for Protocol Buffers. (Protocols)
- fucking-awesome-elixir - protox - Elixir implementation for Protocol Buffers. (Protocols)
- awesome-elixir - protox - Elixir implementation for Protocol Buffers. (Protocols)
README
# Protox
[![Elixir CI](https://github.com/ahamez/protox/actions/workflows/elixir.yml/badge.svg)](https://github.com/ahamez/protox/actions/workflows/elixir.yml) [![Coverage Status](https://coveralls.io/repos/github/ahamez/protox/badge.svg?branch=master)](https://coveralls.io/github/ahamez/protox?branch=master) [![Hex.pm Version](http://img.shields.io/hexpm/v/protox.svg)](https://hex.pm/packages/protox) [![Hex Docs](https://img.shields.io/badge/hex-docs-brightgreen.svg)](https://hexdocs.pm/protox/) [![License](https://img.shields.io/hexpm/l/protox.svg)](https://github.com/ahamez/protox/blob/master/LICENSE)
`protox` is an Elixir library to work with [Google's Protocol Buffers](https://developers.google.com/protocol-buffers), versions 2 and 3. It supports both binary and JSON encoding and decoding ([well-known types](https://developers.google.com/protocol-buffers/docs/reference/google.protobuf) included, except the Any type for the time being).
The primary objective of `protox` is **reliability**: it uses [property based testing](https://github.com/alfert/propcheck) and has a [near 100% code coverage](https://coveralls.io/github/ahamez/protox?branch=master). Also, using [mutation testing](https://en.wikipedia.org/wiki/Mutation_testing) with the invaluable help of [Muzak pro](https://devonestes.com/muzak), the quality of the `protox` test suite has been validated. Therefore, `protox` [passes all the tests](#conformance) of the conformance checker provided by Google.
It's also easy to use: just point to the `*.proto` files or give the schema to the `Protox` macro, no need to generate any file! However, should you need to generate files, a mix task is [available](#files-generation).
Given the following protobuf definition, `protox` will generate a `Msg` struct:
```proto
message Msg{
int32 a = 1;
map b = 2;
}
```You can then interact with `Msg` like any Elixir structure:
```elixir
iex> msg = %Msg{a: 42, b: %{1 => "a map entry"}}
iex> {:ok, iodata} = Msg.encode(msg)
iex> {:ok, iodata} = Msg.json_encode(msg)iex> binary = # read binary from a socket, a file, etc.
iex> {:ok, msg} = Msg.decode(binary)
iex> json = # read json from a socket, file, etc.
iex> {:ok, msg} = Msg.json_decode(json)
```You can find [here](https://github.com/ahamez/protox/blob/master/test/example_test.exs) a more involved example with most types.
## Table of contents
- [Prerequisites](#prerequisites)
- [Installation](#installation)
- [Usage with a textual description](#usage-with-a-textual-description)
- [Usage with files](#usage-with-files)
- [Protobuf binary format](#protobuf-binary-format)
- [Protobuf JSON format](#protobuf-json-format)
- [Packages and namespaces](#packages-and--namespaces)
- [Specify import path](#specify-import-path)
- [Unknown fields](#unknown-fields)
- [Unsupported features](#unsupported-features)
- [Implementation choices](#implementation-choices)
- [Generated code reference](#generated-code-reference)
- [Files generation](#files-generation)
- [Conformance](#conformance)
- [Types mapping](#types-mapping)
- [Benchmarks](#benchmarks)
- [Development](#development)
- [Credits](#credits)## Prerequisites
- Elixir >= 1.12
- protoc >= 3.0 *This dependency is only required at compile-time*
`protox` uses Google's `protoc` (>= 3.0) to parse `.proto` files. It must be available in `$PATH`.đ You can download it [here](https://github.com/google/protobuf) or you can install it with your favorite package manager (`brew install protobuf`, `apt install protobuf-compiler`, etc.).
âšī¸ If you choose to generate files, `protoc` won't be needed to compile these files.
## Installation
Add `:protox` to your list of dependencies in `mix.exs`:
```elixir
def deps do
[{:protox, "~> 1.7"}]
end
```If you plan to use the JSON encoding, you'll need to add [`Jason`](https://github.com/michalmuskala/jason) to your dependencies:
```elixir
def deps do
[
{:protox, "~> 1.7"},
{:jason, "~> 1.2"}
]
end
```## Usage with an inlined textual description
The following example generates two modules: `Baz` and `Foo` from a textual description:
```elixir
defmodule MyModule do
use Protox, schema: """
syntax = "proto3";message Baz {
}message Foo {
int32 a = 1;
map b = 2;
}
"""
end
```âšī¸ The module in which the `Protox` macro is called is completely ignored and therefore does not appear in the names of the generated modules.
## Usage with files
Here's how to generate the modules from a set of files:
```elixir
defmodule MyModule do
use Protox, files: [
"./defs/foo.proto",
"./defs/bar.proto",
"./defs/baz/fiz.proto",
]
end
```## Protobuf binary format
### Encode
Here's how to create and encode a new message to binary protobuf:
```elixir
iex> msg = %Fiz.Foo{a: 3, b: %{1 => %Fiz.Baz{}}}
iex> {:ok, iodata} = Protox.encode(msg)
```
Or, with throwing style:
```elixir
iex> iodata = Protox.encode!(msg)
```It's also possible to call `encode/1` and `encode!/1` directly on the generated structures:
```elixir
iex> {:ok, iodata} = Fiz.Foo.encode(msg)
iex> iodata = Fiz.Foo.encode!(msg)
```âšī¸ Note that `encode/1` returns an [IO data](https://hexdocs.pm/elixir/IO.html#module-use-cases-for-io-data) for efficiency reasons. Such IO data can be used
directly with files or sockets write operations:
```elixir
iex> {:ok, iodata} = Protox.encode(%Fiz.Foo{a: 3, b: %{1 => %Fiz.Baz{}}})
[[[], <<18>>, <<4>>, "\b", <<1>>, <<18>>, <<0>>], "\b", <<3>>]
iex> {:ok, file} = File.open("msg.bin", [:write])
{:ok, #PID<0.1023.0>}
iex> IO.binwrite(file, iodata)
:ok
```đ You can use [`:binary.list_to_bin/1`](https://erlang.org/doc/man/binary.html#list_to_bin-1) or [`IO.iodata_to_binary`](https://hexdocs.pm/elixir/IO.html#iodata_to_binary/1) to get a binary:
```elixir
iex> %Fiz.Foo{a: 3, b: %{1 => %Fiz.Baz{}}} |> Protox.encode!() |> :binary.list_to_bin()
<<8, 3, 18, 4, 8, 1, 18, 0>>
```### Decode
Here's how to decode a message from binary protobuf:
```elixir
iex> {:ok, msg} = Protox.decode(<<8, 3, 18, 4, 8, 1, 18, 0>>, Fiz.Foo)
```
Or, with throwing style:
```elixir
iex> msg = Protox.decode!(<<8, 3, 18, 4, 8, 1, 18, 0>>, Fiz.Foo)
```It's also possible to call `decode/1` and `decode!/1` directly on the generated structures:
```elixir
iex> {:ok, msg} = Fiz.Foo.decode(<<8, 3, 18, 4, 8, 1, 18, 0>>)
iex> msg = Fiz.Foo.decode!(<<8, 3, 18, 4, 8, 1, 18, 0>>)
```## Protobuf JSON format
`protox` implements the [Google's JSON specification](https://developers.google.com/protocol-buffers/docs/proto3#json).
### Encode
Here's how to encode a message to JSON, exported as IO data:
```elixir
iex> msg = %Fiz.Foo{a: 42}
iex> {:ok, iodata} = Protox.json_encode(msg)
{:ok, ["{", ["\"a\"", ":", "42"], "}"]}
```Or, with throwing style:
```elixir
iex> msg = %Fiz.Foo{a: 42}
iex> iodata = Protox.json_encode!(msg)
["{", ["\"a\"", ":", "42"], "}"]
```It's also possible to call `json_encode` and `json_encode!` directly on the generated structures:
```elixir
iex> {:ok, iodata} = Fiz.Foo.json_encode(msg)
iex> iodata = Fiz.Foo.json_encode!(msg)
```### Decode
Here's how to decode JSON to a message:
```elixir
iex> Protox.json_decode("{\"a\":42}", Fiz.Foo)
{:ok, %Fiz.Foo{__uf__: [], a: 42, b: %{}}}
```Or, with throwing style:
```elixir
iex> Protox.json_decode!("{\"a\":42}", Fiz.Foo)
%Fiz.Foo{__uf__: [], a: 42, b: %{}}
```It's also possible to call `json_decode` and `json_decode!` directly on the generated structures:
```elixir
iex> Fiz.Foo.json_decode("{\"a\":42}")
iex> Fiz.Foo.json_decode!("{\"a\":42}")
```### JSON library configuration
By default, `protox` uses [`Jason`](https://github.com/michalmuskala/jason) to encode values to JSON (mostly to escape strings). You can also use [`Poison`](https://github.com/devinus/poison):
```elixir
iex> Protox.json_decode!(iodata, Fiz.Foo, json_library: Protox.Poison)
iex> Protox.json_encode!(msg, json_library: Protox.Poison)
```âšī¸ You can use any other library by implementing the `Protox.JsonLibrary` behaviour.
đ Don't forget to add the chosen library to the list of dependencies in `mix.exs`.
### Well-known types
Note that `protox` does not completely support the [Any](https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#any) well-know type: it will be encoded and decoded like a regular message, rather than with the custom encoding specified in the [JSON specification](https://developers.google.com/protocol-buffers/docs/proto3#json).
## Packages and namespaces
### Packages
Protobuf provides a `package` [directive](https://developers.google.com/protocol-buffers/docs/proto#packages):
```proto
package abc.def;
message Baz {}
```Modules generated by protox will include this package declaration. Thus, the example above will be translated to `Abc.Def.Baz` (note the [camelization](#implementation-choices) of package `abc.def` to `Abc.Def`).
### Prepend namespaces
In addition, protox provides the possibility to prepend a namespace with the `namespace` option to all generated modules:```elixir
defmodule Bar do
use Protox, schema: """
syntax = "proto3";package abc;
message Msg {
int32 a = 1;
}
""",
namespace: MyApp
end
```In this example, the module `MyApp.Abc.Msg` is generated:
```elixir
iex> msg = %MyApp.Abc.Msg{a: 42}
```## Specify import path
An import path can be specified using the `path:` or `paths:` options that respectively specify the directory or directories in which to search for imports:
```elixir
defmodule Baz do
use Protox,
files: [
"./defs/prefix/foo.proto",
"./defs/prefix/bar/bar.proto",
],
path: "./defs"
end
```If multiple search paths are needed:
```elixir
defmodule Baz do
use Protox,
files: [
"./defs1/prefix/foo.proto",
"./defs1/prefix/bar.proto",
"./defs2/prefix/baz/baz.proto"
],
paths: [
"./defs1",
"./defs2"
]
end
```It corresponds to the `-I` option of `protoc`.
## Unknown fields
[Unknown fields](https://developers.google.com/protocol-buffers/docs/proto3#unknowns) are fields that are present on the wire but which do not correspond to an entry in the protobuf definition. Typically, it occurs when the sender has a newer version of the protobuf definition. It enables backwards compatibility as the receiver with an old version of the protobuf definition will still be able to decode old fields.
When unknown fields are encountered at decoding time, they are kept in the decoded message. It's possible to access them with the `unknown_fields/1` function defined with the message.
```elixir
iex> msg = Msg.decode!(<<8, 42, 42, 4, 121, 97, 121, 101, 136, 241, 4, 83>>)
%Msg{a: 42, b: "", z: -42, __uf__: [{5, 2, <<121, 97, 121, 101>>}]}iex> Msg.unknown_fields(msg)
[{5, 2, <<121, 97, 121, 101>>}]
```You must always use `unknown_fields/1` as the name of the field (e.g. `__uf__` in the above example) is generated at compile-time to avoid collision with the actual fields of the Protobuf message. This function returns a list of tuples `{tag, wire_type, bytes}`. For more information, please see [protobuf encoding guide](https://developers.google.com/protocol-buffers/docs/encoding).
When you encode a message that contains unknown fields, they will be reencoded in the serialized output.
### Disable support of unknown fields
You can deactivate the support of unknown fields by setting the `:keep_unknown_fields` option to `false`:
```elixir
defmodule Baz do
use Protox,
schema: """
syntax = "proto3";message Sub {
int32 a = 1;
string b = 2;
}
""",
keep_unknown_fields: false
end
```
âšī¸ `protox` will still correctly parse unknown fields, they just won't be added to the structure and you won't be able to access them. This also means that unkown fields won't be serialized back.## Unsupported features
* The [Any](https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#any) well-known type is partially supported: you can manually unpack the embedded message right after decoding and conversely pack it right before encoding;
* Groups ([deprecated in protobuf](https://developers.google.com/protocol-buffers/docs/proto#groups));
* All [options](https://developers.google.com/protocol-buffers/docs/proto3#options) other than `packed` and `default` are ignored as they concern other languages implementation details.## Implementation choices
* This library enforces the presence of required fields (Protobuf 2). Therefore an error is raised when encoding or decoding a message with a missing required field:
```elixir
defmodule Bar do
use Protox, schema: """
syntax = "proto2";message Required {
required int32 a = 1;
}
"""
endiex> Protox.encode!(%Required{})
** (Protox.RequiredFieldsError) Some required fields are not set: [:a]iex> Required.decode!(<<>>)
** (Protox.RequiredFieldsError) Some required fields are not set: [:a]
```* When decoding enum aliases, the last encountered constant is used. For instance, in the following example, `:BAR` is always used if the value `1` is read on the wire:
```protobuf
enum E {
option allow_alias = true;
FOO = 0;
BAZ = 1;
BAR = 1;
}
```* Unset optionals
* For Protobuf 2, unset optional fields are mapped to `nil`. You can use the generated `default/1` function to get the default value of a field:
```elixir
defmodule Bar do
use Protox,
schema: """
syntax = "proto2";message Foo {
optional int32 a = 1 [default = 42];
}
"""
endiex> Foo.default(:a)
{:ok, 42}iex> %Foo{}.a
nil```
It means that if you need to know if a field has been set by the sender, you just have to test if its value is `nil` or not.* For Protobuf 3, unset fields are mapped to their [default values](https://developers.google.com/protocol-buffers/docs/proto3#default). However, if you use the `optional` keyword (available in protoc version 3.15 and higher), then unset fields will be mapped to `nil`:
```elixir
defmodule Bar do
use Protox,
schema: """
syntax = "proto3";message Foo {
int32 a = 1;
optional int32 b = 2;
}
"""
endiex> Foo.default(:a)
{:ok, 0}iex> %Foo{}.a
0iex> Foo.default(:b)
{:error, :no_default_value}iex> %Foo{}.b
nil
```* Messages and enums names: they are converted using the [`Macro.camelize/1`](https://hexdocs.pm/elixir/Macro.html#camelize/1) function.
Thus, in the following example, `non_camel_message` becomes `NonCamelMessage`, but the field `non_camel_field` is left unchanged:
```elixir
defmodule Bar do
use Protox,
schema: """
syntax = "proto3";message non_camel_message {
}message CamelMessage {
int32 non_camel_field = 1;
}
"""
endiex> msg = %NonCamelMessage{}
%NonCamelMessage{__uf__: []}iex> msg = %CamelMessage{}
%CamelMessage{__uf__: [], non_camel_field: 0}
```## Generated code reference
The detailed reference of the generated code is available [here](documentation/reference.md).
## Files generation
It's possible to generate a file that will contain all code corresponding to the protobuf messages:
```shell
MIX_ENV=prod mix protox.generate --output-path=/path/to/message.ex --include-path=./test/samples test/samples/messages.proto test/samples/proto2.proto
```The generated file will be usable in any project as long as `protox` is declared in the dependencies as it needs functions from the protox runtime.
### Options
* `--output-path`
The path to the file to be generated or to the destination folder when generating multiple files.* `--include-path`
Specifies the [import path](#specify-import-path). If multiple include paths are needed, add more `--include-path` options.* `--multiple-files`
Generates one file per module. In this case, `--output-path` must point to a directory. It's useful for definitions with a lot of messages as Elixir will be able to parallelize the compilation of the generated modules.* `--namespace`
Prepends a [namespace](#prepend-namespaces) to all generated modules.* `--keep-unknown-fields=[true|false]`
Toggles support of [unknown fields](#unknown-fields). Default to `true`.* `--generate-defs-funs=[true|false]`
Generates deprecated functions `defs/0` and `defs_by_name/0`. Default to `true` for backward compatibility, but it's safe to set it to `false` if you don't use these functions (and it reduces the size of the generated code).## Conformance
The protox library has been thoroughly tested using the conformance checker [provided by Google](https://github.com/protocolbuffers/protobuf/tree/master/conformance).
Here's how to launch the conformance tests:
* Get conformance-test-runner [sources](https://github.com/protocolbuffers/protobuf/archive/refs/tags/v3.18.0.tar.gz).
* Compile conformance-test-runner ([macOS and Linux only](https://github.com/protocolbuffers/protobuf/tree/master/conformance#portability)):
```
tar xf protobuf-3.18.0.tar.gz && cd protobuf-3.18.0 && ./autogen.sh && ./configure && make -j && cd conformance && make -j
```* Launch the conformance tests:
```
mix protox.conformance --runner=/path/to/protobuf-3.18.0/conformance/conformance-test-runner
```
* A report will be generated in the directory `conformance_report` and the following text should be displayed:```
CONFORMANCE TEST BEGIN ====================================CONFORMANCE SUITE PASSED: 1996 successes, 0 skipped, 21 expected failures, 0 unexpected failures.
CONFORMANCE TEST BEGIN ====================================
CONFORMANCE SUITE PASSED: 0 successes, 120 skipped, 0 expected failures, 0 unexpected failures.
```* You can alternatively launch these conformance tests with `mix test` by setting the `PROTOBUF_CONFORMANCE_RUNNER` environment variable and including the `conformance` tag:
```
PROTOBUF_CONFORMANCE_RUNNER=/path/to/conformance-test-runner MIX_ENV=test mix test --include conformance
```### Skipped conformance tests
You may have noticed that there are `21 expected failures`. Indeed, we removed on purpose some conformance tests that `protox` can't currently pass. Here are the reasons why:
- [Any](https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#any) is not yet supported by `protox`;
- We could not find the specification for the protobuf2 case of field name extensions when decoding from JSON.The exact list of skipped tests is [here](https://github.com/ahamez/protox/blob/master/conformance/failure_list.txt).
## Types mapping
The following table shows how Protobuf types are mapped to Elixir's ones.
Protobuf | Elixir
---------- | -------------
`int32` | `integer()`
`int64` | `integer()`
`uint32` | `integer()`
`uint64` | `integer()`
`sint32` | `integer()`
`sint64` | `integer()`
`fixed32` | `integer()`
`fixed64` | `integer()`
`sfixed32` | `integer()`
`sfixed64` | `integer()`
`float` | `float() \| :infinity \| :'-infinity' \| :nan`
`double` | `float() \| :infinity \| :'-infinity' \| :nan`
`bool` | `boolean()`
`string` | `String.t()`
`bytes` | `binary()`
`repeated` | `list(value_type)` where `value_type` is the type of the repeated field
`map` | `map()`
`oneof` | `{atom(), value_type}` where `atom()` is the type of the set field and where `value_type` is the type of the set field
`enum` | `atom() \| integer()`
`message` | `struct()`## Benchmarks
You can launch benchmarks to see how `protox` perform:
```
mix run ./benchmarks/generate_payloads.exs # first time only, generates random payloads
mix run ./benchmarks/run.exs --lib=./benchmarks/protox.exs
mix run ./benchmarks/load.exs
```## Development
`protox` uses [`pre-commit`](https://pre-commit.com/#install) to launch git hooks. Thus, it's strongly recommended to install it, and then to install hooks as follows:
```sh
pre-commit install && pre-commit install -t pre-push
```## Credits
Both [gpb](https://github.com/tomas-abrahamsson/gpb) and [exprotobuf](https://github.com/bitwalker/exprotobuf) were very useful in understanding how to implement Protocol Buffers.