A simple benchmark of SVM entrypoints.
The purpose of eisodos
is to offer a simple benchmark of different program entrypoint implementations. An entrypoint is used to parse the SBF input for a program, providing the information of an instruction input in a "friendly" way. The SBF loader passes the input parameters as a byte array and the entrypoint then transforms the input into separate typed entities — program id
, accounts
array and instruction data
.
Entrypoint implementation currently included in the benchmark:
Note
Previous benchmark included the solana-nostd-entrypoint
– the project has not been archived and therefore ommitted.
Benchmark | pinocchio |
solana-program |
jiminy |
---|---|---|---|
Ping | 🟩 14 | 🟨 98 (+84) | 🟩 16 (+2) |
Log | 🟩 118 | 🟨 202 (+84) | 🟩 120 (+3) |
Account (1) | 🟩 22 | 🟥 268 (+246) | 🟨 37 (+15) |
Account (3) | 🟩 44 | 🟥 546 (+502) | 🟨 67 (+23) |
Account (5) | 🟩 59 | 🟥 824 (+765) | 🟨 97 (+38) |
Account (10) | 🟩 101 | 🟥 1,519 (+1,418) | 🟨 172 (+71) |
Account (20) | 🟩 177 | 🟥 2,909 (+2,732) | 🟥 322 (+145) |
Account (32) | 🟩 269 | 🟥 4,577 (+4,308) | 🟥 502 (+233) |
Account (64) | 🟩 512 | 🟥 9,025 (+8,513) | 🟥 982 (+470) |
Important
Values correspond to compute units (CUs) consumed by the entrypoint. The delta in relation to the lowest consumption is shown in brackets.
- 🟩 (green): value within 10 CUs of the best value (
value < best value + 10
) - 🟨 (yellow): value within 100 CUs of the best value (
value < best value + 100
) - 🟥 (red): value over 100 CUs of the best value (
value >= best value + 100
)
Solana platform tools v1.51
with LTO
enabled was used in the bench tests.
There are also benchmarks for CPI and binary size produced by the different entrypoints libraries. Note that these actually measure how efficient the helpers of the library are instead of the entrypoint efficiency, since it is generally possible to improve/re-write the helpers.
Benchmark (CPI) | pinocchio |
solana-program |
jiminy |
---|---|---|---|
system::create_account | 🟩 1,291 | 🟥 2,592 (+1,301) | 🟨 1,307 (+13) |
system::transfer | 🟩 1,287 | 🟥 2,189 (+902) | 🟨 1,301 (+14) |
pinocchio |
solana-program |
jiminy |
|
---|---|---|---|
Binary size (bytes) | 5,824 (+2,144) | 64,784 (+61,104) | 3,680 |
The benchmark uses a simple program with multiple instructions to measure the compute units (CUs) consumed by the entrypoint. Note that the intention is not to write the most efficient program, instead to reflect an "average" program implementation. The aim is to use the exactly same program implementation, replacing the entrypoint to determine the impact on the CUs consumed.
Warning
This does not apply to instructions that use CPIs since these involve using library specific helpers.
The program used has the following instructions:
pub enum Instruction {
Ping,
Log,
Account {
expected: u64,
},
CreateAccount,
Transfer
}
This instruction has an empty processor and does not expect any account. The only data passed to the program is the instruction discriminator (0
in this case).
Similar to the Ping
instruction, this instruction does not expect any account and only logs a static message.
This instruction receives an u64
value as part of the instruction data, which specifies the number of accounts expected by the processor. The processor only asserts that the number of accounts received is the same as the expected
value. This in essence measures how much CUs the entrypoint comsumes to parse the input accounts.
This instruction receives 3 accounts (from
, account
and system_program
) and performs a CPI to the System program to create the account
with 500_000_000
lamports and 10
bytes of account data. These values are fixed on the processor.
This instruction receives 3 accounts (from
, to
and system_program
) and performs a CPI to the System program to transfer 1_000_000_000
lamports. The lamports amount is fixed.
The program is structure in 4 different source files:
-
entrypoint.rs
: includes the entrypoint definition and "dispatches" the instruction to the corresponding processor. -
instruction.rs
: defines the instructions available on the program and the parsing logic for the input instruction data. -
lib.rs
: defines the modules of the program and the program ID. -
processor.rs
: includes the processor for each instruction.
The implementation across all different entrypoint programs is as similar as possible. In most cases, the only differences are on the types import, since each entrypoint defines their own AccountInfo
and/or Pubkey
types.
The evaluation is performed using mollusk
.
To run the benchmarks, you will need to build the programs. After cloning the repository, run:
pnpm install
This will install the required packages. Then all programs can be buiit using:
RUSTFLAGS="-C embed-bitcode=yes -C lto=fat" pnpm programs:build --tools-version v1.51
After this, you are ready to run individual benchmarks by using:
cargo bench --bench <ENTRYPOINT_NAME>
The ENTRYPOINT_NAME
will be one of pinocchio
, solana_nostd_entrypoint
or solana_program
.
The results are written to ./target/benches/compute_units.md
. Each execution is described by 3 columns:
-
Name
: name of the benchmark; this will specify the name of the instruction and the parameters used. -
CUs
: number of compute units consumed by the execution. -
Delta
: the difference in compute units between latest benchmark and the previous; this will provide a quick way to assess the differences between entrypoints.
The results of an execution are compared to the previous one (if there is one), with delta differences shown after a +
and -
symbol.