The aim of this project is to measure the performances of actor model frameworks such as Akka.Net, Proto.Actor, Orleans. At the moment there are only Akka.Net and Proto.Actor benchmarks.
Skynet is a benchmark measuring actor creation performance.
Creates an actor (goroutine, whatever), which spawns 10 new actors, each of them spawns 10 more actors, etc. until one million actors are created on the final level. Then, each of them returns back its ordinal number (from 0 to 999999), which are summed on the previous level and sent back upstream, until reaching the root actor. (The answer should be 499999500000).
Scala akka codes in the skynet repository are adapted to Akka.Net and Proto.Actor.
Creates receive and echo actors as many as the number of cores, maps them to each other and sends messages between them. Measures how long to create N number of messages and how long it takes to send messages between receive and echo actors.
There are two kinds of ping pong benchmark and both of them have both in process and remote benchmarks. The benchmark named InProc is adapted from Proto.Actor repository and PingPong is adepted from Akka.Net repository.
Like the ping pong benchmark, actors and n number of messages are created. Actors calculates pi number as many as messages and send it back. It is aimed to simulate real world scenarios by making a calculation in Actors. There are both in process and remote benchmark examples.
You can use the dotnet
CLI commands, or use Visual Studio 2017, or use Visual Studio Code. For VS Code user both launch.json
and tasks.json
files created and configurations added for all different benchmarks.
- Solution strucutered under two solution folders, Akka.Net and Proto.Actor.
- All benchmarks were run on a computer with Intel i7-6700HQ @ 2.6 GHz and 16 GB RAM configuration.
- Benchmark settings are separated into sections in the
benchmark-settings.json
file that added as link to all benchmark projects. Changes made in this file apply to both the Akka.Net and Proto.Actor benchmarks for the same benchmark. - There is also seperate .hocon files for Akka.Net benchmarks.
You can change SkynetBenchmarkSettings.TimesToRun
value in benchmark-settings.json
for benchmark repeats.
19,269.76 ms avg for 3 repeats.
1,808 ms avg for 3 repeats.
You can change PingPongSettings.TimesToRun
value for benchmark repeats and PingPongSettings.Throughputs
for set throughput values in benchmark-settings.json
.
The benchmark actor was separately tested, derived from both ActorBase and ReceiveActor.
ActorBase first start time: 15.12 ms ReceiveActor first start time: 42.80 ms
ActorBase | ReceiveActor | |||||
---|---|---|---|---|---|---|
Throughput | Msgs/sec | Start [ms] | Total [ms] | Msgs/sec | Start [ms] | Total [ms] |
20 | 19,710,000 | 10.58 | 1532.90 | 19,749,000 | 7.78 | 1527.66 |
30 | 22,058,000 | 7.13 | 1367.89 | 22,075,000 | 5.36 | 1364.61 |
40 | 23,183,000 | 5.76 | 1300.53 | 21,786,000 | 5.53 | 1383.11 |
50 | 22,865,000 | 5.20 | 1317.86 | 20,632,000 | 5.24 | 1459.46 |
60 | 19,986,000 | 5.08 | 1506.14 | 19,736,000 | 6.41 | 1526.61 |
70 | 19,379,000 | 7.35 | 1556.23 | 19,788,000 | 7.72 | 1524.71 |
80 | 20,120,000 | 7.38 | 1499.06 | 21,246,000 | 6.21 | 1419.08 |
90 | 21,291,000 | 7.23 | 1417.23 | 22,658,000 | 4.99 | 1330.00 |
100 | 23,148,000 | 6.34 | 1303.22 | 22,271,000 | 4.85 | 1352.40 |
200 | 24,115,000 | 7.03 | 1251.97 | 24,038,000 | 4.76 | 1252.97 |
300 | 23,809,000 | 4.29 | 1264.37 | 23,059,000 | 5.25 | 1306.93 |
400 | 24,896,000 | 6.40 | 1211.94 | 23,980,000 | 4.86 | 1256.52 |
500 | 24,370,000 | 4.87 | 1236.63 | 24,174,000 | 4.53 | 1245.97 |
600 | 24,529,000 | 4.98 | 1228.14 | 23,566,000 | 7.30 | 1280.65 |
700 | 23,734,000 | 4.35 | 1269.08 | 23,310,000 | 7.11 | 1294.48 |
800 | 24,650,000 | 6.41 | 1223.73 | 23,273,000 | 5.08 | 1294.85 |
900 | 24,291,000 | 4.36 | 1239.76 | 24,671,000 | 4.40 | 1221.06 |
The benchmark actor is derived from IActor.
Actor first start time: 24.10 ms
Throughput | Msgs/sec | Start [ms] | Total [ms] |
---|---|---|---|
20 | 33,783,000 | 0.26 | 888.66 |
30 | 37,546,000 | 0.12 | 799.86 |
40 | 37,974,000 | 0.09 | 790.13 |
50 | 35,756,000 | 0.11 | 839.64 |
60 | 40,650,000 | 0.09 | 738.41 |
70 | 42,075,000 | 0.15 | 713.34 |
80 | 42,613,000 | 0.10 | 704.40 |
90 | 37,641,000 | 0.54 | 797.63 |
100 | 42,253,000 | 0.10 | 710.52 |
200 | 44,313,000 | 0.10 | 678.06 |
300 | 45,592,000 | 0.09 | 658.12 |
400 | 45,045,000 | 0.10 | 666.70 |
500 | 44,709,000 | 0.08 | 671.28 |
600 | 44,444,000 | 0.09 | 675.65 |
700 | 44,776,000 | 0.09 | 670.84 |
800 | 44,910,000 | 0.10 | 668.99 |
900 | 45,248,000 | 0.10 | 663.41 |
The same settings were used in the Akka.Net benchmark since this benchmark was adapted from the Proto.Actor repository. In other words, 2048-capacity BoundedMailbox was used as a mailbox. Normally Akka.Net's default mailbox is unbounded mailbox. When unbounded mailbox was used on Akka.Net, two times increase in msg/sec was observed. You can also get different results by using ForkJoinDispatcher on Akka.Net and changing the thread numbers. I could'nt benchmark it for Proto.Actor because I could not find ForkJoinDispatcher.
In benchmark-settings.json
file you can change dispatcher and mailbox types, for capacity and thread count, you need to change them in hocon file for Akka.Net.
"InprocBenchmarkSettings": {
"MessageCount": 1000000,
"BatchSize": 100,
"DispatcherType": "custom-dispatcher",
"MailboxType": "bounded-mailbox",
//"DispatcherType": "custom-fork-join-dispatcher",
//"MailboxType": "unbounded-mailbox",
"Throughputs": [ 100, 200, 300, 400, 500, 600, 700, 800, 900 ]
//"Throughputs": [ 10, 20, 30, 40, 50, 60, 70, 80, 90, 100, 200, 300, 400, 500, 600, 700, 800, 900 ]
//"Throughputs": [ 100, 100, 100, 100, 100, 100, 100, 100, 100, 100, 100, 100 ]
},
Is Server GC False
Dispatcher | Elapsed | Msg/sec |
---|---|---|
100 | 712 | 22,471,910 |
200 | 722 | 22,160,664 |
300 | 701 | 22,824,536 |
400 | 698 | 22,922,636 |
500 | 689 | 23,222,060 |
600 | 698 | 22,922,636 |
700 | 686 | 23,323,615 |
800 | 692 | 23,121,387 |
900 | 691 | 23,154,848 |
Avg Msg/sec : 22,902,699.1111111
Is Server GC False
Dispatcher | Elapsed | Msg/sec |
---|---|---|
100 | 197 | 81,218,274 |
200 | 148 | 108,108,108 |
300 | 155 | 103,225,806 |
400 | 146 | 109,589,041 |
500 | 154 | 103,896,103 |
600 | 149 | 107,382,550 |
700 | 144 | 111,111,111 |
800 | 154 | 103,896,103 |
900 | 149 | 107,382,550 |
Avg Msg/sec : 103,978,849.555556
Is Server GC False
Dispatcher | Elapsed | Msg/sec |
---|---|---|
100 | 315 | 50,793,650 |
200 | 275 | 58,181,818 |
300 | 276 | 57,971,014 |
400 | 276 | 57,971,014 |
500 | 275 | 58,181,818 |
600 | 275 | 58,181,818 |
700 | 279 | 57,347,670 |
800 | 296 | 54,054,054 |
900 | 279 | 57,347,670 |
Avg Msg/sec : 56,670,058.4444444
Is Server GC False
Dispatcher | Elapsed | Msg/sec |
---|---|---|
100 | 200 | 80,000,000 |
200 | 158 | 101,265,822 |
300 | 160 | 100,000,000 |
400 | 158 | 101,265,822 |
500 | 154 | 103,896,103 |
600 | 155 | 103,225,806 |
700 | 157 | 101,910,828 |
800 | 159 | 100,628,930 |
900 | 153 | 104,575,163 |
Avg Msg/sec : 99,640,941.5555556
This benchmark is adaptation of ping pong benchmark to the remote. The Benchmarks were prepared for both Proto.Actor and Akka.Net using both Google.Protobuff and Wire serializers. Hyperion serializer (fork of the Wire serializer) for Akka.Net, was used.
Within the Messages project there are both common messages and their Protobuf generated versions. All Protobuf messages are automatically generated on the prebuild event of the Messages project.
In the benchmark-settings.json
file you can change the message count (that will send between nodes) and the ip and port values for Node1, Node2. For Akka.Net you also need to change ip and ports from hocon files for Node1 and Node2.
"RemoteBenchmarkSettings": {
"MessageCount": 300000,
"Node1Ip": "127.0.0.1",
"Node2Ip": "127.0.0.1",
"Node1Port": 8090,
"Node2Port": 8091
}
In order to switch to Protobuff serializer in Akka.Net benchmarks, You need to comment the line using ActorModelBenchmarks.Messages;
and uncomment the line using ActorModelBenchmarks.Messages.Protobuf;
inside Program.cs for both Node1 and Node2.
You need to do the same think for the Proto.Actor benchmarks, additionally you need to replace the SwitchToWire
method with the SwitchToProtobuf
method inside Program.cs for both Node1 and Node2.
Elapsed 00:00:07.0944259
Throughput 84,573.4395506196 msg / sec
Elapsed 00:00:00.5927615
Throughput 1,012,211.48809428 msg / sec
Elapsed 00:00:06.7227722
Throughput 89,248.8964597075 msg / sec
Elapsed 00:00:00.2773321
Throughput 2,163,471.15966742 msg / sec
All these results were observed in the local environment (127.0.0.1). To make the benchmarks a bit more realistic, Node2 of both Akka.Net and Proto.actor has moved a Lubuntu 17.10 virtual machine with 2.6 GHz four core 2 GB RAM configuration on VMWare and benchmarks were repeated.
Elapsed 00:00:17.1707912
Throughput 34,943.0607484179 msg / sec
Elapsed 00:00:09.3672863
Throughput 64,052.7022217737 msg / sec
Elapsed 00:00:16.3266224
Throughput 36,749.7933926615 msg / sec
Elapsed 00:00:08.3608386
Throughput 71,763.1362959213 msg / sec
On the remote Proto.Actor is still twice as fast, but it has not reached million msg/sec like it was in the local environment.
Also, if you set the logging level to WARNING on Akka.Net, you get a warning as follows;
[WARNING][3/1/2018 5:38:45 PM][Thread 0013][[akka://remote-sys/system/endpointManager/reliableEndpointWriter-akka.tcp%3A%2F%2Fremote-sys%40192.168.135.128%3A8091-1/endpointWriter#1364445640]] [143971] buffered messages in EndpointWriter for [akka.tcp://[email protected]:8091]. You should probably implement flow control to avoid flooding the remote connection.
Messages are being buffered. You can tweak the DotNetty pool-size, socket-worker-pool
values to achieve different results.