Welcome to TechEmpower Framework Benchmarks (TFB)
This project provides representative performance measures across a wide field of web
application frameworks. With much help from the community, coverage is quite broad and
we are happy to broaden it further with contributions. The project presently includes
frameworks on many languages including Go
, Python
, Java
, Ruby
, PHP
, C#
, Clojure
,
Groovy
, Dart
, JavaScript
, Erlang
, Haskell
, Scala
, Perl
, Lua
, C
, and others. The
current tests exercise plaintext responses, JSON seralization, database reads
and writes via the object-relational mapper (ORM), collections, sorting, server-side templates,
and XSS counter-measures. Future tests will exercise other components and greater computation.
Read more and see the results of our tests on Amazon EC2 and physical hardware. For descriptions of the test types that we run, see the test requirements section.
If you find yourself in a directory or file that you're not sure what the purpose is, checkout our file structure in our documenation, which will briefly explain the use of relevant directories and files.
Get started developing quickly by utilizing vagrant with TFB. Git, Virtualbox and vagrant are required.
-
Clone TFB.
$ git clone https://github.com/TechEmpower/FrameworkBenchmarks.git
-
Move into the vagrant-development directory.
$ cd FrameworkBenchmarks/deployment/vagrant-development
-
Turn on the VM (takes at least 20 minutes).
$ vagrant up
-
Enter the VM.
$ vagrant ssh
-
Move into the FrameworkBenchmarks directory in the vm.
vagrant@TFB-all:~$ cd ~/FrameworkBenchmarks
-
Run a test.
vagrant@TFB-all:~/FrameworkBenchmarks$ toolset/run-tests.py --install server --mode verify --test beego
Note: --install server
only needs to be added the first time that a tests is run.
The dependencies should already be installed for subsequent runs.
Note: In order to run the same test again the results (located at FrameworkBenchmarks/results) directory needs to be removed prior to running the test again. This directory is automatically added when tests are run.
Our official documentation can be found at frameworkbenchmarks.readthedocs.org. If you find any errors or areas for improvement within the docs, feel free to either submit a pull request or issue at the documentation repository.
The community has consistently helped in making these tests better, and we welcome any and all changes. Reviewing our contribution practices and guidelines will help to keep us all on the same page. The contribution guide can be found in the TFB documentation.
Join in the conversation at our
Google Group,
or chat with us on Freenode at #techempower-fwbm
.