Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Coyote v2 RDMA fails under certain benchmark #71

Open
zhenhaohe opened this issue Aug 6, 2024 · 1 comment
Open

Coyote v2 RDMA fails under certain benchmark #71

zhenhaohe opened this issue Aug 6, 2024 · 1 comment
Assignees

Comments

@zhenhaohe
Copy link
Collaborator

zhenhaohe commented Aug 6, 2024

I was testing the coyote v2 with rdma perf hw design and rdma services sw application.

  1. The RDMA read benchmark is unstable and fails under default amount of repetitions specified in the sw.The experiment below does not return.

./bin/test -d 0 -i 0 -t 10.1.212.177 -x 2048
Queue pair:
Local : QPN 0x000000, PSN 0x22b267, VADDR 00007fe912200000, SIZE 00010000, IP 0x0afd4a60
Remote: QPN 0x000000, PSN 0x30c5c7, VADDR 00007feefbc00000, SIZE 00010000, IP 0x0afd4a5c
Client registered
Sent payload

RDMA BENCHMARK
1024 [bytes], thoughput: 19.94 [MB/s], latency: 33100.42 [ns]
2048 [bytes], thoughput: 2124.81 [MB/s], latency: 8167.80 [ns]

  1. The RDMA write benchmark does not scale beyond 4K message size:

./bin/test -d 0 -i 0 -t 10.1.212.175 -x 1024 -r 10 -l 10 -w 1
Queue pair:
Local : QPN 0x000000, PSN 0x9bd652, VADDR 00007fbc23e00000, SIZE 00010000, IP 0x0afd4a58
Remote: QPN 0x000000, PSN 0xa03ec3, VADDR 00007fe9b5400000, SIZE 00010000, IP 0x0afd4a54
Client registered
Sent payload

RDMA BENCHMARK
1024 [bytes], thoughput: 870.19 [MB/s], latency: 5824.05 [ns]
2048 [bytes], thoughput: 1976.83 [MB/s], latency: 6007.90 [ns]
4096 [bytes], thoughput: 3813.60 [MB/s], latency: 6559.50 [ns]
^Cterminate called after throwing an instance of 'std::runtime_error'
what(): Stalled, SIGINT caught
Aborted

@JonasDann
Copy link
Collaborator

Yes, this is a known issue with RDMA at the moment. @maximilianheer is working on a fix that is hopefully coming soon.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants