Use _Float16 as native float16 type (untested) #57
CI.yml
on: push
build-deeploy
53s
generic-kernels
/
test-runner-generic
1m 31s
generic-models
/
test-runner-generic
1m 57s
cortexm-kernels
/
test-runner-cortexm
1m 25s
cortexm-models
/
test-runner-cortexm
1m 57s
mempool-kernels
/
test-runner-mempool
2m 31s
mempool-models
/
test-runner-mempool
4m 23s
siracusa-kernels
/
test-runner-siracusa
2m 12s
siracusa-models
/
test-runner-siracusa
1m 47s
siracusa-kernels-tiled-singlebuffer-L2
/
test-runner-siracusa-tiled
2m 44s
siracusa-kernels-tiled-doublebuffer-L2
/
test-runner-siracusa-tiled
2m 24s
siracusa-neureka-kernels-tiled-singlebuffer-L2
/
test-runner-siracusa-neureka-tiled
1m 36s
siracusa-neureka-kernels-tiled-doublebuffer-L2
/
test-runner-siracusa-neureka-tiled
1m 52s
siracusa-neureka-kernels-tiled-singlebuffer-L2-wmem
/
test-runner-siracusa-neureka-tiled
1m 49s
deeploy-state-serialization
1m 6s
deeploy-memory-level-extension
57s
deeploy-tiler-extension
1m 12s
deeploy-memory-allocation-extension
1m 48s
deeploy-typing
1m 15s
deeploy-regex-matching
1m 2s
linting
1m 8s
Matrix: siracusa-models-tiled-doublebuffer-L3
Matrix: siracusa-models-tiled-singlebuffer-L2
Matrix: siracusa-models-tiled-singlebuffer-L3
Matrix: siracusa-neureka-models-tiled-doublebuffer-L3-wmem
Matrix: siracusa-neureka-models-tiled-doublebuffer-L3
Matrix: siracusa-neureka-models-tiled-singlebuffer-L3