Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix build failures #178

Merged
merged 1 commit into from
Jun 24, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion experimental/iree/compile_workloads.py
Original file line number Diff line number Diff line change
Expand Up @@ -67,7 +67,7 @@
],
"pixel-8-pro": [
"--iree-llvmcpu-target-triple=aarch64-none-linux-android34",
"--iree-llvmcpu-target-cpu-features=+v9a,+fullfp16,fp-armv8,+neon,+aes,+sha2,+crc,+lse,+rdm,+complxnum,+rcpc,+sha3,+sm4,+dotprod,+fp16fml,+dit,+flagm,+ssbs,+sb,+sve2-aes,+sve2-bitperm,+sve2-sha3,+sve2-sm4,+altnzcv,+fptoint,+bf16,+i8mm,+bti,+mte,+pauth,+perfmon,+predres,+spe,+ras"
"--iree-llvmcpu-target-cpu-features=+v9.2a,+fullfp16,+fp-armv8,+neon,+aes,+sha2,+crc,+lse,+rdm,+complxnum,+rcpc,+sha3,+sm4,+dotprod,+fp16fml,+dit,+flagm,+ssbs,+sb,+sve2-aes,+sve2-bitperm,+sve2-sha3,+sve2-sm4,+altnzcv,+fptoint,+bf16,+i8mm,+bti,+mte,+pauth,+perfmon,+predres,+spe,+ras"
],
}

Expand Down
4 changes: 3 additions & 1 deletion experimental/tflite/build_tflite.sh
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,9 @@ pushd tensorflow
# Log the git version of Tensorflow repo.
git log --oneline --graph --max-count=1

bazel build -c opt --define xnnpack_use_latest_ops=true //tensorflow/lite/tools/benchmark:benchmark_model
# Disable avx512fp16 on build servers due to a gcc compiler error:
# gcc: error: unrecognized command-line option '-mavx512fp16'
bazel build -c opt --define xnnpack_use_latest_ops=true --define xnn_enable_avx512fp16=false //tensorflow/lite/tools/benchmark:benchmark_model
cp "$(realpath bazel-bin/tensorflow/lite/tools/benchmark/benchmark_model)" "${X86_OUTPUT_DIR}/"

# The compiler flags used here were retrieved by running ./configure in the root Tensorflow repo.
Expand Down
Loading