Skip to content

Commit

Permalink
Add suffixes to cmake flags related to building backends. (#3485)
Browse files Browse the repository at this point in the history
Summary:
Pull Request resolved: #3485

.

Differential Revision: D56869549
  • Loading branch information
shoumikhin authored and facebook-github-bot committed May 3, 2024
1 parent b6f8983 commit e138ff2
Show file tree
Hide file tree
Showing 41 changed files with 120 additions and 116 deletions.
2 changes: 1 addition & 1 deletion .ci/scripts/build_llama_android.sh
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ install_executorch_and_backend_lib() {
-DCMAKE_BUILD_TYPE=Release \
-DEXECUTORCH_BUILD_EXTENSION_MODULE=ON \
-DEXECUTORCH_BUILD_EXTENSION_DATA_LOADER=ON \
-DEXECUTORCH_BUILD_XNNPACK=ON \
-DEXECUTORCH_BUILD_BACKEND_XNNPACK=ON \
-DEXECUTORCH_BUILD_KERNELS_OPTIMIZED=ON \
-DEXECUTORCH_BUILD_KERNELS_QUANTIZED=ON \
-DXNNPACK_ENABLE_ARM_BF16=OFF \
Expand Down
2 changes: 1 addition & 1 deletion .ci/scripts/test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -85,7 +85,7 @@ build_cmake_xnn_executor_runner() {
&& mkdir ${CMAKE_OUTPUT_DIR} \
&& cd ${CMAKE_OUTPUT_DIR} \
&& retry cmake -DCMAKE_BUILD_TYPE=Release \
-DEXECUTORCH_BUILD_XNNPACK=ON \
-DEXECUTORCH_BUILD_BACKEND_XNNPACK=ON \
-DCMAKE_PREFIX_PATH="$CMAKE_PREFIX_PATH" \
-DPYTHON_EXECUTABLE="$PYTHON_EXECUTABLE" ..)

Expand Down
4 changes: 2 additions & 2 deletions .ci/scripts/test_llama.sh
Original file line number Diff line number Diff line change
Expand Up @@ -76,7 +76,7 @@ cmake_install_executorch_libraries() {
-DEXECUTORCH_BUILD_KERNELS_CUSTOM="$CUSTOM" \
-DEXECUTORCH_BUILD_KERNELS_OPTIMIZED=ON \
-DEXECUTORCH_BUILD_KERNELS_QUANTIZED=ON \
-DEXECUTORCH_BUILD_XNNPACK="$XNNPACK" \
-DEXECUTORCH_BUILD_BACKEND_XNNPACK="$XNNPACK" \
-DPYTHON_EXECUTABLE="$PYTHON_EXECUTABLE" \
-Bcmake-out .
cmake --build cmake-out -j9 --target install --config Debug
Expand All @@ -90,7 +90,7 @@ cmake_build_llama_runner() {
-DCMAKE_BUILD_TYPE=Debug \
-DEXECUTORCH_BUILD_KERNELS_CUSTOM="$CUSTOM" \
-DEXECUTORCH_BUILD_KERNELS_OPTIMIZED=ON \
-DEXECUTORCH_BUILD_XNNPACK="$XNNPACK" \
-DEXECUTORCH_BUILD_BACKEND_XNNPACK="$XNNPACK" \
-DPYTHON_EXECUTABLE="$PYTHON_EXECUTABLE" \
-Bcmake-out/${dir} \
${dir}
Expand Down
4 changes: 2 additions & 2 deletions .github/workflows/_unittest.yml
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ jobs:
# Setup MacOS dependencies as there is no Docker support on MacOS atm
PYTHON_EXECUTABLE=python \
EXECUTORCH_BUILD_PYBIND=ON \
EXECUTORCH_BUILD_XNNPACK=ON \
EXECUTORCH_BUILD_BACKEND_XNNPACK=ON \
.ci/scripts/setup-linux.sh "${BUILD_TOOL}"
# Run pytest with coverage
Expand All @@ -64,7 +64,7 @@ jobs:
# Setup MacOS dependencies as there is no Docker support on MacOS atm
PYTHON_EXECUTABLE=python \
EXECUTORCH_BUILD_PYBIND=ON \
CMAKE_ARGS="-DEXECUTORCH_BUILD_COREML=ON -DEXECUTORCH_BUILD_MPS=ON -DEXECUTORCH_BUILD_XNNPACK=ON" \
CMAKE_ARGS="-DEXECUTORCH_BUILD_BACKEND_COREML=ON -DEXECUTORCH_BUILD_BACKEND_MPS=ON -DEXECUTORCH_BUILD_BACKEND_XNNPACK=ON" \
${CONDA_RUN} --no-capture-output \
.ci/scripts/setup-macos.sh "${BUILD_TOOL}"
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/pull.yml
Original file line number Diff line number Diff line change
Expand Up @@ -232,7 +232,7 @@ jobs:
# build module for executorch.extension.pybindings.portable_lib
BUILD_TOOL=${{ matrix.build-tool }}
PYTHON_EXECUTABLE=python \
EXECUTORCH_BUILD_XNNPACK=ON \
EXECUTORCH_BUILD_BACKEND_XNNPACK=ON \
EXECUTORCH_BUILD_PYBIND=ON \
bash .ci/scripts/setup-linux.sh "${BUILD_TOOL}"
Expand Down
54 changes: 27 additions & 27 deletions CMakeLists.txt
Original file line number Diff line number Diff line change
Expand Up @@ -14,8 +14,8 @@
# it again until you pull from the upstream repo again.
#
# NOTE: Build options can be configured by passing arguments to cmake. For
# example, to enable the EXECUTORCH_BUILD_XNNPACK option, change the cmake
# command to 'cmake -DEXECUTORCH_BUILD_XNNPACK=ON ..'.
# example, to enable the EXECUTORCH_BUILD_BACKEND_XNNPACK option, change the
# cmake command to 'cmake -DEXECUTORCH_BUILD_BACKEND_XNNPACK=ON ..'.
#[[
(rm -rf cmake-out \
&& mkdir cmake-out \
Expand Down Expand Up @@ -143,17 +143,19 @@ set(CMAKE_CXX_FLAGS_DEBUG "-O0 -g")

option(EXECUTORCH_BUILD_ANDROID_JNI "Build Android JNI" OFF)

option(EXECUTORCH_BUILD_ARM_BAREMETAL
option(EXECUTORCH_BUILD_BACKEND_ARM_BAREMETAL
"Build the Arm Baremetal flow for Cortex-M and Ethos-U" OFF
)

option(EXECUTORCH_BUILD_COREML "Build the Core ML backend" OFF)
option(EXECUTORCH_BUILD_BACKEND_COREML "Build the Core ML backend" OFF)

option(EXECUTORCH_BUILD_KERNELS_CUSTOM "Build the custom kernels" OFF)
option(EXECUTORCH_BUILD_BACKEND_MPS "Build the MPS backend" OFF)

option(EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT "Build the custom ops lib for AOT"
OFF
)
option(EXECUTORCH_BUILD_BACKEND_QNN "Build the Qualcomm backend" OFF)

option(EXECUTORCH_BUILD_BACKEND_XNNPACK "Build the XNNPACK backend" OFF)

option(EXECUTORCH_BUILD_BACKEND_VULKAN "Build the Vulkan backend" OFF)

option(EXECUTORCH_BUILD_EXTENSION_DATA_LOADER "Build the Data Loader extension"
OFF
Expand All @@ -167,38 +169,36 @@ option(EXECUTORCH_BUILD_EXTENSION_RUNNER_UTIL "Build the Runner Util extension"

option(EXECUTORCH_BUILD_GTESTS "Build googletest based test binaries" OFF)

option(EXECUTORCH_BUILD_MPS "Build the MPS backend" OFF)

option(EXECUTORCH_BUILD_PYBIND "Build the Python Bindings" OFF)
option(EXECUTORCH_BUILD_KERNELS_CUSTOM "Build the custom kernels" OFF)

option(EXECUTORCH_BUILD_QNN "Build the Qualcomm backend" OFF)
option(EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT "Build the custom ops lib for AOT"
OFF
)

option(EXECUTORCH_BUILD_KERNELS_OPTIMIZED "Build the optimized kernels" OFF)

option(EXECUTORCH_BUILD_KERNELS_QUANTIZED "Build the quantized kernels" OFF)

option(EXECUTORCH_BUILD_PYBIND "Build the Python Bindings" OFF)

option(EXECUTORCH_BUILD_SDK "Build the ExecuTorch SDK")

option(EXECUTORCH_BUILD_SIZE_TEST "Build the size test" OFF)

option(EXECUTORCH_BUILD_XNNPACK "Build the XNNPACK backend" OFF)

option(EXECUTORCH_BUILD_VULKAN "Build the Vulkan backend" OFF)

#
# pthreadpool: build pthreadpool library. Disable on unsupported platforms
#
cmake_dependent_option(
EXECUTORCH_BUILD_PTHREADPOOL "Build pthreadpool library." ON
"NOT EXECUTORCH_BUILD_ARM_BAREMETAL" OFF
"NOT EXECUTORCH_BUILD_BACKEND_ARM_BAREMETAL" OFF
)

#
# cpuinfo: build cpuinfo library. Disable on unsupported platforms
#
cmake_dependent_option(
EXECUTORCH_BUILD_CPUINFO "Build cpuinfo library." ON
"NOT EXECUTORCH_BUILD_ARM_BAREMETAL" OFF
"NOT EXECUTORCH_BUILD_BACKEND_ARM_BAREMETAL" OFF
)

if(EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT)
Expand Down Expand Up @@ -542,27 +542,27 @@ if(EXECUTORCH_BUILD_EXTENSION_RUNNER_UTIL)
add_subdirectory(${CMAKE_CURRENT_SOURCE_DIR}/extension/runner_util)
endif()

if(EXECUTORCH_BUILD_XNNPACK)
if(EXECUTORCH_BUILD_BACKEND_XNNPACK)
add_subdirectory(${CMAKE_CURRENT_SOURCE_DIR}/backends/xnnpack)
endif()

if(EXECUTORCH_BUILD_VULKAN)
if(EXECUTORCH_BUILD_BACKEND_VULKAN)
add_subdirectory(${CMAKE_CURRENT_SOURCE_DIR}/backends/vulkan)
endif()

if(EXECUTORCH_BUILD_QNN)
if(EXECUTORCH_BUILD_BACKEND_QNN)
add_subdirectory(${CMAKE_CURRENT_SOURCE_DIR}/backends/qualcomm)
endif()

if(EXECUTORCH_BUILD_ARM_BAREMETAL)
if(EXECUTORCH_BUILD_BACKEND_ARM_BAREMETAL)
add_subdirectory(${CMAKE_CURRENT_SOURCE_DIR}/backends/arm)
endif()

if(EXECUTORCH_BUILD_MPS)
if(EXECUTORCH_BUILD_BACKEND_MPS)
add_subdirectory(${CMAKE_CURRENT_SOURCE_DIR}/backends/apple/mps)
endif()

if(EXECUTORCH_BUILD_COREML)
if(EXECUTORCH_BUILD_BACKEND_COREML)
add_subdirectory(${CMAKE_CURRENT_SOURCE_DIR}/backends/apple/coreml)
endif()

Expand Down Expand Up @@ -594,15 +594,15 @@ if(EXECUTORCH_BUILD_PYBIND)
torch
)

if(EXECUTORCH_BUILD_COREML)
if(EXECUTORCH_BUILD_BACKEND_COREML)
list(APPEND _dep_libs coremldelegate)
endif()

if(EXECUTORCH_BUILD_MPS)
if(EXECUTORCH_BUILD_BACKEND_MPS)
list(APPEND _dep_libs mpsdelegate)
endif()

if(EXECUTORCH_BUILD_XNNPACK)
if(EXECUTORCH_BUILD_BACKEND_XNNPACK)
# need to explicitly specify XNNPACK here otherwise uses XNNPACK symbols
# from libtorch_cpu
list(APPEND _dep_libs xnnpack_backend XNNPACK)
Expand Down
2 changes: 1 addition & 1 deletion backends/apple/coreml/scripts/build_tests.sh
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ cmake "$EXECUTORCH_ROOT_PATH" -B"$CMAKE_EXECUTORCH_BUILD_DIR_PATH" \
-DDEPLOYMENT_TARGET=13.0 \
-DFLATC_EXECUTABLE="$(which flatc)" \
-DEXECUTORCH_BUILD_EXECUTOR_RUNNER=OFF \
-DEXECUTORCH_BUILD_XNNPACK=OFF \
-DEXECUTORCH_BUILD_BACKEND_XNNPACK=OFF \
-DEXECUTORCH_BUILD_GFLAGS=OFF

cmake --build "$CMAKE_EXECUTORCH_BUILD_DIR_PATH" -j9 -t executorch
Expand Down
4 changes: 2 additions & 2 deletions backends/qualcomm/scripts/build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -69,7 +69,7 @@ if [ "$BUILD_AARCH64" = true ]; then
# If we build debug type, we need to change flatcc to flatcc_d
cmake .. \
-DCMAKE_INSTALL_PREFIX=$BUILD_ROOT \
-DEXECUTORCH_BUILD_QNN=ON \
-DEXECUTORCH_BUILD_BACKEND_QNN=ON \
-DEXECUTORCH_BUILD_SDK=ON \
-DEXECUTORCH_ENABLE_EVENT_TRACER=ON \
-DQNN_SDK_ROOT=$QNN_SDK_ROOT \
Expand Down Expand Up @@ -108,7 +108,7 @@ if [ "$BUILD_X86_64" = true ]; then
cmake \
-DCMAKE_BUILD_TYPE=RelWithDebInfo \
-DQNN_SDK_ROOT=${QNN_SDK_ROOT} \
-DEXECUTORCH_BUILD_QNN=ON \
-DEXECUTORCH_BUILD_BACKEND_QNN=ON \
-DPYTHON_EXECUTABLE=$PYTHON_EXECUTABLE \
-DBUCK2=$BUCK2 \
-S $PRJ_ROOT \
Expand Down
4 changes: 2 additions & 2 deletions backends/qualcomm/setup.md
Original file line number Diff line number Diff line change
Expand Up @@ -67,7 +67,7 @@ Make sure `buck2` is under a directory in `PATH`.
cd $EXECUTORCH_ROOT
mkdir build_x86_64
cd build_x86_64
cmake .. -DEXECUTORCH_BUILD_QNN=ON -DQNN_SDK_ROOT=${QNN_SDK_ROOT}
cmake .. -DEXECUTORCH_BUILD_BACKEND_QNN=ON -DQNN_SDK_ROOT=${QNN_SDK_ROOT}
cmake --build . -t "PyQnnManagerAdaptor" "PyQnnWrapperAdaptor" -j8

# install Python APIs to correct import path
Expand All @@ -94,7 +94,7 @@ cd build_android
# build executorch & qnn_executorch_backend
cmake .. \
-DCMAKE_INSTALL_PREFIX=$PWD \
-DEXECUTORCH_BUILD_QNN=ON \
-DEXECUTORCH_BUILD_BACKEND_QNN=ON \
-DQNN_SDK_ROOT=$QNN_SDK_ROOT \
-DCMAKE_TOOLCHAIN_FILE=$ANDROID_NDK/build/cmake/android.toolchain.cmake \
-DANDROID_ABI='arm64-v8a' \
Expand Down
3 changes: 2 additions & 1 deletion backends/vulkan/CMakeLists.txt
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,8 @@
# ~~~
# It should also be cmake-lint clean.
#
# The targets in this file will be built if EXECUTORCH_BUILD_VULKAN is ON
# The targets in this file will be built if EXECUTORCH_BUILD_BACKEND_VULKAN is
# ON

cmake_minimum_required(VERSION 3.19)

Expand Down
4 changes: 2 additions & 2 deletions backends/vulkan/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -134,7 +134,7 @@ and test on a local Android device. Android devices have built in support for
Vulkan, and the Android NDK ships with a GLSL compiler, which is needed to
compile the Vulkan Compute Library's GLSL compute shaders.

The Vulkan Delegate libraries can be built by setting `-DEXECUTORCH_BUILD_VULKAN=ON`
The Vulkan Delegate libraries can be built by setting `-DEXECUTORCH_BUILD_BACKEND_VULKAN=ON`
when building with CMake.

First, make sure that you have the Android NDK installed - Android NDK r25c is
Expand All @@ -161,7 +161,7 @@ Delegate:
pp cmake . -DCMAKE_INSTALL_PREFIX=cmake-android-out \
-DCMAKE_TOOLCHAIN_FILE=$ANDROID_NDK/build/cmake/android.toolchain.cmake \
-DANDROID_ABI=$ANDROID_ABI \
-DEXECUTORCH_BUILD_VULKAN=ON \
-DEXECUTORCH_BUILD_BACKEND_VULKAN=ON \
-DPYTHON_EXECUTABLE=python \
-Bcmake-android-out && \
cmake --build cmake-android-out -j16 --target install)
Expand Down
3 changes: 2 additions & 1 deletion backends/vulkan/cmake/ShaderLibrary.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,8 @@
# ~~~
# It should also be cmake-lint clean.
#
# The targets in this file will be built if EXECUTORCH_BUILD_VULKAN is ON
# The targets in this file will be built if EXECUTORCH_BUILD_BACKEND_VULKAN is
# ON

if(ANDROID)
if(NOT ANDROID_NDK)
Expand Down
4 changes: 2 additions & 2 deletions backends/vulkan/docs/android_demo.md
Original file line number Diff line number Diff line change
Expand Up @@ -96,7 +96,7 @@ binary using the Android NDK toolchain.
-DANDROID_ABI=$ANDROID_ABI \
-DEXECUTORCH_BUILD_EXTENSION_MODULE=ON \
-DEXECUTORCH_BUILD_EXTENSION_DATA_LOADER=ON \
-DEXECUTORCH_BUILD_VULKAN=ON \
-DEXECUTORCH_BUILD_BACKEND_VULKAN=ON \
-DEXECUTORCH_BUILD_KERNELS_OPTIMIZED=ON \
-DPYTHON_EXECUTABLE=python \
-Bcmake-android-out && \
Expand Down Expand Up @@ -140,7 +140,7 @@ the Vulkan backend is built when building and installing ExecuTorch libraries:

```shell
# Run from executorch root directory. You can also edit this in a code editor
sed -i 's/-DEXECUTORCH_BUILD_XNNPACK=ON/-DEXECUTORCH_BUILD_XNNPACK=ON -DEXECUTORCH_BUILD_VULKAN=ON/g' examples/demo-apps/android/LlamaDemo/setup.sh
sed -i 's/-DEXECUTORCH_BUILD_BACKEND_XNNPACK=ON/-DEXECUTORCH_BUILD_BACKEND_XNNPACK=ON -DEXECUTORCH_BUILD_BACKEND_VULKAN=ON/g' examples/demo-apps/android/LlamaDemo/setup.sh
```

Then, Follow the instructions at [**Setting up the ExecuTorch LLaMA Android Demo App**](./llm/llama-demo-android.md)
Expand Down
2 changes: 1 addition & 1 deletion backends/xnnpack/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -106,7 +106,7 @@ cmake \
-DCMAKE_INSTALL_PREFIX=cmake-out \
-DCMAKE_BUILD_TYPE=Release \
-DEXECUTORCH_BUILD_EXTENSION_MODULE=ON \
-DEXECUTORCH_BUILD_XNNPACK=ON \
-DEXECUTORCH_BUILD_BACKEND_XNNPACK=ON \
-DEXECUTORCH_BUILD_EXTENSION_DATA_LOADER=ON \
-DEXECUTORCH_ENABLE_LOGGING=ON \
-DPYTHON_EXECUTABLE=python \
Expand Down
54 changes: 28 additions & 26 deletions build/Utils.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -43,15 +43,32 @@ function(executorch_print_configuration_summary)
message(STATUS " EXECUTORCH_BUILD_ANDROID_JNI : "
"${EXECUTORCH_BUILD_ANDROID_JNI}"
)
message(STATUS " EXECUTORCH_BUILD_ARM_BAREMETAL : "
"${EXECUTORCH_BUILD_ARM_BAREMETAL}"
message(STATUS " EXECUTORCH_BUILD_BACKEND_ARM_BAREMETAL : "
"${EXECUTORCH_BUILD_BACKEND_ARM_BAREMETAL}"
)
message(
STATUS
" EXECUTORCH_BUILD_COREML : ${EXECUTORCH_BUILD_COREML}"
" EXECUTORCH_BUILD_BACKEND_COREML : ${EXECUTORCH_BUILD_BACKEND_COREML}"
)
message(STATUS " EXECUTORCH_BUILD_KERNELS_CUSTOM : "
"${EXECUTORCH_BUILD_KERNELS_CUSTOM}"
message(
STATUS
" EXECUTORCH_BUILD_BACKEND_MPS : ${EXECUTORCH_BUILD_BACKEND_MPS}"
)
message(
STATUS
" EXECUTORCH_BUILD_BACKEND_QNN : ${EXECUTORCH_BUILD_BACKEND_QNN}"
)
message(
STATUS
" EXECUTORCH_BUILD_BACKEND_XNNPACK : ${EXECUTORCH_BUILD_BACKEND_XNNPACK}"
)
message(
STATUS
" EXECUTORCH_BUILD_BACKEND_VULKAN : ${EXECUTORCH_BUILD_BACKEND_VULKAN}"
)
message(
STATUS
" EXECUTORCH_BUILD_CPUINFO : ${EXECUTORCH_BUILD_CPUINFO}"
)
message(STATUS " EXECUTORCH_BUILD_EXECUTOR_RUNNER : "
"${EXECUTORCH_BUILD_EXECUTOR_RUNNER}"
Expand Down Expand Up @@ -80,44 +97,29 @@ function(executorch_print_configuration_summary)
message(STATUS " EXECUTORCH_BUILD_HOST_TARGETS : "
"${EXECUTORCH_BUILD_HOST_TARGETS}"
)
message(
STATUS " EXECUTORCH_BUILD_MPS : ${EXECUTORCH_BUILD_MPS}"
)
message(
STATUS
" EXECUTORCH_BUILD_PYBIND : ${EXECUTORCH_BUILD_PYBIND}"
)
message(
STATUS " EXECUTORCH_BUILD_QNN : ${EXECUTORCH_BUILD_QNN}"
message(STATUS " EXECUTORCH_BUILD_KERNELS_CUSTOM : "
"${EXECUTORCH_BUILD_KERNELS_CUSTOM}"
)
message(STATUS " EXECUTORCH_BUILD_KERNELS_OPTIMIZED : "
"${EXECUTORCH_BUILD_KERNELS_OPTIMIZED}"
)
message(STATUS " EXECUTORCH_BUILD_KERNELS_QUANTIZED : "
"${EXECUTORCH_BUILD_KERNELS_QUANTIZED}"
)
message(
STATUS " EXECUTORCH_BUILD_SDK : ${EXECUTORCH_BUILD_SDK}"
)
message(
STATUS
" EXECUTORCH_BUILD_SIZE_TEST : ${EXECUTORCH_BUILD_SIZE_TEST}"
)
message(
STATUS
" EXECUTORCH_BUILD_XNNPACK : ${EXECUTORCH_BUILD_XNNPACK}"
" EXECUTORCH_BUILD_PTHREADPOOL : ${EXECUTORCH_BUILD_PTHREADPOOL}"
)
message(
STATUS
" EXECUTORCH_BUILD_VULKAN : ${EXECUTORCH_BUILD_VULKAN}"
" EXECUTORCH_BUILD_PYBIND : ${EXECUTORCH_BUILD_PYBIND}"
)
message(
STATUS
" EXECUTORCH_BUILD_PTHREADPOOL : ${EXECUTORCH_BUILD_PTHREADPOOL}"
STATUS " EXECUTORCH_BUILD_SDK : ${EXECUTORCH_BUILD_SDK}"
)
message(
STATUS
" EXECUTORCH_BUILD_CPUINFO : ${EXECUTORCH_BUILD_CPUINFO}"
" EXECUTORCH_BUILD_SIZE_TEST : ${EXECUTORCH_BUILD_SIZE_TEST}"
)

endfunction()
Expand Down

0 comments on commit e138ff2

Please sign in to comment.