Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Windows] misc fixes to enable Windows build #9198

Merged
merged 1 commit into from
Mar 14, 2025
Merged

[Windows] misc fixes to enable Windows build #9198

merged 1 commit into from
Mar 14, 2025

Conversation

SS-JIA
Copy link
Contributor

@SS-JIA SS-JIA commented Mar 12, 2025

Summary:

Context

Implement some small miscellaneous fixes to enable installing ExecuTorch on Windows. Currently, the optimized kernels cannot be built successfully but the installation works in custom ops are not built.

Installation

cd executorch

# Need to disable building custom kernels; there are a bunch of build errors
$env:EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT=0
# For some reason need to explicitly enable tensor extension otherwise I see a linker error
$env:CMAKE_ARGS="-DEXECUTORCH_BUILD_EXTENSION_TENSOR=ON"

python install_executorch.py

# Unset if you want
Remove-Item Env:\EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT
Remove-Item Env:\CMAKE_ARGS

Testing

I am able to follow the Getting Started tutorial successfully.

Outstanding Build Issues

Currently, with CMake I can successfully build with the following settings:

del -Recurse -Force cmake-out; `
  cmake . `
  -DCMAKE_INSTALL_PREFIX=cmake-out `
  -DPYTHON_EXECUTABLE=C:\\Users\\ssjia\\AppData\\Local\\miniconda3\\python.exe `
  -DCMAKE_PREFIX_PATH=C:\\Users\\ssjia\\AppData\\Local\\miniconda3\\Lib\\site-packages `
  -DCMAKE_BUILD_TYPE=Release `
  -DEXECUTORCH_BUILD_EXTENSION_TENSOR=ON `
  -DEXECUTORCH_BUILD_PYBIND=ON `
  -DEXECUTORCH_BUILD_XNNPACK=ON `
  -DEXECUTORCH_BUILD_KERNELS_QUANTIZED_AOT=ON `
  -DEXECUTORCH_BUILD_KERNELS_CUSTOM_AOT=ON `
  -DEXECUTORCH_BUILD_KERNELS_CUSTOM=OFF `
  -T ClangCL `
  -Bcmake-out; `
  cmake --build cmake-out -j64 --target install

If I switch EXECUTORCH_BUILD_KERNELS_CUSTOM to ON, then the build fails. The primary offenders appear to be the gelu. The implementation includes a header from ATen which is causing the majority of the errors.

Copy link

pytorch-bot bot commented Mar 12, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/9198

Note: Links to docs will display an error until the docs builds have been completed.

✅ You can merge normally! (1 Unrelated Failure)

As of commit 86e4837 with merge base 58d3582 (image):

BROKEN TRUNK - The following job failed but were present on the merge base:

👉 Rebase onto the `viable/strict` branch to avoid these failures

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Mar 12, 2025
@@ -49,7 +49,7 @@ if(CMAKE_SYSTEM_PROCESSOR MATCHES "^(aarch64|arm64|armv7)$")
list(APPEND _custom_ops__srcs
"extension/llm/custom_ops/spinquant/third-party/FFHT/fht_neon.c"
)
elseif(CMAKE_SYSTEM_PROCESSOR STREQUAL "x86_64")
elseif(CMAKE_SYSTEM_PROCESSOR MATCHES "^(x86_64|AMD64)")
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

apparently AMD64 is synonymous with x86_64. On my Windows machine this value is AMD64.

@SS-JIA SS-JIA added the release notes: build Changes related to build, including dependency upgrades, build flags, optimizations, etc. label Mar 12, 2025
add_library(xnnpack_schema INTERFACE ${_xnnpack_schema__outputs})
set_target_properties(xnnpack_schema PROPERTIES LINKER_LANGUAGE CXX)
target_include_directories(
xnnpack_schema INTERFACE ${_xnnpack_schema__include_dir}
${EXECUTORCH_ROOT}/third-party/flatbuffers/include
)

set(xnnpack_third_party pthreadpool cpuinfo)
set(xnnpack_third_party pthreadpool extension_threadpool cpuinfo)
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

required due to this recent refactor

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

without it, I see a linking error complaining that pthreadpool symbols cannot be found.


include(cmake/Dependencies.cmake)

list(TRANSFORM _xnnpack_backend__srcs PREPEND "${EXECUTORCH_ROOT}/")
add_library(xnnpack_backend STATIC ${_xnnpack_backend__srcs})
target_link_libraries(
xnnpack_backend PRIVATE ${xnnpack_third_party} executorch_core xnnpack_schema
xnnpack_backend PUBLIC ${xnnpack_third_party} executorch_core xnnpack_schema
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

changing to PUBLIC as suggested by @swolchok

relative to the cmake-out directory. May be an fnmatch-style
glob that matches exactly one file. If the path ends in `.so`,
this class will also look for similarly-named `.dylib` files.
src_dir: The directory of the file to install, relative to the cmake-out
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

these changes are lifed from #6979

@SS-JIA SS-JIA force-pushed the pr9198 branch 2 times, most recently from 566a7ed to b802f6e Compare March 12, 2025 22:55
@@ -73,34 +73,42 @@ foreach(fbs_file ${_xnnpack_schema__srcs})
)
endforeach()

if(WIN32)
set(MV_COMMAND powershell -Command "Move-Item -Path ${_xnnpack_flatbuffer__outputs} -Destination ${_xnnpack_schema__outputs}")

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

When cross-compiling for Windows, WIN32 is set in cmake, but there is no powershell

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

good point. I may address this in another PR though. There are some other places in our CMake files that don't account for cross compiling for windows. So currently I prefer to address it in a follow up.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

do you mind adding this as a comment in this cmake? worthwhile to keep this info for the follow up next time.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

yes, for sure.

SS-JIA added a commit to pytorch/pytorch that referenced this pull request Mar 13, 2025
Summary:
## Context

This PR is mostly to enable ExecuTorch build for Windows: pytorch/executorch#9198

In ExecuTorch, the optimized GeLU kernel calls the ATen implementation. However, on Windows `math.h` needs to be included with `#define _USE_MATH_DEFINES` in order for math constants to be defined.

Test Plan:
Rely on CI to make sure existing tests do not break. Tested separately with ExecuTorch to make sure Windows build is successful.
SS-JIA added a commit to pytorch/pytorch that referenced this pull request Mar 13, 2025
Summary:
## Context

This PR is mostly to enable ExecuTorch build for Windows: pytorch/executorch#9198

In ExecuTorch, the optimized GeLU kernel calls the ATen implementation. However, on Windows `math.h` needs to be included with `#define _USE_MATH_DEFINES` in order for math constants to be defined.

Test Plan:
Rely on CI to make sure existing tests do not break. Tested separately with ExecuTorch to make sure Windows build is successful.
SS-JIA added a commit to pytorch/pytorch that referenced this pull request Mar 14, 2025
Summary:
## Context

This PR is mostly to enable ExecuTorch build for Windows: pytorch/executorch#9198

In ExecuTorch, the optimized GeLU kernel calls the ATen implementation. However, on Windows `math.h` needs to be included with `#define _USE_MATH_DEFINES` in order for math constants to be defined.

Test Plan:
Rely on CI to make sure existing tests do not break. Tested separately with ExecuTorch to make sure Windows build is successful.
)
)
rem Under windows, it's always python
set PYTHON_EXECUTABLE=python
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

these changes are lifed from #6979

Copy link
Contributor

@mcr229 mcr229 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

lgtm stamping to unblock

Summary:
## Context

Implement some small miscellaneous fixes to enable installing ExecuTorch on Windows. Currently, the optimized kernels cannot be built successfully but the installation works in custom ops are not built.

## Installation

```powershell
cd executorch

# Need to disable building custom kernels; there are a bunch of build errors
$env:EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT=0
# For some reason need to explicitly enable tensor extension otherwise I see a linker error
$env:CMAKE_ARGS="-DEXECUTORCH_BUILD_EXTENSION_TENSOR=ON"

python install_executorch.py

# Unset if you want
Remove-Item Env:\EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT
Remove-Item Env:\CMAKE_ARGS
```

## Testing

I am able to follow the [Getting Started tutorial](https://pytorch.org/executorch/stable/getting-started-setup.html#run-your-program) successfully.

## Outstanding Build Issues

Currently, with CMake I can successfully build with the following settings:

```
del -Recurse -Force cmake-out; `
  cmake . `
  -DCMAKE_INSTALL_PREFIX=cmake-out `
  -DPYTHON_EXECUTABLE=C:\\Users\\ssjia\\AppData\\Local\\miniconda3\\python.exe `
  -DCMAKE_PREFIX_PATH=C:\\Users\\ssjia\\AppData\\Local\\miniconda3\\Lib\\site-packages `
  -DCMAKE_BUILD_TYPE=Release `
  -DEXECUTORCH_BUILD_EXTENSION_TENSOR=ON `
  -DEXECUTORCH_BUILD_PYBIND=ON `
  -DEXECUTORCH_BUILD_XNNPACK=ON `
  -DEXECUTORCH_BUILD_KERNELS_QUANTIZED_AOT=ON `
  -DEXECUTORCH_BUILD_KERNELS_CUSTOM_AOT=ON `
  -DEXECUTORCH_BUILD_KERNELS_CUSTOM=OFF `
  -T ClangCL `
  -Bcmake-out; `
  cmake --build cmake-out -j64 --target install
```

If I switch `EXECUTORCH_BUILD_KERNELS_CUSTOM` to `ON`, then the build fails. The primary offenders appear to be the `gelu`. The implementation includes a header from ATen which is causing the majority of the errors.
if(WIN32)
# flatbuffers does not use CMAKE_BUILD_TYPE. Internally, the build forces Release
# config, but from CMake's perspective the build type is always Debug.
set(FLATC_EXECUTABLE ${BINARY_DIR}/$<CONFIG>/flatc.exe)
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@jathu this is the change that was needed

Copy link
Contributor

@jathu jathu left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

flatc changes lgtm

@SS-JIA SS-JIA merged commit 3137889 into main Mar 14, 2025
81 of 154 checks passed
@SS-JIA SS-JIA deleted the pr9198 branch March 14, 2025 23:36
pytorchmergebot pushed a commit to pytorch/pytorch that referenced this pull request Mar 14, 2025
Summary:
## Context

This PR is mostly to enable ExecuTorch build for Windows: pytorch/executorch#9198

In ExecuTorch, the optimized GeLU kernel calls the ATen implementation. However, on Windows `math.h` needs to be included with `#define _USE_MATH_DEFINES` in order for math constants to be defined.

Test Plan:
Rely on CI to make sure existing tests do not break. Tested separately with ExecuTorch to make sure Windows build is successful.

Pull Request resolved: #149164
Approved by: https://github.com/swolchok
DannyYuyang-quic pushed a commit to CodeLinaro/executorch that referenced this pull request Apr 2, 2025
Summary:
## Context

Implement some small miscellaneous fixes to enable installing ExecuTorch
on Windows. Currently, the optimized kernels cannot be built
successfully but the installation works in custom ops are not built.

## Installation

```powershell
cd executorch

# Need to disable building custom kernels; there are a bunch of build errors
$env:EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT=0
# For some reason need to explicitly enable tensor extension otherwise I see a linker error
$env:CMAKE_ARGS="-DEXECUTORCH_BUILD_EXTENSION_TENSOR=ON"

python install_executorch.py

# Unset if you want
Remove-Item Env:\EXECUTORCH_BUILD_KERNELS_CUSTOM_AOT
Remove-Item Env:\CMAKE_ARGS
```

## Testing

I am able to follow the [Getting Started
tutorial](https://pytorch.org/executorch/stable/getting-started-setup.html#run-your-program)
successfully.

## Outstanding Build Issues

Currently, with CMake I can successfully build with the following
settings:

```
del -Recurse -Force cmake-out; `
  cmake . `
  -DCMAKE_INSTALL_PREFIX=cmake-out `
  -DPYTHON_EXECUTABLE=C:\\Users\\ssjia\\AppData\\Local\\miniconda3\\python.exe `
  -DCMAKE_PREFIX_PATH=C:\\Users\\ssjia\\AppData\\Local\\miniconda3\\Lib\\site-packages `
  -DCMAKE_BUILD_TYPE=Release `
  -DEXECUTORCH_BUILD_EXTENSION_TENSOR=ON `
  -DEXECUTORCH_BUILD_PYBIND=ON `
  -DEXECUTORCH_BUILD_XNNPACK=ON `
  -DEXECUTORCH_BUILD_KERNELS_QUANTIZED_AOT=ON `
  -DEXECUTORCH_BUILD_KERNELS_CUSTOM_AOT=ON `
  -DEXECUTORCH_BUILD_KERNELS_CUSTOM=OFF `
  -T ClangCL `
  -Bcmake-out; `
  cmake --build cmake-out -j64 --target install
```

If I switch `EXECUTORCH_BUILD_KERNELS_CUSTOM` to `ON`, then the build
fails. The primary offenders appear to be the `gelu`. The implementation
includes a header from ATen which is causing the majority of the errors.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. release notes: build Changes related to build, including dependency upgrades, build flags, optimizations, etc.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants