commit: 0a916de9da679ab13ef5bbb3a3c006120269b8e0 Author: Michael Mair-Keimberger <mmk <AT> levelnine <DOT> at> AuthorDate: Mon Dec 23 17:57:20 2024 +0000 Commit: Alfredo Tupone <tupone <AT> gentoo <DOT> org> CommitDate: Mon Dec 23 18:39:13 2024 +0000 URL: https://gitweb.gentoo.org/repo/gentoo.git/commit/?id=0a916de9
sci-libs/caffe2: remove unused patches Signed-off-by: Michael Mair-Keimberger <mmk <AT> levelnine.at> Closes: https://github.com/gentoo/gentoo/pull/39829 Signed-off-by: Alfredo Tupone <tupone <AT> gentoo.org> .../files/caffe2-2.5.1-exclude-aotriton.patch | 22 ---------------------- .../files/caffe2-2.5.1-fix-functorch-install.patch | 8 -------- .../caffe2/files/caffe2-2.5.1-install-dirs.patch | 11 ----------- 3 files changed, 41 deletions(-) diff --git a/sci-libs/caffe2/files/caffe2-2.5.1-exclude-aotriton.patch b/sci-libs/caffe2/files/caffe2-2.5.1-exclude-aotriton.patch deleted file mode 100644 index 340d6e45c95a..000000000000 --- a/sci-libs/caffe2/files/caffe2-2.5.1-exclude-aotriton.patch +++ /dev/null @@ -1,22 +0,0 @@ -Disables aotriton download when both USE_FLASH_ATTENTION and USE_MEM_EFF_ATTENTION cmake flags are OFF -Backports upstream PR to 2.3.0: https://github.com/pytorch/pytorch/pull/130197 ---- a/aten/src/ATen/native/transformers/cuda/sdp_utils.cpp -+++ b/aten/src/ATen/native/transformers/cuda/sdp_utils.cpp -@@ -659,7 +659,7 @@ bool can_use_mem_efficient_attention(sdp_params const& params, bool debug) { - array_of<at::ScalarType>(at::kHalf, at::kFloat, at::kBFloat16); - constexpr auto less_than_sm80_mem_efficient_dtypes = - array_of<at::ScalarType>(at::kHalf, at::kFloat); --#ifdef USE_ROCM -+#if defined(USE_ROCM) && defined(USE_MEM_EFF_ATTENTION) - constexpr auto aotriton_mem_efficient_dtypes = - array_of<at::ScalarType>(at::kHalf, at::kFloat, at::kBFloat16); - #endif -@@ -709,7 +709,7 @@ bool can_use_mem_efficient_attention(sdp_params const& params, bool debug) { - } - } - --#ifdef USE_ROCM -+#if defined(USE_ROCM) && defined(USE_MEM_EFF_ATTENTION) - return check_tensor_dtype(params, aotriton_mem_efficient_dtypes, debug); - #else - auto dprop = at::cuda::getCurrentDeviceProperties(); diff --git a/sci-libs/caffe2/files/caffe2-2.5.1-fix-functorch-install.patch b/sci-libs/caffe2/files/caffe2-2.5.1-fix-functorch-install.patch deleted file mode 100644 index ffce0028a556..000000000000 --- a/sci-libs/caffe2/files/caffe2-2.5.1-fix-functorch-install.patch +++ /dev/null @@ -1,8 +0,0 @@ ---- a/functorch/CMakeLists.txt 2024-11-10 11:18:29.151992840 +0100 -+++ b/functorch/CMakeLists.txt 2024-11-10 11:20:19.642389982 +0100 -@@ -42,4 +42,4 @@ - if(NOT ${TORCH_PYTHON_LINK_FLAGS} STREQUAL "") - set_target_properties(${PROJECT_NAME} PROPERTIES LINK_FLAGS ${TORCH_PYTHON_LINK_FLAGS}) - endif() --install(TARGETS ${PROJECT_NAME} DESTINATION "${CMAKE_CURRENT_SOURCE_DIR}") -+install(TARGETS ${PROJECT_NAME} DESTINATION "${CMAKE_INSTALL_LIBDIR}") diff --git a/sci-libs/caffe2/files/caffe2-2.5.1-install-dirs.patch b/sci-libs/caffe2/files/caffe2-2.5.1-install-dirs.patch deleted file mode 100644 index e99b7e59cb5b..000000000000 --- a/sci-libs/caffe2/files/caffe2-2.5.1-install-dirs.patch +++ /dev/null @@ -1,11 +0,0 @@ ---- a/c10/CMakeLists.txt -+++ b/c10/CMakeLists.txt -@@ -162,7 +162,7 @@ if(NOT BUILD_LIBTORCHLESS) - # Note: for now, we will put all export path into one single Caffe2Targets group - # to deal with the cmake deployment need. Inside the Caffe2Targets set, the - # individual libraries like libc10.so and libcaffe2.so are still self-contained. -- install(TARGETS c10 EXPORT Caffe2Targets DESTINATION lib) -+ install(TARGETS c10 EXPORT Caffe2Targets DESTINATION ${CMAKE_INSTALL_LIBDIR}) - endif() - - install(DIRECTORY ${CMAKE_CURRENT_LIST_DIR}
