commit:     0a916de9da679ab13ef5bbb3a3c006120269b8e0
Author:     Michael Mair-Keimberger <mmk <AT> levelnine <DOT> at>
AuthorDate: Mon Dec 23 17:57:20 2024 +0000
Commit:     Alfredo Tupone <tupone <AT> gentoo <DOT> org>
CommitDate: Mon Dec 23 18:39:13 2024 +0000
URL:        https://gitweb.gentoo.org/repo/gentoo.git/commit/?id=0a916de9

sci-libs/caffe2: remove unused patches

Signed-off-by: Michael Mair-Keimberger <mmk <AT> levelnine.at>
Closes: https://github.com/gentoo/gentoo/pull/39829
Signed-off-by: Alfredo Tupone <tupone <AT> gentoo.org>

 .../files/caffe2-2.5.1-exclude-aotriton.patch      | 22 ----------------------
 .../files/caffe2-2.5.1-fix-functorch-install.patch |  8 --------
 .../caffe2/files/caffe2-2.5.1-install-dirs.patch   | 11 -----------
 3 files changed, 41 deletions(-)

diff --git a/sci-libs/caffe2/files/caffe2-2.5.1-exclude-aotriton.patch 
b/sci-libs/caffe2/files/caffe2-2.5.1-exclude-aotriton.patch
deleted file mode 100644
index 340d6e45c95a..000000000000
--- a/sci-libs/caffe2/files/caffe2-2.5.1-exclude-aotriton.patch
+++ /dev/null
@@ -1,22 +0,0 @@
-Disables aotriton download when both USE_FLASH_ATTENTION and 
USE_MEM_EFF_ATTENTION cmake flags are OFF
-Backports upstream PR to 2.3.0: https://github.com/pytorch/pytorch/pull/130197
---- a/aten/src/ATen/native/transformers/cuda/sdp_utils.cpp
-+++ b/aten/src/ATen/native/transformers/cuda/sdp_utils.cpp
-@@ -659,7 +659,7 @@ bool can_use_mem_efficient_attention(sdp_params const& 
params, bool debug) {
-       array_of<at::ScalarType>(at::kHalf, at::kFloat, at::kBFloat16);
-   constexpr auto less_than_sm80_mem_efficient_dtypes =
-       array_of<at::ScalarType>(at::kHalf, at::kFloat);
--#ifdef USE_ROCM
-+#if defined(USE_ROCM) && defined(USE_MEM_EFF_ATTENTION)
-   constexpr auto aotriton_mem_efficient_dtypes =
-       array_of<at::ScalarType>(at::kHalf, at::kFloat, at::kBFloat16);
- #endif
-@@ -709,7 +709,7 @@ bool can_use_mem_efficient_attention(sdp_params const& 
params, bool debug) {
-     }
-   }
- 
--#ifdef USE_ROCM
-+#if defined(USE_ROCM) && defined(USE_MEM_EFF_ATTENTION)
-   return check_tensor_dtype(params, aotriton_mem_efficient_dtypes, debug);
- #else
-   auto dprop = at::cuda::getCurrentDeviceProperties();

diff --git a/sci-libs/caffe2/files/caffe2-2.5.1-fix-functorch-install.patch 
b/sci-libs/caffe2/files/caffe2-2.5.1-fix-functorch-install.patch
deleted file mode 100644
index ffce0028a556..000000000000
--- a/sci-libs/caffe2/files/caffe2-2.5.1-fix-functorch-install.patch
+++ /dev/null
@@ -1,8 +0,0 @@
---- a/functorch/CMakeLists.txt 2024-11-10 11:18:29.151992840 +0100
-+++ b/functorch/CMakeLists.txt 2024-11-10 11:20:19.642389982 +0100
-@@ -42,4 +42,4 @@
- if(NOT ${TORCH_PYTHON_LINK_FLAGS} STREQUAL "")
-   set_target_properties(${PROJECT_NAME} PROPERTIES LINK_FLAGS 
${TORCH_PYTHON_LINK_FLAGS})
- endif()
--install(TARGETS ${PROJECT_NAME} DESTINATION "${CMAKE_CURRENT_SOURCE_DIR}")
-+install(TARGETS ${PROJECT_NAME} DESTINATION "${CMAKE_INSTALL_LIBDIR}")

diff --git a/sci-libs/caffe2/files/caffe2-2.5.1-install-dirs.patch 
b/sci-libs/caffe2/files/caffe2-2.5.1-install-dirs.patch
deleted file mode 100644
index e99b7e59cb5b..000000000000
--- a/sci-libs/caffe2/files/caffe2-2.5.1-install-dirs.patch
+++ /dev/null
@@ -1,11 +0,0 @@
---- a/c10/CMakeLists.txt
-+++ b/c10/CMakeLists.txt
-@@ -162,7 +162,7 @@ if(NOT BUILD_LIBTORCHLESS)
-   # Note: for now, we will put all export path into one single Caffe2Targets 
group
-   # to deal with the cmake deployment need. Inside the Caffe2Targets set, the
-   # individual libraries like libc10.so and libcaffe2.so are still 
self-contained.
--  install(TARGETS c10 EXPORT Caffe2Targets DESTINATION lib)
-+  install(TARGETS c10 EXPORT Caffe2Targets DESTINATION 
${CMAKE_INSTALL_LIBDIR})
- endif()
- 
- install(DIRECTORY ${CMAKE_CURRENT_LIST_DIR}

Reply via email to