[arch-commits] Commit in python-pytorch/repos (3 files)
Konstantin Gizdov
kgizdov at archlinux.org
Wed Jun 26 19:52:31 UTC 2019
Date: Wednesday, June 26, 2019 @ 19:52:30
Author: kgizdov
Revision: 484604
archrelease: copy trunk to community-testing-x86_64
Added:
python-pytorch/repos/community-testing-x86_64/
python-pytorch/repos/community-testing-x86_64/PKGBUILD
(from rev 484603, python-pytorch/trunk/PKGBUILD)
python-pytorch/repos/community-testing-x86_64/change_default_config.patch
(from rev 484603, python-pytorch/trunk/change_default_config.patch)
-----------------------------+
PKGBUILD | 158 ++++++++++++++++++++++++++++++++++++++++++
change_default_config.patch | 28 +++++++
2 files changed, 186 insertions(+)
Copied: python-pytorch/repos/community-testing-x86_64/PKGBUILD (from rev 484603, python-pytorch/trunk/PKGBUILD)
===================================================================
--- community-testing-x86_64/PKGBUILD (rev 0)
+++ community-testing-x86_64/PKGBUILD 2019-06-26 19:52:30 UTC (rev 484604)
@@ -0,0 +1,158 @@
+# Maintainer: Sven-Hendrik Haase <sh at lutzhaase.com>
+# Contributor: Stephen Zhang <zsrkmyn at gmail dot com>
+
+pkgbase="python-pytorch"
+pkgname=("python-pytorch" "python-pytorch-opt" "python-pytorch-cuda" "python-pytorch-opt-cuda")
+_pkgname="pytorch"
+pkgver=1.1.0
+pkgrel=7
+pkgdesc="Tensors and Dynamic neural networks in Python with strong GPU acceleration"
+arch=('x86_64')
+url="https://pytorch.org"
+license=('BSD')
+depends=('google-glog' 'gflags' 'opencv' 'openmp' 'nccl' 'pybind11' 'python' 'python-yaml' 'python-numpy' 'protobuf' 'ffmpeg')
+makedepends=('python' 'python-setuptools' 'python-yaml' 'python-numpy' 'cmake' 'cuda' 'cudnn' 'git')
+source=("${_pkgname}-${pkgver}::git+https://github.com/pytorch/pytorch.git#tag=v$pkgver")
+sha256sums=('SKIP')
+
+get_pyver () {
+ python -c 'import sys; print(str(sys.version_info[0]) + "." + str(sys.version_info[1]))'
+}
+
+prepare() {
+ cd "${_pkgname}-${pkgver}"
+
+ # This is the lazy way since pytorch has sooo many submodules and they keep
+ # changing them around but we've run into more problems so far doing it the
+ # manual than the lazy way. This lazy way (not explicitly specifying all
+ # submodules) will make building inefficient but for now I'll take it.
+ # It will result in the same package, don't worry.
+ git submodule update --init --recursive
+
+ cd ..
+
+ cp -a "${_pkgname}-${pkgver}" "${_pkgname}-${pkgver}-opt"
+ cp -a "${_pkgname}-${pkgver}" "${_pkgname}-${pkgver}-cuda"
+ cp -a "${_pkgname}-${pkgver}" "${_pkgname}-${pkgver}-opt-cuda"
+
+ export CC=gcc-8
+ export CXX=g++-8
+ export VERBOSE=1
+ export PYTORCH_BUILD_VERSION="${pkgver}"
+ export PYTORCH_BUILD_NUMBER=1
+
+ # Check tools/setup_helpers/cmake.py, setup.py and CMakeLists.txt for a list of flags that can be set via env vars.
+ export USE_MKLDNN=OFF
+ # export BUILD_CUSTOM_PROTOBUF=OFF
+ # export BUILD_SHARED_LIBS=OFF
+ export USE_FFMPEG=ON
+ export USE_GFLAGS=ON
+ export USE_GLOG=ON
+ export BUILD_BINARY=ON
+ export USE_OPENCV=ON
+ export USE_SYSTEM_NCCL=ON
+ export CUDAHOSTCXX=g++-8
+ export CUDA_HOME=/opt/cuda
+ export CUDNN_LIB_DIR=/usr/lib
+ export CUDNN_INCLUDE_DIR=/usr/include
+ export TORCH_NVCC_FLAGS="-Xfatbin -compress-all"
+ export TORCH_CUDA_ARCH_LIST="3.0;3.2;3.5;3.7;5.0;5.2;5.3;6.0;6.1;6.2;7.0;7.2;7.5"
+}
+
+build() {
+ echo "Building without cuda and without non-x86-64 optimizations"
+ export NO_CUDA=1
+ export WITH_CUDNN=0
+ cd "${srcdir}/${_pkgname}-${pkgver}"
+ python setup.py build
+
+
+ echo "Building without cuda and with non-x86-64 optimizations"
+ export NO_CUDA=1
+ export WITH_CUDNN=0
+ cd "${srcdir}/${_pkgname}-${pkgver}-opt"
+ echo "add_definitions(-march=haswell)" >> cmake/MiscCheck.cmake
+ python setup.py build
+
+
+ echo "Building with cuda and without non-x86-64 optimizations"
+ export NO_CUDA=0
+ export WITH_CUDNN=1
+ cd "${srcdir}/${_pkgname}-${pkgver}-cuda"
+ python setup.py build
+
+
+ echo "Building with cuda and with non-x86-64 optimizations"
+ export NO_CUDA=0
+ export WITH_CUDNN=1
+ cd "${srcdir}/${_pkgname}-${pkgver}-opt-cuda"
+ echo "add_definitions(-march=haswell)" >> cmake/MiscCheck.cmake
+ python setup.py build
+}
+
+_package() {
+ # Prevent setup.py from re-running CMake and rebuilding
+ sed -e 's/RUN_BUILD_DEPS = True/RUN_BUILD_DEPS = False/g' -i setup.py
+
+ python setup.py install --root="${pkgdir}"/ --optimize=1 --skip-build
+
+ install -Dm644 LICENSE "${pkgdir}/usr/share/licenses/${pkgname}/LICENSE"
+
+ pytorchpath="usr/lib/python$(get_pyver)/site-packages/torch"
+ install -d "${pkgdir}/usr/lib"
+
+ # put CMake files in correct place
+ mv "${pkgdir}/${pytorchpath}/share/cmake" "${pkgdir}/usr/lib/cmake"
+
+ # put C++ API in correct place
+ mv "${pkgdir}/${pytorchpath}/include" "${pkgdir}/usr/include"
+ mv "${pkgdir}/${pytorchpath}/lib"/*.so* "${pkgdir}/usr/lib/"
+
+ # clean up duplicates
+ # TODO: move towards direct shared library dependecy of:
+ # c10, caffe2, libcpuinfo, CUDA RT, gloo, GTest, Intel MKL,
+ # NVRTC, ONNX, protobuf, libthreadpool, QNNPACK
+ rm -rf "${pkgdir}/usr/include/pybind11"
+
+ # python module is hardcoded to look there at runtime
+ ln -s /usr/include "${pkgdir}/${pytorchpath}/include"
+ find "${pkgdir}"/usr/lib -type f -name "*.so*" -print0 | while read -rd $'\0' _lib; do
+ ln -s ${_lib#"$pkgdir"} "${pkgdir}/${pytorchpath}/lib/"
+ done
+}
+
+package_python-pytorch() {
+ cd "${srcdir}/${_pkgname}-${pkgver}"
+ _package
+}
+
+package_python-pytorch-opt() {
+ pkgdesc="Tensors and Dynamic neural networks in Python with strong GPU acceleration (with CPU optimizations)"
+ conflicts=(python-pytorch)
+ provides=(python-pytorch)
+
+ cd "${srcdir}/${_pkgname}-${pkgver}-opt"
+ _package
+}
+
+package_python-pytorch-cuda() {
+ pkgdesc="Tensors and Dynamic neural networks in Python with strong GPU acceleration (with CUDA)"
+ depends+=(cuda cudnn)
+ conflicts=(python-pytorch)
+ provides=(python-pytorch)
+
+ cd "${srcdir}/${_pkgname}-${pkgver}-cuda"
+ _package
+}
+
+package_python-pytorch-opt-cuda() {
+ pkgdesc="Tensors and Dynamic neural networks in Python with strong GPU acceleration (with CUDA and CPU optimizations)"
+ depends+=(cuda cudnn)
+ conflicts=(python-pytorch)
+ provides=(python-pytorch python-pytorch-cuda)
+
+ cd "${srcdir}/${_pkgname}-${pkgver}-opt-cuda"
+ _package
+}
+
+# vim:set ts=2 sw=2 et:
Copied: python-pytorch/repos/community-testing-x86_64/change_default_config.patch (from rev 484603, python-pytorch/trunk/change_default_config.patch)
===================================================================
--- community-testing-x86_64/change_default_config.patch (rev 0)
+++ community-testing-x86_64/change_default_config.patch 2019-06-26 19:52:30 UTC (rev 484604)
@@ -0,0 +1,28 @@
+diff --git a/CMakeLists.txt b/CMakeLists.txt
+index 71f8b693a..07e564919 100644
+--- a/CMakeLists.txt
++++ b/CMakeLists.txt
+@@ -67,10 +67,10 @@ option(BUILD_ATEN_MOBILE "Build ATen for Android and iOS" OFF)
+ option(BUILD_ATEN_ONLY "Build only a subset focused on ATen only" OFF)
+ option(BUILD_BINARY "Build C++ binaries" OFF)
+ option(BUILD_DOCS "Build Caffe2 documentation" OFF)
+-option(BUILD_CUSTOM_PROTOBUF "Build and use Caffe2's own protobuf under third_party" ON)
++option(BUILD_CUSTOM_PROTOBUF "Build and use Caffe2's own protobuf under third_party" OFF)
+ option(BUILD_PYTHON "Build Python binaries" ON)
+ option(BUILD_CAFFE2_OPS "Build Caffe2 operators" ON)
+-option(BUILD_SHARED_LIBS "Build libcaffe2.so" ON)
++option(BUILD_SHARED_LIBS "Build libcaffe2.so" OFF)
+ cmake_dependent_option(
+ CAFFE2_LINK_LOCAL_PROTOBUF "If set, build protobuf inside libcaffe2.so." ON
+ "BUILD_SHARED_LIBS AND BUILD_CUSTOM_PROTOBUF" OFF)
+@@ -90,8 +90,8 @@ cmake_dependent_option(
+ "USE_CUDA" OFF)
+ option(USE_FBGEMM "Use FBGEMM (quantized 8-bit server operators)" OFF)
+ option(USE_FFMPEG "Use ffmpeg" OFF)
+-option(USE_GFLAGS "Use GFLAGS" OFF)
+-option(USE_GLOG "Use GLOG" OFF)
++option(USE_GFLAGS "Use GFLAGS" ON)
++option(USE_GLOG "Use GLOG" ON)
+ option(USE_LEVELDB "Use LEVELDB" OFF)
+ option(USE_LITE_PROTO "Use lite protobuf instead of full." OFF)
+ option(USE_LMDB "Use LMDB" OFF)
More information about the arch-commits
mailing list