[arch-commits] Commit in libvpx/repos/extra-x86_64 (3 files)
Maxime Gauduin
alucryd at archlinux.org
Wed Jul 17 12:16:34 UTC 2019
Date: Wednesday, July 17, 2019 @ 12:16:34
Author: alucryd
Revision: 358095
archrelease: copy trunk to extra-x86_64
Added:
libvpx/repos/extra-x86_64/PKGBUILD
(from rev 358094, libvpx/trunk/PKGBUILD)
Deleted:
libvpx/repos/extra-x86_64/0001-Fix-for-issue-1114-compile-error.patch
libvpx/repos/extra-x86_64/PKGBUILD
---------------------------------------------+
0001-Fix-for-issue-1114-compile-error.patch | 317 --------------------------
PKGBUILD | 90 +++----
2 files changed, 45 insertions(+), 362 deletions(-)
Deleted: 0001-Fix-for-issue-1114-compile-error.patch
===================================================================
--- 0001-Fix-for-issue-1114-compile-error.patch 2019-07-17 12:16:28 UTC (rev 358094)
+++ 0001-Fix-for-issue-1114-compile-error.patch 2019-07-17 12:16:34 UTC (rev 358095)
@@ -1,317 +0,0 @@
-From 7361ef732b432e153496c30da66081d7e530c7f6 Mon Sep 17 00:00:00 2001
-From: Peter de Rivaz <peter.derivaz at argondesign.com>
-Date: Mon, 14 Dec 2015 16:35:29 +0000
-Subject: [PATCH] Fix for issue 1114 compile error
-
-In 32-bit build with --enable-shared, there is a lot of
-register pressure and register src_strideq is reused.
-The code needs to use the stack based version of src_stride,
-but this doesn't compile when used in an lea instruction.
-
-This patch also fixes a related segmentation fault caused by the
-implementation using src_strideq even though it has been
-reused.
-
-This patch also fixes the HBD subpel variance tests that fail
-when compiled without disable-optimizations.
-These failures were caused by local variables in the assembler
-routines colliding with the caller's stack frame.
-
-Change-Id: Ice9d4dafdcbdc6038ad5ee7c1c09a8f06deca362
----
- vpx_dsp/x86/highbd_subpel_variance_impl_sse2.asm | 18 +++----
- vpx_dsp/x86/highbd_variance_sse2.c | 64 ++++++++++++++----------
- 2 files changed, 44 insertions(+), 38 deletions(-)
-
-diff --git a/vpx_dsp/x86/highbd_subpel_variance_impl_sse2.asm b/vpx_dsp/x86/highbd_subpel_variance_impl_sse2.asm
-index 22d52a2..30ee81b 100644
---- a/vpx_dsp/x86/highbd_subpel_variance_impl_sse2.asm
-+++ b/vpx_dsp/x86/highbd_subpel_variance_impl_sse2.asm
-@@ -79,20 +79,13 @@ SECTION .text
-
- %macro INC_SRC_BY_SRC_STRIDE 0
- %if ARCH_X86=1 && CONFIG_PIC=1
-- lea srcq, [srcq + src_stridemp*2]
-+ add srcq, src_stridemp
-+ add srcq, src_stridemp
- %else
- lea srcq, [srcq + src_strideq*2]
- %endif
- %endmacro
-
--%macro INC_SRC_BY_SRC_2STRIDE 0
--%if ARCH_X86=1 && CONFIG_PIC=1
-- lea srcq, [srcq + src_stridemp*4]
--%else
-- lea srcq, [srcq + src_strideq*4]
--%endif
--%endmacro
--
- %macro SUBPEL_VARIANCE 1-2 0 ; W
- %define bilin_filter_m bilin_filter_m_sse2
- %define filter_idx_shift 5
-@@ -984,8 +977,9 @@ SECTION .text
- .x_other_y_other_loop:
- movu m2, [srcq]
- movu m4, [srcq+2]
-- movu m3, [srcq+src_strideq*2]
-- movu m5, [srcq+src_strideq*2+2]
-+ INC_SRC_BY_SRC_STRIDE
-+ movu m3, [srcq]
-+ movu m5, [srcq+2]
- pmullw m2, filter_x_a
- pmullw m4, filter_x_b
- paddw m2, filter_rnd
-@@ -1018,7 +1012,7 @@ SECTION .text
- SUM_SSE m0, m2, m4, m3, m6, m7
- mova m0, m5
-
-- INC_SRC_BY_SRC_2STRIDE
-+ INC_SRC_BY_SRC_STRIDE
- lea dstq, [dstq + dst_strideq * 4]
- %if %2 == 1 ; avg
- add secq, sec_str
-diff --git a/vpx_dsp/x86/highbd_variance_sse2.c b/vpx_dsp/x86/highbd_variance_sse2.c
-index b45331c..81ec5db 100644
---- a/vpx_dsp/x86/highbd_variance_sse2.c
-+++ b/vpx_dsp/x86/highbd_variance_sse2.c
-@@ -243,13 +243,18 @@ unsigned int vpx_highbd_12_mse8x8_sse2(const uint8_t *src8, int src_stride,
- }
-
- #if CONFIG_USE_X86INC
-+// The 2 unused parameters are place holders for PIC enabled build.
-+// These definitions are for functions defined in
-+// highbd_subpel_variance_impl_sse2.asm
- #define DECL(w, opt) \
- int vpx_highbd_sub_pixel_variance##w##xh_##opt(const uint16_t *src, \
- ptrdiff_t src_stride, \
- int x_offset, int y_offset, \
- const uint16_t *dst, \
- ptrdiff_t dst_stride, \
-- int height, unsigned int *sse);
-+ int height, \
-+ unsigned int *sse, \
-+ void *unused0, void *unused);
- #define DECLS(opt1, opt2) \
- DECL(8, opt1); \
- DECL(16, opt1)
-@@ -274,7 +279,7 @@ uint32_t vpx_highbd_8_sub_pixel_variance##w##x##h##_##opt(const uint8_t *src8, \
- int se = vpx_highbd_sub_pixel_variance##wf##xh_##opt(src, src_stride, \
- x_offset, y_offset, \
- dst, dst_stride, h, \
-- &sse); \
-+ &sse, NULL, NULL); \
- if (w > wf) { \
- unsigned int sse2; \
- int se2 = vpx_highbd_sub_pixel_variance##wf##xh_##opt(src + 16, \
-@@ -282,19 +287,20 @@ uint32_t vpx_highbd_8_sub_pixel_variance##w##x##h##_##opt(const uint8_t *src8, \
- x_offset, y_offset, \
- dst + 16, \
- dst_stride, \
-- h, &sse2); \
-+ h, &sse2, \
-+ NULL, NULL); \
- se += se2; \
- sse += sse2; \
- if (w > wf * 2) { \
- se2 = vpx_highbd_sub_pixel_variance##wf##xh_##opt(src + 32, src_stride, \
- x_offset, y_offset, \
- dst + 32, dst_stride, \
-- h, &sse2); \
-+ h, &sse2, NULL, NULL); \
- se += se2; \
- sse += sse2; \
- se2 = vpx_highbd_sub_pixel_variance##wf##xh_##opt( \
- src + 48, src_stride, x_offset, y_offset, \
-- dst + 48, dst_stride, h, &sse2); \
-+ dst + 48, dst_stride, h, &sse2, NULL, NULL); \
- se += se2; \
- sse += sse2; \
- } \
-@@ -312,7 +318,7 @@ uint32_t vpx_highbd_10_sub_pixel_variance##w##x##h##_##opt( \
- int se = vpx_highbd_sub_pixel_variance##wf##xh_##opt(src, src_stride, \
- x_offset, y_offset, \
- dst, dst_stride, \
-- h, &sse); \
-+ h, &sse, NULL, NULL); \
- if (w > wf) { \
- uint32_t sse2; \
- int se2 = vpx_highbd_sub_pixel_variance##wf##xh_##opt(src + 16, \
-@@ -320,20 +326,21 @@ uint32_t vpx_highbd_10_sub_pixel_variance##w##x##h##_##opt( \
- x_offset, y_offset, \
- dst + 16, \
- dst_stride, \
-- h, &sse2); \
-+ h, &sse2, \
-+ NULL, NULL); \
- se += se2; \
- sse += sse2; \
- if (w > wf * 2) { \
- se2 = vpx_highbd_sub_pixel_variance##wf##xh_##opt(src + 32, src_stride, \
- x_offset, y_offset, \
- dst + 32, dst_stride, \
-- h, &sse2); \
-+ h, &sse2, NULL, NULL); \
- se += se2; \
- sse += sse2; \
- se2 = vpx_highbd_sub_pixel_variance##wf##xh_##opt(src + 48, src_stride, \
- x_offset, y_offset, \
- dst + 48, dst_stride, \
-- h, &sse2); \
-+ h, &sse2, NULL, NULL); \
- se += se2; \
- sse += sse2; \
- } \
-@@ -359,27 +366,27 @@ uint32_t vpx_highbd_12_sub_pixel_variance##w##x##h##_##opt( \
- int se2 = vpx_highbd_sub_pixel_variance##wf##xh_##opt( \
- src + (start_row * src_stride), src_stride, \
- x_offset, y_offset, dst + (start_row * dst_stride), \
-- dst_stride, height, &sse2); \
-+ dst_stride, height, &sse2, NULL, NULL); \
- se += se2; \
- long_sse += sse2; \
- if (w > wf) { \
- se2 = vpx_highbd_sub_pixel_variance##wf##xh_##opt( \
- src + 16 + (start_row * src_stride), src_stride, \
- x_offset, y_offset, dst + 16 + (start_row * dst_stride), \
-- dst_stride, height, &sse2); \
-+ dst_stride, height, &sse2, NULL, NULL); \
- se += se2; \
- long_sse += sse2; \
- if (w > wf * 2) { \
- se2 = vpx_highbd_sub_pixel_variance##wf##xh_##opt( \
- src + 32 + (start_row * src_stride), src_stride, \
- x_offset, y_offset, dst + 32 + (start_row * dst_stride), \
-- dst_stride, height, &sse2); \
-+ dst_stride, height, &sse2, NULL, NULL); \
- se += se2; \
- long_sse += sse2; \
- se2 = vpx_highbd_sub_pixel_variance##wf##xh_##opt( \
- src + 48 + (start_row * src_stride), src_stride, \
- x_offset, y_offset, dst + 48 + (start_row * dst_stride), \
-- dst_stride, height, &sse2); \
-+ dst_stride, height, &sse2, NULL, NULL); \
- se += se2; \
- long_sse += sse2; \
- }\
-@@ -410,6 +417,7 @@ FNS(sse2, sse);
- #undef FNS
- #undef FN
-
-+// The 2 unused parameters are place holders for PIC enabled build.
- #define DECL(w, opt) \
- int vpx_highbd_sub_pixel_avg_variance##w##xh_##opt(const uint16_t *src, \
- ptrdiff_t src_stride, \
-@@ -419,7 +427,8 @@ int vpx_highbd_sub_pixel_avg_variance##w##xh_##opt(const uint16_t *src, \
- const uint16_t *sec, \
- ptrdiff_t sec_stride, \
- int height, \
-- unsigned int *sse);
-+ unsigned int *sse, \
-+ void *unused0, void *unused);
- #define DECLS(opt1) \
- DECL(16, opt1) \
- DECL(8, opt1)
-@@ -439,23 +448,23 @@ uint32_t vpx_highbd_8_sub_pixel_avg_variance##w##x##h##_##opt( \
- uint16_t *sec = CONVERT_TO_SHORTPTR(sec8); \
- int se = vpx_highbd_sub_pixel_avg_variance##wf##xh_##opt( \
- src, src_stride, x_offset, \
-- y_offset, dst, dst_stride, sec, w, h, &sse); \
-+ y_offset, dst, dst_stride, sec, w, h, &sse, NULL, NULL); \
- if (w > wf) { \
- uint32_t sse2; \
- int se2 = vpx_highbd_sub_pixel_avg_variance##wf##xh_##opt( \
- src + 16, src_stride, x_offset, y_offset, \
-- dst + 16, dst_stride, sec + 16, w, h, &sse2); \
-+ dst + 16, dst_stride, sec + 16, w, h, &sse2, NULL, NULL); \
- se += se2; \
- sse += sse2; \
- if (w > wf * 2) { \
- se2 = vpx_highbd_sub_pixel_avg_variance##wf##xh_##opt( \
- src + 32, src_stride, x_offset, y_offset, \
-- dst + 32, dst_stride, sec + 32, w, h, &sse2); \
-+ dst + 32, dst_stride, sec + 32, w, h, &sse2, NULL, NULL); \
- se += se2; \
- sse += sse2; \
- se2 = vpx_highbd_sub_pixel_avg_variance##wf##xh_##opt( \
- src + 48, src_stride, x_offset, y_offset, \
-- dst + 48, dst_stride, sec + 48, w, h, &sse2); \
-+ dst + 48, dst_stride, sec + 48, w, h, &sse2, NULL, NULL); \
- se += se2; \
- sse += sse2; \
- } \
-@@ -475,14 +484,15 @@ uint32_t vpx_highbd_10_sub_pixel_avg_variance##w##x##h##_##opt( \
- int se = vpx_highbd_sub_pixel_avg_variance##wf##xh_##opt( \
- src, src_stride, x_offset, \
- y_offset, dst, dst_stride, \
-- sec, w, h, &sse); \
-+ sec, w, h, &sse, NULL, NULL); \
- if (w > wf) { \
- uint32_t sse2; \
- int se2 = vpx_highbd_sub_pixel_avg_variance##wf##xh_##opt( \
- src + 16, src_stride, \
- x_offset, y_offset, \
- dst + 16, dst_stride, \
-- sec + 16, w, h, &sse2); \
-+ sec + 16, w, h, &sse2, \
-+ NULL, NULL); \
- se += se2; \
- sse += sse2; \
- if (w > wf * 2) { \
-@@ -490,14 +500,16 @@ uint32_t vpx_highbd_10_sub_pixel_avg_variance##w##x##h##_##opt( \
- src + 32, src_stride, \
- x_offset, y_offset, \
- dst + 32, dst_stride, \
-- sec + 32, w, h, &sse2); \
-+ sec + 32, w, h, &sse2, \
-+ NULL, NULL); \
- se += se2; \
- sse += sse2; \
- se2 = vpx_highbd_sub_pixel_avg_variance##wf##xh_##opt( \
- src + 48, src_stride, \
- x_offset, y_offset, \
- dst + 48, dst_stride, \
-- sec + 48, w, h, &sse2); \
-+ sec + 48, w, h, &sse2, \
-+ NULL, NULL); \
- se += se2; \
- sse += sse2; \
- } \
-@@ -525,7 +537,7 @@ uint32_t vpx_highbd_12_sub_pixel_avg_variance##w##x##h##_##opt( \
- int se2 = vpx_highbd_sub_pixel_avg_variance##wf##xh_##opt( \
- src + (start_row * src_stride), src_stride, x_offset, \
- y_offset, dst + (start_row * dst_stride), dst_stride, \
-- sec + (start_row * w), w, height, &sse2); \
-+ sec + (start_row * w), w, height, &sse2, NULL, NULL); \
- se += se2; \
- long_sse += sse2; \
- if (w > wf) { \
-@@ -533,7 +545,7 @@ uint32_t vpx_highbd_12_sub_pixel_avg_variance##w##x##h##_##opt( \
- src + 16 + (start_row * src_stride), src_stride, \
- x_offset, y_offset, \
- dst + 16 + (start_row * dst_stride), dst_stride, \
-- sec + 16 + (start_row * w), w, height, &sse2); \
-+ sec + 16 + (start_row * w), w, height, &sse2, NULL, NULL); \
- se += se2; \
- long_sse += sse2; \
- if (w > wf * 2) { \
-@@ -541,14 +553,14 @@ uint32_t vpx_highbd_12_sub_pixel_avg_variance##w##x##h##_##opt( \
- src + 32 + (start_row * src_stride), src_stride, \
- x_offset, y_offset, \
- dst + 32 + (start_row * dst_stride), dst_stride, \
-- sec + 32 + (start_row * w), w, height, &sse2); \
-+ sec + 32 + (start_row * w), w, height, &sse2, NULL, NULL); \
- se += se2; \
- long_sse += sse2; \
- se2 = vpx_highbd_sub_pixel_avg_variance##wf##xh_##opt( \
- src + 48 + (start_row * src_stride), src_stride, \
- x_offset, y_offset, \
- dst + 48 + (start_row * dst_stride), dst_stride, \
-- sec + 48 + (start_row * w), w, height, &sse2); \
-+ sec + 48 + (start_row * w), w, height, &sse2, NULL, NULL); \
- se += se2; \
- long_sse += sse2; \
- } \
---
-2.7.0
-
Deleted: PKGBUILD
===================================================================
--- PKGBUILD 2019-07-17 12:16:28 UTC (rev 358094)
+++ PKGBUILD 2019-07-17 12:16:34 UTC (rev 358095)
@@ -1,45 +0,0 @@
-# Maintainer: Maxime Gauduin <alucryd at archlinux.org>
-# Contributor: Ionut Biru <ibiru at archlinux.org>
-# Contributor: Bartłomiej Piotrowski <bpiotrowski at archlinux.org>
-
-pkgname=libvpx
-pkgver=1.8.0
-pkgrel=1
-pkgdesc='VP8 and VP9 codec'
-arch=(x86_64)
-url=http://www.webmproject.org/
-license=(BSD)
-makedepends=(
- git
- nasm
-)
-provides=(libvpx.so)
-source=(git+https://chromium.googlesource.com/webm/libvpx#tag=v${pkgver})
-sha256sums=(SKIP)
-
-build() {
- cd libvpx
-
- ./configure \
- --prefix='/usr' \
- --disable-install-docs \
- --disable-install-srcs \
- --enable-pic \
- --enable-postproc \
- --enable-runtime-cpu-detect \
- --enable-shared \
- --enable-vp8 \
- --enable-vp9 \
- --enable-vp9-highbitdepth \
- --enable-vp9-temporal-denoising
- make
-}
-
-package() {
- cd libvpx
-
- make DIST_DIR="${pkgdir}"/usr install
- install -Dm 644 LICENSE -t "${pkgdir}"/usr/share/licenses/libvpx/
-}
-
-# vim: ts=2 sw=2 et:
Copied: libvpx/repos/extra-x86_64/PKGBUILD (from rev 358094, libvpx/trunk/PKGBUILD)
===================================================================
--- PKGBUILD (rev 0)
+++ PKGBUILD 2019-07-17 12:16:34 UTC (rev 358095)
@@ -0,0 +1,45 @@
+# Maintainer: Maxime Gauduin <alucryd at archlinux.org>
+# Contributor: Ionut Biru <ibiru at archlinux.org>
+# Contributor: Bartłomiej Piotrowski <bpiotrowski at archlinux.org>
+
+pkgname=libvpx
+pkgver=1.8.1
+pkgrel=1
+pkgdesc='VP8 and VP9 codec'
+arch=(x86_64)
+url=https://www.webmproject.org/
+license=(BSD)
+makedepends=(
+ git
+ nasm
+)
+provides=(libvpx.so)
+source=(git+https://chromium.googlesource.com/webm/libvpx#tag=v${pkgver})
+sha256sums=('SKIP')
+
+build() {
+ cd libvpx
+
+ ./configure \
+ --prefix='/usr' \
+ --disable-install-docs \
+ --disable-install-srcs \
+ --enable-pic \
+ --enable-postproc \
+ --enable-runtime-cpu-detect \
+ --enable-shared \
+ --enable-vp8 \
+ --enable-vp9 \
+ --enable-vp9-highbitdepth \
+ --enable-vp9-temporal-denoising
+ make
+}
+
+package() {
+ cd libvpx
+
+ make DIST_DIR="${pkgdir}"/usr install
+ install -Dm 644 LICENSE -t "${pkgdir}"/usr/share/licenses/libvpx/
+}
+
+# vim: ts=2 sw=2 et:
More information about the arch-commits
mailing list