188d58eaa9
Add optimizations for blocks of width 4 BUG=webm:1423 Change-Id: Idfb458d36db3014d48fbfbe7f5462aa6eb249938
170 lines
5.1 KiB
C
170 lines
5.1 KiB
C
/*
|
|
* Copyright (c) 2017 The WebM project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#ifndef VPX_DSP_ARM_MEM_NEON_H_
|
|
#define VPX_DSP_ARM_MEM_NEON_H_
|
|
|
|
#include <arm_neon.h>
|
|
#include <assert.h>
|
|
#include <string.h>
|
|
|
|
#include "./vpx_config.h"
|
|
#include "vpx/vpx_integer.h"
|
|
#include "vpx_dsp/vpx_dsp_common.h"
|
|
|
|
// Helper functions used to load tran_low_t into int16, narrowing if necessary.
|
|
static INLINE int16x8x2_t load_tran_low_to_s16x2q(const tran_low_t *buf) {
|
|
#if CONFIG_VP9_HIGHBITDEPTH
|
|
const int32x4x2_t v0 = vld2q_s32(buf);
|
|
const int32x4x2_t v1 = vld2q_s32(buf + 8);
|
|
const int16x4_t s0 = vmovn_s32(v0.val[0]);
|
|
const int16x4_t s1 = vmovn_s32(v0.val[1]);
|
|
const int16x4_t s2 = vmovn_s32(v1.val[0]);
|
|
const int16x4_t s3 = vmovn_s32(v1.val[1]);
|
|
int16x8x2_t res;
|
|
res.val[0] = vcombine_s16(s0, s2);
|
|
res.val[1] = vcombine_s16(s1, s3);
|
|
return res;
|
|
#else
|
|
return vld2q_s16(buf);
|
|
#endif
|
|
}
|
|
|
|
static INLINE int16x8_t load_tran_low_to_s16q(const tran_low_t *buf) {
|
|
#if CONFIG_VP9_HIGHBITDEPTH
|
|
const int32x4_t v0 = vld1q_s32(buf);
|
|
const int32x4_t v1 = vld1q_s32(buf + 4);
|
|
const int16x4_t s0 = vmovn_s32(v0);
|
|
const int16x4_t s1 = vmovn_s32(v1);
|
|
return vcombine_s16(s0, s1);
|
|
#else
|
|
return vld1q_s16(buf);
|
|
#endif
|
|
}
|
|
|
|
static INLINE int16x4_t load_tran_low_to_s16d(const tran_low_t *buf) {
|
|
#if CONFIG_VP9_HIGHBITDEPTH
|
|
const int32x4_t v0 = vld1q_s32(buf);
|
|
return vmovn_s32(v0);
|
|
#else
|
|
return vld1_s16(buf);
|
|
#endif
|
|
}
|
|
|
|
static INLINE void store_s16q_to_tran_low(tran_low_t *buf, const int16x8_t a) {
|
|
#if CONFIG_VP9_HIGHBITDEPTH
|
|
const int32x4_t v0 = vmovl_s16(vget_low_s16(a));
|
|
const int32x4_t v1 = vmovl_s16(vget_high_s16(a));
|
|
vst1q_s32(buf, v0);
|
|
vst1q_s32(buf + 4, v1);
|
|
#else
|
|
vst1q_s16(buf, a);
|
|
#endif
|
|
}
|
|
|
|
// Propagate type information to the compiler. Without this the compiler may
|
|
// assume the required alignment of uint32_t (4 bytes) and add alignment hints
|
|
// to the memory access.
|
|
//
|
|
// This is used for functions operating on uint8_t which wish to load or store 4
|
|
// values at a time but which may not be on 4 byte boundaries.
|
|
static INLINE void uint32_to_mem(uint8_t *buf, uint32_t a) {
|
|
memcpy(buf, &a, 4);
|
|
}
|
|
|
|
// Load 2 sets of 4 bytes when alignment is not guaranteed.
|
|
static INLINE uint8x8_t load_unaligned_u8(const uint8_t *buf, int stride) {
|
|
uint32_t a;
|
|
uint32x2_t a_u32 = vdup_n_u32(0);
|
|
if (stride == 4) return vld1_u8(buf);
|
|
memcpy(&a, buf, 4);
|
|
buf += stride;
|
|
a_u32 = vld1_lane_u32(&a, a_u32, 0);
|
|
memcpy(&a, buf, 4);
|
|
a_u32 = vld1_lane_u32(&a, a_u32, 1);
|
|
return vreinterpret_u8_u32(a_u32);
|
|
}
|
|
|
|
// Store 2 sets of 4 bytes when alignment is not guaranteed.
|
|
static INLINE void store_unaligned_u8(uint8_t *buf, int stride,
|
|
const uint8x8_t a) {
|
|
const uint32x2_t a_u32 = vreinterpret_u32_u8(a);
|
|
if (stride == 4) {
|
|
vst1_u8(buf, a);
|
|
return;
|
|
}
|
|
uint32_to_mem(buf, vget_lane_u32(a_u32, 0));
|
|
buf += stride;
|
|
uint32_to_mem(buf, vget_lane_u32(a_u32, 1));
|
|
}
|
|
|
|
// Load 4 sets of 4 bytes when alignment is not guaranteed.
|
|
static INLINE uint8x16_t load_unaligned_u8q(const uint8_t *buf, int stride) {
|
|
uint32_t a;
|
|
uint32x4_t a_u32 = vdupq_n_u32(0);
|
|
if (stride == 4) return vld1q_u8(buf);
|
|
memcpy(&a, buf, 4);
|
|
buf += stride;
|
|
a_u32 = vld1q_lane_u32(&a, a_u32, 0);
|
|
memcpy(&a, buf, 4);
|
|
buf += stride;
|
|
a_u32 = vld1q_lane_u32(&a, a_u32, 1);
|
|
memcpy(&a, buf, 4);
|
|
buf += stride;
|
|
a_u32 = vld1q_lane_u32(&a, a_u32, 2);
|
|
memcpy(&a, buf, 4);
|
|
buf += stride;
|
|
a_u32 = vld1q_lane_u32(&a, a_u32, 3);
|
|
return vreinterpretq_u8_u32(a_u32);
|
|
}
|
|
|
|
// Store 4 sets of 4 bytes when alignment is not guaranteed.
|
|
static INLINE void store_unaligned_u8q(uint8_t *buf, int stride,
|
|
const uint8x16_t a) {
|
|
const uint32x4_t a_u32 = vreinterpretq_u32_u8(a);
|
|
if (stride == 4) {
|
|
vst1q_u8(buf, a);
|
|
return;
|
|
}
|
|
uint32_to_mem(buf, vgetq_lane_u32(a_u32, 0));
|
|
buf += stride;
|
|
uint32_to_mem(buf, vgetq_lane_u32(a_u32, 1));
|
|
buf += stride;
|
|
uint32_to_mem(buf, vgetq_lane_u32(a_u32, 2));
|
|
buf += stride;
|
|
uint32_to_mem(buf, vgetq_lane_u32(a_u32, 3));
|
|
}
|
|
|
|
// Load 2 sets of 4 bytes when alignment is guaranteed.
|
|
static INLINE uint8x8_t load_u8(const uint8_t *buf, int stride) {
|
|
uint32x2_t a = vdup_n_u32(0);
|
|
|
|
assert(!((intptr_t)buf % sizeof(uint32_t)));
|
|
assert(!(stride % sizeof(uint32_t)));
|
|
|
|
a = vld1_lane_u32((const uint32_t *)buf, a, 0);
|
|
buf += stride;
|
|
a = vld1_lane_u32((const uint32_t *)buf, a, 1);
|
|
return vreinterpret_u8_u32(a);
|
|
}
|
|
|
|
// Store 2 sets of 4 bytes when alignment is guaranteed.
|
|
static INLINE void store_u8(uint8_t *buf, int stride, const uint8x8_t a) {
|
|
uint32x2_t a_u32 = vreinterpret_u32_u8(a);
|
|
|
|
assert(!((intptr_t)buf % sizeof(uint32_t)));
|
|
assert(!(stride % sizeof(uint32_t)));
|
|
|
|
vst1_lane_u32((uint32_t *)buf, a_u32, 0);
|
|
buf += stride;
|
|
vst1_lane_u32((uint32_t *)buf, a_u32, 1);
|
|
}
|
|
#endif // VPX_DSP_ARM_MEM_NEON_H_
|