diff options
-rw-r--r-- | README.chromium | 2 | ||||
-rw-r--r-- | include/libyuv/row.h | 16 | ||||
-rw-r--r-- | include/libyuv/version.h | 2 | ||||
-rw-r--r-- | source/convert_argb.cc | 50 | ||||
-rw-r--r-- | source/convert_from.cc | 16 | ||||
-rw-r--r-- | source/planar_functions.cc | 6 | ||||
-rw-r--r-- | source/row_common.cc | 36 | ||||
-rw-r--r-- | source/row_neon.cc | 8 | ||||
-rw-r--r-- | source/row_neon64.cc | 8 |
9 files changed, 72 insertions, 72 deletions
diff --git a/README.chromium b/README.chromium index ad006667..cee698ff 100644 --- a/README.chromium +++ b/README.chromium @@ -1,6 +1,6 @@ Name: libyuv URL: http://code.google.com/p/libyuv/ -Version: 1529 +Version: 1530 License: BSD License File: LICENSE diff --git a/include/libyuv/row.h b/include/libyuv/row.h index 08e1ff64..4e02a212 100644 --- a/include/libyuv/row.h +++ b/include/libyuv/row.h @@ -447,14 +447,14 @@ struct YuvConstants { #endif // Conversion matrix for YUV to RGB -extern const struct YuvConstants kYuvIConstants; // BT.601 -extern const struct YuvConstants kYuvJConstants; // JPeg color space -extern const struct YuvConstants kYuvHConstants; // BT.709 - -// Conversion matrix for YVU to BGR. -extern const struct YuvConstants kYvuIConstants; // BT.601 -extern const struct YuvConstants kYvuJConstants; // JPeg color space -extern const struct YuvConstants kYvuHConstants; // BT.709 +extern const struct YuvConstants kYuvI601Constants; // BT.601 +extern const struct YuvConstants kYuvJPEGConstants; // JPeg color space +extern const struct YuvConstants kYuvH709Constants; // BT.709 + +// Conversion matrix for YVU to BGR +extern const struct YuvConstants kYvuI601Constants; // BT.601 +extern const struct YuvConstants kYvuJPEGConstants; // JPeg color space +extern const struct YuvConstants kYvuH709Constants; // BT.709 #if defined(__APPLE__) || defined(__x86_64__) || defined(__llvm__) #define OMITFP diff --git a/include/libyuv/version.h b/include/libyuv/version.h index f3f99d74..368d7da0 100644 --- a/include/libyuv/version.h +++ b/include/libyuv/version.h @@ -11,6 +11,6 @@ #ifndef INCLUDE_LIBYUV_VERSION_H_ // NOLINT #define INCLUDE_LIBYUV_VERSION_H_ -#define LIBYUV_VERSION 1529 +#define LIBYUV_VERSION 1530 #endif // INCLUDE_LIBYUV_VERSION_H_ NOLINT diff --git a/source/convert_argb.cc b/source/convert_argb.cc index 3b5a25d6..cf3d7228 100644 --- a/source/convert_argb.cc +++ b/source/convert_argb.cc @@ -125,7 +125,7 @@ int I420ToARGB(const uint8* src_y, int src_stride_y, src_u, src_stride_u, src_v, src_stride_v, dst_argb, dst_stride_argb, - &kYuvIConstants, + &kYuvI601Constants, width, height); } @@ -140,7 +140,7 @@ int I420ToABGR(const uint8* src_y, int src_stride_y, src_v, src_stride_v, // Swap U and V src_u, src_stride_u, dst_abgr, dst_stride_abgr, - &kYvuIConstants, // Use Yvu matrix + &kYvuI601Constants, // Use Yvu matrix width, height); } @@ -155,7 +155,7 @@ int J420ToARGB(const uint8* src_y, int src_stride_y, src_u, src_stride_u, src_v, src_stride_v, dst_argb, dst_stride_argb, - &kYuvJConstants, + &kYuvJPEGConstants, width, height); } @@ -170,7 +170,7 @@ int J420ToABGR(const uint8* src_y, int src_stride_y, src_v, src_stride_v, // Swap U and V src_u, src_stride_u, dst_abgr, dst_stride_abgr, - &kYvuJConstants, // Use Yvu matrix + &kYvuJPEGConstants, // Use Yvu matrix width, height); } @@ -185,7 +185,7 @@ int H420ToARGB(const uint8* src_y, int src_stride_y, src_u, src_stride_u, src_v, src_stride_v, dst_argb, dst_stride_argb, - &kYuvHConstants, + &kYuvH709Constants, width, height); } @@ -200,7 +200,7 @@ int H420ToABGR(const uint8* src_y, int src_stride_y, src_v, src_stride_v, // Swap U and V src_u, src_stride_u, dst_abgr, dst_stride_abgr, - &kYvuHConstants, // Use Yvu matrix + &kYvuH709Constants, // Use Yvu matrix width, height); } @@ -293,7 +293,7 @@ int I422ToARGB(const uint8* src_y, int src_stride_y, src_u, src_stride_u, src_v, src_stride_v, dst_argb, dst_stride_argb, - &kYuvIConstants, + &kYuvI601Constants, width, height); } @@ -308,7 +308,7 @@ int I422ToABGR(const uint8* src_y, int src_stride_y, src_v, src_stride_v, // Swap U and V src_u, src_stride_u, dst_abgr, dst_stride_abgr, - &kYvuIConstants, // Use Yvu matrix + &kYvuI601Constants, // Use Yvu matrix width, height); } @@ -323,7 +323,7 @@ int J422ToARGB(const uint8* src_y, int src_stride_y, src_u, src_stride_u, src_v, src_stride_v, dst_argb, dst_stride_argb, - &kYuvJConstants, + &kYuvJPEGConstants, width, height); } @@ -338,7 +338,7 @@ int J422ToABGR(const uint8* src_y, int src_stride_y, src_v, src_stride_v, // Swap U and V src_u, src_stride_u, dst_abgr, dst_stride_abgr, - &kYvuJConstants, // Use Yvu matrix + &kYvuJPEGConstants, // Use Yvu matrix width, height); } @@ -353,7 +353,7 @@ int H422ToARGB(const uint8* src_y, int src_stride_y, src_u, src_stride_u, src_v, src_stride_v, dst_argb, dst_stride_argb, - &kYuvHConstants, + &kYuvH709Constants, width, height); } @@ -368,7 +368,7 @@ int H422ToABGR(const uint8* src_y, int src_stride_y, src_v, src_stride_v, // Swap U and V src_u, src_stride_u, dst_abgr, dst_stride_abgr, - &kYvuHConstants, // Use Yvu matrix + &kYvuH709Constants, // Use Yvu matrix width, height); } @@ -452,7 +452,7 @@ int I444ToARGB(const uint8* src_y, int src_stride_y, src_u, src_stride_u, src_v, src_stride_v, dst_argb, dst_stride_argb, - &kYuvIConstants, + &kYuvI601Constants, width, height); } @@ -467,7 +467,7 @@ int I444ToABGR(const uint8* src_y, int src_stride_y, src_v, src_stride_v, // Swap U and V src_u, src_stride_u, dst_abgr, dst_stride_abgr, - &kYvuIConstants, // Use Yvu matrix + &kYvuI601Constants, // Use Yvu matrix width, height); } @@ -482,7 +482,7 @@ int J444ToARGB(const uint8* src_y, int src_stride_y, src_u, src_stride_u, src_v, src_stride_v, dst_argb, dst_stride_argb, - &kYuvJConstants, + &kYuvJPEGConstants, width, height); } @@ -546,7 +546,7 @@ int I411ToARGB(const uint8* src_y, int src_stride_y, #endif for (y = 0; y < height; ++y) { - I411ToARGBRow(src_y, src_u, src_v, dst_argb, &kYuvIConstants, width); + I411ToARGBRow(src_y, src_u, src_v, dst_argb, &kYuvI601Constants, width); dst_argb += dst_stride_argb; src_y += src_stride_y; src_u += src_stride_u; @@ -671,7 +671,7 @@ int I420AlphaToARGB(const uint8* src_y, int src_stride_y, src_v, src_stride_v, src_a, src_stride_a, dst_argb, dst_stride_argb, - &kYuvIConstants, + &kYuvI601Constants, width, height, attenuate); } @@ -688,7 +688,7 @@ int I420AlphaToABGR(const uint8* src_y, int src_stride_y, src_u, src_stride_u, src_a, src_stride_a, dst_abgr, dst_stride_abgr, - &kYvuIConstants, // Use Yvu matrix + &kYvuI601Constants, // Use Yvu matrix width, height, attenuate); } @@ -1200,7 +1200,7 @@ int NV12ToARGB(const uint8* src_y, int src_stride_y, #endif for (y = 0; y < height; ++y) { - NV12ToARGBRow(src_y, src_uv, dst_argb, &kYuvIConstants, width); + NV12ToARGBRow(src_y, src_uv, dst_argb, &kYuvI601Constants, width); dst_argb += dst_stride_argb; src_y += src_stride_y; if (y & 1) { @@ -1258,7 +1258,7 @@ int NV21ToARGB(const uint8* src_y, int src_stride_y, #endif for (y = 0; y < height; ++y) { - NV21ToARGBRow(src_y, src_uv, dst_argb, &kYuvIConstants, width); + NV21ToARGBRow(src_y, src_uv, dst_argb, &kYuvI601Constants, width); dst_argb += dst_stride_argb; src_y += src_stride_y; if (y & 1) { @@ -1316,15 +1316,15 @@ int M420ToARGB(const uint8* src_m420, int src_stride_m420, for (y = 0; y < height - 1; y += 2) { NV12ToARGBRow(src_m420, src_m420 + src_stride_m420 * 2, dst_argb, - &kYuvIConstants, width); + &kYuvI601Constants, width); NV12ToARGBRow(src_m420 + src_stride_m420, src_m420 + src_stride_m420 * 2, - dst_argb + dst_stride_argb, &kYuvIConstants, width); + dst_argb + dst_stride_argb, &kYuvI601Constants, width); dst_argb += dst_stride_argb * 2; src_m420 += src_stride_m420 * 3; } if (height & 1) { NV12ToARGBRow(src_m420, src_m420 + src_stride_m420 * 2, dst_argb, - &kYuvIConstants, width); + &kYuvI601Constants, width); } return 0; } @@ -1382,7 +1382,7 @@ int YUY2ToARGB(const uint8* src_yuy2, int src_stride_yuy2, } #endif for (y = 0; y < height; ++y) { - YUY2ToARGBRow(src_yuy2, dst_argb, &kYuvIConstants, width); + YUY2ToARGBRow(src_yuy2, dst_argb, &kYuvI601Constants, width); src_yuy2 += src_stride_yuy2; dst_argb += dst_stride_argb; } @@ -1442,7 +1442,7 @@ int UYVYToARGB(const uint8* src_uyvy, int src_stride_uyvy, } #endif for (y = 0; y < height; ++y) { - UYVYToARGBRow(src_uyvy, dst_argb, &kYuvIConstants, width); + UYVYToARGBRow(src_uyvy, dst_argb, &kYuvI601Constants, width); src_uyvy += src_stride_uyvy; dst_argb += dst_stride_argb; } diff --git a/source/convert_from.cc b/source/convert_from.cc index 0043b640..9c138d93 100644 --- a/source/convert_from.cc +++ b/source/convert_from.cc @@ -531,7 +531,7 @@ int I420ToRGBA(const uint8* src_y, int src_stride_y, src_u, src_stride_u, src_v, src_stride_v, dst_rgba, dst_stride_rgba, - &kYuvIConstants, + &kYuvI601Constants, width, height); } @@ -546,7 +546,7 @@ int I420ToBGRA(const uint8* src_y, int src_stride_y, src_v, src_stride_v, // Swap U and V src_u, src_stride_u, dst_bgra, dst_stride_bgra, - &kYvuIConstants, // Use Yvu matrix + &kYvuI601Constants, // Use Yvu matrix width, height); } @@ -622,7 +622,7 @@ int I420ToRGB24(const uint8* src_y, int src_stride_y, src_u, src_stride_u, src_v, src_stride_v, dst_rgb24, dst_stride_rgb24, - &kYuvIConstants, + &kYuvI601Constants, width, height); } @@ -637,7 +637,7 @@ int I420ToRAW(const uint8* src_y, int src_stride_y, src_v, src_stride_v, // Swap U and V src_u, src_stride_u, dst_raw, dst_stride_raw, - &kYvuIConstants, // Use Yvu matrix + &kYvuI601Constants, // Use Yvu matrix width, height); } @@ -691,7 +691,7 @@ int I420ToARGB1555(const uint8* src_y, int src_stride_y, #endif for (y = 0; y < height; ++y) { - I422ToARGB1555Row(src_y, src_u, src_v, dst_argb1555, &kYuvIConstants, + I422ToARGB1555Row(src_y, src_u, src_v, dst_argb1555, &kYuvI601Constants, width); dst_argb1555 += dst_stride_argb1555; src_y += src_stride_y; @@ -754,7 +754,7 @@ int I420ToARGB4444(const uint8* src_y, int src_stride_y, #endif for (y = 0; y < height; ++y) { - I422ToARGB4444Row(src_y, src_u, src_v, dst_argb4444, &kYuvIConstants, + I422ToARGB4444Row(src_y, src_u, src_v, dst_argb4444, &kYuvI601Constants, width); dst_argb4444 += dst_stride_argb4444; src_y += src_stride_y; @@ -816,7 +816,7 @@ int I420ToRGB565(const uint8* src_y, int src_stride_y, #endif for (y = 0; y < height; ++y) { - I422ToRGB565Row(src_y, src_u, src_v, dst_rgb565, &kYuvIConstants, width); + I422ToRGB565Row(src_y, src_u, src_v, dst_rgb565, &kYuvI601Constants, width); dst_rgb565 += dst_stride_rgb565; src_y += src_stride_y; if (y & 1) { @@ -924,7 +924,7 @@ int I420ToRGB565Dither(const uint8* src_y, int src_stride_y, // Allocate a row of argb. align_buffer_64(row_argb, width * 4); for (y = 0; y < height; ++y) { - I422ToARGBRow(src_y, src_u, src_v, row_argb, &kYuvIConstants, width); + I422ToARGBRow(src_y, src_u, src_v, row_argb, &kYuvI601Constants, width); ARGBToRGB565DitherRow(row_argb, dst_rgb565, *(uint32*)(dither4x4 + ((y & 3) << 2)), width); dst_rgb565 += dst_stride_rgb565; diff --git a/source/planar_functions.cc b/source/planar_functions.cc index 900a35e7..63792178 100644 --- a/source/planar_functions.cc +++ b/source/planar_functions.cc @@ -843,7 +843,7 @@ int I422ToRGBA(const uint8* src_y, int src_stride_y, src_u, src_stride_u, src_v, src_stride_v, dst_rgba, dst_stride_rgba, - &kYuvIConstants, + &kYuvI601Constants, width, height); } @@ -858,7 +858,7 @@ int I422ToBGRA(const uint8* src_y, int src_stride_y, src_v, src_stride_v, // Swap U and V src_u, src_stride_u, dst_bgra, dst_stride_bgra, - &kYvuIConstants, // Use Yvu matrix + &kYvuI601Constants, // Use Yvu matrix width, height); } @@ -910,7 +910,7 @@ int NV12ToRGB565(const uint8* src_y, int src_stride_y, #endif for (y = 0; y < height; ++y) { - NV12ToRGB565Row(src_y, src_uv, dst_rgb565, &kYuvIConstants, width); + NV12ToRGB565Row(src_y, src_uv, dst_rgb565, &kYuvI601Constants, width); dst_rgb565 += dst_stride_rgb565; src_y += src_stride_y; if (y & 1) { diff --git a/source/row_common.cc b/source/row_common.cc index c0cb3919..4cc7591c 100644 --- a/source/row_common.cc +++ b/source/row_common.cc @@ -1033,7 +1033,7 @@ void J400ToARGBRow_C(const uint8* src_y, uint8* dst_argb, int width) { #define BR (VR * 128 + YGB) #if defined(__aarch64__) -const YuvConstants SIMD_ALIGNED(kYuvIConstants) = { +const YuvConstants SIMD_ALIGNED(kYuvI601Constants) = { { -UB, -VR, -UB, -VR, -UB, -VR, -UB, -VR }, { -UB, -VR, -UB, -VR, -UB, -VR, -UB, -VR }, { UG, VG, UG, VG, UG, VG, UG, VG }, @@ -1041,7 +1041,7 @@ const YuvConstants SIMD_ALIGNED(kYuvIConstants) = { { BB, BG, BR, 0, 0, 0, 0, 0 }, { 0x0101 * YG, 0, 0, 0 } }; -const YuvConstants SIMD_ALIGNED(kYvuIConstants) = { +const YuvConstants SIMD_ALIGNED(kYvuI601Constants) = { { -VR, -UB, -VR, -UB, -VR, -UB, -VR, -UB }, { -VR, -UB, -VR, -UB, -VR, -UB, -VR, -UB }, { VG, UG, VG, UG, VG, UG, VG, UG }, @@ -1050,20 +1050,20 @@ const YuvConstants SIMD_ALIGNED(kYvuIConstants) = { { 0x0101 * YG, 0, 0, 0 } }; #elif defined(__arm__) -const YuvConstants SIMD_ALIGNED(kYuvIConstants) = { +const YuvConstants SIMD_ALIGNED(kYuvI601Constants) = { { -UB, -UB, -UB, -UB, -VR, -VR, -VR, -VR, 0, 0, 0, 0, 0, 0, 0, 0 }, { UG, UG, UG, UG, VG, VG, VG, VG, 0, 0, 0, 0, 0, 0, 0, 0 }, { BB, BG, BR, 0, 0, 0, 0, 0 }, { 0x0101 * YG, 0, 0, 0 } }; -const YuvConstants SIMD_ALIGNED(kYvuIConstants) = { +const YuvConstants SIMD_ALIGNED(kYvuI601Constants) = { { -VR, -VR, -VR, -VR, -UB, -UB, -UB, -UB, 0, 0, 0, 0, 0, 0, 0, 0 }, { VG, VG, VG, VG, UG, UG, UG, UG, 0, 0, 0, 0, 0, 0, 0, 0 }, { BR, BG, BB, 0, 0, 0, 0, 0 }, { 0x0101 * YG, 0, 0, 0 } }; #else -const YuvConstants SIMD_ALIGNED(kYuvIConstants) = { +const YuvConstants SIMD_ALIGNED(kYuvI601Constants) = { { UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0 }, { UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, @@ -1075,7 +1075,7 @@ const YuvConstants SIMD_ALIGNED(kYuvIConstants) = { { BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR }, { YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG } }; -const YuvConstants SIMD_ALIGNED(kYvuIConstants) = { +const YuvConstants SIMD_ALIGNED(kYvuI601Constants) = { { VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, UB, 0 }, { VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, @@ -1120,7 +1120,7 @@ const YuvConstants SIMD_ALIGNED(kYvuIConstants) = { #define BR (VR * 128 + YGB) #if defined(__aarch64__) -const YuvConstants SIMD_ALIGNED(kYuvJConstants) = { +const YuvConstants SIMD_ALIGNED(kYuvJPEGConstants) = { { -UB, -VR, -UB, -VR, -UB, -VR, -UB, -VR }, { -UB, -VR, -UB, -VR, -UB, -VR, -UB, -VR }, { UG, VG, UG, VG, UG, VG, UG, VG }, @@ -1128,7 +1128,7 @@ const YuvConstants SIMD_ALIGNED(kYuvJConstants) = { { BB, BG, BR, 0, 0, 0, 0, 0 }, { 0x0101 * YG, 0, 0, 0 } }; -const YuvConstants SIMD_ALIGNED(kYvuJConstants) = { +const YuvConstants SIMD_ALIGNED(kYvuJPEGConstants) = { { -VR, -UB, -VR, -UB, -VR, -UB, -VR, -UB }, { -VR, -UB, -VR, -UB, -VR, -UB, -VR, -UB }, { VG, UG, VG, UG, VG, UG, VG, UG }, @@ -1137,20 +1137,20 @@ const YuvConstants SIMD_ALIGNED(kYvuJConstants) = { { 0x0101 * YG, 0, 0, 0 } }; #elif defined(__arm__) -const YuvConstants SIMD_ALIGNED(kYuvJConstants) = { +const YuvConstants SIMD_ALIGNED(kYuvJPEGConstants) = { { -UB, -UB, -UB, -UB, -VR, -VR, -VR, -VR, 0, 0, 0, 0, 0, 0, 0, 0 }, { UG, UG, UG, UG, VG, VG, VG, VG, 0, 0, 0, 0, 0, 0, 0, 0 }, { BB, BG, BR, 0, 0, 0, 0, 0 }, { 0x0101 * YG, 0, 0, 0 } }; -const YuvConstants SIMD_ALIGNED(kYvuJConstants) = { +const YuvConstants SIMD_ALIGNED(kYvuJPEGConstants) = { { -VR, -VR, -VR, -VR, -UB, -UB, -UB, -UB, 0, 0, 0, 0, 0, 0, 0, 0 }, { VG, VG, VG, VG, UG, UG, UG, UG, 0, 0, 0, 0, 0, 0, 0, 0 }, { BR, BG, BB, 0, 0, 0, 0, 0 }, { 0x0101 * YG, 0, 0, 0 } }; #else -const YuvConstants SIMD_ALIGNED(kYuvJConstants) = { +const YuvConstants SIMD_ALIGNED(kYuvJPEGConstants) = { { UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0 }, { UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, @@ -1162,7 +1162,7 @@ const YuvConstants SIMD_ALIGNED(kYuvJConstants) = { { BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR }, { YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG } }; -const YuvConstants SIMD_ALIGNED(kYvuJConstants) = { +const YuvConstants SIMD_ALIGNED(kYvuJPEGConstants) = { { VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, UB, 0 }, { VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, @@ -1208,7 +1208,7 @@ const YuvConstants SIMD_ALIGNED(kYvuJConstants) = { #define BR (VR * 128 + YGB) #if defined(__aarch64__) -const YuvConstants SIMD_ALIGNED(kYuvHConstants) = { +const YuvConstants SIMD_ALIGNED(kYuvH709Constants) = { { -UB, -VR, -UB, -VR, -UB, -VR, -UB, -VR }, { -UB, -VR, -UB, -VR, -UB, -VR, -UB, -VR }, { UG, VG, UG, VG, UG, VG, UG, VG }, @@ -1216,7 +1216,7 @@ const YuvConstants SIMD_ALIGNED(kYuvHConstants) = { { BB, BG, BR, 0, 0, 0, 0, 0 }, { 0x0101 * YG, 0, 0, 0 } }; -const YuvConstants SIMD_ALIGNED(kYvuHConstants) = { +const YuvConstants SIMD_ALIGNED(kYvuH709Constants) = { { -VR, -UB, -VR, -UB, -VR, -UB, -VR, -UB }, { -VR, -UB, -VR, -UB, -VR, -UB, -VR, -UB }, { VG, UG, VG, UG, VG, UG, VG, UG }, @@ -1225,20 +1225,20 @@ const YuvConstants SIMD_ALIGNED(kYvuHConstants) = { { 0x0101 * YG, 0, 0, 0 } }; #elif defined(__arm__) -const YuvConstants SIMD_ALIGNED(kYuvHConstants) = { +const YuvConstants SIMD_ALIGNED(kYuvH709Constants) = { { -UB, -UB, -UB, -UB, -VR, -VR, -VR, -VR, 0, 0, 0, 0, 0, 0, 0, 0 }, { UG, UG, UG, UG, VG, VG, VG, VG, 0, 0, 0, 0, 0, 0, 0, 0 }, { BB, BG, BR, 0, 0, 0, 0, 0 }, { 0x0101 * YG, 0, 0, 0 } }; -const YuvConstants SIMD_ALIGNED(kYvuHConstants) = { +const YuvConstants SIMD_ALIGNED(kYvuH709Constants) = { { -VR, -VR, -VR, -VR, -UB, -UB, -UB, -UB, 0, 0, 0, 0, 0, 0, 0, 0 }, { VG, VG, VG, VG, UG, UG, UG, UG, 0, 0, 0, 0, 0, 0, 0, 0 }, { BR, BG, BB, 0, 0, 0, 0, 0 }, { 0x0101 * YG, 0, 0, 0 } }; #else -const YuvConstants SIMD_ALIGNED(kYuvHConstants) = { +const YuvConstants SIMD_ALIGNED(kYuvH709Constants) = { { UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0 }, { UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, @@ -1250,7 +1250,7 @@ const YuvConstants SIMD_ALIGNED(kYuvHConstants) = { { BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR }, { YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG } }; -const YuvConstants SIMD_ALIGNED(kYvuHConstants) = { +const YuvConstants SIMD_ALIGNED(kYvuH709Constants) = { { VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, UB, 0 }, { VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, diff --git a/source/row_neon.cc b/source/row_neon.cc index 24457d53..ae8c1506 100644 --- a/source/row_neon.cc +++ b/source/row_neon.cc @@ -428,10 +428,10 @@ void I400ToARGBRow_NEON(const uint8* src_y, : "+r"(src_y), // %0 "+r"(dst_argb), // %1 "+r"(width) // %2 - : [kUVToRB]"r"(&kYuvIConstants.kUVToRB), - [kUVToG]"r"(&kYuvIConstants.kUVToG), - [kUVBiasBGR]"r"(&kYuvIConstants.kUVBiasBGR), - [kYToRgb]"r"(&kYuvIConstants.kYToRgb) + : [kUVToRB]"r"(&kYuvI601Constants.kUVToRB), + [kUVToG]"r"(&kYuvI601Constants.kUVToG), + [kUVBiasBGR]"r"(&kYuvI601Constants.kUVBiasBGR), + [kYToRgb]"r"(&kYuvI601Constants.kYToRgb) : "cc", "memory", "q0", "q1", "q2", "q3", "q4", "q8", "q9", "q10", "q11", "q12", "q13", "q14", "q15" ); diff --git a/source/row_neon64.cc b/source/row_neon64.cc index a067bc3f..194a566b 100644 --- a/source/row_neon64.cc +++ b/source/row_neon64.cc @@ -439,10 +439,10 @@ void I400ToARGBRow_NEON(const uint8* src_y, : "+r"(src_y), // %0 "+r"(dst_argb), // %1 "+r"(width64) // %2 - : [kUVToRB]"r"(&kYuvIConstants.kUVToRB), - [kUVToG]"r"(&kYuvIConstants.kUVToG), - [kUVBiasBGR]"r"(&kYuvIConstants.kUVBiasBGR), - [kYToRgb]"r"(&kYuvIConstants.kYToRgb) + : [kUVToRB]"r"(&kYuvI601Constants.kUVToRB), + [kUVToG]"r"(&kYuvI601Constants.kUVToG), + [kUVBiasBGR]"r"(&kYuvI601Constants.kUVBiasBGR), + [kYToRgb]"r"(&kYuvI601Constants.kYToRgb) : "cc", "memory", "v0", "v1", "v2", "v3", "v4", "v5", "v6", "v7", "v20", "v21", "v22", "v23", "v24", "v25", "v26", "v27", "v28", "v29", "v30" ); |