aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--README.chromium2
-rw-r--r--include/libyuv/row.h16
-rw-r--r--include/libyuv/version.h2
-rw-r--r--source/convert_argb.cc50
-rw-r--r--source/convert_from.cc16
-rw-r--r--source/planar_functions.cc6
-rw-r--r--source/row_common.cc36
-rw-r--r--source/row_neon.cc8
-rw-r--r--source/row_neon64.cc8
9 files changed, 72 insertions, 72 deletions
diff --git a/README.chromium b/README.chromium
index ad006667..cee698ff 100644
--- a/README.chromium
+++ b/README.chromium
@@ -1,6 +1,6 @@
Name: libyuv
URL: http://code.google.com/p/libyuv/
-Version: 1529
+Version: 1530
License: BSD
License File: LICENSE
diff --git a/include/libyuv/row.h b/include/libyuv/row.h
index 08e1ff64..4e02a212 100644
--- a/include/libyuv/row.h
+++ b/include/libyuv/row.h
@@ -447,14 +447,14 @@ struct YuvConstants {
#endif
// Conversion matrix for YUV to RGB
-extern const struct YuvConstants kYuvIConstants; // BT.601
-extern const struct YuvConstants kYuvJConstants; // JPeg color space
-extern const struct YuvConstants kYuvHConstants; // BT.709
-
-// Conversion matrix for YVU to BGR.
-extern const struct YuvConstants kYvuIConstants; // BT.601
-extern const struct YuvConstants kYvuJConstants; // JPeg color space
-extern const struct YuvConstants kYvuHConstants; // BT.709
+extern const struct YuvConstants kYuvI601Constants; // BT.601
+extern const struct YuvConstants kYuvJPEGConstants; // JPeg color space
+extern const struct YuvConstants kYuvH709Constants; // BT.709
+
+// Conversion matrix for YVU to BGR
+extern const struct YuvConstants kYvuI601Constants; // BT.601
+extern const struct YuvConstants kYvuJPEGConstants; // JPeg color space
+extern const struct YuvConstants kYvuH709Constants; // BT.709
#if defined(__APPLE__) || defined(__x86_64__) || defined(__llvm__)
#define OMITFP
diff --git a/include/libyuv/version.h b/include/libyuv/version.h
index f3f99d74..368d7da0 100644
--- a/include/libyuv/version.h
+++ b/include/libyuv/version.h
@@ -11,6 +11,6 @@
#ifndef INCLUDE_LIBYUV_VERSION_H_ // NOLINT
#define INCLUDE_LIBYUV_VERSION_H_
-#define LIBYUV_VERSION 1529
+#define LIBYUV_VERSION 1530
#endif // INCLUDE_LIBYUV_VERSION_H_ NOLINT
diff --git a/source/convert_argb.cc b/source/convert_argb.cc
index 3b5a25d6..cf3d7228 100644
--- a/source/convert_argb.cc
+++ b/source/convert_argb.cc
@@ -125,7 +125,7 @@ int I420ToARGB(const uint8* src_y, int src_stride_y,
src_u, src_stride_u,
src_v, src_stride_v,
dst_argb, dst_stride_argb,
- &kYuvIConstants,
+ &kYuvI601Constants,
width, height);
}
@@ -140,7 +140,7 @@ int I420ToABGR(const uint8* src_y, int src_stride_y,
src_v, src_stride_v, // Swap U and V
src_u, src_stride_u,
dst_abgr, dst_stride_abgr,
- &kYvuIConstants, // Use Yvu matrix
+ &kYvuI601Constants, // Use Yvu matrix
width, height);
}
@@ -155,7 +155,7 @@ int J420ToARGB(const uint8* src_y, int src_stride_y,
src_u, src_stride_u,
src_v, src_stride_v,
dst_argb, dst_stride_argb,
- &kYuvJConstants,
+ &kYuvJPEGConstants,
width, height);
}
@@ -170,7 +170,7 @@ int J420ToABGR(const uint8* src_y, int src_stride_y,
src_v, src_stride_v, // Swap U and V
src_u, src_stride_u,
dst_abgr, dst_stride_abgr,
- &kYvuJConstants, // Use Yvu matrix
+ &kYvuJPEGConstants, // Use Yvu matrix
width, height);
}
@@ -185,7 +185,7 @@ int H420ToARGB(const uint8* src_y, int src_stride_y,
src_u, src_stride_u,
src_v, src_stride_v,
dst_argb, dst_stride_argb,
- &kYuvHConstants,
+ &kYuvH709Constants,
width, height);
}
@@ -200,7 +200,7 @@ int H420ToABGR(const uint8* src_y, int src_stride_y,
src_v, src_stride_v, // Swap U and V
src_u, src_stride_u,
dst_abgr, dst_stride_abgr,
- &kYvuHConstants, // Use Yvu matrix
+ &kYvuH709Constants, // Use Yvu matrix
width, height);
}
@@ -293,7 +293,7 @@ int I422ToARGB(const uint8* src_y, int src_stride_y,
src_u, src_stride_u,
src_v, src_stride_v,
dst_argb, dst_stride_argb,
- &kYuvIConstants,
+ &kYuvI601Constants,
width, height);
}
@@ -308,7 +308,7 @@ int I422ToABGR(const uint8* src_y, int src_stride_y,
src_v, src_stride_v, // Swap U and V
src_u, src_stride_u,
dst_abgr, dst_stride_abgr,
- &kYvuIConstants, // Use Yvu matrix
+ &kYvuI601Constants, // Use Yvu matrix
width, height);
}
@@ -323,7 +323,7 @@ int J422ToARGB(const uint8* src_y, int src_stride_y,
src_u, src_stride_u,
src_v, src_stride_v,
dst_argb, dst_stride_argb,
- &kYuvJConstants,
+ &kYuvJPEGConstants,
width, height);
}
@@ -338,7 +338,7 @@ int J422ToABGR(const uint8* src_y, int src_stride_y,
src_v, src_stride_v, // Swap U and V
src_u, src_stride_u,
dst_abgr, dst_stride_abgr,
- &kYvuJConstants, // Use Yvu matrix
+ &kYvuJPEGConstants, // Use Yvu matrix
width, height);
}
@@ -353,7 +353,7 @@ int H422ToARGB(const uint8* src_y, int src_stride_y,
src_u, src_stride_u,
src_v, src_stride_v,
dst_argb, dst_stride_argb,
- &kYuvHConstants,
+ &kYuvH709Constants,
width, height);
}
@@ -368,7 +368,7 @@ int H422ToABGR(const uint8* src_y, int src_stride_y,
src_v, src_stride_v, // Swap U and V
src_u, src_stride_u,
dst_abgr, dst_stride_abgr,
- &kYvuHConstants, // Use Yvu matrix
+ &kYvuH709Constants, // Use Yvu matrix
width, height);
}
@@ -452,7 +452,7 @@ int I444ToARGB(const uint8* src_y, int src_stride_y,
src_u, src_stride_u,
src_v, src_stride_v,
dst_argb, dst_stride_argb,
- &kYuvIConstants,
+ &kYuvI601Constants,
width, height);
}
@@ -467,7 +467,7 @@ int I444ToABGR(const uint8* src_y, int src_stride_y,
src_v, src_stride_v, // Swap U and V
src_u, src_stride_u,
dst_abgr, dst_stride_abgr,
- &kYvuIConstants, // Use Yvu matrix
+ &kYvuI601Constants, // Use Yvu matrix
width, height);
}
@@ -482,7 +482,7 @@ int J444ToARGB(const uint8* src_y, int src_stride_y,
src_u, src_stride_u,
src_v, src_stride_v,
dst_argb, dst_stride_argb,
- &kYuvJConstants,
+ &kYuvJPEGConstants,
width, height);
}
@@ -546,7 +546,7 @@ int I411ToARGB(const uint8* src_y, int src_stride_y,
#endif
for (y = 0; y < height; ++y) {
- I411ToARGBRow(src_y, src_u, src_v, dst_argb, &kYuvIConstants, width);
+ I411ToARGBRow(src_y, src_u, src_v, dst_argb, &kYuvI601Constants, width);
dst_argb += dst_stride_argb;
src_y += src_stride_y;
src_u += src_stride_u;
@@ -671,7 +671,7 @@ int I420AlphaToARGB(const uint8* src_y, int src_stride_y,
src_v, src_stride_v,
src_a, src_stride_a,
dst_argb, dst_stride_argb,
- &kYuvIConstants,
+ &kYuvI601Constants,
width, height, attenuate);
}
@@ -688,7 +688,7 @@ int I420AlphaToABGR(const uint8* src_y, int src_stride_y,
src_u, src_stride_u,
src_a, src_stride_a,
dst_abgr, dst_stride_abgr,
- &kYvuIConstants, // Use Yvu matrix
+ &kYvuI601Constants, // Use Yvu matrix
width, height, attenuate);
}
@@ -1200,7 +1200,7 @@ int NV12ToARGB(const uint8* src_y, int src_stride_y,
#endif
for (y = 0; y < height; ++y) {
- NV12ToARGBRow(src_y, src_uv, dst_argb, &kYuvIConstants, width);
+ NV12ToARGBRow(src_y, src_uv, dst_argb, &kYuvI601Constants, width);
dst_argb += dst_stride_argb;
src_y += src_stride_y;
if (y & 1) {
@@ -1258,7 +1258,7 @@ int NV21ToARGB(const uint8* src_y, int src_stride_y,
#endif
for (y = 0; y < height; ++y) {
- NV21ToARGBRow(src_y, src_uv, dst_argb, &kYuvIConstants, width);
+ NV21ToARGBRow(src_y, src_uv, dst_argb, &kYuvI601Constants, width);
dst_argb += dst_stride_argb;
src_y += src_stride_y;
if (y & 1) {
@@ -1316,15 +1316,15 @@ int M420ToARGB(const uint8* src_m420, int src_stride_m420,
for (y = 0; y < height - 1; y += 2) {
NV12ToARGBRow(src_m420, src_m420 + src_stride_m420 * 2, dst_argb,
- &kYuvIConstants, width);
+ &kYuvI601Constants, width);
NV12ToARGBRow(src_m420 + src_stride_m420, src_m420 + src_stride_m420 * 2,
- dst_argb + dst_stride_argb, &kYuvIConstants, width);
+ dst_argb + dst_stride_argb, &kYuvI601Constants, width);
dst_argb += dst_stride_argb * 2;
src_m420 += src_stride_m420 * 3;
}
if (height & 1) {
NV12ToARGBRow(src_m420, src_m420 + src_stride_m420 * 2, dst_argb,
- &kYuvIConstants, width);
+ &kYuvI601Constants, width);
}
return 0;
}
@@ -1382,7 +1382,7 @@ int YUY2ToARGB(const uint8* src_yuy2, int src_stride_yuy2,
}
#endif
for (y = 0; y < height; ++y) {
- YUY2ToARGBRow(src_yuy2, dst_argb, &kYuvIConstants, width);
+ YUY2ToARGBRow(src_yuy2, dst_argb, &kYuvI601Constants, width);
src_yuy2 += src_stride_yuy2;
dst_argb += dst_stride_argb;
}
@@ -1442,7 +1442,7 @@ int UYVYToARGB(const uint8* src_uyvy, int src_stride_uyvy,
}
#endif
for (y = 0; y < height; ++y) {
- UYVYToARGBRow(src_uyvy, dst_argb, &kYuvIConstants, width);
+ UYVYToARGBRow(src_uyvy, dst_argb, &kYuvI601Constants, width);
src_uyvy += src_stride_uyvy;
dst_argb += dst_stride_argb;
}
diff --git a/source/convert_from.cc b/source/convert_from.cc
index 0043b640..9c138d93 100644
--- a/source/convert_from.cc
+++ b/source/convert_from.cc
@@ -531,7 +531,7 @@ int I420ToRGBA(const uint8* src_y, int src_stride_y,
src_u, src_stride_u,
src_v, src_stride_v,
dst_rgba, dst_stride_rgba,
- &kYuvIConstants,
+ &kYuvI601Constants,
width, height);
}
@@ -546,7 +546,7 @@ int I420ToBGRA(const uint8* src_y, int src_stride_y,
src_v, src_stride_v, // Swap U and V
src_u, src_stride_u,
dst_bgra, dst_stride_bgra,
- &kYvuIConstants, // Use Yvu matrix
+ &kYvuI601Constants, // Use Yvu matrix
width, height);
}
@@ -622,7 +622,7 @@ int I420ToRGB24(const uint8* src_y, int src_stride_y,
src_u, src_stride_u,
src_v, src_stride_v,
dst_rgb24, dst_stride_rgb24,
- &kYuvIConstants,
+ &kYuvI601Constants,
width, height);
}
@@ -637,7 +637,7 @@ int I420ToRAW(const uint8* src_y, int src_stride_y,
src_v, src_stride_v, // Swap U and V
src_u, src_stride_u,
dst_raw, dst_stride_raw,
- &kYvuIConstants, // Use Yvu matrix
+ &kYvuI601Constants, // Use Yvu matrix
width, height);
}
@@ -691,7 +691,7 @@ int I420ToARGB1555(const uint8* src_y, int src_stride_y,
#endif
for (y = 0; y < height; ++y) {
- I422ToARGB1555Row(src_y, src_u, src_v, dst_argb1555, &kYuvIConstants,
+ I422ToARGB1555Row(src_y, src_u, src_v, dst_argb1555, &kYuvI601Constants,
width);
dst_argb1555 += dst_stride_argb1555;
src_y += src_stride_y;
@@ -754,7 +754,7 @@ int I420ToARGB4444(const uint8* src_y, int src_stride_y,
#endif
for (y = 0; y < height; ++y) {
- I422ToARGB4444Row(src_y, src_u, src_v, dst_argb4444, &kYuvIConstants,
+ I422ToARGB4444Row(src_y, src_u, src_v, dst_argb4444, &kYuvI601Constants,
width);
dst_argb4444 += dst_stride_argb4444;
src_y += src_stride_y;
@@ -816,7 +816,7 @@ int I420ToRGB565(const uint8* src_y, int src_stride_y,
#endif
for (y = 0; y < height; ++y) {
- I422ToRGB565Row(src_y, src_u, src_v, dst_rgb565, &kYuvIConstants, width);
+ I422ToRGB565Row(src_y, src_u, src_v, dst_rgb565, &kYuvI601Constants, width);
dst_rgb565 += dst_stride_rgb565;
src_y += src_stride_y;
if (y & 1) {
@@ -924,7 +924,7 @@ int I420ToRGB565Dither(const uint8* src_y, int src_stride_y,
// Allocate a row of argb.
align_buffer_64(row_argb, width * 4);
for (y = 0; y < height; ++y) {
- I422ToARGBRow(src_y, src_u, src_v, row_argb, &kYuvIConstants, width);
+ I422ToARGBRow(src_y, src_u, src_v, row_argb, &kYuvI601Constants, width);
ARGBToRGB565DitherRow(row_argb, dst_rgb565,
*(uint32*)(dither4x4 + ((y & 3) << 2)), width);
dst_rgb565 += dst_stride_rgb565;
diff --git a/source/planar_functions.cc b/source/planar_functions.cc
index 900a35e7..63792178 100644
--- a/source/planar_functions.cc
+++ b/source/planar_functions.cc
@@ -843,7 +843,7 @@ int I422ToRGBA(const uint8* src_y, int src_stride_y,
src_u, src_stride_u,
src_v, src_stride_v,
dst_rgba, dst_stride_rgba,
- &kYuvIConstants,
+ &kYuvI601Constants,
width, height);
}
@@ -858,7 +858,7 @@ int I422ToBGRA(const uint8* src_y, int src_stride_y,
src_v, src_stride_v, // Swap U and V
src_u, src_stride_u,
dst_bgra, dst_stride_bgra,
- &kYvuIConstants, // Use Yvu matrix
+ &kYvuI601Constants, // Use Yvu matrix
width, height);
}
@@ -910,7 +910,7 @@ int NV12ToRGB565(const uint8* src_y, int src_stride_y,
#endif
for (y = 0; y < height; ++y) {
- NV12ToRGB565Row(src_y, src_uv, dst_rgb565, &kYuvIConstants, width);
+ NV12ToRGB565Row(src_y, src_uv, dst_rgb565, &kYuvI601Constants, width);
dst_rgb565 += dst_stride_rgb565;
src_y += src_stride_y;
if (y & 1) {
diff --git a/source/row_common.cc b/source/row_common.cc
index c0cb3919..4cc7591c 100644
--- a/source/row_common.cc
+++ b/source/row_common.cc
@@ -1033,7 +1033,7 @@ void J400ToARGBRow_C(const uint8* src_y, uint8* dst_argb, int width) {
#define BR (VR * 128 + YGB)
#if defined(__aarch64__)
-const YuvConstants SIMD_ALIGNED(kYuvIConstants) = {
+const YuvConstants SIMD_ALIGNED(kYuvI601Constants) = {
{ -UB, -VR, -UB, -VR, -UB, -VR, -UB, -VR },
{ -UB, -VR, -UB, -VR, -UB, -VR, -UB, -VR },
{ UG, VG, UG, VG, UG, VG, UG, VG },
@@ -1041,7 +1041,7 @@ const YuvConstants SIMD_ALIGNED(kYuvIConstants) = {
{ BB, BG, BR, 0, 0, 0, 0, 0 },
{ 0x0101 * YG, 0, 0, 0 }
};
-const YuvConstants SIMD_ALIGNED(kYvuIConstants) = {
+const YuvConstants SIMD_ALIGNED(kYvuI601Constants) = {
{ -VR, -UB, -VR, -UB, -VR, -UB, -VR, -UB },
{ -VR, -UB, -VR, -UB, -VR, -UB, -VR, -UB },
{ VG, UG, VG, UG, VG, UG, VG, UG },
@@ -1050,20 +1050,20 @@ const YuvConstants SIMD_ALIGNED(kYvuIConstants) = {
{ 0x0101 * YG, 0, 0, 0 }
};
#elif defined(__arm__)
-const YuvConstants SIMD_ALIGNED(kYuvIConstants) = {
+const YuvConstants SIMD_ALIGNED(kYuvI601Constants) = {
{ -UB, -UB, -UB, -UB, -VR, -VR, -VR, -VR, 0, 0, 0, 0, 0, 0, 0, 0 },
{ UG, UG, UG, UG, VG, VG, VG, VG, 0, 0, 0, 0, 0, 0, 0, 0 },
{ BB, BG, BR, 0, 0, 0, 0, 0 },
{ 0x0101 * YG, 0, 0, 0 }
};
-const YuvConstants SIMD_ALIGNED(kYvuIConstants) = {
+const YuvConstants SIMD_ALIGNED(kYvuI601Constants) = {
{ -VR, -VR, -VR, -VR, -UB, -UB, -UB, -UB, 0, 0, 0, 0, 0, 0, 0, 0 },
{ VG, VG, VG, VG, UG, UG, UG, UG, 0, 0, 0, 0, 0, 0, 0, 0 },
{ BR, BG, BB, 0, 0, 0, 0, 0 },
{ 0x0101 * YG, 0, 0, 0 }
};
#else
-const YuvConstants SIMD_ALIGNED(kYuvIConstants) = {
+const YuvConstants SIMD_ALIGNED(kYuvI601Constants) = {
{ UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0,
UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0 },
{ UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG,
@@ -1075,7 +1075,7 @@ const YuvConstants SIMD_ALIGNED(kYuvIConstants) = {
{ BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR },
{ YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG }
};
-const YuvConstants SIMD_ALIGNED(kYvuIConstants) = {
+const YuvConstants SIMD_ALIGNED(kYvuI601Constants) = {
{ VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0,
VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, UB, 0 },
{ VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG,
@@ -1120,7 +1120,7 @@ const YuvConstants SIMD_ALIGNED(kYvuIConstants) = {
#define BR (VR * 128 + YGB)
#if defined(__aarch64__)
-const YuvConstants SIMD_ALIGNED(kYuvJConstants) = {
+const YuvConstants SIMD_ALIGNED(kYuvJPEGConstants) = {
{ -UB, -VR, -UB, -VR, -UB, -VR, -UB, -VR },
{ -UB, -VR, -UB, -VR, -UB, -VR, -UB, -VR },
{ UG, VG, UG, VG, UG, VG, UG, VG },
@@ -1128,7 +1128,7 @@ const YuvConstants SIMD_ALIGNED(kYuvJConstants) = {
{ BB, BG, BR, 0, 0, 0, 0, 0 },
{ 0x0101 * YG, 0, 0, 0 }
};
-const YuvConstants SIMD_ALIGNED(kYvuJConstants) = {
+const YuvConstants SIMD_ALIGNED(kYvuJPEGConstants) = {
{ -VR, -UB, -VR, -UB, -VR, -UB, -VR, -UB },
{ -VR, -UB, -VR, -UB, -VR, -UB, -VR, -UB },
{ VG, UG, VG, UG, VG, UG, VG, UG },
@@ -1137,20 +1137,20 @@ const YuvConstants SIMD_ALIGNED(kYvuJConstants) = {
{ 0x0101 * YG, 0, 0, 0 }
};
#elif defined(__arm__)
-const YuvConstants SIMD_ALIGNED(kYuvJConstants) = {
+const YuvConstants SIMD_ALIGNED(kYuvJPEGConstants) = {
{ -UB, -UB, -UB, -UB, -VR, -VR, -VR, -VR, 0, 0, 0, 0, 0, 0, 0, 0 },
{ UG, UG, UG, UG, VG, VG, VG, VG, 0, 0, 0, 0, 0, 0, 0, 0 },
{ BB, BG, BR, 0, 0, 0, 0, 0 },
{ 0x0101 * YG, 0, 0, 0 }
};
-const YuvConstants SIMD_ALIGNED(kYvuJConstants) = {
+const YuvConstants SIMD_ALIGNED(kYvuJPEGConstants) = {
{ -VR, -VR, -VR, -VR, -UB, -UB, -UB, -UB, 0, 0, 0, 0, 0, 0, 0, 0 },
{ VG, VG, VG, VG, UG, UG, UG, UG, 0, 0, 0, 0, 0, 0, 0, 0 },
{ BR, BG, BB, 0, 0, 0, 0, 0 },
{ 0x0101 * YG, 0, 0, 0 }
};
#else
-const YuvConstants SIMD_ALIGNED(kYuvJConstants) = {
+const YuvConstants SIMD_ALIGNED(kYuvJPEGConstants) = {
{ UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0,
UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0 },
{ UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG,
@@ -1162,7 +1162,7 @@ const YuvConstants SIMD_ALIGNED(kYuvJConstants) = {
{ BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR },
{ YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG }
};
-const YuvConstants SIMD_ALIGNED(kYvuJConstants) = {
+const YuvConstants SIMD_ALIGNED(kYvuJPEGConstants) = {
{ VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0,
VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, UB, 0 },
{ VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG,
@@ -1208,7 +1208,7 @@ const YuvConstants SIMD_ALIGNED(kYvuJConstants) = {
#define BR (VR * 128 + YGB)
#if defined(__aarch64__)
-const YuvConstants SIMD_ALIGNED(kYuvHConstants) = {
+const YuvConstants SIMD_ALIGNED(kYuvH709Constants) = {
{ -UB, -VR, -UB, -VR, -UB, -VR, -UB, -VR },
{ -UB, -VR, -UB, -VR, -UB, -VR, -UB, -VR },
{ UG, VG, UG, VG, UG, VG, UG, VG },
@@ -1216,7 +1216,7 @@ const YuvConstants SIMD_ALIGNED(kYuvHConstants) = {
{ BB, BG, BR, 0, 0, 0, 0, 0 },
{ 0x0101 * YG, 0, 0, 0 }
};
-const YuvConstants SIMD_ALIGNED(kYvuHConstants) = {
+const YuvConstants SIMD_ALIGNED(kYvuH709Constants) = {
{ -VR, -UB, -VR, -UB, -VR, -UB, -VR, -UB },
{ -VR, -UB, -VR, -UB, -VR, -UB, -VR, -UB },
{ VG, UG, VG, UG, VG, UG, VG, UG },
@@ -1225,20 +1225,20 @@ const YuvConstants SIMD_ALIGNED(kYvuHConstants) = {
{ 0x0101 * YG, 0, 0, 0 }
};
#elif defined(__arm__)
-const YuvConstants SIMD_ALIGNED(kYuvHConstants) = {
+const YuvConstants SIMD_ALIGNED(kYuvH709Constants) = {
{ -UB, -UB, -UB, -UB, -VR, -VR, -VR, -VR, 0, 0, 0, 0, 0, 0, 0, 0 },
{ UG, UG, UG, UG, VG, VG, VG, VG, 0, 0, 0, 0, 0, 0, 0, 0 },
{ BB, BG, BR, 0, 0, 0, 0, 0 },
{ 0x0101 * YG, 0, 0, 0 }
};
-const YuvConstants SIMD_ALIGNED(kYvuHConstants) = {
+const YuvConstants SIMD_ALIGNED(kYvuH709Constants) = {
{ -VR, -VR, -VR, -VR, -UB, -UB, -UB, -UB, 0, 0, 0, 0, 0, 0, 0, 0 },
{ VG, VG, VG, VG, UG, UG, UG, UG, 0, 0, 0, 0, 0, 0, 0, 0 },
{ BR, BG, BB, 0, 0, 0, 0, 0 },
{ 0x0101 * YG, 0, 0, 0 }
};
#else
-const YuvConstants SIMD_ALIGNED(kYuvHConstants) = {
+const YuvConstants SIMD_ALIGNED(kYuvH709Constants) = {
{ UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0,
UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0, UB, 0 },
{ UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG,
@@ -1250,7 +1250,7 @@ const YuvConstants SIMD_ALIGNED(kYuvHConstants) = {
{ BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR, BR },
{ YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG, YG }
};
-const YuvConstants SIMD_ALIGNED(kYvuHConstants) = {
+const YuvConstants SIMD_ALIGNED(kYvuH709Constants) = {
{ VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0,
VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, VR, 0, UB, 0 },
{ VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG, VG, UG,
diff --git a/source/row_neon.cc b/source/row_neon.cc
index 24457d53..ae8c1506 100644
--- a/source/row_neon.cc
+++ b/source/row_neon.cc
@@ -428,10 +428,10 @@ void I400ToARGBRow_NEON(const uint8* src_y,
: "+r"(src_y), // %0
"+r"(dst_argb), // %1
"+r"(width) // %2
- : [kUVToRB]"r"(&kYuvIConstants.kUVToRB),
- [kUVToG]"r"(&kYuvIConstants.kUVToG),
- [kUVBiasBGR]"r"(&kYuvIConstants.kUVBiasBGR),
- [kYToRgb]"r"(&kYuvIConstants.kYToRgb)
+ : [kUVToRB]"r"(&kYuvI601Constants.kUVToRB),
+ [kUVToG]"r"(&kYuvI601Constants.kUVToG),
+ [kUVBiasBGR]"r"(&kYuvI601Constants.kUVBiasBGR),
+ [kYToRgb]"r"(&kYuvI601Constants.kYToRgb)
: "cc", "memory", "q0", "q1", "q2", "q3", "q4",
"q8", "q9", "q10", "q11", "q12", "q13", "q14", "q15"
);
diff --git a/source/row_neon64.cc b/source/row_neon64.cc
index a067bc3f..194a566b 100644
--- a/source/row_neon64.cc
+++ b/source/row_neon64.cc
@@ -439,10 +439,10 @@ void I400ToARGBRow_NEON(const uint8* src_y,
: "+r"(src_y), // %0
"+r"(dst_argb), // %1
"+r"(width64) // %2
- : [kUVToRB]"r"(&kYuvIConstants.kUVToRB),
- [kUVToG]"r"(&kYuvIConstants.kUVToG),
- [kUVBiasBGR]"r"(&kYuvIConstants.kUVBiasBGR),
- [kYToRgb]"r"(&kYuvIConstants.kYToRgb)
+ : [kUVToRB]"r"(&kYuvI601Constants.kUVToRB),
+ [kUVToG]"r"(&kYuvI601Constants.kUVToG),
+ [kUVBiasBGR]"r"(&kYuvI601Constants.kUVBiasBGR),
+ [kYToRgb]"r"(&kYuvI601Constants.kYToRgb)
: "cc", "memory", "v0", "v1", "v2", "v3", "v4", "v5", "v6", "v7", "v20",
"v21", "v22", "v23", "v24", "v25", "v26", "v27", "v28", "v29", "v30"
);