Lines Matching refs:dstW
91 static void yuv2plane1_8_u(const int16_t *src, uint8_t *dest, int dstW,
95 for (i = start; i < dstW; i++) {
101 static void yuv2plane1_8_vsx(const int16_t *src, uint8_t *dest, int dstW,
120 for (i = dst_u; i < dstW - 15; i += 16) {
134 yuv2plane1_8_u(src, dest, dstW, dither, offset, i);
146 static void yuv2plane1_nbps_u(const int16_t *src, uint16_t *dest, int dstW,
152 for (i = start; i < dstW; i++) {
159 uint16_t *dest, int dstW,
176 for (i = dst_u; i < dstW - 7; i += 8) {
185 yuv2plane1_nbps_u(src, dest, dstW, big_endian, output_bits, i);
189 const int16_t **src, uint16_t *dest, int dstW,
195 for (i = start; i < dstW; i++) {
207 const int16_t **src, uint16_t *dest, int dstW,
233 for (i = dst_u; i < dstW - 7; i += 8) {
254 yuv2planeX_nbps_u(filter, filterSize, src, dest, dstW, big_endian, output_bits, i);
267 static void yuv2plane1_16_u(const int32_t *src, uint16_t *dest, int dstW,
273 for (i = start; i < dstW; i++) {
280 uint16_t *dest, int dstW,
296 for (i = dst_u; i < dstW - 7; i += 8) {
311 yuv2plane1_16_u(src, dest, dstW, big_endian, output_bits, i);
317 const int32_t **src, uint16_t *dest, int dstW,
323 for (i = start; i < dstW; i++) {
341 const int32_t **src, uint16_t *dest, int dstW,
365 for (i = dst_u; i < dstW - 7; i += 8) {
386 yuv2planeX_16_u(filter, filterSize, src, dest, dstW, big_endian, output_bits, i);
397 uint8_t *dest, int dstW, \
401 (uint16_t *) dest, dstW, is_be, bits); \
406 const int16_t **src, uint8_t *dest, int dstW, \
411 (uint16_t *) dest, dstW, is_be, bits); \
534 int dstW, int y, enum AVPixelFormat target, int hasAlpha)
586 for (i = 0; i < dstW; i += 8) {
682 const int16_t *abuf[2], uint8_t *dest, int dstW,
743 for (i = 0; i < dstW; i += 8) {
804 const int16_t *abuf[2], uint8_t *dest, int dstW,
873 for (i = 0; i < (dstW + 1) >> 1; i += 8) {
980 const int16_t *abuf0, uint8_t *dest, int dstW,
1032 for (i = 0; i < dstW; i += 8) { // The x86 asm also overwrites padding bytes.
1109 const int16_t *abuf0, uint8_t *dest, int dstW,
1170 for (i = 0; i < (dstW + 1) >> 1; i += 8) { // The x86 asm also overwrites padding bytes.
1297 const int16_t **alpSrc, uint8_t *dest, int dstW, \
1302 alpSrc, dest, dstW, y, fmt, hasAlpha); \
1308 const int16_t *abuf[2], uint8_t *dest, int dstW, \
1312 dest, dstW, yalpha, uvalpha, y, fmt, hasAlpha); \
1318 const int16_t *abuf0, uint8_t *dest, int dstW, \
1322 dstW, uvalpha, y, fmt, hasAlpha); \
1432 const int16_t **alpSrc, uint8_t *dest, int dstW,
1447 for (i = 0; i < ((dstW + 1) >> 1); i += 8) {
1538 const int16_t *abuf[2], uint8_t *dest, int dstW,
1556 for (i = 0; i < ((dstW + 1) >> 1); i += 8) {
1572 const int16_t *abuf0, uint8_t *dest, int dstW,
1584 for (i = 0; i < ((dstW + 1) >> 1); i += 8) {
1604 for (i = 0; i < ((dstW + 1) >> 1); i += 8) {
1634 const int16_t **alpSrc, uint8_t *dest, int dstW, \
1639 alpSrc, dest, dstW, y, fmt); \
1646 const int16_t *abuf[2], uint8_t *dest, int dstW, \
1650 dest, dstW, yalpha, uvalpha, y, fmt); \
1657 const int16_t *abuf0, uint8_t *dest, int dstW, \
1661 abuf0, dest, dstW, uvalpha, \
1861 static void hScale8To19_vsx(SwsContext *c, int16_t *_dst, int dstW,
1892 for (i = 0; i < dstW; i++) {
1901 for (i = 0; i < dstW; i++) {
1919 static void hScale16To19_vsx(SwsContext *c, int16_t *_dst, int dstW,
1959 for (i = 0; i < dstW; i++) {
1970 for (i = 0; i < dstW; i++) {
1997 static void hScale16To15_vsx(SwsContext *c, int16_t *dst, int dstW,
2035 for (i = 0; i < dstW; i++) {
2046 for (i = 0; i < dstW; i++) {
2091 if (c->flags & SWS_FAST_BILINEAR && c->dstW >= c->srcW && c->chrDstW >= c->chrSrcW) {