123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388 |
- /*
- * Loongson LSX optimized swscale
- *
- * Copyright (c) 2023 Loongson Technology Corporation Limited
- * Contributed by Lu Wang <wanglu@loongson.cn>
- *
- * This file is part of FFmpeg.
- *
- * FFmpeg is free software; you can redistribute it and/or
- * modify it under the terms of the GNU Lesser General Public
- * License as published by the Free Software Foundation; either
- * version 2.1 of the License, or (at your option) any later version.
- *
- * FFmpeg is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- * Lesser General Public License for more details.
- *
- * You should have received a copy of the GNU Lesser General Public
- * License along with FFmpeg; if not, write to the Free Software
- * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- */
- #include "libavcodec/loongarch/loongson_asm.S"
- /* static void yuv2planeX_8_lsx(const int16_t *filter, int filterSize,
- * const int16_t **src, uint8_t *dest, int dstW,
- * const uint8_t *dither, int offset)
- */
- function yuv2planeX_8_lsx
- addi.w t1, a6, 1
- addi.w t2, a6, 2
- addi.w t3, a6, 3
- addi.w t4, a6, 4
- addi.w t5, a6, 5
- addi.w t6, a6, 6
- addi.w t7, a6, 7
- andi t0, a6, 7
- andi t1, t1, 7
- andi t2, t2, 7
- andi t3, t3, 7
- andi t4, t4, 7
- andi t5, t5, 7
- andi t6, t6, 7
- andi t7, t7, 7
- ldx.bu t0, a5, t0
- ldx.bu t1, a5, t1
- ldx.bu t2, a5, t2
- ldx.bu t3, a5, t3
- ldx.bu t4, a5, t4
- ldx.bu t5, a5, t5
- ldx.bu t6, a5, t6
- ldx.bu t7, a5, t7
- vreplgr2vr.w vr0, t0
- vreplgr2vr.w vr1, t1
- vreplgr2vr.w vr2, t2
- vreplgr2vr.w vr3, t3
- vreplgr2vr.w vr4, t4
- vreplgr2vr.w vr5, t5
- vreplgr2vr.w vr6, t6
- vreplgr2vr.w vr7, t7
- vilvl.w vr0, vr2, vr0
- vilvl.w vr4, vr6, vr4
- vilvl.w vr1, vr3, vr1
- vilvl.w vr5, vr7, vr5
- vilvl.d vr12, vr4, vr0
- vilvl.d vr13, vr5, vr1
- li.w t5, 0
- li.w t8, 8
- bge a4, t8, .WIDTH8
- blt zero, a4, .WIDTH
- b .END
- .WIDTH8:
- li.d t1, 0
- li.d t4, 0
- vslli.w vr2, vr12, 12
- vslli.w vr3, vr13, 12
- move t3, a0
- .FILTERSIZE8:
- ldx.d t2, a2, t1
- vldx vr4, t2, t5
- vldrepl.h vr5, t3, 0
- vmaddwev.w.h vr2, vr4, vr5
- vmaddwod.w.h vr3, vr4, vr5
- addi.d t1, t1, 8
- addi.d t3, t3, 2
- addi.d t4, t4, 1
- blt t4, a1, .FILTERSIZE8
- vsrai.w vr2, vr2, 19
- vsrai.w vr3, vr3, 19
- vclip255.w vr2, vr2
- vclip255.w vr3, vr3
- vpickev.h vr2, vr3, vr2
- vpickev.b vr2, vr2, vr2
- vbsrl.v vr3, vr2, 4
- vilvl.b vr2, vr3, vr2
- fst.d f2, a3, 0
- addi.d t5, t5, 16
- addi.d a4, a4, -8
- addi.d a3, a3, 8
- bge a4, t8, .WIDTH8
- blt zero, a4, .WIDTH
- b .END
- .WIDTH:
- li.d t1, 0
- li.d t4, 0
- vslli.w vr2, vr12, 12
- vslli.w vr3, vr13, 12
- .FILTERSIZE:
- ldx.d t2, a2, t1
- vldx vr4, t2, t5
- vldrepl.h vr5, a0, 0
- vmaddwev.w.h vr2, vr4, vr5
- vmaddwod.w.h vr3, vr4, vr5
- addi.d t1, t1, 8
- addi.d a0, a0, 2
- addi.d t4, t4, 1
- blt t4, a1, .FILTERSIZE
- vsrai.w vr2, vr2, 19
- vsrai.w vr3, vr3, 19
- vclip255.w vr2, vr2
- vclip255.w vr3, vr3
- vpickev.h vr2, vr3, vr2
- vpickev.b vr2, vr2, vr2
- vbsrl.v vr3, vr2, 4
- vilvl.b vr2, vr3, vr2
- .DEST:
- vstelm.b vr2, a3, 0, 0
- vbsrl.v vr2, vr2, 1
- addi.d a4, a4, -1
- addi.d a3, a3, 1
- blt zero, a4, .DEST
- .END:
- endfunc
- /*
- * void yuv2plane1_8_lsx(const int16_t *src, uint8_t *dest, int dstW,
- * const uint8_t *dither, int offset)
- */
- function yuv2plane1_8_lsx
- addi.w t1, a4, 1
- addi.w t2, a4, 2
- addi.w t3, a4, 3
- addi.w t4, a4, 4
- addi.w t5, a4, 5
- addi.w t6, a4, 6
- addi.w t7, a4, 7
- andi t0, a4, 7
- andi t1, t1, 7
- andi t2, t2, 7
- andi t3, t3, 7
- andi t4, t4, 7
- andi t5, t5, 7
- andi t6, t6, 7
- andi t7, t7, 7
- ldx.bu t0, a3, t0
- ldx.bu t1, a3, t1
- ldx.bu t2, a3, t2
- ldx.bu t3, a3, t3
- ldx.bu t4, a3, t4
- ldx.bu t5, a3, t5
- ldx.bu t6, a3, t6
- ldx.bu t7, a3, t7
- vinsgr2vr.h vr1, t0, 0
- vinsgr2vr.h vr1, t1, 1
- vinsgr2vr.h vr1, t2, 2
- vinsgr2vr.h vr1, t3, 3
- vinsgr2vr.h vr1, t4, 4
- vinsgr2vr.h vr1, t5, 5
- vinsgr2vr.h vr1, t6, 6
- vinsgr2vr.h vr1, t7, 7
- vsub.h vr0, vr0, vr0
- vilvl.h vr2, vr0, vr1
- vilvh.h vr3, vr0, vr1
- andi t8, a2, 7
- srli.d a2, a2, 3
- beqz a2, 2f
- 1:
- vld vr1, a0, 0
- addi.d a0, a0, 16
- vshuf4i.d vr0, vr1, 8
- vexth.w.h vr4, vr0
- vexth.w.h vr5, vr1
- vadd.w vr4, vr2, vr4
- vadd.w vr5, vr3, vr5
- vsrai.w vr4, vr4, 7
- vsrai.w vr5, vr5, 7
- vclip255.w vr4, vr4
- vclip255.w vr5, vr5
- vpickev.h vr1, vr5, vr4
- vpickev.b vr1, vr1, vr1
- fst.d f1, a1, 0
- addi.d a1, a1, 8
- addi.d a2, a2, -1
- bnez a2, 1b
- 2:
- beqz t8, 4f
- 3:
- add.w a4, a4, t8
- addi.w t1, a4, 1
- addi.w t2, a4, 2
- addi.w t3, a4, 3
- addi.w t4, a4, 4
- addi.w t5, a4, 5
- addi.w t6, a4, 6
- addi.w t7, a4, 7
- andi t0, a4, 7
- andi t1, t1, 7
- andi t2, t2, 7
- andi t3, t3, 7
- andi t4, t4, 7
- andi t5, t5, 7
- andi t6, t6, 7
- andi t7, t7, 7
- ldx.bu t0, a3, t0
- ldx.bu t1, a3, t1
- ldx.bu t2, a3, t2
- ldx.bu t3, a3, t3
- ldx.bu t4, a3, t4
- ldx.bu t5, a3, t5
- ldx.bu t6, a3, t6
- ldx.bu t7, a3, t7
- vinsgr2vr.h vr1, t0, 0
- vinsgr2vr.h vr1, t1, 1
- vinsgr2vr.h vr1, t2, 2
- vinsgr2vr.h vr1, t3, 3
- vinsgr2vr.h vr1, t4, 4
- vinsgr2vr.h vr1, t5, 5
- vinsgr2vr.h vr1, t6, 6
- vinsgr2vr.h vr1, t7, 7
- vsub.h vr0, vr0, vr0
- vilvl.h vr2, vr0, vr1
- vilvh.h vr3, vr0, vr1
- addi.d a0, a0, -16
- add.d a0, a0, t8
- add.d a0, a0, t8
- addi.d a1, a1, -8
- add.d a1, a1, t8
- vld vr1, a0, 0
- vshuf4i.d vr0, vr1, 8
- vexth.w.h vr4, vr0
- vexth.w.h vr5, vr1
- vadd.w vr4, vr2, vr4
- vadd.w vr5, vr3, vr5
- vsrai.w vr4, vr4, 7
- vsrai.w vr5, vr5, 7
- vclip255.w vr4, vr4
- vclip255.w vr5, vr5
- vpickev.h vr1, vr5, vr4
- vpickev.b vr1, vr1, vr1
- fst.d f1, a1, 0
- 4:
- endfunc
- function yuv2plane1_8_lasx
- addi.w t1, a4, 1
- addi.w t2, a4, 2
- addi.w t3, a4, 3
- addi.w t4, a4, 4
- addi.w t5, a4, 5
- addi.w t6, a4, 6
- addi.w t7, a4, 7
- andi t0, a4, 7
- andi t1, t1, 7
- andi t2, t2, 7
- andi t3, t3, 7
- andi t4, t4, 7
- andi t5, t5, 7
- andi t6, t6, 7
- andi t7, t7, 7
- ldx.bu t0, a3, t0
- ldx.bu t1, a3, t1
- ldx.bu t2, a3, t2
- ldx.bu t3, a3, t3
- ldx.bu t4, a3, t4
- ldx.bu t5, a3, t5
- ldx.bu t6, a3, t6
- ldx.bu t7, a3, t7
- vinsgr2vr.h vr1, t0, 0
- vinsgr2vr.h vr1, t1, 1
- vinsgr2vr.h vr1, t2, 2
- vinsgr2vr.h vr1, t3, 3
- vinsgr2vr.h vr1, t4, 4
- vinsgr2vr.h vr1, t5, 5
- vinsgr2vr.h vr1, t6, 6
- vinsgr2vr.h vr1, t7, 7
- xvpermi.q xr1, xr1, 0
- xvsub.h xr0, xr0, xr0
- xvilvl.h xr2, xr0, xr1
- xvilvh.h xr3, xr0, xr1
- andi t8, a2, 15
- srli.d a2, a2, 4
- beqz a2, 2f
- 1:
- xvld xr1, a0, 0
- addi.d a0, a0, 32
- xvpermi.d xr0, xr1, 0xa0
- xvexth.w.h xr4, xr0
- xvexth.w.h xr5, xr1
- xvadd.w xr4, xr2, xr4
- xvadd.w xr5, xr3, xr5
- xvsrai.w xr4, xr4, 7
- xvsrai.w xr5, xr5, 7
- xvclip255.w xr4, xr4
- xvclip255.w xr5, xr5
- xvpickev.h xr1, xr5, xr4
- xvpickev.b xr0, xr1, xr1
- xvpermi.q xr1, xr0, 1
- fst.d f0, a1, 0
- fst.d f1, a1, 8
- addi.d a1, a1, 16
- addi.d a2, a2, -1
- bnez a2, 1b
- 2:
- beqz t8, 4f
- 3:
- add.w a4, a4, t8
- addi.w t1, a4, 1
- addi.w t2, a4, 2
- addi.w t3, a4, 3
- addi.w t4, a4, 4
- addi.w t5, a4, 5
- addi.w t6, a4, 6
- addi.w t7, a4, 7
- andi t0, a4, 7
- andi t1, t1, 7
- andi t2, t2, 7
- andi t3, t3, 7
- andi t4, t4, 7
- andi t5, t5, 7
- andi t6, t6, 7
- andi t7, t7, 7
- ldx.bu t0, a3, t0
- ldx.bu t1, a3, t1
- ldx.bu t2, a3, t2
- ldx.bu t3, a3, t3
- ldx.bu t4, a3, t4
- ldx.bu t5, a3, t5
- ldx.bu t6, a3, t6
- ldx.bu t7, a3, t7
- vinsgr2vr.h vr1, t0, 0
- vinsgr2vr.h vr1, t1, 1
- vinsgr2vr.h vr1, t2, 2
- vinsgr2vr.h vr1, t3, 3
- vinsgr2vr.h vr1, t4, 4
- vinsgr2vr.h vr1, t5, 5
- vinsgr2vr.h vr1, t6, 6
- vinsgr2vr.h vr1, t7, 7
- xvpermi.q xr1, xr1, 0
- xvsub.h xr0, xr0, xr0
- xvilvl.h xr2, xr0, xr1
- xvilvh.h xr3, xr0, xr1
- addi.d a0, a0, -32
- add.d a0, a0, t8
- add.d a0, a0, t8
- addi.d a1, a1, -16
- add.d a1, a1, t8
- xvld xr1, a0, 0
- xvpermi.d xr0, xr1, 0xa0
- xvexth.w.h xr4, xr0
- xvexth.w.h xr5, xr1
- xvadd.w xr4, xr2, xr4
- xvadd.w xr5, xr3, xr5
- xvsrai.w xr4, xr4, 7
- xvsrai.w xr5, xr5, 7
- xvclip255.w xr4, xr4
- xvclip255.w xr5, xr5
- xvpickev.h xr1, xr5, xr4
- xvpickev.b xr0, xr1, xr1
- xvpermi.q xr1, xr0, 1
- fst.d f0, a1, 0
- fst.d f1, a1, 8
- 4:
- endfunc
|