blob: 54566a43c0013c7b7ed378069a44867cca93231f [file] [log] [blame]
Gregory Maxwellae231142011-07-30 08:18:48 -04001/***********************************************************************
2Copyright (c) 2006-2011, Skype Limited. All rights reserved.
3Redistribution and use in source and binary forms, with or without
Jean-Marc Valinae00e602012-04-20 16:31:04 -04004modification, are permitted provided that the following conditions
5are met:
Gregory Maxwellae231142011-07-30 08:18:48 -04006- Redistributions of source code must retain the above copyright notice,
7this list of conditions and the following disclaimer.
8- Redistributions in binary form must reproduce the above copyright
9notice, this list of conditions and the following disclaimer in the
10documentation and/or other materials provided with the distribution.
Ralph Gilesf2446c22013-09-16 14:40:04 -070011- Neither the name of Internet Society, IETF or IETF Trust, nor the
Jean-Marc Valinae00e602012-04-20 16:31:04 -040012names of specific contributors, may be used to endorse or promote
13products derived from this software without specific prior written
14permission.
Timothy B. Terriberry80ad3832013-05-19 18:00:39 -070015THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
Jean-Marc Valinae00e602012-04-20 16:31:04 -040016AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
19LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
20CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
21SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
22INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
23CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
24ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
25POSSIBILITY OF SUCH DAMAGE.
Gregory Maxwellae231142011-07-30 08:18:48 -040026***********************************************************************/
27
28/* *
29 * silk_biquad_alt.c *
30 * *
31 * Second order ARMA filter *
32 * Can handle slowly varying filter coefficients *
33 * */
Jean-Marc Valin5a484122011-08-15 10:49:53 -040034
35#ifdef HAVE_CONFIG_H
36#include "config.h"
37#endif
38
Jean-Marc Valin1c2f5632011-09-16 01:16:53 -070039#include "SigProc_FIX.h"
Gregory Maxwellae231142011-07-30 08:18:48 -040040
Gregory Maxwellae231142011-07-30 08:18:48 -040041/* Second order ARMA filter, alternative implementation */
Linfeng Zhang60eb7d82016-09-01 13:44:11 -070042void silk_biquad_alt_stride1(
Koen Vosacc7a6c2011-10-28 19:44:26 -040043 const opus_int16 *in, /* I input signal */
44 const opus_int32 *B_Q28, /* I MA coefficients [3] */
45 const opus_int32 *A_Q28, /* I AR coefficients [2] */
46 opus_int32 *S, /* I/O State vector [2] */
47 opus_int16 *out, /* O output signal */
Linfeng Zhang60eb7d82016-09-01 13:44:11 -070048 const opus_int32 len /* I signal length (must be even) */
Gregory Maxwellae231142011-07-30 08:18:48 -040049)
50{
51 /* DIRECT FORM II TRANSPOSED (uses 2 element state vector) */
52 opus_int k;
53 opus_int32 inval, A0_U_Q28, A0_L_Q28, A1_U_Q28, A1_L_Q28, out32_Q14;
54
55 /* Negate A_Q28 values and split in two parts */
56 A0_L_Q28 = ( -A_Q28[ 0 ] ) & 0x00003FFF; /* lower part */
Koen Vosacc7a6c2011-10-28 19:44:26 -040057 A0_U_Q28 = silk_RSHIFT( -A_Q28[ 0 ], 14 ); /* upper part */
Gregory Maxwellae231142011-07-30 08:18:48 -040058 A1_L_Q28 = ( -A_Q28[ 1 ] ) & 0x00003FFF; /* lower part */
Koen Vosacc7a6c2011-10-28 19:44:26 -040059 A1_U_Q28 = silk_RSHIFT( -A_Q28[ 1 ], 14 ); /* upper part */
Gregory Maxwellae231142011-07-30 08:18:48 -040060
61 for( k = 0; k < len; k++ ) {
62 /* S[ 0 ], S[ 1 ]: Q12 */
Linfeng Zhang60eb7d82016-09-01 13:44:11 -070063 inval = in[ k ];
Jean-Marc Valinfb3a4372011-09-16 00:58:26 -070064 out32_Q14 = silk_LSHIFT( silk_SMLAWB( S[ 0 ], B_Q28[ 0 ], inval ), 2 );
Gregory Maxwellae231142011-07-30 08:18:48 -040065
Jean-Marc Valinfb3a4372011-09-16 00:58:26 -070066 S[ 0 ] = S[1] + silk_RSHIFT_ROUND( silk_SMULWB( out32_Q14, A0_L_Q28 ), 14 );
67 S[ 0 ] = silk_SMLAWB( S[ 0 ], out32_Q14, A0_U_Q28 );
68 S[ 0 ] = silk_SMLAWB( S[ 0 ], B_Q28[ 1 ], inval);
Gregory Maxwellae231142011-07-30 08:18:48 -040069
Jean-Marc Valinfb3a4372011-09-16 00:58:26 -070070 S[ 1 ] = silk_RSHIFT_ROUND( silk_SMULWB( out32_Q14, A1_L_Q28 ), 14 );
71 S[ 1 ] = silk_SMLAWB( S[ 1 ], out32_Q14, A1_U_Q28 );
72 S[ 1 ] = silk_SMLAWB( S[ 1 ], B_Q28[ 2 ], inval );
Gregory Maxwellae231142011-07-30 08:18:48 -040073
74 /* Scale back to Q0 and saturate */
Linfeng Zhang60eb7d82016-09-01 13:44:11 -070075 out[ k ] = (opus_int16)silk_SAT16( silk_RSHIFT( out32_Q14 + (1<<14) - 1, 14 ) );
76 }
77}
78
Linfeng Zhang43db5622016-09-01 14:20:57 -070079void silk_biquad_alt_stride2_c(
Linfeng Zhang60eb7d82016-09-01 13:44:11 -070080 const opus_int16 *in, /* I input signal */
81 const opus_int32 *B_Q28, /* I MA coefficients [3] */
82 const opus_int32 *A_Q28, /* I AR coefficients [2] */
83 opus_int32 *S, /* I/O State vector [4] */
84 opus_int16 *out, /* O output signal */
85 const opus_int32 len /* I signal length (must be even) */
86)
87{
88 /* DIRECT FORM II TRANSPOSED (uses 2 element state vector) */
89 opus_int k;
90 opus_int32 A0_U_Q28, A0_L_Q28, A1_U_Q28, A1_L_Q28, out32_Q14[ 2 ];
91
92 /* Negate A_Q28 values and split in two parts */
93 A0_L_Q28 = ( -A_Q28[ 0 ] ) & 0x00003FFF; /* lower part */
94 A0_U_Q28 = silk_RSHIFT( -A_Q28[ 0 ], 14 ); /* upper part */
95 A1_L_Q28 = ( -A_Q28[ 1 ] ) & 0x00003FFF; /* lower part */
96 A1_U_Q28 = silk_RSHIFT( -A_Q28[ 1 ], 14 ); /* upper part */
97
98 for( k = 0; k < len; k++ ) {
99 /* S[ 0 ], S[ 1 ], S[ 2 ], S[ 3 ]: Q12 */
100 out32_Q14[ 0 ] = silk_LSHIFT( silk_SMLAWB( S[ 0 ], B_Q28[ 0 ], in[ 2 * k + 0 ] ), 2 );
101 out32_Q14[ 1 ] = silk_LSHIFT( silk_SMLAWB( S[ 2 ], B_Q28[ 0 ], in[ 2 * k + 1 ] ), 2 );
102
103 S[ 0 ] = S[ 1 ] + silk_RSHIFT_ROUND( silk_SMULWB( out32_Q14[ 0 ], A0_L_Q28 ), 14 );
104 S[ 2 ] = S[ 3 ] + silk_RSHIFT_ROUND( silk_SMULWB( out32_Q14[ 1 ], A0_L_Q28 ), 14 );
105 S[ 0 ] = silk_SMLAWB( S[ 0 ], out32_Q14[ 0 ], A0_U_Q28 );
106 S[ 2 ] = silk_SMLAWB( S[ 2 ], out32_Q14[ 1 ], A0_U_Q28 );
107 S[ 0 ] = silk_SMLAWB( S[ 0 ], B_Q28[ 1 ], in[ 2 * k + 0 ] );
108 S[ 2 ] = silk_SMLAWB( S[ 2 ], B_Q28[ 1 ], in[ 2 * k + 1 ] );
109
110 S[ 1 ] = silk_RSHIFT_ROUND( silk_SMULWB( out32_Q14[ 0 ], A1_L_Q28 ), 14 );
111 S[ 3 ] = silk_RSHIFT_ROUND( silk_SMULWB( out32_Q14[ 1 ], A1_L_Q28 ), 14 );
112 S[ 1 ] = silk_SMLAWB( S[ 1 ], out32_Q14[ 0 ], A1_U_Q28 );
113 S[ 3 ] = silk_SMLAWB( S[ 3 ], out32_Q14[ 1 ], A1_U_Q28 );
114 S[ 1 ] = silk_SMLAWB( S[ 1 ], B_Q28[ 2 ], in[ 2 * k + 0 ] );
115 S[ 3 ] = silk_SMLAWB( S[ 3 ], B_Q28[ 2 ], in[ 2 * k + 1 ] );
116
117 /* Scale back to Q0 and saturate */
118 out[ 2 * k + 0 ] = (opus_int16)silk_SAT16( silk_RSHIFT( out32_Q14[ 0 ] + (1<<14) - 1, 14 ) );
119 out[ 2 * k + 1 ] = (opus_int16)silk_SAT16( silk_RSHIFT( out32_Q14[ 1 ] + (1<<14) - 1, 14 ) );
Gregory Maxwellae231142011-07-30 08:18:48 -0400120 }
121}