31 static void imdct36_blocks_ ## CPU(float *out, float *buf, float *in, int count, int switch_point, int block_type);\
32 void ff_imdct36_float_ ## CPU(float *out, float *buf, float *in, float *win);
51 #if HAVE_6REGS && HAVE_SSE_INLINE
53 #define MACS(rt, ra, rb) rt+=(ra)*(rb)
54 #define MLSS(rt, ra, rb) rt-=(ra)*(rb)
56 #define SUM8(op, sum, w, p) \
58 op(sum, (w)[0 * 64], (p)[0 * 64]); \
59 op(sum, (w)[1 * 64], (p)[1 * 64]); \
60 op(sum, (w)[2 * 64], (p)[2 * 64]); \
61 op(sum, (w)[3 * 64], (p)[3 * 64]); \
62 op(sum, (w)[4 * 64], (p)[4 * 64]); \
63 op(sum, (w)[5 * 64], (p)[5 * 64]); \
64 op(sum, (w)[6 * 64], (p)[6 * 64]); \
65 op(sum, (w)[7 * 64], (p)[7 * 64]); \
68 static void apply_window(
const float *buf,
const float *win1,
69 const float *win2,
float *sum1,
float *sum2,
int len)
72 const float *win1a = win1+
len;
73 const float *win2a = win2+
len;
74 const float *bufa = buf+
len;
75 float *sum1a = sum1+
len;
76 float *sum2a = sum2+
len;
80 "movaps " #a "(%1,%0), %%xmm1 \n\t" \
81 "movaps " #a "(%3,%0), %%xmm2 \n\t" \
82 "mulps %%xmm2, %%xmm1 \n\t" \
83 "subps %%xmm1, %%xmm0 \n\t" \
84 "mulps " #b "(%2,%0), %%xmm2 \n\t" \
85 "subps %%xmm2, %%xmm4 \n\t" \
89 "xorps %%xmm0, %%xmm0 \n\t"
90 "xorps %%xmm4, %%xmm4 \n\t"
101 "movaps %%xmm0, (%4,%0) \n\t"
102 "movaps %%xmm4, (%5,%0) \n\t"
106 :
"r"(win1a),
"r"(win2a),
"r"(bufa),
"r"(sum1a),
"r"(sum2a)
112 static void apply_window_mp3(
float *
in,
float *
win,
int *unused,
float *
out,
124 "movaps 0(%0), %%xmm0 \n\t" \
125 "movaps 16(%0), %%xmm1 \n\t" \
126 "movaps 32(%0), %%xmm2 \n\t" \
127 "movaps 48(%0), %%xmm3 \n\t" \
128 "movaps %%xmm0, 0(%1) \n\t" \
129 "movaps %%xmm1, 16(%1) \n\t" \
130 "movaps %%xmm2, 32(%1) \n\t" \
131 "movaps %%xmm3, 48(%1) \n\t" \
132 "movaps 64(%0), %%xmm0 \n\t" \
133 "movaps 80(%0), %%xmm1 \n\t" \
134 "movaps 96(%0), %%xmm2 \n\t" \
135 "movaps 112(%0), %%xmm3 \n\t" \
136 "movaps %%xmm0, 64(%1) \n\t" \
137 "movaps %%xmm1, 80(%1) \n\t" \
138 "movaps %%xmm2, 96(%1) \n\t" \
139 "movaps %%xmm3, 112(%1) \n\t"
140 ::
"r"(
in),
"r"(
in+512)
153 #define SUMS(suma, sumb, sumc, sumd, out1, out2) \
154 "movups " #sumd "(%4), %%xmm0 \n\t" \
155 "shufps $0x1b, %%xmm0, %%xmm0 \n\t" \
156 "subps " #suma "(%1), %%xmm0 \n\t" \
157 "movaps %%xmm0," #out1 "(%0) \n\t" \
159 "movups " #sumc "(%3), %%xmm0 \n\t" \
160 "shufps $0x1b, %%xmm0, %%xmm0 \n\t" \
161 "addps " #sumb "(%2), %%xmm0 \n\t" \
162 "movaps %%xmm0," #out2 "(%0) \n\t"
166 SUMS( 0, 48, 4, 52, 0, 112)
167 SUMS(16, 32, 20, 36, 16, 96)
168 SUMS(32, 16, 36, 20, 32, 80)
169 SUMS(48, 0, 52, 4, 48, 64)
172 :"
r"(&suma[0]), "
r"(&sumb[0]), "
r"(&sumc[0]), "
r"(&sumd[0])
178 float *out2 =
out + 32 * incr;
183 *
out = -suma[ j] + sumd[16-j];
184 *out2 = sumb[16-j] + sumc[ j];
198 #define DECL_IMDCT_BLOCKS(CPU1, CPU2) \
199 static void imdct36_blocks_ ## CPU1(float *out, float *buf, float *in, \
200 int count, int switch_point, int block_type) \
202 int align_end = count - (count & 3); \
204 for (j = 0; j < align_end; j+= 4) { \
205 LOCAL_ALIGNED_16(float, tmpbuf, [1024]); \
206 float *win = mdct_win_sse[switch_point && j < 4][block_type]; \
210 ff_four_imdct36_float_ ## CPU2(out, buf, in, win, tmpbuf); \
215 for (; j < count; j++) { \
219 int win_idx = (switch_point && j < 2) ? 0 : block_type; \
220 float *win = ff_mdct_win_float[win_idx + (4 & -(j & 1))]; \
222 ff_imdct36_float_ ## CPU1(out, buf, in, win); \
232 DECL_IMDCT_BLOCKS(
sse,
sse)
234 DECL_IMDCT_BLOCKS(sse2,
sse)
235 DECL_IMDCT_BLOCKS(sse3,
sse)
236 DECL_IMDCT_BLOCKS(ssse3,
sse)
238 #if HAVE_AVX_EXTERNAL
239 DECL_IMDCT_BLOCKS(avx,avx)
246 for (j = 0; j < 4; j++) {
247 for (
i = 0;
i < 40;
i ++) {
264 #if HAVE_6REGS && HAVE_SSE_INLINE
266 s->apply_window_float = apply_window_mp3;
274 s->imdct36_blocks_float = imdct36_blocks_sse;
278 s->imdct36_blocks_float = imdct36_blocks_sse2;
281 s->imdct36_blocks_float = imdct36_blocks_sse3;
284 s->imdct36_blocks_float = imdct36_blocks_ssse3;
287 #if HAVE_AVX_EXTERNAL
289 s->imdct36_blocks_float = imdct36_blocks_avx;
static void(*const apply_window[4])(AVFloatDSPContext *fdsp, SingleChannelElement *sce, const float *audio)
static float win(SuperEqualizerContext *s, float n, int N)
__asm__(".macro parse_r var r\n\t" "\\var = -1\n\t" _IFC_REG(0) _IFC_REG(1) _IFC_REG(2) _IFC_REG(3) _IFC_REG(4) _IFC_REG(5) _IFC_REG(6) _IFC_REG(7) _IFC_REG(8) _IFC_REG(9) _IFC_REG(10) _IFC_REG(11) _IFC_REG(12) _IFC_REG(13) _IFC_REG(14) _IFC_REG(15) _IFC_REG(16) _IFC_REG(17) _IFC_REG(18) _IFC_REG(19) _IFC_REG(20) _IFC_REG(21) _IFC_REG(22) _IFC_REG(23) _IFC_REG(24) _IFC_REG(25) _IFC_REG(26) _IFC_REG(27) _IFC_REG(28) _IFC_REG(29) _IFC_REG(30) _IFC_REG(31) ".iflt \\var\n\t" ".error \"Unable to parse register name \\r\"\n\t" ".endif\n\t" ".endm")
Macro definitions for various function/variable attributes.
uint8_t pi<< 24) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_U8, uint8_t,(*(const uint8_t *) pi - 0x80) *(1.0f/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_U8, uint8_t,(*(const uint8_t *) pi - 0x80) *(1.0/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S16, int16_t,(*(const int16_t *) pi >> 8)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S16, int16_t, *(const int16_t *) pi *(1.0f/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S16, int16_t, *(const int16_t *) pi *(1.0/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S32, int32_t,(*(const int32_t *) pi >> 24)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S32, int32_t, *(const int32_t *) pi *(1.0f/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S32, int32_t, *(const int32_t *) pi *(1.0/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_FLT, float, av_clip_uint8(lrintf(*(const float *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_FLT, float, av_clip_int16(lrintf(*(const float *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_FLT, float, av_clipl_int32(llrintf(*(const float *) pi *(1U<< 31)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_DBL, double, av_clip_uint8(lrint(*(const double *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_DBL, double, av_clip_int16(lrint(*(const double *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_DBL, double, av_clipl_int32(llrint(*(const double *) pi *(1U<< 31)))) #define SET_CONV_FUNC_GROUP(ofmt, ifmt) static void set_generic_function(AudioConvert *ac) { } void ff_audio_convert_free(AudioConvert **ac) { if(! *ac) return;ff_dither_free(&(*ac) ->dc);av_freep(ac);} AudioConvert *ff_audio_convert_alloc(AVAudioResampleContext *avr, enum AVSampleFormat out_fmt, enum AVSampleFormat in_fmt, int channels, int sample_rate, int apply_map) { AudioConvert *ac;int in_planar, out_planar;ac=av_mallocz(sizeof(*ac));if(!ac) return NULL;ac->avr=avr;ac->out_fmt=out_fmt;ac->in_fmt=in_fmt;ac->channels=channels;ac->apply_map=apply_map;if(avr->dither_method !=AV_RESAMPLE_DITHER_NONE &&av_get_packed_sample_fmt(out_fmt)==AV_SAMPLE_FMT_S16 &&av_get_bytes_per_sample(in_fmt) > 2) { ac->dc=ff_dither_alloc(avr, out_fmt, in_fmt, channels, sample_rate, apply_map);if(!ac->dc) { av_free(ac);return NULL;} return ac;} in_planar=ff_sample_fmt_is_planar(in_fmt, channels);out_planar=ff_sample_fmt_is_planar(out_fmt, channels);if(in_planar==out_planar) { ac->func_type=CONV_FUNC_TYPE_FLAT;ac->planes=in_planar ? ac->channels :1;} else if(in_planar) ac->func_type=CONV_FUNC_TYPE_INTERLEAVE;else ac->func_type=CONV_FUNC_TYPE_DEINTERLEAVE;set_generic_function(ac);if(ARCH_AARCH64) ff_audio_convert_init_aarch64(ac);if(ARCH_ARM) ff_audio_convert_init_arm(ac);if(ARCH_X86) ff_audio_convert_init_x86(ac);return ac;} int ff_audio_convert(AudioConvert *ac, AudioData *out, AudioData *in) { int use_generic=1;int len=in->nb_samples;int p;if(ac->dc) { av_log(ac->avr, AV_LOG_TRACE, "%d samples - audio_convert: %s to %s (dithered)\n", len, av_get_sample_fmt_name(ac->in_fmt), av_get_sample_fmt_name(ac->out_fmt));return ff_convert_dither(ac-> in
static atomic_int cpu_flags
int av_get_cpu_flags(void)
Return the flags which specify extensions supported by the CPU.
#define DECLARE_ALIGNED(n, t, v)
Declare a variable that is aligned in memory.
common internal API header
#define LOCAL_ALIGNED_16(t, v,...)
float ff_mdct_win_float[8][MDCT_BUF_SIZE]
#define SUM8(op, sum, w, p)
static int sse(MpegEncContext *s, uint8_t *src1, uint8_t *src2, int w, int h, int stride)
#define EXTERNAL_SSE(flags)
#define EXTERNAL_AVX(flags)
#define INLINE_SSE(flags)
#define EXTERNAL_SSE3(flags)
#define EXTERNAL_SSSE3(flags)
#define EXTERNAL_SSE2(flags)
av_cold void ff_mpadsp_init_x86(MPADSPContext *s)
static float mdct_win_sse[2][4][4 *40]
void ff_four_imdct36_float_sse(float *out, float *buf, float *in, float *win, float *tmpbuf)
av_cold void ff_mpadsp_init_x86_tabs(void)
void ff_four_imdct36_float_avx(float *out, float *buf, float *in, float *win, float *tmpbuf)