|
@@ -2579,11 +2579,11 @@ void dsputil_init_mmx(DSPContext* c, AVCodecContext *avctx)
|
|
|
c->add_hfyu_median_prediction = ff_add_hfyu_median_prediction_mmx2;
|
|
|
#endif
|
|
|
#if HAVE_7REGS
|
|
|
- if( mm_flags&AV_CPU_FLAG_3DNOW )
|
|
|
+ if (HAVE_AMD3DNOW && (mm_flags & AV_CPU_FLAG_3DNOW))
|
|
|
c->add_hfyu_median_prediction = add_hfyu_median_prediction_cmov;
|
|
|
#endif
|
|
|
|
|
|
- } else if (mm_flags & AV_CPU_FLAG_3DNOW) {
|
|
|
+ } else if (HAVE_AMD3DNOW && (mm_flags & AV_CPU_FLAG_3DNOW)) {
|
|
|
c->prefetch = prefetch_3dnow;
|
|
|
|
|
|
if (!high_bit_depth) {
|
|
@@ -2732,11 +2732,11 @@ void dsputil_init_mmx(DSPContext* c, AVCodecContext *avctx)
|
|
|
}
|
|
|
#endif
|
|
|
|
|
|
- if(mm_flags & AV_CPU_FLAG_3DNOW){
|
|
|
+ if (HAVE_AMD3DNOW && (mm_flags & AV_CPU_FLAG_3DNOW)) {
|
|
|
c->vorbis_inverse_coupling = vorbis_inverse_coupling_3dnow;
|
|
|
c->vector_fmul = vector_fmul_3dnow;
|
|
|
}
|
|
|
- if(mm_flags & AV_CPU_FLAG_3DNOWEXT){
|
|
|
+ if (HAVE_AMD3DNOWEXT && (mm_flags & AV_CPU_FLAG_3DNOWEXT)) {
|
|
|
c->vector_fmul_reverse = vector_fmul_reverse_3dnow2;
|
|
|
#if HAVE_6REGS
|
|
|
c->vector_fmul_window = vector_fmul_window_3dnow2;
|
|
@@ -2767,7 +2767,7 @@ void dsputil_init_mmx(DSPContext* c, AVCodecContext *avctx)
|
|
|
c->scalarproduct_float = ff_scalarproduct_float_sse;
|
|
|
#endif
|
|
|
}
|
|
|
- if(mm_flags & AV_CPU_FLAG_3DNOW)
|
|
|
+ if (HAVE_AMD3DNOW && (mm_flags & AV_CPU_FLAG_3DNOW))
|
|
|
c->vector_fmul_add = vector_fmul_add_3dnow; // faster than sse
|
|
|
if(mm_flags & AV_CPU_FLAG_SSE2){
|
|
|
#if HAVE_YASM
|