[FFmpeg-cvslog] lavu/tx: rewrite internal code as a tree-based codelet constructor

Lynne git at videolan.org
Wed Jan 26 05:17:05 EET 2022


ffmpeg | branch: master | Lynne <dev at lynne.ee> | Thu Jan 20 07:14:46 2022 +0100| [ef4bd8161575a79f0ac247ad0aa2f05b8c20052b] | committer: Lynne

lavu/tx: rewrite internal code as a tree-based codelet constructor

This commit rewrites the internal transform code into a constructor
that stitches transforms (codelets).
This allows for transforms to reuse arbitrary parts of other
transforms, and allows transforms to be stacked onto one
another (such as a full iMDCT using a half-iMDCT which in turn
uses an FFT). It also permits for each step to be individually
replaced by assembly or a custom implementation (such as an ASIC).

> http://git.videolan.org/gitweb.cgi/ffmpeg.git/?a=commit;h=ef4bd8161575a79f0ac247ad0aa2f05b8c20052b
---

 libavutil/tx.c                |  596 ++++++++++++++---
 libavutil/tx.h                |    4 +
 libavutil/tx_priv.h           |  277 +++++---
 libavutil/tx_template.c       | 1407 +++++++++++++++++++++++++----------------
 libavutil/x86/tx_float.asm    |   85 ++-
 libavutil/x86/tx_float_init.c |  160 +++--
 6 files changed, 1673 insertions(+), 856 deletions(-)

diff --git a/libavutil/tx.c b/libavutil/tx.c
index fa81ada2f1..c34d3d94fe 100644
--- a/libavutil/tx.c
+++ b/libavutil/tx.c
@@ -16,19 +16,16 @@
  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  */
 
+#include "cpu.h"
+#include "qsort.h"
+#include "bprint.h"
+
 #include "tx_priv.h"
 
-int ff_tx_type_is_mdct(enum AVTXType type)
-{
-    switch (type) {
-    case AV_TX_FLOAT_MDCT:
-    case AV_TX_DOUBLE_MDCT:
-    case AV_TX_INT32_MDCT:
-        return 1;
-    default:
-        return 0;
-    }
-}
+#define TYPE_IS(type, x)               \
+    (((x) == AV_TX_FLOAT_ ## type)  || \
+     ((x) == AV_TX_DOUBLE_ ## type) || \
+     ((x) == AV_TX_INT32_ ## type))
 
 /* Calculates the modular multiplicative inverse */
 static av_always_inline int mulinv(int n, int m)
@@ -42,22 +39,26 @@ static av_always_inline int mulinv(int n, int m)
 }
 
 /* Guaranteed to work for any n, m where gcd(n, m) == 1 */
-int ff_tx_gen_compound_mapping(AVTXContext *s)
+int ff_tx_gen_compound_mapping(AVTXContext *s, int n, int m)
 {
     int *in_map, *out_map;
-    const int n     = s->n;
-    const int m     = s->m;
-    const int inv   = s->inv;
-    const int len   = n*m;
-    const int m_inv = mulinv(m, n);
-    const int n_inv = mulinv(n, m);
-    const int mdct  = ff_tx_type_is_mdct(s->type);
-
-    if (!(s->pfatab = av_malloc(2*len*sizeof(*s->pfatab))))
+    const int inv = s->inv;
+    const int len = n*m;    /* Will not be equal to s->len for MDCTs */
+    const int mdct = TYPE_IS(MDCT, s->type);
+    int m_inv, n_inv;
+
+    /* Make sure the numbers are coprime */
+    if (av_gcd(n, m) != 1)
+        return AVERROR(EINVAL);
+
+    m_inv = mulinv(m, n);
+    n_inv = mulinv(n, m);
+
+    if (!(s->map = av_malloc(2*len*sizeof(*s->map))))
         return AVERROR(ENOMEM);
 
-    in_map  = s->pfatab;
-    out_map = s->pfatab + n*m;
+    in_map  = s->map;
+    out_map = s->map + len;
 
     /* Ruritanian map for input, CRT map for output, can be swapped */
     for (int j = 0; j < m; j++) {
@@ -92,48 +93,50 @@ int ff_tx_gen_compound_mapping(AVTXContext *s)
     return 0;
 }
 
-static inline int split_radix_permutation(int i, int m, int inverse)
+static inline int split_radix_permutation(int i, int len, int inv)
 {
-    m >>= 1;
-    if (m <= 1)
+    len >>= 1;
+    if (len <= 1)
         return i & 1;
-    if (!(i & m))
-        return split_radix_permutation(i, m, inverse) * 2;
-    m >>= 1;
-    return split_radix_permutation(i, m, inverse) * 4 + 1 - 2*(!(i & m) ^ inverse);
+    if (!(i & len))
+        return split_radix_permutation(i, len, inv) * 2;
+    len >>= 1;
+    return split_radix_permutation(i, len, inv) * 4 + 1 - 2*(!(i & len) ^ inv);
 }
 
 int ff_tx_gen_ptwo_revtab(AVTXContext *s, int invert_lookup)
 {
-    const int m = s->m, inv = s->inv;
+    int len = s->len;
 
-    if (!(s->revtab = av_malloc(s->m*sizeof(*s->revtab))))
-        return AVERROR(ENOMEM);
-    if (!(s->revtab_c = av_malloc(m*sizeof(*s->revtab_c))))
+    if (!(s->map = av_malloc(len*sizeof(*s->map))))
         return AVERROR(ENOMEM);
 
-    /* Default */
-    for (int i = 0; i < m; i++) {
-        int k = -split_radix_permutation(i, m, inv) & (m - 1);
-        if (invert_lookup)
-            s->revtab[i] = s->revtab_c[i] = k;
-        else
-            s->revtab[i] = s->revtab_c[k] = i;
+    if (invert_lookup) {
+        for (int i = 0; i < s->len; i++)
+            s->map[i] = -split_radix_permutation(i, len, s->inv) & (len - 1);
+    } else {
+        for (int i = 0; i < s->len; i++)
+            s->map[-split_radix_permutation(i, len, s->inv) & (len - 1)] = i;
     }
 
     return 0;
 }
 
-int ff_tx_gen_ptwo_inplace_revtab_idx(AVTXContext *s, int *revtab)
+int ff_tx_gen_ptwo_inplace_revtab_idx(AVTXContext *s)
 {
-    int nb_inplace_idx = 0;
+    int *src_map, out_map_idx = 0, len = s->len;
 
-    if (!(s->inplace_idx = av_malloc(s->m*sizeof(*s->inplace_idx))))
+    if (!s->sub || !s->sub->map)
+        return AVERROR(EINVAL);
+
+    if (!(s->map = av_mallocz(len*sizeof(*s->map))))
         return AVERROR(ENOMEM);
 
+    src_map = s->sub->map;
+
     /* The first coefficient is always already in-place */
-    for (int src = 1; src < s->m; src++) {
-        int dst = revtab[src];
+    for (int src = 1; src < s->len; src++) {
+        int dst = src_map[src];
         int found = 0;
 
         if (dst <= src)
@@ -143,48 +146,53 @@ int ff_tx_gen_ptwo_inplace_revtab_idx(AVTXContext *s, int *revtab)
          * and if so, skips it, since to fully permute a loop we must only
          * enter it once. */
         do {
-            for (int j = 0; j < nb_inplace_idx; j++) {
-                if (dst == s->inplace_idx[j]) {
+            for (int j = 0; j < out_map_idx; j++) {
+                if (dst == s->map[j]) {
                     found = 1;
                     break;
                 }
             }
-            dst = revtab[dst];
+            dst = src_map[dst];
         } while (dst != src && !found);
 
         if (!found)
-            s->inplace_idx[nb_inplace_idx++] = src;
+            s->map[out_map_idx++] = src;
     }
 
-    s->inplace_idx[nb_inplace_idx++] = 0;
+    s->map[out_map_idx++] = 0;
 
     return 0;
 }
 
 static void parity_revtab_generator(int *revtab, int n, int inv, int offset,
                                     int is_dual, int dual_high, int len,
-                                    int basis, int dual_stride)
+                                    int basis, int dual_stride, int inv_lookup)
 {
     len >>= 1;
 
     if (len <= basis) {
-        int k1, k2, *even, *odd, stride;
+        int k1, k2, stride, even_idx, odd_idx;
 
         is_dual = is_dual && dual_stride;
         dual_high = is_dual & dual_high;
         stride = is_dual ? FFMIN(dual_stride, len) : 0;
 
-        even = &revtab[offset + dual_high*(stride - 2*len)];
-        odd  = &even[len + (is_dual && !dual_high)*len + dual_high*len];
+        even_idx = offset + dual_high*(stride - 2*len);
+        odd_idx  = even_idx + len + (is_dual && !dual_high)*len + dual_high*len;
 
         for (int i = 0; i < len; i++) {
             k1 = -split_radix_permutation(offset + i*2 + 0, n, inv) & (n - 1);
             k2 = -split_radix_permutation(offset + i*2 + 1, n, inv) & (n - 1);
-            *even++ = k1;
-            *odd++  = k2;
+            if (inv_lookup) {
+                revtab[even_idx++] = k1;
+                revtab[odd_idx++]  = k2;
+            } else {
+                revtab[k1] = even_idx++;
+                revtab[k2] = odd_idx++;
+            }
             if (stride && !((i + 1) % stride)) {
-                even += stride;
-                odd  += stride;
+                even_idx += stride;
+                odd_idx  += stride;
             }
         }
 
@@ -192,22 +200,52 @@ static void parity_revtab_generator(int *revtab, int n, int inv, int offset,
     }
 
     parity_revtab_generator(revtab, n, inv, offset,
-                            0, 0, len >> 0, basis, dual_stride);
+                            0, 0, len >> 0, basis, dual_stride, inv_lookup);
     parity_revtab_generator(revtab, n, inv, offset + (len >> 0),
-                            1, 0, len >> 1, basis, dual_stride);
+                            1, 0, len >> 1, basis, dual_stride, inv_lookup);
     parity_revtab_generator(revtab, n, inv, offset + (len >> 0) + (len >> 1),
-                            1, 1, len >> 1, basis, dual_stride);
+                            1, 1, len >> 1, basis, dual_stride, inv_lookup);
 }
 
-void ff_tx_gen_split_radix_parity_revtab(int *revtab, int len, int inv,
-                                         int basis, int dual_stride)
+int ff_tx_gen_split_radix_parity_revtab(AVTXContext *s, int invert_lookup,
+                                        int basis, int dual_stride)
 {
+    int len = s->len;
+    int inv = s->inv;
+
+    if (!(s->map = av_mallocz(len*sizeof(*s->map))))
+        return AVERROR(ENOMEM);
+
     basis >>= 1;
     if (len < basis)
-        return;
+        return AVERROR(EINVAL);
+
     av_assert0(!dual_stride || !(dual_stride & (dual_stride - 1)));
     av_assert0(dual_stride <= basis);
-    parity_revtab_generator(revtab, len, inv, 0, 0, 0, len, basis, dual_stride);
+    parity_revtab_generator(s->map, len, inv, 0, 0, 0, len,
+                            basis, dual_stride, invert_lookup);
+
+    return 0;
+}
+
+static void reset_ctx(AVTXContext *s)
+{
+    if (!s)
+        return;
+
+    if (s->sub)
+        for (int i = 0; i < s->nb_sub; i++)
+            reset_ctx(&s->sub[i]);
+
+    if (s->cd_self->uninit)
+        s->cd_self->uninit(s);
+
+    av_freep(&s->sub);
+    av_freep(&s->map);
+    av_freep(&s->exp);
+    av_freep(&s->tmp);
+
+    memset(s, 0, sizeof(*s));
 }
 
 av_cold void av_tx_uninit(AVTXContext **ctx)
@@ -215,53 +253,401 @@ av_cold void av_tx_uninit(AVTXContext **ctx)
     if (!(*ctx))
         return;
 
-    av_free((*ctx)->pfatab);
-    av_free((*ctx)->exptab);
-    av_free((*ctx)->revtab);
-    av_free((*ctx)->revtab_c);
-    av_free((*ctx)->inplace_idx);
-    av_free((*ctx)->tmp);
-
+    reset_ctx(*ctx);
     av_freep(ctx);
 }
 
+static av_cold int ff_tx_null_init(AVTXContext *s, const FFTXCodelet *cd,
+                                   uint64_t flags, FFTXCodeletOptions *opts,
+                                   int len, int inv, const void *scale)
+{
+    /* Can only handle one sample+type to one sample+type transforms */
+    if (TYPE_IS(MDCT, s->type))
+        return AVERROR(EINVAL);
+    return 0;
+}
+
+/* Null transform when the length is 1 */
+static void ff_tx_null(AVTXContext *s, void *_out, void *_in, ptrdiff_t stride)
+{
+    memcpy(_out, _in, stride);
+}
+
+static const FFTXCodelet ff_tx_null_def = {
+    .name       = "null",
+    .function   = ff_tx_null,
+    .type       = TX_TYPE_ANY,
+    .flags      = AV_TX_UNALIGNED | FF_TX_ALIGNED |
+                  FF_TX_OUT_OF_PLACE | AV_TX_INPLACE,
+    .factors[0] = TX_FACTOR_ANY,
+    .min_len    = 1,
+    .max_len    = 1,
+    .init       = ff_tx_null_init,
+    .cpu_flags  = FF_TX_CPU_FLAGS_ALL,
+    .prio       = FF_TX_PRIO_MAX,
+};
+
+static const FFTXCodelet * const ff_tx_null_list[] = {
+    &ff_tx_null_def,
+    NULL,
+};
+
+static void print_flags(AVBPrint *bp, uint64_t f)
+{
+    int prev = 0;
+    const char *sep = ", ";
+    av_bprintf(bp, "flags: [");
+    if ((f & FF_TX_ALIGNED) && ++prev)
+        av_bprintf(bp, "aligned");
+    if ((f & AV_TX_UNALIGNED) && ++prev)
+        av_bprintf(bp, "%sunaligned", prev > 1 ? sep : "");
+    if ((f & AV_TX_INPLACE) && ++prev)
+        av_bprintf(bp, "%sinplace", prev > 1 ? sep : "");
+    if ((f & FF_TX_OUT_OF_PLACE) && ++prev)
+        av_bprintf(bp, "%sout_of_place", prev > 1 ? sep : "");
+    if ((f & FF_TX_FORWARD_ONLY) && ++prev)
+        av_bprintf(bp, "%sfwd_only", prev > 1 ? sep : "");
+    if ((f & FF_TX_INVERSE_ONLY) && ++prev)
+        av_bprintf(bp, "%sinv_only", prev > 1 ? sep : "");
+    if ((f & FF_TX_PRESHUFFLE) && ++prev)
+        av_bprintf(bp, "%spreshuf", prev > 1 ? sep : "");
+    if ((f & AV_TX_FULL_IMDCT) && ++prev)
+        av_bprintf(bp, "%simdct_full", prev > 1 ? sep : "");
+    av_bprintf(bp, "]");
+}
+
+static void print_type(AVBPrint *bp, enum AVTXType type)
+{
+    av_bprintf(bp, "%s",
+               type == TX_TYPE_ANY       ? "any"         :
+               type == AV_TX_FLOAT_FFT   ? "fft_float"   :
+               type == AV_TX_FLOAT_MDCT  ? "mdct_float"  :
+               type == AV_TX_DOUBLE_FFT  ? "fft_double"  :
+               type == AV_TX_DOUBLE_MDCT ? "mdct_double" :
+               type == AV_TX_INT32_FFT   ? "fft_int32"   :
+               type == AV_TX_INT32_MDCT  ? "mdct_int32"  :
+               "unknown");
+}
+
+static void print_cd_info(const FFTXCodelet *cd, int prio, int print_prio)
+{
+    AVBPrint bp = { 0 };
+    av_bprint_init(&bp, 0, AV_BPRINT_SIZE_AUTOMATIC);
+
+    av_bprintf(&bp, "%s - type: ", cd->name);
+
+    print_type(&bp, cd->type);
+
+    av_bprintf(&bp, ", len: ");
+    if (cd->min_len != cd->max_len)
+        av_bprintf(&bp, "[%i, ", cd->min_len);
+
+    if (cd->max_len == TX_LEN_UNLIMITED)
+        av_bprintf(&bp, "∞");
+    else
+        av_bprintf(&bp, "%i", cd->max_len);
+
+    av_bprintf(&bp, "%s, factors: [", cd->min_len != cd->max_len ? "]" : "");
+    for (int i = 0; i < TX_MAX_SUB; i++) {
+        if (i && cd->factors[i])
+            av_bprintf(&bp, ", ");
+        if (cd->factors[i] == TX_FACTOR_ANY)
+            av_bprintf(&bp, "any");
+        else if (cd->factors[i])
+            av_bprintf(&bp, "%i", cd->factors[i]);
+        else
+            break;
+    }
+
+    av_bprintf(&bp, "], ");
+    print_flags(&bp, cd->flags);
+
+    if (print_prio)
+        av_bprintf(&bp, ", prio: %i", prio);
+
+    av_log(NULL, AV_LOG_VERBOSE, "%s\n", bp.str);
+}
+
+typedef struct TXCodeletMatch {
+    const FFTXCodelet *cd;
+    int prio;
+} TXCodeletMatch;
+
+static int cmp_matches(TXCodeletMatch *a, TXCodeletMatch *b)
+{
+    return FFDIFFSIGN(b->prio, a->prio);
+}
+
+/* We want all factors to completely cover the length */
+static inline int check_cd_factors(const FFTXCodelet *cd, int len)
+{
+    int all_flag = 0;
+
+    for (int i = 0; i < TX_MAX_SUB; i++) {
+        int factor = cd->factors[i];
+
+        /* Conditions satisfied */
+        if (len == 1)
+            return 1;
+
+        /* No more factors */
+        if (!factor) {
+            break;
+        } else if (factor == TX_FACTOR_ANY) {
+            all_flag = 1;
+            continue;
+        }
+
+        if (factor == 2) { /* Fast path */
+            int bits_2 = ff_ctz(len);
+            if (!bits_2)
+                return 0; /* Factor not supported */
+
+            len >>= bits_2;
+        } else {
+            int res = len % factor;
+            if (res)
+                return 0; /* Factor not supported */
+
+            while (!res) {
+                len /= factor;
+                res = len % factor;
+            }
+        }
+    }
+
+    return all_flag || (len == 1);
+}
+
+av_cold int ff_tx_init_subtx(AVTXContext *s, enum AVTXType type,
+                             uint64_t flags, FFTXCodeletOptions *opts,
+                             int len, int inv, const void *scale)
+{
+    int ret = 0;
+    AVTXContext *sub = NULL;
+    TXCodeletMatch *cd_tmp, *cd_matches = NULL;
+    unsigned int cd_matches_size = 0;
+    int nb_cd_matches = 0;
+    AVBPrint bp = { 0 };
+
+    /* Array of all compiled codelet lists. Order is irrelevant. */
+    const FFTXCodelet * const * const codelet_list[] = {
+        ff_tx_codelet_list_float_c,
+        ff_tx_codelet_list_double_c,
+        ff_tx_codelet_list_int32_c,
+        ff_tx_null_list,
+#if ARCH_X86
+        ff_tx_codelet_list_float_x86,
+#endif
+    };
+    int codelet_list_num = FF_ARRAY_ELEMS(codelet_list);
+
+    /* We still accept functions marked with SLOW, even if the CPU is
+     * marked with the same flag, but we give them lower priority. */
+    const int cpu_flags = av_get_cpu_flags();
+    const int slow_mask = AV_CPU_FLAG_SSE2SLOW | AV_CPU_FLAG_SSE3SLOW  |
+                          AV_CPU_FLAG_ATOM     | AV_CPU_FLAG_SSSE3SLOW |
+                          AV_CPU_FLAG_AVXSLOW  | AV_CPU_FLAG_SLOW_GATHER;
+
+    /* Flags the transform wants */
+    uint64_t req_flags = flags;
+
+    /* Unaligned codelets are compatible with the aligned flag */
+    if (req_flags & FF_TX_ALIGNED)
+        req_flags |= AV_TX_UNALIGNED;
+
+    /* If either flag is set, both are okay, so don't check for an exact match */
+    if ((req_flags & AV_TX_INPLACE) && (req_flags & FF_TX_OUT_OF_PLACE))
+        req_flags &= ~(AV_TX_INPLACE | FF_TX_OUT_OF_PLACE);
+    if ((req_flags & FF_TX_ALIGNED) && (req_flags & AV_TX_UNALIGNED))
+        req_flags &= ~(FF_TX_ALIGNED | AV_TX_UNALIGNED);
+
+    /* Flags the codelet may require to be present */
+    uint64_t inv_req_mask = AV_TX_FULL_IMDCT | FF_TX_PRESHUFFLE;
+
+    /* Loop through all codelets in all codelet lists to find matches
+     * to the requirements */
+    while (codelet_list_num--) {
+        const FFTXCodelet * const * list = codelet_list[codelet_list_num];
+        const FFTXCodelet *cd = NULL;
+
+        while ((cd = *list++)) {
+            int max_factor = 0;
+
+            /* Check if the type matches */
+            if (cd->type != TX_TYPE_ANY && type != cd->type)
+                continue;
+
+            /* Check direction for non-orthogonal codelets */
+            if (((cd->flags & FF_TX_FORWARD_ONLY) && inv) ||
+                ((cd->flags & (FF_TX_INVERSE_ONLY | AV_TX_FULL_IMDCT)) && !inv))
+                continue;
+
+            /* Check if the requested flags match from both sides */
+            if (((req_flags    & cd->flags) != (req_flags)) ||
+                ((inv_req_mask & cd->flags) != (req_flags & inv_req_mask)))
+                continue;
+
+            /* Check if length is supported */
+            if ((len < cd->min_len) || (cd->max_len != -1 && (len > cd->max_len)))
+                continue;
+
+            /* Check if the CPU supports the required ISA */
+            if (!(!cd->cpu_flags || (cpu_flags & (cd->cpu_flags & ~slow_mask))))
+                continue;
+
+            /* Check for factors */
+            if (!check_cd_factors(cd, len))
+                continue;
+
+            /* Realloc array and append */
+            cd_tmp = av_fast_realloc(cd_matches, &cd_matches_size,
+                                     sizeof(*cd_tmp) * (nb_cd_matches + 1));
+            if (!cd_tmp) {
+                av_free(cd_matches);
+                return AVERROR(ENOMEM);
+            }
+
+            cd_matches                     = cd_tmp;
+            cd_matches[nb_cd_matches].cd   = cd;
+            cd_matches[nb_cd_matches].prio = cd->prio;
+
+            /* If the CPU has a SLOW flag, and the instruction is also flagged
+             * as being slow for such, reduce its priority */
+            if ((cpu_flags & cd->cpu_flags) & slow_mask)
+                cd_matches[nb_cd_matches].prio -= 64;
+
+            /* Prioritize aligned-only codelets */
+            if ((cd->flags & FF_TX_ALIGNED) && !(cd->flags & AV_TX_UNALIGNED))
+                cd_matches[nb_cd_matches].prio += 64;
+
+            /* Codelets for specific lengths are generally faster */
+            if ((len == cd->min_len) && (len == cd->max_len))
+                cd_matches[nb_cd_matches].prio += 64;
+
+            /* Forward-only or inverse-only transforms are generally better */
+            if ((cd->flags & (FF_TX_FORWARD_ONLY | FF_TX_INVERSE_ONLY)))
+                cd_matches[nb_cd_matches].prio += 64;
+
+            /* Larger factors are generally better */
+            for (int i = 0; i < TX_MAX_SUB; i++)
+                max_factor = FFMAX(cd->factors[i], max_factor);
+            if (max_factor)
+                cd_matches[nb_cd_matches].prio += 16*max_factor;
+
+            nb_cd_matches++;
+        }
+    }
+
+    /* No matches found */
+    if (!nb_cd_matches)
+        return AVERROR(ENOSYS);
+
+    /* Sort the list */
+    AV_QSORT(cd_matches, nb_cd_matches, TXCodeletMatch, cmp_matches);
+
+    /* Print debugging info */
+    av_bprint_init(&bp, 0, AV_BPRINT_SIZE_AUTOMATIC);
+    av_bprintf(&bp, "For transform of length %i, %s, ", len,
+               inv ? "inverse" : "forward");
+    print_type(&bp, type);
+    av_bprintf(&bp, ", ");
+    print_flags(&bp, flags);
+    av_bprintf(&bp, ", found %i matches:", nb_cd_matches);
+    av_log(NULL, AV_LOG_VERBOSE, "%s\n", bp.str);
+
+    for (int i = 0; i < nb_cd_matches; i++) {
+        av_log(NULL, AV_LOG_VERBOSE, "    %i: ", i + 1);
+        print_cd_info(cd_matches[i].cd, cd_matches[i].prio, 1);
+    }
+
+    if (!s->sub)
+        s->sub = sub = av_mallocz(TX_MAX_SUB*sizeof(*sub));
+
+    /* Attempt to initialize each */
+    for (int i = 0; i < nb_cd_matches; i++) {
+        const FFTXCodelet *cd = cd_matches[i].cd;
+        AVTXContext *sctx = &s->sub[s->nb_sub];
+
+        sctx->len        = len;
+        sctx->inv        = inv;
+        sctx->type       = type;
+        sctx->flags      = flags;
+        sctx->cd_self    = cd;
+
+        s->fn[s->nb_sub] = cd->function;
+        s->cd[s->nb_sub] = cd;
+
+        ret = 0;
+        if (cd->init)
+            ret = cd->init(sctx, cd, flags, opts, len, inv, scale);
+
+        if (ret >= 0) {
+            s->nb_sub++;
+            goto end;
+        }
+
+        s->fn[s->nb_sub] = NULL;
+        s->cd[s->nb_sub] = NULL;
+
+        reset_ctx(sctx);
+        if (ret == AVERROR(ENOMEM))
+            break;
+    }
+
+    av_free(sub);
+
+    if (ret >= 0)
+        ret = AVERROR(ENOSYS);
+
+end:
+    av_free(cd_matches);
+    return ret;
+}
+
+static void print_tx_structure(AVTXContext *s, int depth)
+{
+    const FFTXCodelet *cd = s->cd_self;
+
+    for (int i = 0; i <= depth; i++)
+        av_log(NULL, AV_LOG_VERBOSE, "    ");
+
+    print_cd_info(cd, cd->prio, 0);
+
+    for (int i = 0; i < s->nb_sub; i++)
+        print_tx_structure(&s->sub[i], depth + 1);
+}
+
 av_cold int av_tx_init(AVTXContext **ctx, av_tx_fn *tx, enum AVTXType type,
                        int inv, int len, const void *scale, uint64_t flags)
 {
-    int err;
-    AVTXContext *s = av_mallocz(sizeof(*s));
-    if (!s)
-        return AVERROR(ENOMEM);
+    int ret;
+    AVTXContext tmp = { 0 };
+    const double default_scale_d = 1.0;
+    const float  default_scale_f = 1.0f;
 
-    switch (type) {
-    case AV_TX_FLOAT_FFT:
-    case AV_TX_FLOAT_MDCT:
-        if ((err = ff_tx_init_mdct_fft_float(s, tx, type, inv, len, scale, flags)))
-            goto fail;
-        if (ARCH_X86)
-            ff_tx_init_float_x86(s, tx);
-        break;
-    case AV_TX_DOUBLE_FFT:
-    case AV_TX_DOUBLE_MDCT:
-        if ((err = ff_tx_init_mdct_fft_double(s, tx, type, inv, len, scale, flags)))
-            goto fail;
-        break;
-    case AV_TX_INT32_FFT:
-    case AV_TX_INT32_MDCT:
-        if ((err = ff_tx_init_mdct_fft_int32(s, tx, type, inv, len, scale, flags)))
-            goto fail;
-        break;
-    default:
-        err = AVERROR(EINVAL);
-        goto fail;
-    }
+    if (!len || type >= AV_TX_NB || !ctx || !tx)
+        return AVERROR(EINVAL);
 
-    *ctx = s;
+    if (!(flags & AV_TX_UNALIGNED))
+        flags |= FF_TX_ALIGNED;
+    if (!(flags & AV_TX_INPLACE))
+        flags |= FF_TX_OUT_OF_PLACE;
 
-    return 0;
+    if (!scale && ((type == AV_TX_FLOAT_MDCT) || (type == AV_TX_INT32_MDCT)))
+        scale = &default_scale_f;
+    else if (!scale && (type == AV_TX_DOUBLE_MDCT))
+        scale = &default_scale_d;
+
+    ret = ff_tx_init_subtx(&tmp, type, flags, NULL, len, inv, scale);
+    if (ret < 0)
+        return ret;
+
+    *ctx = &tmp.sub[0];
+    *tx  = tmp.fn[0];
+
+    av_log(NULL, AV_LOG_VERBOSE, "Transform tree:\n");
+    print_tx_structure(*ctx, 0);
 
-fail:
-    av_tx_uninit(&s);
-    *tx = NULL;
-    return err;
+    return ret;
 }
diff --git a/libavutil/tx.h b/libavutil/tx.h
index 94626e709f..0e57874376 100644
--- a/libavutil/tx.h
+++ b/libavutil/tx.h
@@ -52,6 +52,7 @@ enum AVTXType {
      * Standard MDCT with a sample data type of float, double or int32_t,
      * respecively. For the float and int32 variants, the scale type is
      * 'float', while for the double variant, it's 'double'.
+     * If scale is NULL, 1.0 will be used as a default.
      *
      * Length is the frame size, not the window size (which is 2x frame).
      * For forward transforms, the stride specifies the spacing between each
@@ -67,6 +68,9 @@ enum AVTXType {
     AV_TX_FLOAT_MDCT  = 1,
     AV_TX_DOUBLE_MDCT = 3,
     AV_TX_INT32_MDCT  = 5,
+
+    /* Not part of the API, do not use */
+    AV_TX_NB,
 };
 
 /**
diff --git a/libavutil/tx_priv.h b/libavutil/tx_priv.h
index 63dc6bbe6d..47e53fcc26 100644
--- a/libavutil/tx_priv.h
+++ b/libavutil/tx_priv.h
@@ -25,36 +25,48 @@
 #include "attributes.h"
 
 #ifdef TX_FLOAT
-#define TX_NAME(x) x ## _float
+#define TX_TAB(x) x ## _float
+#define TX_NAME(x) x ## _float_c
+#define TX_NAME_STR(x) x "_float_c"
+#define TX_TYPE(x) AV_TX_FLOAT_ ## x
+#define MULT(x, m) ((x) * (m))
 #define SCALE_TYPE float
-typedef float FFTSample;
-typedef AVComplexFloat FFTComplex;
+typedef float TXSample;
+typedef AVComplexFloat TXComplex;
 #elif defined(TX_DOUBLE)
-#define TX_NAME(x) x ## _double
+#define TX_TAB(x) x ## _double
+#define TX_NAME(x) x ## _double_c
+#define TX_NAME_STR(x) x "_double_c"
+#define TX_TYPE(x) AV_TX_DOUBLE_ ## x
+#define MULT(x, m) ((x) * (m))
 #define SCALE_TYPE double
-typedef double FFTSample;
-typedef AVComplexDouble FFTComplex;
+typedef double TXSample;
+typedef AVComplexDouble TXComplex;
 #elif defined(TX_INT32)
-#define TX_NAME(x) x ## _int32
+#define TX_TAB(x) x ## _int32
+#define TX_NAME(x) x ## _int32_c
+#define TX_NAME_STR(x) x "_int32_c"
+#define TX_TYPE(x) AV_TX_INT32_ ## x
+#define MULT(x, m) (((((int64_t)(x)) * (int64_t)(m)) + 0x40000000) >> 31)
 #define SCALE_TYPE float
-typedef int32_t FFTSample;
-typedef AVComplexInt32 FFTComplex;
+typedef int32_t TXSample;
+typedef AVComplexInt32 TXComplex;
 #else
-typedef void FFTComplex;
+typedef void TXComplex;
 #endif
 
 #if defined(TX_FLOAT) || defined(TX_DOUBLE)
 
-#define CMUL(dre, dim, are, aim, bre, bim)                                     \
-    do {                                                                       \
-        (dre) = (are) * (bre) - (aim) * (bim);                                 \
-        (dim) = (are) * (bim) + (aim) * (bre);                                 \
+#define CMUL(dre, dim, are, aim, bre, bim)      \
+    do {                                        \
+        (dre) = (are) * (bre) - (aim) * (bim);  \
+        (dim) = (are) * (bim) + (aim) * (bre);  \
     } while (0)
 
-#define SMUL(dre, dim, are, aim, bre, bim)                                     \
-    do {                                                                       \
-        (dre) = (are) * (bre) - (aim) * (bim);                                 \
-        (dim) = (are) * (bim) - (aim) * (bre);                                 \
+#define SMUL(dre, dim, are, aim, bre, bim)      \
+    do {                                        \
+        (dre) = (are) * (bre) - (aim) * (bim);  \
+        (dim) = (are) * (bim) - (aim) * (bre);  \
     } while (0)
 
 #define UNSCALE(x) (x)
@@ -65,91 +77,167 @@ typedef void FFTComplex;
 #elif defined(TX_INT32)
 
 /* Properly rounds the result */
-#define CMUL(dre, dim, are, aim, bre, bim)                                     \
-    do {                                                                       \
-        int64_t accu;                                                          \
-        (accu)  = (int64_t)(bre) * (are);                                      \
-        (accu) -= (int64_t)(bim) * (aim);                                      \
-        (dre)   = (int)(((accu) + 0x40000000) >> 31);                          \
-        (accu)  = (int64_t)(bim) * (are);                                      \
-        (accu) += (int64_t)(bre) * (aim);                                      \
-        (dim)   = (int)(((accu) + 0x40000000) >> 31);                          \
+#define CMUL(dre, dim, are, aim, bre, bim)             \
+    do {                                               \
+        int64_t accu;                                  \
+        (accu)  = (int64_t)(bre) * (are);              \
+        (accu) -= (int64_t)(bim) * (aim);              \
+        (dre)   = (int)(((accu) + 0x40000000) >> 31);  \
+        (accu)  = (int64_t)(bim) * (are);              \
+        (accu) += (int64_t)(bre) * (aim);              \
+        (dim)   = (int)(((accu) + 0x40000000) >> 31);  \
     } while (0)
 
-#define SMUL(dre, dim, are, aim, bre, bim)                                     \
-    do {                                                                       \
-        int64_t accu;                                                          \
-        (accu)  = (int64_t)(bre) * (are);                                      \
-        (accu) -= (int64_t)(bim) * (aim);                                      \
-        (dre)   = (int)(((accu) + 0x40000000) >> 31);                          \
-        (accu)  = (int64_t)(bim) * (are);                                      \
-        (accu) -= (int64_t)(bre) * (aim);                                      \
-        (dim)   = (int)(((accu) + 0x40000000) >> 31);                          \
+#define SMUL(dre, dim, are, aim, bre, bim)             \
+    do {                                               \
+        int64_t accu;                                  \
+        (accu)  = (int64_t)(bre) * (are);              \
+        (accu) -= (int64_t)(bim) * (aim);              \
+        (dre)   = (int)(((accu) + 0x40000000) >> 31);  \
+        (accu)  = (int64_t)(bim) * (are);              \
+        (accu) -= (int64_t)(bre) * (aim);              \
+        (dim)   = (int)(((accu) + 0x40000000) >> 31);  \
     } while (0)
 
-#define UNSCALE(x) ((double)x/2147483648.0)
+#define UNSCALE(x) ((double)(x)/2147483648.0)
 #define RESCALE(x) (av_clip64(lrintf((x) * 2147483648.0), INT32_MIN, INT32_MAX))
 
-#define FOLD(x, y) ((int)((x) + (unsigned)(y) + 32) >> 6)
+#define FOLD(x, y) ((int32_t)((x) + (unsigned)(y) + 32) >> 6)
 
-#endif
+#endif /* TX_INT32 */
 
-#define BF(x, y, a, b)                                                         \
-    do {                                                                       \
-        x = (a) - (b);                                                         \
-        y = (a) + (b);                                                         \
+#define BF(x, y, a, b)  \
+    do {                \
+        x = (a) - (b);  \
+        y = (a) + (b);  \
     } while (0)
 
-#define CMUL3(c, a, b)                                                         \
-    CMUL((c).re, (c).im, (a).re, (a).im, (b).re, (b).im)
+#define CMUL3(c, a, b) CMUL((c).re, (c).im, (a).re, (a).im, (b).re, (b).im)
+
+/* Codelet flags, used to pick codelets. Must be a superset of enum AVTXFlags,
+ * but if it runs out of bits, it can be made separate. */
+typedef enum FFTXCodeletFlags {
+    FF_TX_OUT_OF_PLACE  = (1ULL << 63), /* Can be OR'd with AV_TX_INPLACE             */
+    FF_TX_ALIGNED       = (1ULL << 62), /* Cannot be OR'd with AV_TX_UNALIGNED        */
+    FF_TX_PRESHUFFLE    = (1ULL << 61), /* Codelet expects permuted coeffs            */
+    FF_TX_INVERSE_ONLY  = (1ULL << 60), /* For non-orthogonal inverse-only transforms */
+    FF_TX_FORWARD_ONLY  = (1ULL << 59), /* For non-orthogonal forward-only transforms */
+} FFTXCodeletFlags;
+
+typedef enum FFTXCodeletPriority {
+    FF_TX_PRIO_BASE = 0,               /* Baseline priority */
+
+    /* For SIMD, set base prio to the register size in bits and increment in
+     * steps of 64 depending on faster/slower features, like FMA. */
+
+    FF_TX_PRIO_MIN          = -131072, /* For naive implementations */
+    FF_TX_PRIO_MAX          =  32768,  /* For custom implementations/ASICs */
+} FFTXCodeletPriority;
+
+/* Codelet options */
+typedef struct FFTXCodeletOptions {
+    int invert_lookup;     /* If codelet is flagged as FF_TX_CODELET_PRESHUFFLE,
+                              invert the lookup direction for the map generated */
+} FFTXCodeletOptions;
+
+/* Maximum amount of subtransform functions, subtransforms and factors. Arbitrary. */
+#define TX_MAX_SUB 4
+
+typedef struct FFTXCodelet {
+    const char    *name;          /* Codelet name, for debugging */
+    av_tx_fn       function;      /* Codelet function, != NULL */
+    enum AVTXType  type;          /* Type of codelet transform */
+#define TX_TYPE_ANY INT32_MAX     /* Special type to allow all types */
+
+    uint64_t flags;               /* A combination of AVTXFlags and FFTXCodeletFlags
+                                   * that describe the codelet's properties. */
+
+    int factors[TX_MAX_SUB];      /* Length factors */
+#define TX_FACTOR_ANY -1          /* When used alone, signals that the codelet
+                                   * supports all factors. Otherwise, if other
+                                   * factors are present, it signals that whatever
+                                   * remains will be supported, as long as the
+                                   * other factors are a component of the length */
+
+    int min_len;                  /* Minimum length of transform, must be >= 1 */
+    int max_len;                  /* Maximum length of transform */
+#define TX_LEN_UNLIMITED -1       /* Special length value to permit all lengths */
 
-#define COSTABLE(size)                                                         \
-    DECLARE_ALIGNED(32, FFTSample, TX_NAME(ff_cos_##size))[size/4 + 1]
+    int (*init)(AVTXContext *s,   /* Optional callback for current context initialization. */
+                const struct FFTXCodelet *cd,
+                uint64_t flags,
+                FFTXCodeletOptions *opts,
+                int len, int inv,
+                const void *scale);
+
+    int (*uninit)(AVTXContext *s); /* Optional callback for uninitialization. */
+
+    int cpu_flags;                 /* CPU flags. If any negative flags like
+                                    * SLOW are present, will avoid picking.
+                                    * 0x0 to signal it's a C codelet */
+#define FF_TX_CPU_FLAGS_ALL 0x0    /* Special CPU flag for C */
+
+    int prio;                      /* < 0 = least, 0 = no pref, > 0 = prefer */
+} FFTXCodelet;
 
-/* Used by asm, reorder with care */
 struct AVTXContext {
-    int n;              /* Non-power-of-two part */
-    int m;              /* Power-of-two part */
-    int inv;            /* Is inverse */
-    int type;           /* Type */
-    uint64_t flags;     /* Flags */
-    double scale;       /* Scale */
-
-    FFTComplex *exptab; /* MDCT exptab */
-    FFTComplex    *tmp; /* Temporary buffer needed for all compound transforms */
-    int        *pfatab; /* Input/Output mapping for compound transforms */
-    int        *revtab; /* Input mapping for power of two transforms */
-    int   *inplace_idx; /* Required indices to revtab for in-place transforms */
-
-    int      *revtab_c; /* Revtab for only the C transforms, needed because
-                         * checkasm makes us reuse the same context. */
-
-    av_tx_fn    top_tx; /* Used for computing transforms derived from other
-                         * transforms, like full-length iMDCTs and RDFTs.
-                         * NOTE: Do NOT use this to mix assembly with C code. */
+    /* Fields the root transform and subtransforms use or may use.
+     * NOTE: This section is used by assembly, do not reorder or change */
+    int                len;             /* Length of the transform */
+    int                inv;             /* If transform is inverse */
+    int               *map;             /* Lookup table(s) */
+    TXComplex         *exp;             /* Any non-pre-baked multiplication factors needed */
+    TXComplex         *tmp;             /* Temporary buffer, if needed */
+
+    AVTXContext       *sub;             /* Subtransform context(s), if needed */
+    av_tx_fn           fn[TX_MAX_SUB];  /* Function(s) for the subtransforms */
+    int                nb_sub;          /* Number of subtransforms.
+                                         * The reason all of these are set here
+                                         * rather than in each separate context
+                                         * is to eliminate extra pointer
+                                         * dereferences. */
+
+    /* Fields mainly useul/applicable for the root transform or initialization.
+     * Fields below are not used by assembly code. */
+    const FFTXCodelet *cd[TX_MAX_SUB];  /* Subtransform codelets */
+    const FFTXCodelet *cd_self;         /* Codelet for the current context */
+    enum AVTXType      type;            /* Type of transform */
+    uint64_t           flags;           /* A combination of AVTXFlags
+                                           and FFTXCodeletFlags flags
+                                           used when creating */
+    float              scale_f;
+    double             scale_d;
+    void              *opaque;          /* Free to use by implementations */
 };
 
-/* Checks if type is an MDCT */
-int ff_tx_type_is_mdct(enum AVTXType type);
+/* Create a subtransform in the current context with the given parameters.
+ * The flags parameter from FFTXCodelet.init() should be preserved as much
+ * as that's possible.
+ * MUST be called during the sub() callback of each codelet. */
+int ff_tx_init_subtx(AVTXContext *s, enum AVTXType type,
+                     uint64_t flags, FFTXCodeletOptions *opts,
+                     int len, int inv, const void *scale);
 
 /*
  * Generates the PFA permutation table into AVTXContext->pfatab. The end table
  * is appended to the start table.
  */
-int ff_tx_gen_compound_mapping(AVTXContext *s);
+int ff_tx_gen_compound_mapping(AVTXContext *s, int n, int m);
 
 /*
  * Generates a standard-ish (slightly modified) Split-Radix revtab into
- * AVTXContext->revtab
+ * AVTXContext->map. Invert lookup changes how the mapping needs to be applied.
+ * If it's set to 0, it has to be applied like out[map[i]] = in[i], otherwise
+ * if it's set to 1, has to be applied as out[i] = in[map[i]]
  */
 int ff_tx_gen_ptwo_revtab(AVTXContext *s, int invert_lookup);
 
 /*
  * Generates an index into AVTXContext->inplace_idx that if followed in the
- * specific order,  allows the revtab to be done in-place. AVTXContext->revtab
- * must already exist.
+ * specific order, allows the revtab to be done in-place. The sub-transform
+ * and its map should already be initialized.
  */
-int ff_tx_gen_ptwo_inplace_revtab_idx(AVTXContext *s, int *revtab);
+int ff_tx_gen_ptwo_inplace_revtab_idx(AVTXContext *s);
 
 /*
  * This generates a parity-based revtab of length len and direction inv.
@@ -179,25 +267,26 @@ int ff_tx_gen_ptwo_inplace_revtab_idx(AVTXContext *s, int *revtab);
  *
  * If length is smaller than basis/2 this function will not do anything.
  */
-void ff_tx_gen_split_radix_parity_revtab(int *revtab, int len, int inv,
-                                         int basis, int dual_stride);
-
-/* Templated init functions */
-int ff_tx_init_mdct_fft_float(AVTXContext *s, av_tx_fn *tx,
-                              enum AVTXType type, int inv, int len,
-                              const void *scale, uint64_t flags);
-int ff_tx_init_mdct_fft_double(AVTXContext *s, av_tx_fn *tx,
-                               enum AVTXType type, int inv, int len,
-                               const void *scale, uint64_t flags);
-int ff_tx_init_mdct_fft_int32(AVTXContext *s, av_tx_fn *tx,
-                              enum AVTXType type, int inv, int len,
-                              const void *scale, uint64_t flags);
-
-typedef struct CosTabsInitOnce {
-    void (*func)(void);
-    AVOnce control;
-} CosTabsInitOnce;
-
-void ff_tx_init_float_x86(AVTXContext *s, av_tx_fn *tx);
+int ff_tx_gen_split_radix_parity_revtab(AVTXContext *s, int invert_lookup,
+                                        int basis, int dual_stride);
+
+/* Typed init function to initialize shared tables. Will initialize all tables
+ * for all factors of a length. */
+void ff_tx_init_tabs_float (int len);
+void ff_tx_init_tabs_double(int len);
+void ff_tx_init_tabs_int32 (int len);
+
+/* Typed init function to initialize an MDCT exptab in a context. */
+int  ff_tx_mdct_gen_exp_float (AVTXContext *s);
+int  ff_tx_mdct_gen_exp_double(AVTXContext *s);
+int  ff_tx_mdct_gen_exp_int32 (AVTXContext *s);
+
+/* Lists of codelets */
+extern const FFTXCodelet * const ff_tx_codelet_list_float_c       [];
+extern const FFTXCodelet * const ff_tx_codelet_list_float_x86     [];
+
+extern const FFTXCodelet * const ff_tx_codelet_list_double_c      [];
+
+extern const FFTXCodelet * const ff_tx_codelet_list_int32_c       [];
 
 #endif /* AVUTIL_TX_PRIV_H */
diff --git a/libavutil/tx_template.c b/libavutil/tx_template.c
index cad66a8bc0..cee8458970 100644
--- a/libavutil/tx_template.c
+++ b/libavutil/tx_template.c
@@ -24,134 +24,160 @@
  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  */
 
-/* All costabs for a type are defined here */
-COSTABLE(16);
-COSTABLE(32);
-COSTABLE(64);
-COSTABLE(128);
-COSTABLE(256);
-COSTABLE(512);
-COSTABLE(1024);
-COSTABLE(2048);
-COSTABLE(4096);
-COSTABLE(8192);
-COSTABLE(16384);
-COSTABLE(32768);
-COSTABLE(65536);
-COSTABLE(131072);
-DECLARE_ALIGNED(32, FFTComplex, TX_NAME(ff_cos_53))[4];
-DECLARE_ALIGNED(32, FFTComplex, TX_NAME(ff_cos_7))[3];
-DECLARE_ALIGNED(32, FFTComplex, TX_NAME(ff_cos_9))[4];
-
-static FFTSample * const cos_tabs[18] = {
-    NULL,
-    NULL,
-    NULL,
-    NULL,
-    TX_NAME(ff_cos_16),
-    TX_NAME(ff_cos_32),
-    TX_NAME(ff_cos_64),
-    TX_NAME(ff_cos_128),
-    TX_NAME(ff_cos_256),
-    TX_NAME(ff_cos_512),
-    TX_NAME(ff_cos_1024),
-    TX_NAME(ff_cos_2048),
-    TX_NAME(ff_cos_4096),
-    TX_NAME(ff_cos_8192),
-    TX_NAME(ff_cos_16384),
-    TX_NAME(ff_cos_32768),
-    TX_NAME(ff_cos_65536),
-    TX_NAME(ff_cos_131072),
-};
-
-static av_always_inline void init_cos_tabs_idx(int index)
-{
-    int m = 1 << index;
-    double freq = 2*M_PI/m;
-    FFTSample *tab = cos_tabs[index];
-
-    for (int i = 0; i < m/4; i++)
-        *tab++ = RESCALE(cos(i*freq));
-
-    *tab = 0;
+#define TABLE_DEF(name, size) \
+    DECLARE_ALIGNED(32, TXSample, TX_TAB(ff_tx_tab_ ##name))[size]
+
+#define SR_TABLE(len) \
+    TABLE_DEF(len, len/4 + 1)
+
+/* Power of two tables */
+SR_TABLE(8);
+SR_TABLE(16);
+SR_TABLE(32);
+SR_TABLE(64);
+SR_TABLE(128);
+SR_TABLE(256);
+SR_TABLE(512);
+SR_TABLE(1024);
+SR_TABLE(2048);
+SR_TABLE(4096);
+SR_TABLE(8192);
+SR_TABLE(16384);
+SR_TABLE(32768);
+SR_TABLE(65536);
+SR_TABLE(131072);
+
+/* Other factors' tables */
+TABLE_DEF(53, 8);
+TABLE_DEF( 7, 6);
+TABLE_DEF( 9, 8);
+
+typedef struct FFSRTabsInitOnce {
+    void (*func)(void);
+    AVOnce control;
+    int factors[TX_MAX_SUB]; /* Must be sorted high -> low */
+} FFSRTabsInitOnce;
+
+#define INIT_FF_SR_TAB(len)                                        \
+static av_cold void TX_TAB(ff_tx_init_tab_ ##len)(void)            \
+{                                                                  \
+    double freq = 2*M_PI/len;                                      \
+    TXSample *tab = TX_TAB(ff_tx_tab_ ##len);                      \
+                                                                   \
+    for (int i = 0; i < len/4; i++)                                \
+        *tab++ = RESCALE(cos(i*freq));                             \
+                                                                   \
+    *tab = 0;                                                      \
 }
 
-#define INIT_FF_COS_TABS_FUNC(index, size)                                     \
-static av_cold void init_cos_tabs_ ## size (void)                              \
-{                                                                              \
-    init_cos_tabs_idx(index);                                                  \
-}
+INIT_FF_SR_TAB(8)
+INIT_FF_SR_TAB(16)
+INIT_FF_SR_TAB(32)
+INIT_FF_SR_TAB(64)
+INIT_FF_SR_TAB(128)
+INIT_FF_SR_TAB(256)
+INIT_FF_SR_TAB(512)
+INIT_FF_SR_TAB(1024)
+INIT_FF_SR_TAB(2048)
+INIT_FF_SR_TAB(4096)
+INIT_FF_SR_TAB(8192)
+INIT_FF_SR_TAB(16384)
+INIT_FF_SR_TAB(32768)
+INIT_FF_SR_TAB(65536)
+INIT_FF_SR_TAB(131072)
+
+static FFSRTabsInitOnce sr_tabs_init_once[] = {
+    { TX_TAB(ff_tx_init_tab_8),      AV_ONCE_INIT },
+    { TX_TAB(ff_tx_init_tab_16),     AV_ONCE_INIT },
+    { TX_TAB(ff_tx_init_tab_32),     AV_ONCE_INIT },
+    { TX_TAB(ff_tx_init_tab_64),     AV_ONCE_INIT },
+    { TX_TAB(ff_tx_init_tab_128),    AV_ONCE_INIT },
+    { TX_TAB(ff_tx_init_tab_256),    AV_ONCE_INIT },
+    { TX_TAB(ff_tx_init_tab_512),    AV_ONCE_INIT },
+    { TX_TAB(ff_tx_init_tab_1024),   AV_ONCE_INIT },
+    { TX_TAB(ff_tx_init_tab_2048),   AV_ONCE_INIT },
+    { TX_TAB(ff_tx_init_tab_4096),   AV_ONCE_INIT },
+    { TX_TAB(ff_tx_init_tab_8192),   AV_ONCE_INIT },
+    { TX_TAB(ff_tx_init_tab_16384),  AV_ONCE_INIT },
+    { TX_TAB(ff_tx_init_tab_32768),  AV_ONCE_INIT },
+    { TX_TAB(ff_tx_init_tab_65536),  AV_ONCE_INIT },
+    { TX_TAB(ff_tx_init_tab_131072), AV_ONCE_INIT },
+};
 
-INIT_FF_COS_TABS_FUNC(4, 16)
-INIT_FF_COS_TABS_FUNC(5, 32)
-INIT_FF_COS_TABS_FUNC(6, 64)
-INIT_FF_COS_TABS_FUNC(7, 128)
-INIT_FF_COS_TABS_FUNC(8, 256)
-INIT_FF_COS_TABS_FUNC(9, 512)
-INIT_FF_COS_TABS_FUNC(10, 1024)
-INIT_FF_COS_TABS_FUNC(11, 2048)
-INIT_FF_COS_TABS_FUNC(12, 4096)
-INIT_FF_COS_TABS_FUNC(13, 8192)
-INIT_FF_COS_TABS_FUNC(14, 16384)
-INIT_FF_COS_TABS_FUNC(15, 32768)
-INIT_FF_COS_TABS_FUNC(16, 65536)
-INIT_FF_COS_TABS_FUNC(17, 131072)
-
-static av_cold void ff_init_53_tabs(void)
+static av_cold void TX_TAB(ff_tx_init_tab_53)(void)
 {
-    TX_NAME(ff_cos_53)[0] = (FFTComplex){ RESCALE(cos(2 * M_PI / 12)), RESCALE(cos(2 * M_PI / 12)) };
-    TX_NAME(ff_cos_53)[1] = (FFTComplex){ RESCALE(cos(2 * M_PI /  6)), RESCALE(cos(2 * M_PI /  6)) };
-    TX_NAME(ff_cos_53)[2] = (FFTComplex){ RESCALE(cos(2 * M_PI /  5)), RESCALE(sin(2 * M_PI /  5)) };
-    TX_NAME(ff_cos_53)[3] = (FFTComplex){ RESCALE(cos(2 * M_PI / 10)), RESCALE(sin(2 * M_PI / 10)) };
+    TX_TAB(ff_tx_tab_53)[0] = RESCALE(cos(2 * M_PI / 12));
+    TX_TAB(ff_tx_tab_53)[1] = RESCALE(cos(2 * M_PI / 12));
+    TX_TAB(ff_tx_tab_53)[2] = RESCALE(cos(2 * M_PI /  6));
+    TX_TAB(ff_tx_tab_53)[3] = RESCALE(cos(2 * M_PI /  6));
+    TX_TAB(ff_tx_tab_53)[4] = RESCALE(cos(2 * M_PI /  5));
+    TX_TAB(ff_tx_tab_53)[5] = RESCALE(sin(2 * M_PI /  5));
+    TX_TAB(ff_tx_tab_53)[6] = RESCALE(cos(2 * M_PI / 10));
+    TX_TAB(ff_tx_tab_53)[7] = RESCALE(sin(2 * M_PI / 10));
 }
 
-static av_cold void ff_init_7_tabs(void)
+static av_cold void TX_TAB(ff_tx_init_tab_7)(void)
 {
-    TX_NAME(ff_cos_7)[0] = (FFTComplex){ RESCALE(cos(2 * M_PI /  7)), RESCALE(sin(2 * M_PI /  7)) };
-    TX_NAME(ff_cos_7)[1] = (FFTComplex){ RESCALE(sin(2 * M_PI / 28)), RESCALE(cos(2 * M_PI / 28)) };
-    TX_NAME(ff_cos_7)[2] = (FFTComplex){ RESCALE(cos(2 * M_PI / 14)), RESCALE(sin(2 * M_PI / 14)) };
+    TX_TAB(ff_tx_tab_7)[0] = RESCALE(cos(2 * M_PI /  7));
+    TX_TAB(ff_tx_tab_7)[1] = RESCALE(sin(2 * M_PI /  7));
+    TX_TAB(ff_tx_tab_7)[2] = RESCALE(sin(2 * M_PI / 28));
+    TX_TAB(ff_tx_tab_7)[3] = RESCALE(cos(2 * M_PI / 28));
+    TX_TAB(ff_tx_tab_7)[4] = RESCALE(cos(2 * M_PI / 14));
+    TX_TAB(ff_tx_tab_7)[5] = RESCALE(sin(2 * M_PI / 14));
 }
 
-static av_cold void ff_init_9_tabs(void)
+static av_cold void TX_TAB(ff_tx_init_tab_9)(void)
 {
-    TX_NAME(ff_cos_9)[0] = (FFTComplex){ RESCALE(cos(2 * M_PI /  3)), RESCALE( sin(2 * M_PI /  3)) };
-    TX_NAME(ff_cos_9)[1] = (FFTComplex){ RESCALE(cos(2 * M_PI /  9)), RESCALE( sin(2 * M_PI /  9)) };
-    TX_NAME(ff_cos_9)[2] = (FFTComplex){ RESCALE(cos(2 * M_PI / 36)), RESCALE( sin(2 * M_PI / 36)) };
-    TX_NAME(ff_cos_9)[3] = (FFTComplex){ TX_NAME(ff_cos_9)[1].re + TX_NAME(ff_cos_9)[2].im,
-                                         TX_NAME(ff_cos_9)[1].im - TX_NAME(ff_cos_9)[2].re };
+    TX_TAB(ff_tx_tab_9)[0] = RESCALE(cos(2 * M_PI /  3));
+    TX_TAB(ff_tx_tab_9)[1] = RESCALE(sin(2 * M_PI /  3));
+    TX_TAB(ff_tx_tab_9)[2] = RESCALE(cos(2 * M_PI /  9));
+    TX_TAB(ff_tx_tab_9)[3] = RESCALE(sin(2 * M_PI /  9));
+    TX_TAB(ff_tx_tab_9)[4] = RESCALE(cos(2 * M_PI / 36));
+    TX_TAB(ff_tx_tab_9)[5] = RESCALE(sin(2 * M_PI / 36));
+    TX_TAB(ff_tx_tab_9)[6] = TX_TAB(ff_tx_tab_9)[2] + TX_TAB(ff_tx_tab_9)[5];
+    TX_TAB(ff_tx_tab_9)[7] = TX_TAB(ff_tx_tab_9)[3] - TX_TAB(ff_tx_tab_9)[4];
 }
 
-static CosTabsInitOnce cos_tabs_init_once[] = {
-    { ff_init_53_tabs, AV_ONCE_INIT },
-    { ff_init_7_tabs, AV_ONCE_INIT },
-    { ff_init_9_tabs, AV_ONCE_INIT },
-    { NULL },
-    { init_cos_tabs_16, AV_ONCE_INIT },
-    { init_cos_tabs_32, AV_ONCE_INIT },
-    { init_cos_tabs_64, AV_ONCE_INIT },
-    { init_cos_tabs_128, AV_ONCE_INIT },
-    { init_cos_tabs_256, AV_ONCE_INIT },
-    { init_cos_tabs_512, AV_ONCE_INIT },
-    { init_cos_tabs_1024, AV_ONCE_INIT },
-    { init_cos_tabs_2048, AV_ONCE_INIT },
-    { init_cos_tabs_4096, AV_ONCE_INIT },
-    { init_cos_tabs_8192, AV_ONCE_INIT },
-    { init_cos_tabs_16384, AV_ONCE_INIT },
-    { init_cos_tabs_32768, AV_ONCE_INIT },
-    { init_cos_tabs_65536, AV_ONCE_INIT },
-    { init_cos_tabs_131072, AV_ONCE_INIT },
+static FFSRTabsInitOnce nptwo_tabs_init_once[] = {
+    { TX_TAB(ff_tx_init_tab_53),      AV_ONCE_INIT, { 15, 5, 3 } },
+    { TX_TAB(ff_tx_init_tab_9),       AV_ONCE_INIT, {  9 }       },
+    { TX_TAB(ff_tx_init_tab_7),       AV_ONCE_INIT, {  7 }       },
 };
 
-static av_cold void init_cos_tabs(int index)
+av_cold void TX_TAB(ff_tx_init_tabs)(int len)
 {
-    ff_thread_once(&cos_tabs_init_once[index].control,
-                    cos_tabs_init_once[index].func);
+    int factor_2 = ff_ctz(len);
+    if (factor_2) {
+        int idx = factor_2 - 3;
+        for (int i = 0; i <= idx; i++)
+            ff_thread_once(&sr_tabs_init_once[i].control,
+                            sr_tabs_init_once[i].func);
+        len >>= factor_2;
+    }
+
+    for (int i = 0; i < FF_ARRAY_ELEMS(nptwo_tabs_init_once); i++) {
+        int f, f_idx = 0;
+
+        if (len <= 1)
+            return;
+
+        while ((f = nptwo_tabs_init_once[i].factors[f_idx++])) {
+            if (f % len)
+                continue;
+
+            ff_thread_once(&nptwo_tabs_init_once[i].control,
+                            nptwo_tabs_init_once[i].func);
+            len /= f;
+            break;
+        }
+    }
 }
 
-static av_always_inline void fft3(FFTComplex *out, FFTComplex *in,
+static av_always_inline void fft3(TXComplex *out, TXComplex *in,
                                   ptrdiff_t stride)
 {
-    FFTComplex tmp[2];
+    TXComplex tmp[2];
+    const TXSample *tab = TX_TAB(ff_tx_tab_53);
 #ifdef TX_INT32
     int64_t mtmp[4];
 #endif
@@ -163,19 +189,19 @@ static av_always_inline void fft3(FFTComplex *out, FFTComplex *in,
     out[0*stride].im = in[0].im + tmp[1].im;
 
 #ifdef TX_INT32
-    mtmp[0] = (int64_t)TX_NAME(ff_cos_53)[0].re * tmp[0].re;
-    mtmp[1] = (int64_t)TX_NAME(ff_cos_53)[0].im * tmp[0].im;
-    mtmp[2] = (int64_t)TX_NAME(ff_cos_53)[1].re * tmp[1].re;
-    mtmp[3] = (int64_t)TX_NAME(ff_cos_53)[1].re * tmp[1].im;
+    mtmp[0] = (int64_t)tab[0] * tmp[0].re;
+    mtmp[1] = (int64_t)tab[1] * tmp[0].im;
+    mtmp[2] = (int64_t)tab[2] * tmp[1].re;
+    mtmp[3] = (int64_t)tab[2] * tmp[1].im;
     out[1*stride].re = in[0].re - (mtmp[2] + mtmp[0] + 0x40000000 >> 31);
     out[1*stride].im = in[0].im - (mtmp[3] - mtmp[1] + 0x40000000 >> 31);
     out[2*stride].re = in[0].re - (mtmp[2] - mtmp[0] + 0x40000000 >> 31);
     out[2*stride].im = in[0].im - (mtmp[3] + mtmp[1] + 0x40000000 >> 31);
 #else
-    tmp[0].re = TX_NAME(ff_cos_53)[0].re * tmp[0].re;
-    tmp[0].im = TX_NAME(ff_cos_53)[0].im * tmp[0].im;
-    tmp[1].re = TX_NAME(ff_cos_53)[1].re * tmp[1].re;
-    tmp[1].im = TX_NAME(ff_cos_53)[1].re * tmp[1].im;
+    tmp[0].re = tab[0] * tmp[0].re;
+    tmp[0].im = tab[1] * tmp[0].im;
+    tmp[1].re = tab[2] * tmp[1].re;
+    tmp[1].im = tab[2] * tmp[1].im;
     out[1*stride].re = in[0].re - tmp[1].re + tmp[0].re;
     out[1*stride].im = in[0].im - tmp[1].im - tmp[0].im;
     out[2*stride].re = in[0].re - tmp[1].re - tmp[0].re;
@@ -183,38 +209,39 @@ static av_always_inline void fft3(FFTComplex *out, FFTComplex *in,
 #endif
 }
 
-#define DECL_FFT5(NAME, D0, D1, D2, D3, D4)                                                       \
-static av_always_inline void NAME(FFTComplex *out, FFTComplex *in,                                \
-                                  ptrdiff_t stride)                                               \
-{                                                                                                 \
-    FFTComplex z0[4], t[6];                                                                       \
-                                                                                                  \
-    BF(t[1].im, t[0].re, in[1].re, in[4].re);                                                     \
-    BF(t[1].re, t[0].im, in[1].im, in[4].im);                                                     \
-    BF(t[3].im, t[2].re, in[2].re, in[3].re);                                                     \
-    BF(t[3].re, t[2].im, in[2].im, in[3].im);                                                     \
-                                                                                                  \
-    out[D0*stride].re = in[0].re + t[0].re + t[2].re;                                             \
-    out[D0*stride].im = in[0].im + t[0].im + t[2].im;                                             \
-                                                                                                  \
-    SMUL(t[4].re, t[0].re, TX_NAME(ff_cos_53)[2].re, TX_NAME(ff_cos_53)[3].re, t[2].re, t[0].re); \
-    SMUL(t[4].im, t[0].im, TX_NAME(ff_cos_53)[2].re, TX_NAME(ff_cos_53)[3].re, t[2].im, t[0].im); \
-    CMUL(t[5].re, t[1].re, TX_NAME(ff_cos_53)[2].im, TX_NAME(ff_cos_53)[3].im, t[3].re, t[1].re); \
-    CMUL(t[5].im, t[1].im, TX_NAME(ff_cos_53)[2].im, TX_NAME(ff_cos_53)[3].im, t[3].im, t[1].im); \
-                                                                                                  \
-    BF(z0[0].re, z0[3].re, t[0].re, t[1].re);                                                     \
-    BF(z0[0].im, z0[3].im, t[0].im, t[1].im);                                                     \
-    BF(z0[2].re, z0[1].re, t[4].re, t[5].re);                                                     \
-    BF(z0[2].im, z0[1].im, t[4].im, t[5].im);                                                     \
-                                                                                                  \
-    out[D1*stride].re = in[0].re + z0[3].re;                                                      \
-    out[D1*stride].im = in[0].im + z0[0].im;                                                      \
-    out[D2*stride].re = in[0].re + z0[2].re;                                                      \
-    out[D2*stride].im = in[0].im + z0[1].im;                                                      \
-    out[D3*stride].re = in[0].re + z0[1].re;                                                      \
-    out[D3*stride].im = in[0].im + z0[2].im;                                                      \
-    out[D4*stride].re = in[0].re + z0[0].re;                                                      \
-    out[D4*stride].im = in[0].im + z0[3].im;                                                      \
+#define DECL_FFT5(NAME, D0, D1, D2, D3, D4)                         \
+static av_always_inline void NAME(TXComplex *out, TXComplex *in,    \
+                                  ptrdiff_t stride)                 \
+{                                                                   \
+    TXComplex z0[4], t[6];                                          \
+    const TXSample *tab = TX_TAB(ff_tx_tab_53);                     \
+                                                                    \
+    BF(t[1].im, t[0].re, in[1].re, in[4].re);                       \
+    BF(t[1].re, t[0].im, in[1].im, in[4].im);                       \
+    BF(t[3].im, t[2].re, in[2].re, in[3].re);                       \
+    BF(t[3].re, t[2].im, in[2].im, in[3].im);                       \
+                                                                    \
+    out[D0*stride].re = in[0].re + t[0].re + t[2].re;               \
+    out[D0*stride].im = in[0].im + t[0].im + t[2].im;               \
+                                                                    \
+    SMUL(t[4].re, t[0].re, tab[4], tab[6], t[2].re, t[0].re);       \
+    SMUL(t[4].im, t[0].im, tab[4], tab[6], t[2].im, t[0].im);       \
+    CMUL(t[5].re, t[1].re, tab[5], tab[7], t[3].re, t[1].re);       \
+    CMUL(t[5].im, t[1].im, tab[5], tab[7], t[3].im, t[1].im);       \
+                                                                    \
+    BF(z0[0].re, z0[3].re, t[0].re, t[1].re);                       \
+    BF(z0[0].im, z0[3].im, t[0].im, t[1].im);                       \
+    BF(z0[2].re, z0[1].re, t[4].re, t[5].re);                       \
+    BF(z0[2].im, z0[1].im, t[4].im, t[5].im);                       \
+                                                                    \
+    out[D1*stride].re = in[0].re + z0[3].re;                        \
+    out[D1*stride].im = in[0].im + z0[0].im;                        \
+    out[D2*stride].re = in[0].re + z0[2].re;                        \
+    out[D2*stride].im = in[0].im + z0[1].im;                        \
+    out[D3*stride].re = in[0].re + z0[1].re;                        \
+    out[D3*stride].im = in[0].im + z0[2].im;                        \
+    out[D4*stride].re = in[0].re + z0[0].re;                        \
+    out[D4*stride].im = in[0].im + z0[3].im;                        \
 }
 
 DECL_FFT5(fft5,     0,  1,  2,  3,  4)
@@ -222,11 +249,11 @@ DECL_FFT5(fft5_m1,  0,  6, 12,  3,  9)
 DECL_FFT5(fft5_m2, 10,  1,  7, 13,  4)
 DECL_FFT5(fft5_m3,  5, 11,  2,  8, 14)
 
-static av_always_inline void fft7(FFTComplex *out, FFTComplex *in,
+static av_always_inline void fft7(TXComplex *out, TXComplex *in,
                                   ptrdiff_t stride)
 {
-    FFTComplex t[6], z[3];
-    const FFTComplex *tab = TX_NAME(ff_cos_7);
+    TXComplex t[6], z[3];
+    const TXComplex *tab = (const TXComplex *)TX_TAB(ff_tx_tab_7);
 #ifdef TX_INT32
     int64_t mtmp[12];
 #endif
@@ -309,11 +336,11 @@ static av_always_inline void fft7(FFTComplex *out, FFTComplex *in,
     out[6*stride].im = in[0].im + z[0].im;
 }
 
-static av_always_inline void fft9(FFTComplex *out, FFTComplex *in,
+static av_always_inline void fft9(TXComplex *out, TXComplex *in,
                                   ptrdiff_t stride)
 {
-    const FFTComplex *tab = TX_NAME(ff_cos_9);
-    FFTComplex t[16], w[4], x[5], y[5], z[2];
+    const TXComplex *tab = (const TXComplex *)TX_TAB(ff_tx_tab_9);
+    TXComplex t[16], w[4], x[5], y[5], z[2];
 #ifdef TX_INT32
     int64_t mtmp[12];
 #endif
@@ -423,20 +450,20 @@ static av_always_inline void fft9(FFTComplex *out, FFTComplex *in,
     y[4].re = y[0].re - y[4].re;
     y[4].im = y[0].im - y[4].im;
 
-    out[1*stride] = (FFTComplex){ x[1].re + y[1].im, x[1].im - y[1].re };
-    out[2*stride] = (FFTComplex){ x[2].re + y[2].im, x[2].im - y[2].re };
-    out[3*stride] = (FFTComplex){ x[3].re + y[3].im, x[3].im - y[3].re };
-    out[4*stride] = (FFTComplex){ x[4].re + y[4].im, x[4].im - y[4].re };
-    out[5*stride] = (FFTComplex){ x[4].re - y[4].im, x[4].im + y[4].re };
-    out[6*stride] = (FFTComplex){ x[3].re - y[3].im, x[3].im + y[3].re };
-    out[7*stride] = (FFTComplex){ x[2].re - y[2].im, x[2].im + y[2].re };
-    out[8*stride] = (FFTComplex){ x[1].re - y[1].im, x[1].im + y[1].re };
+    out[1*stride] = (TXComplex){ x[1].re + y[1].im, x[1].im - y[1].re };
+    out[2*stride] = (TXComplex){ x[2].re + y[2].im, x[2].im - y[2].re };
+    out[3*stride] = (TXComplex){ x[3].re + y[3].im, x[3].im - y[3].re };
+    out[4*stride] = (TXComplex){ x[4].re + y[4].im, x[4].im - y[4].re };
+    out[5*stride] = (TXComplex){ x[4].re - y[4].im, x[4].im + y[4].re };
+    out[6*stride] = (TXComplex){ x[3].re - y[3].im, x[3].im + y[3].re };
+    out[7*stride] = (TXComplex){ x[2].re - y[2].im, x[2].im + y[2].re };
+    out[8*stride] = (TXComplex){ x[1].re - y[1].im, x[1].im + y[1].re };
 }
 
-static av_always_inline void fft15(FFTComplex *out, FFTComplex *in,
+static av_always_inline void fft15(TXComplex *out, TXComplex *in,
                                    ptrdiff_t stride)
 {
-    FFTComplex tmp[15];
+    TXComplex tmp[15];
 
     for (int i = 0; i < 5; i++)
         fft3(tmp + i, in + i*3, 5);
@@ -446,7 +473,7 @@ static av_always_inline void fft15(FFTComplex *out, FFTComplex *in,
     fft5_m3(out, tmp + 10, stride);
 }
 
-#define BUTTERFLIES(a0,a1,a2,a3)               \
+#define BUTTERFLIES(a0, a1, a2, a3)            \
     do {                                       \
         r0=a0.re;                              \
         i0=a0.im;                              \
@@ -460,7 +487,7 @@ static av_always_inline void fft15(FFTComplex *out, FFTComplex *in,
         BF(a2.im, a0.im, i0, t6);              \
     } while (0)
 
-#define TRANSFORM(a0,a1,a2,a3,wre,wim)         \
+#define TRANSFORM(a0, a1, a2, a3, wre, wim)    \
     do {                                       \
         CMUL(t1, t2, a2.re, a2.im, wre, -wim); \
         CMUL(t5, t6, a3.re, a3.im, wre,  wim); \
@@ -468,15 +495,16 @@ static av_always_inline void fft15(FFTComplex *out, FFTComplex *in,
     } while (0)
 
 /* z[0...8n-1], w[1...2n-1] */
-static void split_radix_combine(FFTComplex *z, const FFTSample *cos, int n)
+static inline void TX_NAME(ff_tx_fft_sr_combine)(TXComplex *z,
+                                                 const TXSample *cos, int len)
 {
-    int o1 = 2*n;
-    int o2 = 4*n;
-    int o3 = 6*n;
-    const FFTSample *wim = cos + o1 - 7;
-    FFTSample t1, t2, t3, t4, t5, t6, r0, i0, r1, i1;
+    int o1 = 2*len;
+    int o2 = 4*len;
+    int o3 = 6*len;
+    const TXSample *wim = cos + o1 - 7;
+    TXSample t1, t2, t3, t4, t5, t6, r0, i0, r1, i1;
 
-    for (int i = 0; i < n; i += 4) {
+    for (int i = 0; i < len; i += 4) {
         TRANSFORM(z[0], z[o1 + 0], z[o2 + 0], z[o3 + 0], cos[0], wim[7]);
         TRANSFORM(z[2], z[o1 + 2], z[o2 + 2], z[o3 + 2], cos[2], wim[5]);
         TRANSFORM(z[4], z[o1 + 4], z[o2 + 4], z[o3 + 4], cos[4], wim[3]);
@@ -493,26 +521,63 @@ static void split_radix_combine(FFTComplex *z, const FFTSample *cos, int n)
     }
 }
 
-#define DECL_FFT(n, n2, n4)                            \
-static void fft##n(FFTComplex *z)                      \
-{                                                      \
-    fft##n2(z);                                        \
-    fft##n4(z + n4*2);                                 \
-    fft##n4(z + n4*3);                                 \
-    split_radix_combine(z, TX_NAME(ff_cos_##n), n4/2); \
+static av_cold int TX_NAME(ff_tx_fft_sr_codelet_init)(AVTXContext *s,
+                                                      const FFTXCodelet *cd,
+                                                      uint64_t flags,
+                                                      FFTXCodeletOptions *opts,
+                                                      int len, int inv,
+                                                      const void *scale)
+{
+    TX_TAB(ff_tx_init_tabs)(len);
+    return ff_tx_gen_ptwo_revtab(s, opts ? opts->invert_lookup : 1);
 }
 
-static void fft2(FFTComplex *z)
+#define DECL_SR_CODELET_DEF(n)                              \
+static const FFTXCodelet TX_NAME(ff_tx_fft##n##_ns_def) = { \
+    .name       = TX_NAME_STR("fft" #n "_ns"),              \
+    .function   = TX_NAME(ff_tx_fft##n##_ns),               \
+    .type       = TX_TYPE(FFT),                             \
+    .flags      = AV_TX_INPLACE | AV_TX_UNALIGNED |         \
+                  FF_TX_PRESHUFFLE,                         \
+    .factors[0] = 2,                                        \
+    .min_len    = n,                                        \
+    .max_len    = n,                                        \
+    .init       = TX_NAME(ff_tx_fft_sr_codelet_init),       \
+    .cpu_flags  = FF_TX_CPU_FLAGS_ALL,                      \
+    .prio       = FF_TX_PRIO_BASE,                          \
+};
+
+#define DECL_SR_CODELET(n, n2, n4)                                   \
+static void TX_NAME(ff_tx_fft##n##_ns)(AVTXContext *s, void *dst,    \
+                                        void *src, ptrdiff_t stride) \
+{                                                                    \
+    TXComplex *z = dst;                                              \
+    const TXSample *cos = TX_TAB(ff_tx_tab_##n);                     \
+                                                                     \
+    TX_NAME(ff_tx_fft##n2##_ns)(s, z,        z,        stride);      \
+    TX_NAME(ff_tx_fft##n4##_ns)(s, z + n4*2, z + n4*2, stride);      \
+    TX_NAME(ff_tx_fft##n4##_ns)(s, z + n4*3, z + n4*3, stride);      \
+    TX_NAME(ff_tx_fft_sr_combine)(z, cos, n4 >> 1);                  \
+}                                                                    \
+                                                                     \
+DECL_SR_CODELET_DEF(n)
+
+static void TX_NAME(ff_tx_fft2_ns)(AVTXContext *s, void *dst,
+                                   void *src, ptrdiff_t stride)
 {
-    FFTComplex tmp;
+    TXComplex *z = dst;
+    TXComplex tmp;
+
     BF(tmp.re, z[0].re, z[0].re, z[1].re);
     BF(tmp.im, z[0].im, z[0].im, z[1].im);
     z[1] = tmp;
 }
 
-static void fft4(FFTComplex *z)
+static void TX_NAME(ff_tx_fft4_ns)(AVTXContext *s, void *dst,
+                                   void *src, ptrdiff_t stride)
 {
-    FFTSample t1, t2, t3, t4, t5, t6, t7, t8;
+    TXComplex *z = dst;
+    TXSample t1, t2, t3, t4, t5, t6, t7, t8;
 
     BF(t3, t1, z[0].re, z[1].re);
     BF(t8, t6, z[3].re, z[2].re);
@@ -524,11 +589,14 @@ static void fft4(FFTComplex *z)
     BF(z[2].im, z[0].im, t2, t5);
 }
 
-static void fft8(FFTComplex *z)
+static void TX_NAME(ff_tx_fft8_ns)(AVTXContext *s, void *dst,
+                                   void *src, ptrdiff_t stride)
 {
-    FFTSample t1, t2, t3, t4, t5, t6, r0, i0, r1, i1;
+    TXComplex *z = dst;
+    TXSample t1, t2, t3, t4, t5, t6, r0, i0, r1, i1;
+    const TXSample cos = TX_TAB(ff_tx_tab_8)[1];
 
-    fft4(z);
+    TX_NAME(ff_tx_fft4_ns)(s, z, z, stride);
 
     BF(t1, z[5].re, z[4].re, -z[5].re);
     BF(t2, z[5].im, z[4].im, -z[5].im);
@@ -536,19 +604,23 @@ static void fft8(FFTComplex *z)
     BF(t6, z[7].im, z[6].im, -z[7].im);
 
     BUTTERFLIES(z[0], z[2], z[4], z[6]);
-    TRANSFORM(z[1], z[3], z[5], z[7], RESCALE(M_SQRT1_2), RESCALE(M_SQRT1_2));
+    TRANSFORM(z[1], z[3], z[5], z[7], cos, cos);
 }
 
-static void fft16(FFTComplex *z)
+static void TX_NAME(ff_tx_fft16_ns)(AVTXContext *s, void *dst,
+                                    void *src, ptrdiff_t stride)
 {
-    FFTSample t1, t2, t3, t4, t5, t6, r0, i0, r1, i1;
-    FFTSample cos_16_1 = TX_NAME(ff_cos_16)[1];
-    FFTSample cos_16_2 = TX_NAME(ff_cos_16)[2];
-    FFTSample cos_16_3 = TX_NAME(ff_cos_16)[3];
+    TXComplex *z = dst;
+    const TXSample *cos = TX_TAB(ff_tx_tab_16);
 
-    fft8(z +  0);
-    fft4(z +  8);
-    fft4(z + 12);
+    TXSample t1, t2, t3, t4, t5, t6, r0, i0, r1, i1;
+    TXSample cos_16_1 = cos[1];
+    TXSample cos_16_2 = cos[2];
+    TXSample cos_16_3 = cos[3];
+
+    TX_NAME(ff_tx_fft8_ns)(s, z +  0, z +  0, stride);
+    TX_NAME(ff_tx_fft4_ns)(s, z +  8, z +  8, stride);
+    TX_NAME(ff_tx_fft4_ns)(s, z + 12, z + 12, stride);
 
     t1 = z[ 8].re;
     t2 = z[ 8].im;
@@ -561,47 +633,217 @@ static void fft16(FFTComplex *z)
     TRANSFORM(z[ 3], z[ 7], z[11], z[15], cos_16_3, cos_16_1);
 }
 
-DECL_FFT(32,16,8)
-DECL_FFT(64,32,16)
-DECL_FFT(128,64,32)
-DECL_FFT(256,128,64)
-DECL_FFT(512,256,128)
-DECL_FFT(1024,512,256)
-DECL_FFT(2048,1024,512)
-DECL_FFT(4096,2048,1024)
-DECL_FFT(8192,4096,2048)
-DECL_FFT(16384,8192,4096)
-DECL_FFT(32768,16384,8192)
-DECL_FFT(65536,32768,16384)
-DECL_FFT(131072,65536,32768)
-
-static void (* const fft_dispatch[])(FFTComplex*) = {
-    NULL, fft2, fft4, fft8, fft16, fft32, fft64, fft128, fft256, fft512,
-    fft1024, fft2048, fft4096, fft8192, fft16384, fft32768, fft65536, fft131072
+DECL_SR_CODELET_DEF(2)
+DECL_SR_CODELET_DEF(4)
+DECL_SR_CODELET_DEF(8)
+DECL_SR_CODELET_DEF(16)
+DECL_SR_CODELET(32,16,8)
+DECL_SR_CODELET(64,32,16)
+DECL_SR_CODELET(128,64,32)
+DECL_SR_CODELET(256,128,64)
+DECL_SR_CODELET(512,256,128)
+DECL_SR_CODELET(1024,512,256)
+DECL_SR_CODELET(2048,1024,512)
+DECL_SR_CODELET(4096,2048,1024)
+DECL_SR_CODELET(8192,4096,2048)
+DECL_SR_CODELET(16384,8192,4096)
+DECL_SR_CODELET(32768,16384,8192)
+DECL_SR_CODELET(65536,32768,16384)
+DECL_SR_CODELET(131072,65536,32768)
+
+static av_cold int TX_NAME(ff_tx_fft_sr_init)(AVTXContext *s,
+                                              const FFTXCodelet *cd,
+                                              uint64_t flags,
+                                              FFTXCodeletOptions *opts,
+                                              int len, int inv,
+                                              const void *scale)
+{
+    int ret;
+    int is_inplace = !!(flags & AV_TX_INPLACE);
+    FFTXCodeletOptions sub_opts = { .invert_lookup = !is_inplace };
+
+    flags &= ~FF_TX_OUT_OF_PLACE; /* We want the subtransform to be */
+    flags |=  AV_TX_INPLACE;      /* in-place */
+    flags |=  FF_TX_PRESHUFFLE;   /* This function handles the permute step */
+
+    if ((ret = ff_tx_init_subtx(s, TX_TYPE(FFT), flags, &sub_opts, len, inv, scale)))
+        return ret;
+
+    if (is_inplace && (ret = ff_tx_gen_ptwo_inplace_revtab_idx(s)))
+        return ret;
+
+    return 0;
+}
+
+static void TX_NAME(ff_tx_fft_sr)(AVTXContext *s, void *_dst,
+                                  void *_src, ptrdiff_t stride)
+{
+    TXComplex *src = _src;
+    TXComplex *dst = _dst;
+    int *map = s->sub[0].map;
+    int len = s->len;
+
+    /* Compilers can't vectorize this anyway without assuming AVX2, which they
+     * generally don't, at least without -march=native -mtune=native */
+    for (int i = 0; i < len; i++)
+        dst[i] = src[map[i]];
+
+    s->fn[0](&s->sub[0], dst, dst, stride);
+}
+
+static void TX_NAME(ff_tx_fft_sr_inplace)(AVTXContext *s, void *_dst,
+                                          void *_src, ptrdiff_t stride)
+{
+    TXComplex *dst = _dst;
+    TXComplex tmp;
+    const int *map = s->sub->map;
+    const int *inplace_idx = s->map;
+    int src_idx, dst_idx;
+
+    src_idx = *inplace_idx++;
+    do {
+        tmp = dst[src_idx];
+        dst_idx = map[src_idx];
+        do {
+            FFSWAP(TXComplex, tmp, dst[dst_idx]);
+            dst_idx = map[dst_idx];
+        } while (dst_idx != src_idx); /* Can be > as well, but was less predictable */
+        dst[dst_idx] = tmp;
+    } while ((src_idx = *inplace_idx++));
+
+    s->fn[0](&s->sub[0], dst, dst, stride);
+}
+
+static const FFTXCodelet TX_NAME(ff_tx_fft_sr_def) = {
+    .name       = TX_NAME_STR("fft_sr"),
+    .function   = TX_NAME(ff_tx_fft_sr),
+    .type       = TX_TYPE(FFT),
+    .flags      = AV_TX_UNALIGNED | FF_TX_OUT_OF_PLACE,
+    .factors[0] = 2,
+    .min_len    = 2,
+    .max_len    = TX_LEN_UNLIMITED,
+    .init       = TX_NAME(ff_tx_fft_sr_init),
+    .cpu_flags  = FF_TX_CPU_FLAGS_ALL,
+    .prio       = FF_TX_PRIO_BASE,
+};
+
+static const FFTXCodelet TX_NAME(ff_tx_fft_sr_inplace_def) = {
+    .name       = TX_NAME_STR("fft_sr_inplace"),
+    .function   = TX_NAME(ff_tx_fft_sr_inplace),
+    .type       = TX_TYPE(FFT),
+    .flags      = AV_TX_UNALIGNED | AV_TX_INPLACE,
+    .factors[0] = 2,
+    .min_len    = 2,
+    .max_len    = TX_LEN_UNLIMITED,
+    .init       = TX_NAME(ff_tx_fft_sr_init),
+    .cpu_flags  = FF_TX_CPU_FLAGS_ALL,
+    .prio       = FF_TX_PRIO_BASE,
+};
+
+static void TX_NAME(ff_tx_fft_naive)(AVTXContext *s, void *_dst, void *_src,
+                                     ptrdiff_t stride)
+{
+    TXComplex *src = _src;
+    TXComplex *dst = _dst;
+    const int n = s->len;
+    double phase = s->inv ? 2.0*M_PI/n : -2.0*M_PI/n;
+
+    for(int i = 0; i < n; i++) {
+        TXComplex tmp = { 0 };
+        for(int j = 0; j < n; j++) {
+            const double factor = phase*i*j;
+            const TXComplex mult = {
+                RESCALE(cos(factor)),
+                RESCALE(sin(factor)),
+            };
+            TXComplex res;
+            CMUL3(res, src[j], mult);
+            tmp.re += res.re;
+            tmp.im += res.im;
+        }
+        dst[i] = tmp;
+    }
+}
+
+static const FFTXCodelet TX_NAME(ff_tx_fft_naive_def) = {
+    .name       = TX_NAME_STR("fft_naive"),
+    .function   = TX_NAME(ff_tx_fft_naive),
+    .type       = TX_TYPE(FFT),
+    .flags      = AV_TX_UNALIGNED | FF_TX_OUT_OF_PLACE,
+    .factors[0] = TX_FACTOR_ANY,
+    .min_len    = 2,
+    .max_len    = TX_LEN_UNLIMITED,
+    .init       = NULL,
+    .cpu_flags  = FF_TX_CPU_FLAGS_ALL,
+    .prio       = FF_TX_PRIO_MIN,
 };
 
+static av_cold int TX_NAME(ff_tx_fft_pfa_init)(AVTXContext *s,
+                                               const FFTXCodelet *cd,
+                                               uint64_t flags,
+                                               FFTXCodeletOptions *opts,
+                                               int len, int inv,
+                                               const void *scale)
+{
+    int ret;
+    int sub_len = len / cd->factors[0];
+    FFTXCodeletOptions sub_opts = { .invert_lookup = 0 };
+
+    flags &= ~FF_TX_OUT_OF_PLACE; /* We want the subtransform to be */
+    flags |=  AV_TX_INPLACE;      /* in-place */
+    flags |=  FF_TX_PRESHUFFLE;   /* This function handles the permute step */
+
+    if ((ret = ff_tx_init_subtx(s, TX_TYPE(FFT), flags, &sub_opts,
+                                sub_len, inv, scale)))
+        return ret;
+
+    if ((ret = ff_tx_gen_compound_mapping(s, cd->factors[0], sub_len)))
+        return ret;
+
+    if (!(s->tmp = av_malloc(len*sizeof(*s->tmp))))
+        return AVERROR(ENOMEM);
+
+    TX_TAB(ff_tx_init_tabs)(len / sub_len);
+
+    return 0;
+}
+
 #define DECL_COMP_FFT(N)                                                       \
-static void compound_fft_##N##xM(AVTXContext *s, void *_out,                   \
-                                 void *_in, ptrdiff_t stride)                  \
+static void TX_NAME(ff_tx_fft_pfa_##N##xM)(AVTXContext *s, void *_out,         \
+                                           void *_in, ptrdiff_t stride)        \
 {                                                                              \
-    const int m = s->m, *in_map = s->pfatab, *out_map = in_map + N*m;          \
-    FFTComplex *in = _in;                                                      \
-    FFTComplex *out = _out;                                                    \
-    FFTComplex fft##N##in[N];                                                  \
-    void (*fftp)(FFTComplex *z) = fft_dispatch[av_log2(m)];                    \
+    const int m = s->sub->len;                                                 \
+    const int *in_map = s->map, *out_map = in_map + s->len;                    \
+    const int *sub_map = s->sub->map;                                          \
+    TXComplex *in = _in;                                                       \
+    TXComplex *out = _out;                                                     \
+    TXComplex fft##N##in[N];                                                   \
                                                                                \
     for (int i = 0; i < m; i++) {                                              \
         for (int j = 0; j < N; j++)                                            \
             fft##N##in[j] = in[in_map[i*N + j]];                               \
-        fft##N(s->tmp + s->revtab_c[i], fft##N##in, m);                        \
+        fft##N(s->tmp + sub_map[i], fft##N##in, m);                            \
     }                                                                          \
                                                                                \
     for (int i = 0; i < N; i++)                                                \
-        fftp(s->tmp + m*i);                                                    \
+        s->fn[0](&s->sub[0], s->tmp + m*i, s->tmp + m*i, sizeof(TXComplex));   \
                                                                                \
     for (int i = 0; i < N*m; i++)                                              \
         out[i] = s->tmp[out_map[i]];                                           \
-}
+}                                                                              \
+                                                                               \
+static const FFTXCodelet TX_NAME(ff_tx_fft_pfa_##N##xM_def) = {                \
+    .name       = TX_NAME_STR("fft_pfa_" #N "xM"),                             \
+    .function   = TX_NAME(ff_tx_fft_pfa_##N##xM),                              \
+    .type       = TX_TYPE(FFT),                                                \
+    .flags      = AV_TX_UNALIGNED | FF_TX_OUT_OF_PLACE,                        \
+    .factors    = { N, TX_FACTOR_ANY },                                        \
+    .min_len    = N*2,                                                         \
+    .max_len    = TX_LEN_UNLIMITED,                                            \
+    .init       = TX_NAME(ff_tx_fft_pfa_init),                                 \
+    .cpu_flags  = FF_TX_CPU_FLAGS_ALL,                                         \
+    .prio       = FF_TX_PRIO_BASE,                                             \
+};
 
 DECL_COMP_FFT(3)
 DECL_COMP_FFT(5)
@@ -609,71 +851,318 @@ DECL_COMP_FFT(7)
 DECL_COMP_FFT(9)
 DECL_COMP_FFT(15)
 
-static void split_radix_fft(AVTXContext *s, void *_out, void *_in,
-                            ptrdiff_t stride)
+static av_cold int TX_NAME(ff_tx_mdct_naive_init)(AVTXContext *s,
+                                                  const FFTXCodelet *cd,
+                                                  uint64_t flags,
+                                                  FFTXCodeletOptions *opts,
+                                                  int len, int inv,
+                                                  const void *scale)
 {
-    FFTComplex *in = _in;
-    FFTComplex *out = _out;
-    int m = s->m, mb = av_log2(m);
+    s->scale_d = *((SCALE_TYPE *)scale);
+    s->scale_f = s->scale_d;
+    return 0;
+}
+
+static void TX_NAME(ff_tx_mdct_naive_fwd)(AVTXContext *s, void *_dst,
+                                          void *_src, ptrdiff_t stride)
+{
+    TXSample *src = _src;
+    TXSample *dst = _dst;
+    double scale = s->scale_d;
+    int len = s->len;
+    const double phase = M_PI/(4.0*len);
 
-    if (s->flags & AV_TX_INPLACE) {
-        FFTComplex tmp;
-        int src, dst, *inplace_idx = s->inplace_idx;
+    stride /= sizeof(*dst);
+
+    for (int i = 0; i < len; i++) {
+        double sum = 0.0;
+        for (int j = 0; j < len*2; j++) {
+            int a = (2*j + 1 + len) * (2*i + 1);
+            sum += UNSCALE(src[j]) * cos(a * phase);
+        }
+        dst[i*stride] = RESCALE(sum*scale);
+    }
+}
 
-        src = *inplace_idx++;
+static void TX_NAME(ff_tx_mdct_naive_inv)(AVTXContext *s, void *_dst,
+                                          void *_src, ptrdiff_t stride)
+{
+    TXSample *src = _src;
+    TXSample *dst = _dst;
+    double scale = s->scale_d;
+    int len = s->len >> 1;
+    int len2 = len*2;
+    const double phase = M_PI/(4.0*len2);
 
-        do {
-            tmp = out[src];
-            dst = s->revtab_c[src];
-            do {
-                FFSWAP(FFTComplex, tmp, out[dst]);
-                dst = s->revtab_c[dst];
-            } while (dst != src); /* Can be > as well, but is less predictable */
-            out[dst] = tmp;
-        } while ((src = *inplace_idx++));
-    } else {
-        for (int i = 0; i < m; i++)
-            out[i] = in[s->revtab_c[i]];
+    stride /= sizeof(*src);
+
+    for (int i = 0; i < len; i++) {
+        double sum_d = 0.0;
+        double sum_u = 0.0;
+        double i_d = phase * (4*len  - 2*i - 1);
+        double i_u = phase * (3*len2 + 2*i + 1);
+        for (int j = 0; j < len2; j++) {
+            double a = (2 * j + 1);
+            double a_d = cos(a * i_d);
+            double a_u = cos(a * i_u);
+            double val = UNSCALE(src[j*stride]);
+            sum_d += a_d * val;
+            sum_u += a_u * val;
+        }
+        dst[i +   0] = RESCALE( sum_d*scale);
+        dst[i + len] = RESCALE(-sum_u*scale);
     }
+}
+
+static const FFTXCodelet TX_NAME(ff_tx_mdct_naive_fwd_def) = {
+    .name       = TX_NAME_STR("mdct_naive_fwd"),
+    .function   = TX_NAME(ff_tx_mdct_naive_fwd),
+    .type       = TX_TYPE(MDCT),
+    .flags      = AV_TX_UNALIGNED | FF_TX_OUT_OF_PLACE | FF_TX_FORWARD_ONLY,
+    .factors    = { 2, TX_FACTOR_ANY }, /* MDCTs need an even length */
+    .min_len    = 2,
+    .max_len    = TX_LEN_UNLIMITED,
+    .init       = TX_NAME(ff_tx_mdct_naive_init),
+    .cpu_flags  = FF_TX_CPU_FLAGS_ALL,
+    .prio       = FF_TX_PRIO_MIN,
+};
+
+static const FFTXCodelet TX_NAME(ff_tx_mdct_naive_inv_def) = {
+    .name       = TX_NAME_STR("mdct_naive_inv"),
+    .function   = TX_NAME(ff_tx_mdct_naive_inv),
+    .type       = TX_TYPE(MDCT),
+    .flags      = AV_TX_UNALIGNED | FF_TX_OUT_OF_PLACE | FF_TX_INVERSE_ONLY,
+    .factors    = { 2, TX_FACTOR_ANY },
+    .min_len    = 2,
+    .max_len    = TX_LEN_UNLIMITED,
+    .init       = TX_NAME(ff_tx_mdct_naive_init),
+    .cpu_flags  = FF_TX_CPU_FLAGS_ALL,
+    .prio       = FF_TX_PRIO_MIN,
+};
+
+static av_cold int TX_NAME(ff_tx_mdct_sr_init)(AVTXContext *s,
+                                               const FFTXCodelet *cd,
+                                               uint64_t flags,
+                                               FFTXCodeletOptions *opts,
+                                               int len, int inv,
+                                               const void *scale)
+{
+    int ret;
+    FFTXCodeletOptions sub_opts = { .invert_lookup = 0 };
+
+    s->scale_d = *((SCALE_TYPE *)scale);
+    s->scale_f = s->scale_d;
 
-    fft_dispatch[mb](out);
+    flags &= ~FF_TX_OUT_OF_PLACE; /* We want the subtransform to be */
+    flags |=  AV_TX_INPLACE;      /* in-place */
+    flags |=  FF_TX_PRESHUFFLE;   /* This function handles the permute step */
+
+    if ((ret = ff_tx_init_subtx(s, TX_TYPE(FFT), flags, &sub_opts, len >> 1,
+                                inv, scale)))
+        return ret;
+
+    if ((ret = TX_TAB(ff_tx_mdct_gen_exp)(s)))
+        return ret;
+
+    return 0;
 }
 
-static void naive_fft(AVTXContext *s, void *_out, void *_in,
-                      ptrdiff_t stride)
+static void TX_NAME(ff_tx_mdct_sr_fwd)(AVTXContext *s, void *_dst, void *_src,
+                                       ptrdiff_t stride)
 {
-    FFTComplex *in = _in;
-    FFTComplex *out = _out;
-    const int n = s->n;
-    double phase = s->inv ? 2.0*M_PI/n : -2.0*M_PI/n;
+    TXSample *src = _src, *dst = _dst;
+    TXComplex *exp = s->exp, tmp, *z = _dst;
+    const int len2 = s->len >> 1;
+    const int len4 = s->len >> 2;
+    const int len3 = len2 * 3;
+    const int *sub_map = s->sub->map;
 
-    for(int i = 0; i < n; i++) {
-        FFTComplex tmp = { 0 };
-        for(int j = 0; j < n; j++) {
-            const double factor = phase*i*j;
-            const FFTComplex mult = {
-                RESCALE(cos(factor)),
-                RESCALE(sin(factor)),
-            };
-            FFTComplex res;
-            CMUL3(res, in[j], mult);
-            tmp.re += res.re;
-            tmp.im += res.im;
+    stride /= sizeof(*dst);
+
+    for (int i = 0; i < len2; i++) { /* Folding and pre-reindexing */
+        const int k = 2*i;
+        const int idx = sub_map[i];
+        if (k < len2) {
+            tmp.re = FOLD(-src[ len2 + k],  src[1*len2 - 1 - k]);
+            tmp.im = FOLD(-src[ len3 + k], -src[1*len3 - 1 - k]);
+        } else {
+            tmp.re = FOLD(-src[ len2 + k], -src[5*len2 - 1 - k]);
+            tmp.im = FOLD( src[-len2 + k], -src[1*len3 - 1 - k]);
         }
-        out[i] = tmp;
+        CMUL(z[idx].im, z[idx].re, tmp.re, tmp.im, exp[i].re, exp[i].im);
+    }
+
+    s->fn[0](&s->sub[0], z, z, sizeof(TXComplex));
+
+    for (int i = 0; i < len4; i++) {
+        const int i0 = len4 + i, i1 = len4 - i - 1;
+        TXComplex src1 = { z[i1].re, z[i1].im };
+        TXComplex src0 = { z[i0].re, z[i0].im };
+
+        CMUL(dst[2*i1*stride + stride], dst[2*i0*stride], src0.re, src0.im,
+             exp[i0].im, exp[i0].re);
+        CMUL(dst[2*i0*stride + stride], dst[2*i1*stride], src1.re, src1.im,
+             exp[i1].im, exp[i1].re);
     }
 }
 
+static void TX_NAME(ff_tx_mdct_sr_inv)(AVTXContext *s, void *_dst, void *_src,
+                                       ptrdiff_t stride)
+{
+    TXComplex *z = _dst, *exp = s->exp;
+    const TXSample *src = _src, *in1, *in2;
+    const int len2 = s->len >> 1;
+    const int len4 = s->len >> 2;
+    const int *sub_map = s->sub->map;
+
+    stride /= sizeof(*src);
+    in1 = src;
+    in2 = src + ((len2*2) - 1) * stride;
+
+    for (int i = 0; i < len2; i++) {
+        TXComplex tmp = { in2[-2*i*stride], in1[2*i*stride] };
+        CMUL3(z[sub_map[i]], tmp, exp[i]);
+    }
+
+    s->fn[0](&s->sub[0], z, z, sizeof(TXComplex));
+
+    for (int i = 0; i < len4; i++) {
+        const int i0 = len4 + i, i1 = len4 - i - 1;
+        TXComplex src1 = { z[i1].im, z[i1].re };
+        TXComplex src0 = { z[i0].im, z[i0].re };
+
+        CMUL(z[i1].re, z[i0].im, src1.re, src1.im, exp[i1].im, exp[i1].re);
+        CMUL(z[i0].re, z[i1].im, src0.re, src0.im, exp[i0].im, exp[i0].re);
+    }
+}
+
+static const FFTXCodelet TX_NAME(ff_tx_mdct_sr_fwd_def) = {
+    .name       = TX_NAME_STR("mdct_sr_fwd"),
+    .function   = TX_NAME(ff_tx_mdct_sr_fwd),
+    .type       = TX_TYPE(MDCT),
+    .flags      = AV_TX_UNALIGNED | FF_TX_OUT_OF_PLACE | FF_TX_FORWARD_ONLY,
+    .factors[0] = 2,
+    .min_len    = 2,
+    .max_len    = TX_LEN_UNLIMITED,
+    .init       = TX_NAME(ff_tx_mdct_sr_init),
+    .cpu_flags  = FF_TX_CPU_FLAGS_ALL,
+    .prio       = FF_TX_PRIO_BASE,
+};
+
+static const FFTXCodelet TX_NAME(ff_tx_mdct_sr_inv_def) = {
+    .name       = TX_NAME_STR("mdct_sr_inv"),
+    .function   = TX_NAME(ff_tx_mdct_sr_inv),
+    .type       = TX_TYPE(MDCT),
+    .flags      = AV_TX_UNALIGNED | FF_TX_OUT_OF_PLACE | FF_TX_INVERSE_ONLY,
+    .factors[0] = 2,
+    .min_len    = 2,
+    .max_len    = TX_LEN_UNLIMITED,
+    .init       = TX_NAME(ff_tx_mdct_sr_init),
+    .cpu_flags  = FF_TX_CPU_FLAGS_ALL,
+    .prio       = FF_TX_PRIO_BASE,
+};
+
+static av_cold int TX_NAME(ff_tx_mdct_inv_full_init)(AVTXContext *s,
+                                                     const FFTXCodelet *cd,
+                                                     uint64_t flags,
+                                                     FFTXCodeletOptions *opts,
+                                                     int len, int inv,
+                                                     const void *scale)
+{
+    int ret;
+
+    s->scale_d = *((SCALE_TYPE *)scale);
+    s->scale_f = s->scale_d;
+
+    flags &= ~AV_TX_FULL_IMDCT;
+
+    if ((ret = ff_tx_init_subtx(s, TX_TYPE(MDCT), flags, NULL, len, 1, scale)))
+        return ret;
+
+    return 0;
+}
+
+static void TX_NAME(ff_tx_mdct_inv_full)(AVTXContext *s, void *_dst,
+                                         void *_src, ptrdiff_t stride)
+{
+    int len  = s->len << 1;
+    int len2 = len >> 1;
+    int len4 = len >> 2;
+    TXSample *dst = _dst;
+
+    s->fn[0](&s->sub[0], dst + len4, _src, stride);
+
+    stride /= sizeof(*dst);
+
+    for (int i = 0; i < len4; i++) {
+        dst[            i*stride] = -dst[(len2 - i - 1)*stride];
+        dst[(len - i - 1)*stride] =  dst[(len2 + i + 0)*stride];
+    }
+}
+
+static const FFTXCodelet TX_NAME(ff_tx_mdct_inv_full_def) = {
+    .name       = TX_NAME_STR("mdct_inv_full"),
+    .function   = TX_NAME(ff_tx_mdct_inv_full),
+    .type       = TX_TYPE(MDCT),
+    .flags      = AV_TX_UNALIGNED | AV_TX_INPLACE |
+                  FF_TX_OUT_OF_PLACE | AV_TX_FULL_IMDCT,
+    .factors    = { 2, TX_FACTOR_ANY },
+    .min_len    = 2,
+    .max_len    = TX_LEN_UNLIMITED,
+    .init       = TX_NAME(ff_tx_mdct_inv_full_init),
+    .cpu_flags  = FF_TX_CPU_FLAGS_ALL,
+    .prio       = FF_TX_PRIO_BASE,
+};
+
+static av_cold int TX_NAME(ff_tx_mdct_pfa_init)(AVTXContext *s,
+                                                const FFTXCodelet *cd,
+                                                uint64_t flags,
+                                                FFTXCodeletOptions *opts,
+                                                int len, int inv,
+                                                const void *scale)
+{
+    int ret, sub_len;
+    FFTXCodeletOptions sub_opts = { .invert_lookup = 0 };
+
+    len >>= 1;
+    sub_len = len / cd->factors[0];
+
+    s->scale_d = *((SCALE_TYPE *)scale);
+    s->scale_f = s->scale_d;
+
+    flags &= ~FF_TX_OUT_OF_PLACE; /* We want the subtransform to be */
+    flags |=  AV_TX_INPLACE;      /* in-place */
+    flags |=  FF_TX_PRESHUFFLE;   /* This function handles the permute step */
+
+    if ((ret = ff_tx_init_subtx(s, TX_TYPE(FFT), flags, &sub_opts,
+                                sub_len, inv, scale)))
+        return ret;
+
+    if ((ret = ff_tx_gen_compound_mapping(s, cd->factors[0], sub_len)))
+        return ret;
+
+    if ((ret = TX_TAB(ff_tx_mdct_gen_exp)(s)))
+        return ret;
+
+    if (!(s->tmp = av_malloc(len*sizeof(*s->tmp))))
+        return AVERROR(ENOMEM);
+
+    TX_TAB(ff_tx_init_tabs)(len / sub_len);
+
+    return 0;
+}
+
 #define DECL_COMP_IMDCT(N)                                                     \
-static void compound_imdct_##N##xM(AVTXContext *s, void *_dst, void *_src,     \
-                                   ptrdiff_t stride)                           \
+static void TX_NAME(ff_tx_mdct_pfa_##N##xM_inv)(AVTXContext *s, void *_dst,    \
+                                                void *_src, ptrdiff_t stride)  \
 {                                                                              \
-    FFTComplex fft##N##in[N];                                                  \
-    FFTComplex *z = _dst, *exp = s->exptab;                                    \
-    const int m = s->m, len8 = N*m >> 1;                                       \
-    const int *in_map = s->pfatab, *out_map = in_map + N*m;                    \
-    const FFTSample *src = _src, *in1, *in2;                                   \
-    void (*fftp)(FFTComplex *) = fft_dispatch[av_log2(m)];                     \
+    TXComplex fft##N##in[N];                                                   \
+    TXComplex *z = _dst, *exp = s->exp;                                        \
+    const TXSample *src = _src, *in1, *in2;                                    \
+    const int len4 = s->len >> 2;                                              \
+    const int m = s->sub->len;                                                 \
+    const int *in_map = s->map, *out_map = in_map + N*m;                       \
+    const int *sub_map = s->sub->map;                                          \
                                                                                \
     stride /= sizeof(*src); /* To convert it from bytes */                     \
     in1 = src;                                                                 \
@@ -682,25 +1171,38 @@ static void compound_imdct_##N##xM(AVTXContext *s, void *_dst, void *_src,     \
     for (int i = 0; i < m; i++) {                                              \
         for (int j = 0; j < N; j++) {                                          \
             const int k = in_map[i*N + j];                                     \
-            FFTComplex tmp = { in2[-k*stride], in1[k*stride] };                \
+            TXComplex tmp = { in2[-k*stride], in1[k*stride] };                 \
             CMUL3(fft##N##in[j], tmp, exp[k >> 1]);                            \
         }                                                                      \
-        fft##N(s->tmp + s->revtab_c[i], fft##N##in, m);                        \
+        fft##N(s->tmp + sub_map[i], fft##N##in, m);                            \
     }                                                                          \
                                                                                \
     for (int i = 0; i < N; i++)                                                \
-        fftp(s->tmp + m*i);                                                    \
+        s->fn[0](&s->sub[0], s->tmp + m*i, s->tmp + m*i, sizeof(TXComplex));   \
                                                                                \
-    for (int i = 0; i < len8; i++) {                                           \
-        const int i0 = len8 + i, i1 = len8 - i - 1;                            \
+    for (int i = 0; i < len4; i++) {                                           \
+        const int i0 = len4 + i, i1 = len4 - i - 1;                            \
         const int s0 = out_map[i0], s1 = out_map[i1];                          \
-        FFTComplex src1 = { s->tmp[s1].im, s->tmp[s1].re };                    \
-        FFTComplex src0 = { s->tmp[s0].im, s->tmp[s0].re };                    \
+        TXComplex src1 = { s->tmp[s1].im, s->tmp[s1].re };                     \
+        TXComplex src0 = { s->tmp[s0].im, s->tmp[s0].re };                     \
                                                                                \
         CMUL(z[i1].re, z[i0].im, src1.re, src1.im, exp[i1].im, exp[i1].re);    \
         CMUL(z[i0].re, z[i1].im, src0.re, src0.im, exp[i0].im, exp[i0].re);    \
     }                                                                          \
-}
+}                                                                              \
+                                                                               \
+static const FFTXCodelet TX_NAME(ff_tx_mdct_pfa_##N##xM_inv_def) = {           \
+    .name       = TX_NAME_STR("mdct_pfa_" #N "xM_inv"),                        \
+    .function   = TX_NAME(ff_tx_mdct_pfa_##N##xM_inv),                         \
+    .type       = TX_TYPE(MDCT),                                               \
+    .flags      = AV_TX_UNALIGNED | FF_TX_OUT_OF_PLACE | FF_TX_INVERSE_ONLY,   \
+    .factors    = { N, TX_FACTOR_ANY },                                        \
+    .min_len    = N*2,                                                         \
+    .max_len    = TX_LEN_UNLIMITED,                                            \
+    .init       = TX_NAME(ff_tx_mdct_pfa_init),                                \
+    .cpu_flags  = FF_TX_CPU_FLAGS_ALL,                                         \
+    .prio       = FF_TX_PRIO_BASE,                                             \
+};
 
 DECL_COMP_IMDCT(3)
 DECL_COMP_IMDCT(5)
@@ -709,14 +1211,18 @@ DECL_COMP_IMDCT(9)
 DECL_COMP_IMDCT(15)
 
 #define DECL_COMP_MDCT(N)                                                      \
-static void compound_mdct_##N##xM(AVTXContext *s, void *_dst, void *_src,      \
-                                  ptrdiff_t stride)                            \
+static void TX_NAME(ff_tx_mdct_pfa_##N##xM_fwd)(AVTXContext *s, void *_dst,    \
+                                                void *_src, ptrdiff_t stride)  \
 {                                                                              \
-    FFTSample *src = _src, *dst = _dst;                                        \
-    FFTComplex *exp = s->exptab, tmp, fft##N##in[N];                           \
-    const int m = s->m, len4 = N*m, len3 = len4 * 3, len8 = len4 >> 1;         \
-    const int *in_map = s->pfatab, *out_map = in_map + N*m;                    \
-    void (*fftp)(FFTComplex *) = fft_dispatch[av_log2(m)];                     \
+    TXComplex fft##N##in[N];                                                   \
+    TXSample *src = _src, *dst = _dst;                                         \
+    TXComplex *exp = s->exp, tmp;                                              \
+    const int m = s->sub->len;                                                 \
+    const int len4 = N*m;                                                      \
+    const int len3 = len4 * 3;                                                 \
+    const int len8 = s->len >> 2;                                              \
+    const int *in_map = s->map, *out_map = in_map + N*m;                       \
+    const int *sub_map = s->sub->map;                                          \
                                                                                \
     stride /= sizeof(*dst);                                                    \
                                                                                \
@@ -733,24 +1239,37 @@ static void compound_mdct_##N##xM(AVTXContext *s, void *_dst, void *_src,      \
             CMUL(fft##N##in[j].im, fft##N##in[j].re, tmp.re, tmp.im,           \
                  exp[k >> 1].re, exp[k >> 1].im);                              \
         }                                                                      \
-        fft##N(s->tmp + s->revtab_c[i], fft##N##in, m);                        \
+        fft##N(s->tmp + sub_map[i], fft##N##in, m);                            \
     }                                                                          \
                                                                                \
     for (int i = 0; i < N; i++)                                                \
-        fftp(s->tmp + m*i);                                                    \
+        s->fn[0](&s->sub[0], s->tmp + m*i, s->tmp + m*i, sizeof(TXComplex));   \
                                                                                \
     for (int i = 0; i < len8; i++) {                                           \
         const int i0 = len8 + i, i1 = len8 - i - 1;                            \
         const int s0 = out_map[i0], s1 = out_map[i1];                          \
-        FFTComplex src1 = { s->tmp[s1].re, s->tmp[s1].im };                    \
-        FFTComplex src0 = { s->tmp[s0].re, s->tmp[s0].im };                    \
+        TXComplex src1 = { s->tmp[s1].re, s->tmp[s1].im };                     \
+        TXComplex src0 = { s->tmp[s0].re, s->tmp[s0].im };                     \
                                                                                \
         CMUL(dst[2*i1*stride + stride], dst[2*i0*stride], src0.re, src0.im,    \
              exp[i0].im, exp[i0].re);                                          \
         CMUL(dst[2*i0*stride + stride], dst[2*i1*stride], src1.re, src1.im,    \
              exp[i1].im, exp[i1].re);                                          \
     }                                                                          \
-}
+}                                                                              \
+                                                                               \
+static const FFTXCodelet TX_NAME(ff_tx_mdct_pfa_##N##xM_fwd_def) = {           \
+    .name       = TX_NAME_STR("mdct_pfa_" #N "xM_fwd"),                        \
+    .function   = TX_NAME(ff_tx_mdct_pfa_##N##xM_fwd),                         \
+    .type       = TX_TYPE(MDCT),                                               \
+    .flags      = AV_TX_UNALIGNED | FF_TX_OUT_OF_PLACE | FF_TX_FORWARD_ONLY,   \
+    .factors    = { N, TX_FACTOR_ANY },                                        \
+    .min_len    = N*2,                                                         \
+    .max_len    = TX_LEN_UNLIMITED,                                            \
+    .init       = TX_NAME(ff_tx_mdct_pfa_init),                                \
+    .cpu_flags  = FF_TX_CPU_FLAGS_ALL,                                         \
+    .prio       = FF_TX_PRIO_BASE,                                             \
+};
 
 DECL_COMP_MDCT(3)
 DECL_COMP_MDCT(5)
@@ -758,267 +1277,69 @@ DECL_COMP_MDCT(7)
 DECL_COMP_MDCT(9)
 DECL_COMP_MDCT(15)
 
-static void monolithic_imdct(AVTXContext *s, void *_dst, void *_src,
-                             ptrdiff_t stride)
-{
-    FFTComplex *z = _dst, *exp = s->exptab;
-    const int m = s->m, len8 = m >> 1;
-    const FFTSample *src = _src, *in1, *in2;
-    void (*fftp)(FFTComplex *) = fft_dispatch[av_log2(m)];
-
-    stride /= sizeof(*src);
-    in1 = src;
-    in2 = src + ((m*2) - 1) * stride;
-
-    for (int i = 0; i < m; i++) {
-        FFTComplex tmp = { in2[-2*i*stride], in1[2*i*stride] };
-        CMUL3(z[s->revtab_c[i]], tmp, exp[i]);
-    }
-
-    fftp(z);
-
-    for (int i = 0; i < len8; i++) {
-        const int i0 = len8 + i, i1 = len8 - i - 1;
-        FFTComplex src1 = { z[i1].im, z[i1].re };
-        FFTComplex src0 = { z[i0].im, z[i0].re };
-
-        CMUL(z[i1].re, z[i0].im, src1.re, src1.im, exp[i1].im, exp[i1].re);
-        CMUL(z[i0].re, z[i1].im, src0.re, src0.im, exp[i0].im, exp[i0].re);
-    }
-}
-
-static void monolithic_mdct(AVTXContext *s, void *_dst, void *_src,
-                            ptrdiff_t stride)
-{
-    FFTSample *src = _src, *dst = _dst;
-    FFTComplex *exp = s->exptab, tmp, *z = _dst;
-    const int m = s->m, len4 = m, len3 = len4 * 3, len8 = len4 >> 1;
-    void (*fftp)(FFTComplex *) = fft_dispatch[av_log2(m)];
-
-    stride /= sizeof(*dst);
-
-    for (int i = 0; i < m; i++) { /* Folding and pre-reindexing */
-        const int k = 2*i;
-        if (k < len4) {
-            tmp.re = FOLD(-src[ len4 + k],  src[1*len4 - 1 - k]);
-            tmp.im = FOLD(-src[ len3 + k], -src[1*len3 - 1 - k]);
-        } else {
-            tmp.re = FOLD(-src[ len4 + k], -src[5*len4 - 1 - k]);
-            tmp.im = FOLD( src[-len4 + k], -src[1*len3 - 1 - k]);
-        }
-        CMUL(z[s->revtab_c[i]].im, z[s->revtab_c[i]].re, tmp.re, tmp.im,
-             exp[i].re, exp[i].im);
-    }
-
-    fftp(z);
-
-    for (int i = 0; i < len8; i++) {
-        const int i0 = len8 + i, i1 = len8 - i - 1;
-        FFTComplex src1 = { z[i1].re, z[i1].im };
-        FFTComplex src0 = { z[i0].re, z[i0].im };
-
-        CMUL(dst[2*i1*stride + stride], dst[2*i0*stride], src0.re, src0.im,
-             exp[i0].im, exp[i0].re);
-        CMUL(dst[2*i0*stride + stride], dst[2*i1*stride], src1.re, src1.im,
-             exp[i1].im, exp[i1].re);
-    }
-}
-
-static void naive_imdct(AVTXContext *s, void *_dst, void *_src,
-                        ptrdiff_t stride)
-{
-    int len = s->n;
-    int len2 = len*2;
-    FFTSample *src = _src;
-    FFTSample *dst = _dst;
-    double scale = s->scale;
-    const double phase = M_PI/(4.0*len2);
-
-    stride /= sizeof(*src);
-
-    for (int i = 0; i < len; i++) {
-        double sum_d = 0.0;
-        double sum_u = 0.0;
-        double i_d = phase * (4*len  - 2*i - 1);
-        double i_u = phase * (3*len2 + 2*i + 1);
-        for (int j = 0; j < len2; j++) {
-            double a = (2 * j + 1);
-            double a_d = cos(a * i_d);
-            double a_u = cos(a * i_u);
-            double val = UNSCALE(src[j*stride]);
-            sum_d += a_d * val;
-            sum_u += a_u * val;
-        }
-        dst[i +   0] = RESCALE( sum_d*scale);
-        dst[i + len] = RESCALE(-sum_u*scale);
-    }
-}
-
-static void naive_mdct(AVTXContext *s, void *_dst, void *_src,
-                       ptrdiff_t stride)
-{
-    int len = s->n*2;
-    FFTSample *src = _src;
-    FFTSample *dst = _dst;
-    double scale = s->scale;
-    const double phase = M_PI/(4.0*len);
-
-    stride /= sizeof(*dst);
-
-    for (int i = 0; i < len; i++) {
-        double sum = 0.0;
-        for (int j = 0; j < len*2; j++) {
-            int a = (2*j + 1 + len) * (2*i + 1);
-            sum += UNSCALE(src[j]) * cos(a * phase);
-        }
-        dst[i*stride] = RESCALE(sum*scale);
-    }
-}
-
-static void full_imdct_wrapper_fn(AVTXContext *s, void *_dst, void *_src,
-                                  ptrdiff_t stride)
-{
-    int len = s->m*s->n*4;
-    int len2 = len >> 1;
-    int len4 = len >> 2;
-    FFTSample *dst = _dst;
-
-    s->top_tx(s, dst + len4, _src, stride);
-
-    stride /= sizeof(*dst);
-
-    for (int i = 0; i < len4; i++) {
-        dst[            i*stride] = -dst[(len2 - i - 1)*stride];
-        dst[(len - i - 1)*stride] =  dst[(len2 + i + 0)*stride];
-    }
-}
-
-static int gen_mdct_exptab(AVTXContext *s, int len4, double scale)
+int TX_TAB(ff_tx_mdct_gen_exp)(AVTXContext *s)
 {
+    int len4 = s->len >> 1;
+    double scale = s->scale_d;
     const double theta = (scale < 0 ? len4 : 0) + 1.0/8.0;
 
-    if (!(s->exptab = av_malloc_array(len4, sizeof(*s->exptab))))
+    if (!(s->exp = av_malloc_array(len4, sizeof(*s->exp))))
         return AVERROR(ENOMEM);
 
     scale = sqrt(fabs(scale));
     for (int i = 0; i < len4; i++) {
         const double alpha = M_PI_2 * (i + theta) / len4;
-        s->exptab[i].re = RESCALE(cos(alpha) * scale);
-        s->exptab[i].im = RESCALE(sin(alpha) * scale);
+        s->exp[i].re = RESCALE(cos(alpha) * scale);
+        s->exp[i].im = RESCALE(sin(alpha) * scale);
     }
 
     return 0;
 }
 
-int TX_NAME(ff_tx_init_mdct_fft)(AVTXContext *s, av_tx_fn *tx,
-                                 enum AVTXType type, int inv, int len,
-                                 const void *scale, uint64_t flags)
-{
-    const int is_mdct = ff_tx_type_is_mdct(type);
-    int err, l, n = 1, m = 1, max_ptwo = 1 << (FF_ARRAY_ELEMS(fft_dispatch) - 1);
-
-    if (is_mdct)
-        len >>= 1;
-
-    l = len;
+const FFTXCodelet * const TX_NAME(ff_tx_codelet_list)[] = {
+    /* Split-Radix codelets */
+    &TX_NAME(ff_tx_fft2_ns_def),
+    &TX_NAME(ff_tx_fft4_ns_def),
+    &TX_NAME(ff_tx_fft8_ns_def),
+    &TX_NAME(ff_tx_fft16_ns_def),
+    &TX_NAME(ff_tx_fft32_ns_def),
+    &TX_NAME(ff_tx_fft64_ns_def),
+    &TX_NAME(ff_tx_fft128_ns_def),
+    &TX_NAME(ff_tx_fft256_ns_def),
+    &TX_NAME(ff_tx_fft512_ns_def),
+    &TX_NAME(ff_tx_fft1024_ns_def),
+    &TX_NAME(ff_tx_fft2048_ns_def),
+    &TX_NAME(ff_tx_fft4096_ns_def),
+    &TX_NAME(ff_tx_fft8192_ns_def),
+    &TX_NAME(ff_tx_fft16384_ns_def),
+    &TX_NAME(ff_tx_fft32768_ns_def),
+    &TX_NAME(ff_tx_fft65536_ns_def),
+    &TX_NAME(ff_tx_fft131072_ns_def),
+
+    /* Standalone transforms */
+    &TX_NAME(ff_tx_fft_sr_def),
+    &TX_NAME(ff_tx_fft_sr_inplace_def),
+    &TX_NAME(ff_tx_fft_pfa_3xM_def),
+    &TX_NAME(ff_tx_fft_pfa_5xM_def),
+    &TX_NAME(ff_tx_fft_pfa_7xM_def),
+    &TX_NAME(ff_tx_fft_pfa_9xM_def),
+    &TX_NAME(ff_tx_fft_pfa_15xM_def),
+    &TX_NAME(ff_tx_fft_naive_def),
+    &TX_NAME(ff_tx_mdct_sr_fwd_def),
+    &TX_NAME(ff_tx_mdct_sr_inv_def),
+    &TX_NAME(ff_tx_mdct_pfa_3xM_fwd_def),
+    &TX_NAME(ff_tx_mdct_pfa_5xM_fwd_def),
+    &TX_NAME(ff_tx_mdct_pfa_7xM_fwd_def),
+    &TX_NAME(ff_tx_mdct_pfa_9xM_fwd_def),
+    &TX_NAME(ff_tx_mdct_pfa_15xM_fwd_def),
+    &TX_NAME(ff_tx_mdct_pfa_3xM_inv_def),
+    &TX_NAME(ff_tx_mdct_pfa_5xM_inv_def),
+    &TX_NAME(ff_tx_mdct_pfa_7xM_inv_def),
+    &TX_NAME(ff_tx_mdct_pfa_9xM_inv_def),
+    &TX_NAME(ff_tx_mdct_pfa_15xM_inv_def),
+    &TX_NAME(ff_tx_mdct_naive_fwd_def),
+    &TX_NAME(ff_tx_mdct_naive_inv_def),
+    &TX_NAME(ff_tx_mdct_inv_full_def),
 
-#define CHECK_FACTOR(DST, FACTOR, SRC)                                         \
-    if (DST == 1 && !(SRC % FACTOR)) {                                         \
-        DST = FACTOR;                                                          \
-        SRC /= FACTOR;                                                         \
-    }
-    CHECK_FACTOR(n, 15, len)
-    CHECK_FACTOR(n,  9, len)
-    CHECK_FACTOR(n,  7, len)
-    CHECK_FACTOR(n,  5, len)
-    CHECK_FACTOR(n,  3, len)
-#undef CHECK_FACTOR
-
-    /* len must be a power of two now */
-    if (!(len & (len - 1)) && len >= 2 && len <= max_ptwo) {
-        m = len;
-        len = 1;
-    }
-
-    s->n = n;
-    s->m = m;
-    s->inv = inv;
-    s->type = type;
-    s->flags = flags;
-
-    /* If we weren't able to split the length into factors we can handle,
-     * resort to using the naive and slow FT. This also filters out
-     * direct 3, 5 and 15 transforms as they're too niche. */
-    if (len > 1 || m == 1) {
-        if (is_mdct && (l & 1)) /* Odd (i)MDCTs are not supported yet */
-            return AVERROR(ENOSYS);
-        if (flags & AV_TX_INPLACE) /* Neither are in-place naive transforms */
-            return AVERROR(ENOSYS);
-        s->n = l;
-        s->m = 1;
-        *tx = naive_fft;
-        if (is_mdct) {
-            s->scale = *((SCALE_TYPE *)scale);
-            *tx = inv ? naive_imdct : naive_mdct;
-            if (inv && (flags & AV_TX_FULL_IMDCT)) {
-                s->top_tx = *tx;
-                *tx = full_imdct_wrapper_fn;
-            }
-        }
-        return 0;
-    }
-
-    if (n > 1 && m > 1) { /* 2D transform case */
-        if ((err = ff_tx_gen_compound_mapping(s)))
-            return err;
-        if (!(s->tmp = av_malloc(n*m*sizeof(*s->tmp))))
-            return AVERROR(ENOMEM);
-        if (!(m & (m - 1))) {
-            *tx = n == 3 ? compound_fft_3xM :
-                  n == 5 ? compound_fft_5xM :
-                  n == 7 ? compound_fft_7xM :
-                  n == 9 ? compound_fft_9xM :
-                           compound_fft_15xM;
-            if (is_mdct)
-                *tx = n == 3 ? inv ? compound_imdct_3xM  : compound_mdct_3xM :
-                      n == 5 ? inv ? compound_imdct_5xM  : compound_mdct_5xM :
-                      n == 7 ? inv ? compound_imdct_7xM  : compound_mdct_7xM :
-                      n == 9 ? inv ? compound_imdct_9xM  : compound_mdct_9xM :
-                               inv ? compound_imdct_15xM : compound_mdct_15xM;
-        }
-    } else { /* Direct transform case */
-        *tx = split_radix_fft;
-        if (is_mdct)
-            *tx = inv ? monolithic_imdct : monolithic_mdct;
-    }
-
-    if (n == 3 || n == 5 || n == 15)
-        init_cos_tabs(0);
-    else if (n == 7)
-        init_cos_tabs(1);
-    else if (n == 9)
-        init_cos_tabs(2);
-
-    if (m != 1 && !(m & (m - 1))) {
-        if ((err = ff_tx_gen_ptwo_revtab(s, n == 1 && !is_mdct && !(flags & AV_TX_INPLACE))))
-            return err;
-        if (flags & AV_TX_INPLACE) {
-            if (is_mdct) /* In-place MDCTs are not supported yet */
-                return AVERROR(ENOSYS);
-            if ((err = ff_tx_gen_ptwo_inplace_revtab_idx(s, s->revtab_c)))
-                return err;
-        }
-        for (int i = 4; i <= av_log2(m); i++)
-            init_cos_tabs(i);
-    }
-
-    if (is_mdct) {
-        if (inv && (flags & AV_TX_FULL_IMDCT)) {
-            s->top_tx = *tx;
-            *tx = full_imdct_wrapper_fn;
-        }
-        return gen_mdct_exptab(s, n*m, *((SCALE_TYPE *)scale));
-    }
-
-    return 0;
-}
+    NULL,
+};
diff --git a/libavutil/x86/tx_float.asm b/libavutil/x86/tx_float.asm
index 4d2283fae1..e3b48d7c1f 100644
--- a/libavutil/x86/tx_float.asm
+++ b/libavutil/x86/tx_float.asm
@@ -31,6 +31,8 @@
 
 %include "x86util.asm"
 
+%define private_prefix ff_tx
+
 %if ARCH_X86_64
 %define ptr resq
 %else
@@ -39,25 +41,22 @@
 
 %assign i 16
 %rep 14
-cextern cos_ %+ i %+ _float ; ff_cos_i_float...
+cextern tab_ %+ i %+ _float ; ff_tab_i_float...
 %assign i (i << 1)
 %endrep
 
 struc AVTXContext
-    .n:           resd 1 ; Non-power-of-two part
-    .m:           resd 1 ; Power-of-two part
-    .inv:         resd 1 ; Is inverse
-    .type:        resd 1 ; Type
-    .flags:       resq 1 ; Flags
-    .scale:       resq 1 ; Scale
-
-    .exptab:       ptr 1 ; MDCT exptab
-    .tmp:          ptr 1 ; Temporary buffer needed for all compound transforms
-    .pfatab:       ptr 1 ; Input/Output mapping for compound transforms
-    .revtab:       ptr 1 ; Input mapping for power of two transforms
-    .inplace_idx:  ptr 1 ; Required indices to revtab for in-place transforms
-
-    .top_tx        ptr 1 ;  Used for transforms derived from other transforms
+    .len:          resd 1 ; Length
+    .inv           resd 1 ; Inverse flag
+    .map:           ptr 1 ; Lookup table(s)
+    .exp:           ptr 1 ; Exponentiation factors
+    .tmp:           ptr 1 ; Temporary data
+
+    .sub:           ptr 1 ; Subcontexts
+    .fn:            ptr 4 ; Subcontext functions
+    .nb_sub:       resd 1 ; Subcontext count
+
+    ; Everything else is inaccessible
 endstruc
 
 SECTION_RODATA 32
@@ -485,8 +484,8 @@ SECTION .text
     movaps [outq + 10*mmsize], tx1_o0
     movaps [outq + 14*mmsize], tx2_o0
 
-    movaps tw_e,           [cos_64_float + mmsize]
-    vperm2f128 tw_o, tw_o, [cos_64_float + 64 - 4*7 - mmsize], 0x23
+    movaps tw_e,           [tab_64_float + mmsize]
+    vperm2f128 tw_o, tw_o, [tab_64_float + 64 - 4*7 - mmsize], 0x23
 
     movaps m0, [outq +  1*mmsize]
     movaps m1, [outq +  3*mmsize]
@@ -710,8 +709,7 @@ FFT4 inv, 1
 
 INIT_XMM sse3
 cglobal fft8_float, 4, 4, 6, ctx, out, in, tmp
-    mov ctxq, [ctxq + AVTXContext.revtab]
-
+    mov ctxq, [ctxq + AVTXContext.map]
     LOAD64_LUT m0, inq, ctxq, (mmsize/2)*0, tmpq
     LOAD64_LUT m1, inq, ctxq, (mmsize/2)*1, tmpq
     LOAD64_LUT m2, inq, ctxq, (mmsize/2)*2, tmpq
@@ -733,8 +731,7 @@ cglobal fft8_float, 4, 4, 6, ctx, out, in, tmp
 
 INIT_YMM avx
 cglobal fft8_float, 4, 4, 4, ctx, out, in, tmp
-    mov ctxq, [ctxq + AVTXContext.revtab]
-
+    mov ctxq, [ctxq + AVTXContext.map]
     LOAD64_LUT m0, inq, ctxq, (mmsize/2)*0, tmpq, m2
     LOAD64_LUT m1, inq, ctxq, (mmsize/2)*1, tmpq, m3
 
@@ -754,7 +751,7 @@ cglobal fft8_float, 4, 4, 4, ctx, out, in, tmp
 %macro FFT16_FN 1
 INIT_YMM %1
 cglobal fft16_float, 4, 4, 8, ctx, out, in, tmp
-    mov ctxq, [ctxq + AVTXContext.revtab]
+    mov ctxq, [ctxq + AVTXContext.map]
 
     LOAD64_LUT m0, inq, ctxq, (mmsize/2)*0, tmpq, m4
     LOAD64_LUT m1, inq, ctxq, (mmsize/2)*1, tmpq, m5
@@ -786,7 +783,7 @@ FFT16_FN fma3
 %macro FFT32_FN 1
 INIT_YMM %1
 cglobal fft32_float, 4, 4, 16, ctx, out, in, tmp
-    mov ctxq, [ctxq + AVTXContext.revtab]
+    mov ctxq, [ctxq + AVTXContext.map]
 
     LOAD64_LUT m4, inq, ctxq, (mmsize/2)*4, tmpq,  m8,  m9
     LOAD64_LUT m5, inq, ctxq, (mmsize/2)*5, tmpq, m10, m11
@@ -800,8 +797,8 @@ cglobal fft32_float, 4, 4, 16, ctx, out, in, tmp
     LOAD64_LUT m2, inq, ctxq, (mmsize/2)*2, tmpq, m12, m13
     LOAD64_LUT m3, inq, ctxq, (mmsize/2)*3, tmpq, m14, m15
 
-    movaps m8,         [cos_32_float]
-    vperm2f128 m9, m9, [cos_32_float + 4*8 - 4*7], 0x23
+    movaps m8,         [tab_32_float]
+    vperm2f128 m9, m9, [tab_32_float + 4*8 - 4*7], 0x23
 
     FFT16 m0, m1, m2, m3, m10, m11, m12, m13
 
@@ -858,8 +855,8 @@ ALIGN 16
     POP lenq
     sub outq, (%1*4) + (%1*2) + (%1/2)
 
-    lea rtabq, [cos_ %+ %1 %+ _float]
-    lea itabq, [cos_ %+ %1 %+ _float + %1 - 4*7]
+    lea rtabq, [tab_ %+ %1 %+ _float]
+    lea itabq, [tab_ %+ %1 %+ _float + %1 - 4*7]
 
 %if %0 > 1
     cmp tgtq, %1
@@ -883,9 +880,9 @@ ALIGN 16
 
 %macro FFT_SPLIT_RADIX_FN 1
 INIT_YMM %1
-cglobal split_radix_fft_float, 4, 8, 16, 272, lut, out, in, len, tmp, itab, rtab, tgt
-    movsxd lenq, dword [lutq + AVTXContext.m]
-    mov lutq, [lutq + AVTXContext.revtab]
+cglobal fft_sr_float, 4, 8, 16, 272, lut, out, in, len, tmp, itab, rtab, tgt
+    movsxd lenq, dword [lutq + AVTXContext.len]
+    mov lutq, [lutq + AVTXContext.map]
     mov tgtq, lenq
 
 ; Bottom-most/32-point transform ===============================================
@@ -903,8 +900,8 @@ ALIGN 16
     LOAD64_LUT m2, inq, lutq, (mmsize/2)*2, tmpq, m12, m13
     LOAD64_LUT m3, inq, lutq, (mmsize/2)*3, tmpq, m14, m15
 
-    movaps m8,         [cos_32_float]
-    vperm2f128 m9, m9, [cos_32_float + 32 - 4*7], 0x23
+    movaps m8,         [tab_32_float]
+    vperm2f128 m9, m9, [tab_32_float + 32 - 4*7], 0x23
 
     FFT16 m0, m1, m2, m3, m10, m11, m12, m13
 
@@ -961,8 +958,8 @@ ALIGN 16
 
     FFT16 tx2_e0, tx2_e1, tx2_o0, tx2_o1, tmp1, tmp2, tw_e, tw_o
 
-    movaps tw_e,           [cos_64_float]
-    vperm2f128 tw_o, tw_o, [cos_64_float + 64 - 4*7], 0x23
+    movaps tw_e,           [tab_64_float]
+    vperm2f128 tw_o, tw_o, [tab_64_float + 64 - 4*7], 0x23
 
     add lutq, (mmsize/2)*8
     cmp tgtq, 64
@@ -989,8 +986,8 @@ ALIGN 16
     POP lenq
     sub outq, 24*mmsize
 
-    lea rtabq, [cos_128_float]
-    lea itabq, [cos_128_float + 128 - 4*7]
+    lea rtabq, [tab_128_float]
+    lea itabq, [tab_128_float + 128 - 4*7]
 
     cmp tgtq, 128
     je .deinterleave
@@ -1016,8 +1013,8 @@ ALIGN 16
     POP lenq
     sub outq, 48*mmsize
 
-    lea rtabq, [cos_256_float]
-    lea itabq, [cos_256_float + 256 - 4*7]
+    lea rtabq, [tab_256_float]
+    lea itabq, [tab_256_float + 256 - 4*7]
 
     cmp tgtq, 256
     je .deinterleave
@@ -1044,8 +1041,8 @@ ALIGN 16
     POP lenq
     sub outq, 96*mmsize
 
-    lea rtabq, [cos_512_float]
-    lea itabq, [cos_512_float + 512 - 4*7]
+    lea rtabq, [tab_512_float]
+    lea itabq, [tab_512_float + 512 - 4*7]
 
     cmp tgtq, 512
     je .deinterleave
@@ -1079,8 +1076,8 @@ ALIGN 16
     POP lenq
     sub outq, 192*mmsize
 
-    lea rtabq, [cos_1024_float]
-    lea itabq, [cos_1024_float + 1024 - 4*7]
+    lea rtabq, [tab_1024_float]
+    lea itabq, [tab_1024_float + 1024 - 4*7]
 
     cmp tgtq, 1024
     je .deinterleave
@@ -1160,8 +1157,8 @@ FFT_SPLIT_RADIX_DEF 131072
     vextractf128 [outq + 13*mmsize +  0], tw_e,   1
     vextractf128 [outq + 13*mmsize + 16], tx2_e0, 1
 
-    movaps tw_e,           [cos_64_float + mmsize]
-    vperm2f128 tw_o, tw_o, [cos_64_float + 64 - 4*7 - mmsize], 0x23
+    movaps tw_e,           [tab_64_float + mmsize]
+    vperm2f128 tw_o, tw_o, [tab_64_float + 64 - 4*7 - mmsize], 0x23
 
     movaps m0, [outq +  1*mmsize]
     movaps m1, [outq +  3*mmsize]
diff --git a/libavutil/x86/tx_float_init.c b/libavutil/x86/tx_float_init.c
index 8b77a5f29f..5c9afa75cc 100644
--- a/libavutil/x86/tx_float_init.c
+++ b/libavutil/x86/tx_float_init.c
@@ -21,86 +21,106 @@
 #include "libavutil/attributes.h"
 #include "libavutil/x86/cpu.h"
 
-void ff_fft2_float_sse3     (AVTXContext *s, void *out, void *in, ptrdiff_t stride);
-void ff_fft4_inv_float_sse2 (AVTXContext *s, void *out, void *in, ptrdiff_t stride);
-void ff_fft4_fwd_float_sse2 (AVTXContext *s, void *out, void *in, ptrdiff_t stride);
-void ff_fft8_float_sse3     (AVTXContext *s, void *out, void *in, ptrdiff_t stride);
-void ff_fft8_float_avx      (AVTXContext *s, void *out, void *in, ptrdiff_t stride);
-void ff_fft16_float_avx     (AVTXContext *s, void *out, void *in, ptrdiff_t stride);
-void ff_fft16_float_fma3    (AVTXContext *s, void *out, void *in, ptrdiff_t stride);
-void ff_fft32_float_avx     (AVTXContext *s, void *out, void *in, ptrdiff_t stride);
-void ff_fft32_float_fma3    (AVTXContext *s, void *out, void *in, ptrdiff_t stride);
+#include "config.h"
 
-void ff_split_radix_fft_float_avx (AVTXContext *s, void *out, void *in, ptrdiff_t stride);
-void ff_split_radix_fft_float_avx2(AVTXContext *s, void *out, void *in, ptrdiff_t stride);
-
-av_cold void ff_tx_init_float_x86(AVTXContext *s, av_tx_fn *tx)
-{
-    int cpu_flags = av_get_cpu_flags();
-    int gen_revtab = 0, basis, revtab_interleave;
+#define DECL_INIT_FN(basis, interleave)                                        \
+static av_cold int                                                             \
+    ff_tx_fft_sr_codelet_init_b ##basis## _i ##interleave## _x86               \
+    (AVTXContext *s,                                                           \
+     const FFTXCodelet *cd,                                                    \
+     uint64_t flags,                                                           \
+     FFTXCodeletOptions *opts,                                                 \
+     int len, int inv,                                                         \
+     const void *scale)                                                        \
+{                                                                              \
+    const int inv_lookup = opts ? opts->invert_lookup : 1;                     \
+    ff_tx_init_tabs_float(len);                                                \
+    return ff_tx_gen_split_radix_parity_revtab(s, inv_lookup,                  \
+                                               basis, interleave);             \
+}
 
-    if (s->flags & AV_TX_UNALIGNED)
-        return;
+#define ff_tx_fft_sr_codelet_init_b0_i0_x86 NULL
+DECL_INIT_FN(8, 0)
+DECL_INIT_FN(8, 2)
 
-    if (ff_tx_type_is_mdct(s->type))
-        return;
+#define DECL_SR_CD_DEF(fn_name, len, init_fn, fn_prio, cpu, fn_flags)          \
+void ff_tx_ ##fn_name(AVTXContext *s, void *out, void *in, ptrdiff_t stride);  \
+static const FFTXCodelet ff_tx_ ##fn_name## _def = {                           \
+    .name       = #fn_name,                                                    \
+    .function   = ff_tx_ ##fn_name,                                            \
+    .type       = TX_TYPE(FFT),                                                \
+    .flags      = FF_TX_OUT_OF_PLACE | FF_TX_ALIGNED | fn_flags,               \
+    .factors[0] = 2,                                                           \
+    .min_len    = len,                                                         \
+    .max_len    = len,                                                         \
+    .init       = ff_tx_fft_sr_codelet_init_ ##init_fn## _x86,                 \
+    .cpu_flags  = AV_CPU_FLAG_ ##cpu,                                          \
+    .prio       = fn_prio,                                                     \
+};
 
-#define TXFN(fn, gentab, sr_basis, interleave) \
-    do {                                       \
-        *tx = fn;                              \
-        gen_revtab = gentab;                   \
-        basis = sr_basis;                      \
-        revtab_interleave = interleave;        \
-    } while (0)
+DECL_SR_CD_DEF(fft2_float_sse3,      2, b0_i0, 128, SSE3, AV_TX_INPLACE)
+DECL_SR_CD_DEF(fft4_fwd_float_sse2,  4, b0_i0, 128, SSE2, AV_TX_INPLACE | FF_TX_FORWARD_ONLY)
+DECL_SR_CD_DEF(fft4_inv_float_sse2,  4, b0_i0, 128, SSE2, AV_TX_INPLACE | FF_TX_INVERSE_ONLY)
+DECL_SR_CD_DEF(fft8_float_sse3,      8, b8_i0, 128, SSE3, AV_TX_INPLACE)
+DECL_SR_CD_DEF(fft8_float_avx,       8, b8_i0, 256, AVX,  AV_TX_INPLACE)
+DECL_SR_CD_DEF(fft16_float_avx,     16, b8_i2, 256, AVX,  AV_TX_INPLACE)
+DECL_SR_CD_DEF(fft16_float_fma3,    16, b8_i2, 288, FMA3, AV_TX_INPLACE)
 
-    if (s->n == 1) {
-        if (EXTERNAL_SSE2(cpu_flags)) {
-            if (s->m == 4 && s->inv)
-                TXFN(ff_fft4_inv_float_sse2, 0, 0, 0);
-            else if (s->m == 4)
-                TXFN(ff_fft4_fwd_float_sse2, 0, 0, 0);
-        }
+#if ARCH_X86_64
+DECL_SR_CD_DEF(fft32_float_avx,     32, b8_i2, 256, AVX,  AV_TX_INPLACE)
+DECL_SR_CD_DEF(fft32_float_fma3,    32, b8_i2, 288, FMA3, AV_TX_INPLACE)
 
-        if (EXTERNAL_SSE3(cpu_flags)) {
-            if (s->m == 2)
-                TXFN(ff_fft2_float_sse3, 0, 0, 0);
-            else if (s->m == 8)
-                TXFN(ff_fft8_float_sse3, 1, 8, 0);
-        }
+void ff_tx_fft_sr_float_avx(AVTXContext *s, void *out, void *in, ptrdiff_t stride);
+const FFTXCodelet ff_tx_fft_sr_float_avx_def = {
+    .name       = "fft_sr_float_avx",
+    .function   = ff_tx_fft_sr_float_avx,
+    .type       = TX_TYPE(FFT),
+    .flags      = FF_TX_ALIGNED | FF_TX_OUT_OF_PLACE,
+    .factors[0] = 2,
+    .min_len    = 64,
+    .max_len    = 131072,
+    .init       = ff_tx_fft_sr_codelet_init_b8_i2_x86,
+    .cpu_flags  = AV_CPU_FLAG_AVX,
+    .prio       = 256,
+};
 
-        if (EXTERNAL_AVX_FAST(cpu_flags)) {
-            if (s->m == 8)
-                TXFN(ff_fft8_float_avx, 1, 8, 0);
-            else if (s->m == 16)
-                TXFN(ff_fft16_float_avx, 1, 8, 2);
-#if ARCH_X86_64
-            else if (s->m == 32)
-                TXFN(ff_fft32_float_avx, 1, 8, 2);
-            else if (s->m >= 64 && s->m <= 131072 && !(s->flags & AV_TX_INPLACE))
-                TXFN(ff_split_radix_fft_float_avx, 1, 8, 2);
+#if HAVE_AVX2_EXTERNAL
+void ff_tx_fft_sr_float_avx2(AVTXContext *s, void *out, void *in, ptrdiff_t stride);
+const FFTXCodelet ff_tx_fft_sr_float_avx2_def = {
+    .name       = "fft_sr_float_avx2",
+    .function   = ff_tx_fft_sr_float_avx2,
+    .type       = TX_TYPE(FFT),
+    .flags      = FF_TX_ALIGNED | FF_TX_OUT_OF_PLACE,
+    .factors[0] = 2,
+    .min_len    = 64,
+    .max_len    = 131072,
+    .init       = ff_tx_fft_sr_codelet_init_b8_i2_x86,
+    .cpu_flags  = AV_CPU_FLAG_AVX2,
+    .prio       = 288,
+};
 #endif
-        }
-
-        if (EXTERNAL_FMA3_FAST(cpu_flags)) {
-            if (s->m == 16)
-                TXFN(ff_fft16_float_fma3, 1, 8, 2);
-#if ARCH_X86_64
-            else if (s->m == 32)
-                TXFN(ff_fft32_float_fma3, 1, 8, 2);
 #endif
-        }
+
+const FFTXCodelet * const ff_tx_codelet_list_float_x86[] = {
+    /* Split-Radix codelets */
+    &ff_tx_fft2_float_sse3_def,
+    &ff_tx_fft4_fwd_float_sse2_def,
+    &ff_tx_fft4_inv_float_sse2_def,
+    &ff_tx_fft8_float_sse3_def,
+    &ff_tx_fft8_float_avx_def,
+    &ff_tx_fft16_float_avx_def,
+    &ff_tx_fft16_float_fma3_def,
 
 #if ARCH_X86_64
-        if (EXTERNAL_AVX2_FAST(cpu_flags)) {
-            if (s->m >= 64 && s->m <= 131072 && !(s->flags & AV_TX_INPLACE))
-                TXFN(ff_split_radix_fft_float_avx2, 1, 8, 2);
-        }
-#endif
-    }
+    &ff_tx_fft32_float_avx_def,
+    &ff_tx_fft32_float_fma3_def,
 
-    if (gen_revtab)
-        ff_tx_gen_split_radix_parity_revtab(s->revtab, s->m, s->inv, basis,
-                                            revtab_interleave);
+    /* Standalone transforms */
+    &ff_tx_fft_sr_float_avx_def,
+#if HAVE_AVX2_EXTERNAL
+    &ff_tx_fft_sr_float_avx2_def,
+#endif
+#endif
 
-#undef TXFN
-}
+    NULL,
+};



More information about the ffmpeg-cvslog mailing list