diff options
author | Lynne <dev@lynne.ee> | 2024-05-16 11:36:12 +0200 |
---|---|---|
committer | Lynne <dev@lynne.ee> | 2024-06-02 18:34:45 +0200 |
commit | eee5fa08083c1df6d0210bf215b658bc3017f98d (patch) | |
tree | 8c222da326d48f19c395d7f631042f03d3dcb726 /libavcodec/aac | |
parent | 23b45d7e20b0f60c8c5a00c631b95aa0f9e19448 (diff) | |
download | ffmpeg-eee5fa08083c1df6d0210bf215b658bc3017f98d.tar.gz |
aacdec: add a decoder for AAC USAC (xHE-AAC)
This commit adds a decoder for the frequency-domain part of USAC.
What works:
- Mono
- Stereo (no prediction)
- Stereo (mid/side coding)
- Stereo (complex prediction)
What's left:
- SBR
- Speech coding
Known issues:
- Desync with certain sequences
- Preroll crossover missing (shouldn't matter, bitrate adaptation only)
Diffstat (limited to 'libavcodec/aac')
-rw-r--r-- | libavcodec/aac/Makefile | 3 | ||||
-rw-r--r-- | libavcodec/aac/aacdec.c | 188 | ||||
-rw-r--r-- | libavcodec/aac/aacdec.h | 187 | ||||
-rw-r--r-- | libavcodec/aac/aacdec_ac.c | 208 | ||||
-rw-r--r-- | libavcodec/aac/aacdec_ac.h | 54 | ||||
-rw-r--r-- | libavcodec/aac/aacdec_dsp_template.c | 4 | ||||
-rw-r--r-- | libavcodec/aac/aacdec_latm.h | 14 | ||||
-rw-r--r-- | libavcodec/aac/aacdec_lpd.c | 198 | ||||
-rw-r--r-- | libavcodec/aac/aacdec_lpd.h | 33 | ||||
-rw-r--r-- | libavcodec/aac/aacdec_usac.c | 1608 | ||||
-rw-r--r-- | libavcodec/aac/aacdec_usac.h | 37 |
11 files changed, 2458 insertions, 76 deletions
diff --git a/libavcodec/aac/Makefile b/libavcodec/aac/Makefile index c3e525d373..70b1dca274 100644 --- a/libavcodec/aac/Makefile +++ b/libavcodec/aac/Makefile @@ -2,6 +2,7 @@ clean:: $(RM) $(CLEANSUFFIXES:%=libavcodec/aac/%) OBJS-$(CONFIG_AAC_DECODER) += aac/aacdec.o aac/aacdec_tab.o \ - aac/aacdec_float.o + aac/aacdec_float.o aac/aacdec_usac.o \ + aac/aacdec_ac.o aac/aacdec_lpd.o OBJS-$(CONFIG_AAC_FIXED_DECODER) += aac/aacdec.o aac/aacdec_tab.o \ aac/aacdec_fixed.o diff --git a/libavcodec/aac/aacdec.c b/libavcodec/aac/aacdec.c index 6f37ac5361..2b8322fc68 100644 --- a/libavcodec/aac/aacdec.c +++ b/libavcodec/aac/aacdec.c @@ -40,6 +40,7 @@ #include "aacdec.h" #include "aacdec_tab.h" +#include "aacdec_usac.h" #include "libavcodec/aac.h" #include "libavcodec/aac_defines.h" @@ -535,6 +536,8 @@ static av_cold void flush(AVCodecContext *avctx) } } } + + ff_aac_usac_reset_state(ac, &ac->oc[1]); } /** @@ -993,13 +996,14 @@ static int decode_eld_specific_config(AACDecContext *ac, AVCodecContext *avctx, */ static int decode_audio_specific_config_gb(AACDecContext *ac, AVCodecContext *avctx, - MPEG4AudioConfig *m4ac, + OutputConfiguration *oc, GetBitContext *gb, int get_bit_alignment, int sync_extension) { int i, ret; GetBitContext gbc = *gb; + MPEG4AudioConfig *m4ac = &oc->m4ac; MPEG4AudioConfig m4ac_bak = *m4ac; if ((i = ff_mpeg4audio_get_config_gb(m4ac, &gbc, sync_extension, avctx)) < 0) { @@ -1033,14 +1037,22 @@ static int decode_audio_specific_config_gb(AACDecContext *ac, case AOT_ER_AAC_LC: case AOT_ER_AAC_LD: if ((ret = decode_ga_specific_config(ac, avctx, gb, get_bit_alignment, - m4ac, m4ac->chan_config)) < 0) + &oc->m4ac, m4ac->chan_config)) < 0) return ret; break; case AOT_ER_AAC_ELD: if ((ret = decode_eld_specific_config(ac, avctx, gb, - m4ac, m4ac->chan_config)) < 0) + &oc->m4ac, m4ac->chan_config)) < 0) + return ret; + break; +#if CONFIG_AAC_DECODER + case AOT_USAC_NOSBR: /* fallthrough */ + case AOT_USAC: + if ((ret = ff_aac_usac_config_decode(ac, avctx, gb, + oc, m4ac->chan_config)) < 0) return ret; break; +#endif default: avpriv_report_missing_feature(avctx, "Audio object type %s%d", @@ -1060,7 +1072,7 @@ static int decode_audio_specific_config_gb(AACDecContext *ac, static int decode_audio_specific_config(AACDecContext *ac, AVCodecContext *avctx, - MPEG4AudioConfig *m4ac, + OutputConfiguration *oc, const uint8_t *data, int64_t bit_size, int sync_extension) { @@ -1080,7 +1092,7 @@ static int decode_audio_specific_config(AACDecContext *ac, if ((ret = init_get_bits(&gb, data, bit_size)) < 0) return ret; - return decode_audio_specific_config_gb(ac, avctx, m4ac, &gb, 0, + return decode_audio_specific_config_gb(ac, avctx, oc, &gb, 0, sync_extension); } @@ -1104,6 +1116,15 @@ static av_cold int decode_close(AVCodecContext *avctx) { AACDecContext *ac = avctx->priv_data; + for (int i = 0; i < 2; i++) { + OutputConfiguration *oc = &ac->oc[i]; + AACUSACConfig *usac = &oc->usac; + for (int j = 0; j < usac->nb_elems; j++) { + AACUsacElemConfig *ec = &usac->elems[i]; + av_freep(&ec->ext.pl_data); + } + } + for (int type = 0; type < FF_ARRAY_ELEMS(ac->che); type++) { for (int i = 0; i < MAX_ELEM_ID; i++) { if (ac->che[type][i]) { @@ -1181,7 +1202,7 @@ av_cold int ff_aac_decode_init(AVCodecContext *avctx) ac->oc[1].m4ac.sample_rate = avctx->sample_rate; if (avctx->extradata_size > 0) { - if ((ret = decode_audio_specific_config(ac, ac->avctx, &ac->oc[1].m4ac, + if ((ret = decode_audio_specific_config(ac, ac->avctx, &ac->oc[1], avctx->extradata, avctx->extradata_size * 8LL, 1)) < 0) @@ -1549,9 +1570,16 @@ static int decode_pulses(Pulse *pulse, GetBitContext *gb, int ff_aac_decode_tns(AACDecContext *ac, TemporalNoiseShaping *tns, GetBitContext *gb, const IndividualChannelStream *ics) { + int tns_max_order = INT32_MAX; + const int is_usac = ac->oc[1].m4ac.object_type == AOT_USAC || + ac->oc[1].m4ac.object_type == AOT_USAC_NOSBR; int w, filt, i, coef_len, coef_res, coef_compress; const int is8 = ics->window_sequence[0] == EIGHT_SHORT_SEQUENCE; - const int tns_max_order = is8 ? 7 : ac->oc[1].m4ac.object_type == AOT_AAC_MAIN ? 20 : 12; + + /* USAC doesn't seem to have a limit */ + if (!is_usac) + tns_max_order = is8 ? 7 : ac->oc[1].m4ac.object_type == AOT_AAC_MAIN ? 20 : 12; + for (w = 0; w < ics->num_windows; w++) { if ((tns->n_filt[w] = get_bits(gb, 2 - is8))) { coef_res = get_bits1(gb); @@ -1560,7 +1588,12 @@ int ff_aac_decode_tns(AACDecContext *ac, TemporalNoiseShaping *tns, int tmp2_idx; tns->length[w][filt] = get_bits(gb, 6 - 2 * is8); - if ((tns->order[w][filt] = get_bits(gb, 5 - 2 * is8)) > tns_max_order) { + if (is_usac) + tns->order[w][filt] = get_bits(gb, 4 - is8); + else + tns->order[w][filt] = get_bits(gb, 5 - (2 * is8)); + + if (tns->order[w][filt] > tns_max_order) { av_log(ac->avctx, AV_LOG_ERROR, "TNS filter order %d is greater than maximum %d.\n", tns->order[w][filt], tns_max_order); @@ -1598,6 +1631,7 @@ static void decode_mid_side_stereo(ChannelElement *cpe, GetBitContext *gb, { int idx; int max_idx = cpe->ch[0].ics.num_window_groups * cpe->ch[0].ics.max_sfb; + cpe->max_sfb_ste = cpe->ch[0].ics.max_sfb; if (ms_present == 1) { for (idx = 0; idx < max_idx; idx++) cpe->ms_mask[idx] = get_bits1(gb); @@ -2182,42 +2216,19 @@ static int aac_decode_er_frame(AVCodecContext *avctx, AVFrame *frame, return 0; } -static int aac_decode_frame_int(AVCodecContext *avctx, AVFrame *frame, - int *got_frame_ptr, GetBitContext *gb, - const AVPacket *avpkt) +static int decode_frame_ga(AVCodecContext *avctx, AACDecContext *ac, + GetBitContext *gb, int *got_frame_ptr) { - AACDecContext *ac = avctx->priv_data; - ChannelElement *che = NULL, *che_prev = NULL; + int err; + int is_dmono; + int elem_id; enum RawDataBlockType elem_type, che_prev_type = TYPE_END; - int err, elem_id; - int samples = 0, multiplier, audio_found = 0, pce_found = 0; - int is_dmono, sce_count = 0; - int payload_alignment; uint8_t che_presence[4][MAX_ELEM_ID] = {{0}}; + ChannelElement *che = NULL, *che_prev = NULL; + int samples = 0, multiplier, audio_found = 0, pce_found = 0, sce_count = 0; + AVFrame *frame = ac->frame; - ac->frame = frame; - - if (show_bits(gb, 12) == 0xfff) { - if ((err = parse_adts_frame_header(ac, gb)) < 0) { - av_log(avctx, AV_LOG_ERROR, "Error decoding AAC frame header.\n"); - goto fail; - } - if (ac->oc[1].m4ac.sampling_index > 12) { - av_log(ac->avctx, AV_LOG_ERROR, "invalid sampling rate index %d\n", ac->oc[1].m4ac.sampling_index); - err = AVERROR_INVALIDDATA; - goto fail; - } - } - - if ((err = frame_configure_elements(avctx)) < 0) - goto fail; - - // The AV_PROFILE_AAC_* defines are all object_type - 1 - // This may lead to an undefined profile being signaled - ac->avctx->profile = ac->oc[1].m4ac.object_type - 1; - - payload_alignment = get_bits_count(gb); - ac->tags_mapped = 0; + int payload_alignment = get_bits_count(gb); // parse while ((elem_type = get_bits(gb, 3)) != TYPE_END) { elem_id = get_bits(gb, 4); @@ -2225,28 +2236,23 @@ static int aac_decode_frame_int(AVCodecContext *avctx, AVFrame *frame, if (avctx->debug & FF_DEBUG_STARTCODE) av_log(avctx, AV_LOG_DEBUG, "Elem type:%x id:%x\n", elem_type, elem_id); - if (!avctx->ch_layout.nb_channels && elem_type != TYPE_PCE) { - err = AVERROR_INVALIDDATA; - goto fail; - } + if (!avctx->ch_layout.nb_channels && elem_type != TYPE_PCE) + return AVERROR_INVALIDDATA; if (elem_type < TYPE_DSE) { if (che_presence[elem_type][elem_id]) { int error = che_presence[elem_type][elem_id] > 1; av_log(ac->avctx, error ? AV_LOG_ERROR : AV_LOG_DEBUG, "channel element %d.%d duplicate\n", elem_type, elem_id); - if (error) { - err = AVERROR_INVALIDDATA; - goto fail; - } + if (error) + return AVERROR_INVALIDDATA; } che_presence[elem_type][elem_id]++; if (!(che=ff_aac_get_che(ac, elem_type, elem_id))) { av_log(ac->avctx, AV_LOG_ERROR, "channel element %d.%d is not allocated\n", elem_type, elem_id); - err = AVERROR_INVALIDDATA; - goto fail; + return AVERROR_INVALIDDATA; } samples = ac->oc[1].m4ac.frame_length_short ? 960 : 1024; che->present = 1; @@ -2283,10 +2289,8 @@ static int aac_decode_frame_int(AVCodecContext *avctx, AVFrame *frame, int tags; int pushed = push_output_configuration(ac); - if (pce_found && !pushed) { - err = AVERROR_INVALIDDATA; - goto fail; - } + if (pce_found && !pushed) + return AVERROR_INVALIDDATA; tags = decode_pce(avctx, &ac->oc[1].m4ac, layout_map, gb, payload_alignment); @@ -2312,8 +2316,7 @@ static int aac_decode_frame_int(AVCodecContext *avctx, AVFrame *frame, elem_id += get_bits(gb, 8) - 1; if (get_bits_left(gb) < 8 * elem_id) { av_log(avctx, AV_LOG_ERROR, "TYPE_FIL: "overread_err); - err = AVERROR_INVALIDDATA; - goto fail; + return AVERROR_INVALIDDATA; } err = 0; while (elem_id > 0) { @@ -2337,19 +2340,16 @@ static int aac_decode_frame_int(AVCodecContext *avctx, AVFrame *frame, } if (err) - goto fail; + return err; if (get_bits_left(gb) < 3) { av_log(avctx, AV_LOG_ERROR, overread_err); - err = AVERROR_INVALIDDATA; - goto fail; + return AVERROR_INVALIDDATA; } } - if (!avctx->ch_layout.nb_channels) { - *got_frame_ptr = 0; + if (!avctx->ch_layout.nb_channels) return 0; - } multiplier = (ac->oc[1].m4ac.sbr == 1) ? ac->oc[1].m4ac.ext_sample_rate > ac->oc[1].m4ac.sample_rate : 0; samples <<= multiplier; @@ -2364,16 +2364,17 @@ static int aac_decode_frame_int(AVCodecContext *avctx, AVFrame *frame, if (!ac->frame->data[0] && samples) { av_log(avctx, AV_LOG_ERROR, "no frame data found\n"); - err = AVERROR_INVALIDDATA; - goto fail; + return AVERROR_INVALIDDATA; } if (samples) { ac->frame->nb_samples = samples; ac->frame->sample_rate = avctx->sample_rate; - } else + *got_frame_ptr = 1; + } else { av_frame_unref(ac->frame); - *got_frame_ptr = !!samples; + *got_frame_ptr = 0; + } /* for dual-mono audio (SCE + SCE) */ is_dmono = ac->dmono_mode && sce_count == 2 && @@ -2387,6 +2388,59 @@ static int aac_decode_frame_int(AVCodecContext *avctx, AVFrame *frame, } return 0; +} + +static int aac_decode_frame_int(AVCodecContext *avctx, AVFrame *frame, + int *got_frame_ptr, GetBitContext *gb, + const AVPacket *avpkt) +{ + int err; + AACDecContext *ac = avctx->priv_data; + + ac->frame = frame; + *got_frame_ptr = 0; + + if (show_bits(gb, 12) == 0xfff) { + if ((err = parse_adts_frame_header(ac, gb)) < 0) { + av_log(avctx, AV_LOG_ERROR, "Error decoding AAC frame header.\n"); + goto fail; + } + if (ac->oc[1].m4ac.sampling_index > 12) { + av_log(ac->avctx, AV_LOG_ERROR, "invalid sampling rate index %d\n", ac->oc[1].m4ac.sampling_index); + err = AVERROR_INVALIDDATA; + goto fail; + } + } + + if ((err = frame_configure_elements(avctx)) < 0) + goto fail; + + // The AV_PROFILE_AAC_* defines are all object_type - 1 + // This may lead to an undefined profile being signaled + ac->avctx->profile = ac->oc[1].m4ac.object_type - 1; + + ac->tags_mapped = 0; + + if ((ac->oc[1].m4ac.object_type == AOT_USAC) || + (ac->oc[1].m4ac.object_type == AOT_USAC_NOSBR)) { + if (ac->is_fixed) { + avpriv_report_missing_feature(ac->avctx, + "AAC USAC fixed-point decoding"); + return AVERROR_PATCHWELCOME; + } +#if CONFIG_AAC_DECODER + err = ff_aac_usac_decode_frame(avctx, ac, gb, got_frame_ptr); + if (err < 0) + goto fail; +#endif + } else { + err = decode_frame_ga(avctx, ac, gb, got_frame_ptr); + if (err < 0) + goto fail; + } + + return err; + fail: pop_output_configuration(ac); return err; @@ -2414,7 +2468,7 @@ static int aac_decode_frame(AVCodecContext *avctx, AVFrame *frame, if (new_extradata) { /* discard previous configuration */ ac->oc[1].status = OC_NONE; - err = decode_audio_specific_config(ac, ac->avctx, &ac->oc[1].m4ac, + err = decode_audio_specific_config(ac, ac->avctx, &ac->oc[1], new_extradata, new_extradata_size * 8LL, 1); if (err < 0) { diff --git a/libavcodec/aac/aacdec.h b/libavcodec/aac/aacdec.h index 8d1eb74066..ee21a94007 100644 --- a/libavcodec/aac/aacdec.h +++ b/libavcodec/aac/aacdec.h @@ -42,6 +42,8 @@ #include "libavcodec/avcodec.h" #include "libavcodec/mpeg4audio.h" +#include "aacdec_ac.h" + typedef struct AACDecContext AACDecContext; /** @@ -69,6 +71,32 @@ enum CouplingPoint { AFTER_IMDCT = 3, }; +enum AACUsacElem { + ID_USAC_SCE = 0, + ID_USAC_CPE = 1, + ID_USAC_LFE = 2, + ID_USAC_EXT = 3, +}; + +enum ExtensionHeaderType { + ID_CONFIG_EXT_FILL = 0, + ID_CONFIG_EXT_LOUDNESS_INFO = 2, + ID_CONFIG_EXT_STREAM_ID = 7, +}; + +enum AACUsacExtension { + ID_EXT_ELE_FILL, + ID_EXT_ELE_MPEGS, + ID_EXT_ELE_SAOC, + ID_EXT_ELE_AUDIOPREROLL, + ID_EXT_ELE_UNI_DRC, +}; + +enum AACUSACLoudnessExt { + UNIDRCLOUDEXT_TERM = 0x0, + UNIDRCLOUDEXT_EQ = 0x1, +}; + // Supposed to be equal to AAC_RENAME() in case of USE_FIXED. #define RENAME_FIXED(name) name ## _fixed @@ -93,6 +121,40 @@ typedef struct LongTermPrediction { int8_t used[MAX_LTP_LONG_SFB]; } LongTermPrediction; +/* Per channel core mode */ +typedef struct AACUsacElemData { + uint8_t core_mode; + uint8_t scale_factor_grouping; + + /* Timewarping ratio */ +#define NUM_TW_NODES 16 + uint8_t tw_ratio[NUM_TW_NODES]; + + struct { + uint8_t acelp_core_mode : 3; + uint8_t lpd_mode : 5; + + uint8_t bpf_control_info : 1; + uint8_t core_mode_last : 1; + uint8_t fac_data_present : 1; + + int last_lpd_mode; + } ldp; + + struct { + unsigned int seed; + uint8_t level : 3; + uint8_t offset : 5; + } noise; + + struct { + uint8_t gain; + uint32_t kv[8 /* (1024 / 16) / 8 */][8]; + } fac; + + AACArithState ac; +} AACUsacElemData; + /** * Individual Channel Stream */ @@ -145,11 +207,13 @@ typedef struct ChannelCoupling { */ typedef struct SingleChannelElement { IndividualChannelStream ics; + AACUsacElemData ue; ///< USAC element data TemporalNoiseShaping tns; enum BandType band_type[128]; ///< band types int sfo[128]; ///< scalefactor offsets INTFLOAT_UNION(sf, [128]); ///< scalefactors (8 windows * 16 sfb max) INTFLOAT_ALIGNED_UNION(32, coeffs, 1024); ///< coefficients for IMDCT, maybe processed + INTFLOAT_ALIGNED_UNION(32, prev_coeffs, 1024); ///< unscaled previous contents of coeffs[] for USAC INTFLOAT_ALIGNED_UNION(32, saved, 1536); ///< overlap INTFLOAT_ALIGNED_UNION(32, ret_buf, 2048); ///< PCM output buffer INTFLOAT_ALIGNED_UNION(16, ltp_state, 3072); ///< time signal for LTP @@ -163,25 +227,148 @@ typedef struct SingleChannelElement { }; } SingleChannelElement; +typedef struct AACUsacStereo { + uint8_t common_window; + uint8_t common_tw; + + uint8_t ms_mask_mode; + uint8_t config_idx; + + /* Complex prediction */ + uint8_t use_prev_frame; + uint8_t pred_dir; + uint8_t complex_coef; + + uint8_t pred_used[128]; + + INTFLOAT_ALIGNED_UNION(32, alpha_q_re, 1024); + INTFLOAT_ALIGNED_UNION(32, alpha_q_im, 1024); + INTFLOAT_ALIGNED_UNION(32, prev_alpha_q_re, 1024); + INTFLOAT_ALIGNED_UNION(32, prev_alpha_q_im, 1024); + + INTFLOAT_ALIGNED_UNION(32, dmix_re, 1024); + INTFLOAT_ALIGNED_UNION(32, prev_dmix_re, 1024); /* Recalculated on every frame */ + INTFLOAT_ALIGNED_UNION(32, dmix_im, 1024); /* Final prediction data */ +} AACUsacStereo; + /** * channel element - generic struct for SCE/CPE/CCE/LFE */ typedef struct ChannelElement { int present; // CPE specific + uint8_t max_sfb_ste; ///< (USAC) Maximum of both max_sfb values uint8_t ms_mask[128]; ///< Set if mid/side stereo is used for each scalefactor window band // shared SingleChannelElement ch[2]; // CCE specific ChannelCoupling coup; + // USAC stereo coupling data + AACUsacStereo us; } ChannelElement; +typedef struct AACUSACLoudnessInfo { + uint8_t drc_set_id : 6; + uint8_t downmix_id : 7; + struct { + uint16_t lvl : 12; + uint8_t present : 1; + } sample_peak; + + struct { + uint16_t lvl : 12; + uint8_t measurement : 4; + uint8_t reliability : 2; + uint8_t present : 1; + } true_peak; + + uint8_t nb_measurements : 4; + struct { + uint8_t method_def : 4; + uint8_t method_val; + uint8_t measurement : 4; + uint8_t reliability : 2; + } measurements[16]; +} AACUSACLoudnessInfo; + +typedef struct AACUsacElemConfig { + enum AACUsacElem type; + + uint8_t tw_mdct : 1; + uint8_t noise_fill : 1; + + uint8_t stereo_config_index; + + struct { + int ratio; + + uint8_t harmonic_sbr : 1; /* harmonicSBR */ + uint8_t bs_intertes : 1; /* bs_interTes */ + uint8_t bs_pvc : 1; /* bs_pvc */ + + struct { + uint8_t start_freq; /* dflt_start_freq */ + uint8_t stop_freq; /* dflt_stop_freq */ + + uint8_t freq_scale; /* dflt_freq_scale */ + uint8_t alter_scale : 1; /* dflt_alter_scale */ + uint8_t noise_scale; /* dflt_noise_scale */ + + uint8_t limiter_bands; /* dflt_limiter_bands */ + uint8_t limiter_gains; /* dflt_limiter_gains */ + uint8_t interpol_freq : 1; /* dflt_interpol_freq */ + uint8_t smoothing_mode : 1; /* dflt_smoothing_mode */ + } dflt; + } sbr; + + struct { + uint8_t freq_res; /* bsFreqRes */ + uint8_t fixed_gain; /* bsFixedGainDMX */ + uint8_t temp_shape_config; /* bsTempShapeConfig */ + uint8_t decorr_config; /* bsDecorrConfig */ + uint8_t high_rate_mode : 1; /* bsHighRateMode */ + uint8_t phase_coding : 1; /* bsPhaseCoding */ + + uint8_t otts_bands_phase; /* bsOttBandsPhase */ + uint8_t residual_coding; /* bsResidualCoding */ + uint8_t residual_bands; /* bsResidualBands */ + uint8_t pseudo_lr : 1; /* bsPseudoLr */ + uint8_t env_quant_mode : 1; /* bsEnvQuantMode */ + } mps; + + struct { + enum AACUsacExtension type; + uint8_t payload_frag; + uint32_t default_len; + uint32_t pl_data_offset; + uint8_t *pl_data; + } ext; +} AACUsacElemConfig; + +typedef struct AACUSACConfig { + uint8_t core_sbr_frame_len_idx; /* coreSbrFrameLengthIndex */ + uint8_t rate_idx; + uint16_t core_frame_len; + uint16_t stream_identifier; + + AACUsacElemConfig elems[64]; + int nb_elems; + + struct { + uint8_t nb_album; + AACUSACLoudnessInfo album_info[64]; + uint8_t nb_info; + AACUSACLoudnessInfo info[64]; + } loudness; +} AACUSACConfig; + typedef struct OutputConfiguration { MPEG4AudioConfig m4ac; uint8_t layout_map[MAX_ELEM_ID*4][3]; int layout_map_tags; AVChannelLayout ch_layout; enum OCStatus status; + AACUSACConfig usac; } OutputConfiguration; /** diff --git a/libavcodec/aac/aacdec_ac.c b/libavcodec/aac/aacdec_ac.c new file mode 100644 index 0000000000..7e5077cd19 --- /dev/null +++ b/libavcodec/aac/aacdec_ac.c @@ -0,0 +1,208 @@ +/* + * AAC definitions and structures + * Copyright (c) 2024 Lynne + * + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + +#include "libavcodec/aactab.h" +#include "aacdec_ac.h" + +uint32_t ff_aac_ac_map_process(AACArithState *state, int reset, int N) +{ + float ratio; + if (reset) { + memset(state->last, 0, sizeof(state->last)); + state->last_len = N; + } else if (state->last_len != N) { + int i; + uint8_t last[512 /* 2048 / 4 */]; + memcpy(last, state->last, sizeof(last)); + + ratio = state->last_len / (float)N; + for (i = 0; i < N/2; i++) { + int k = (int)(i * ratio); + state->last[i] = last[k]; + } + + for (; i < FF_ARRAY_ELEMS(state->last); i++) + state->last[i] = 0; + + state->last_len = N; + } + + state->cur[3] = 0; + state->cur[2] = 0; + state->cur[1] = 0; + state->cur[0] = 1; + + state->state_pre = state->last[0] << 12; + return state->last[0] << 12; +} + +uint32_t ff_aac_ac_get_context(AACArithState *state, uint32_t c, int i, int N) +{ + c = state->state_pre >> 8; + c = c + (state->last[i + 1] << 8); + c = (c << 4); + c += state->cur[1]; + + state->state_pre = c; + + if (i > 3 && + ((state->cur[3] + state->cur[2] + state->cur[1]) < 5)) + return c + 0x10000; + + return c; +} + +uint32_t ff_aac_ac_get_pk(uint32_t c) +{ + int i_min = -1; + int i, j; + int i_max = FF_ARRAY_ELEMS(ff_aac_ac_lookup_m) - 1; + while ((i_max - i_min) > 1) { + i = i_min + ((i_max - i_min) / 2); + j = ff_aac_ac_hash_m[i]; + if (c < (j >> 8)) + i_max = i; + else if (c > (j >> 8)) + i_min = i; + else + return (j & 0xFF); + } + return ff_aac_ac_lookup_m[i_max]; +} + +void ff_aac_ac_update_context(AACArithState *state, int idx, + uint16_t a, uint16_t b) +{ + state->cur[0] = a + b + 1; + if (state->cur[0] > 0xF) + state->cur[0] = 0xF; + + state->cur[3] = state->cur[2]; + state->cur[2] = state->cur[1]; + state->cur[1] = state->cur[0]; + + state->last[idx] = state->cur[0]; +} + +/* Initialize AC */ +void ff_aac_ac_init(AACArith *ac, GetBitContext *gb) +{ + ac->low = 0; + ac->high = UINT16_MAX; + ac->val = get_bits(gb, 16); +} + +uint16_t ff_aac_ac_decode(AACArith *ac, GetBitContext *gb, + const uint16_t *cdf, uint16_t cdf_len) +{ + int val = ac->val; + int low = ac->low; + int high = ac->high; + + int sym; + int rng = high - low + 1; + int c = ((((int)(val - low + 1)) << 14) - ((int)1)); + + const uint16_t *p = cdf - 1; + + /* One for each possible CDF length in the spec */ + switch (cdf_len) { + case 2: + if ((p[1] * rng) > c) + p += 1; + break; + case 4: + if ((p[2] * rng) > c) + p += 2; + if ((p[1] * rng) > c) + p += 1; + break; + case 17: + /* First check if the current probability is even met at all */ + if ((p[1] * rng) <= c) + break; + p += 1; + for (int i = 8; i >= 1; i >>= 1) + if ((p[i] * rng) > c) + p += i; + break; + case 27: + if ((p[16] * rng) > c) + p += 16; + if ((p[8] * rng) > c) + p += 8; + if (p != (cdf - 1 + 24)) + if ((p[4] * rng) > c) + p += 4; + if ((p[2] * rng) > c) + p += 2; + + if (p != (cdf - 1 + 24 + 2)) + if ((p[1] * rng) > c) + p += 1; + break; + default: + /* This should never happen */ + av_assert2(0); + } + + sym = (int)((ptrdiff_t)(p - cdf)) + 1; + if (sym) + high = low + ((rng * cdf[sym - 1]) >> 14) - 1; + low += (rng * cdf[sym]) >> 14; + + /* This loop could be done faster */ + while (1) { + if (high < 32768) { + ; + } else if (low >= 32768) { + val -= 32768; + low -= 32768; + high -= 32768; + } else if (low >= 16384 && high < 49152) { + val -= 16384; + low -= 16384; + high -= 16384; + } else { + break; + } + low += low; + high += high + 1; + val = (val << 1) | get_bits1(gb); + }; + + ac->low = low; + ac->high = high; + ac->val = val; + + return sym; +} + +void ff_aac_ac_finish(AACArithState *state, int offset, int N) +{ + int i; + + for (i = offset; i < N/2; i++) + state->last[i] = 1; + + for (; i < FF_ARRAY_ELEMS(state->last); i++) + state->last[i] = 0; +} diff --git a/libavcodec/aac/aacdec_ac.h b/libavcodec/aac/aacdec_ac.h new file mode 100644 index 0000000000..0b98c0f0d9 --- /dev/null +++ b/libavcodec/aac/aacdec_ac.h @@ -0,0 +1,54 @@ +/* + * AAC definitions and structures + * Copyright (c) 2024 Lynne + * + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + +#ifndef AVCODEC_AAC_AACDEC_AC_H +#define AVCODEC_AAC_AACDEC_AC_H + +#include "libavcodec/get_bits.h" + +typedef struct AACArithState { + uint8_t last[512 /* 2048 / 4 */]; + int last_len; + uint8_t cur[4]; + uint16_t state_pre; +} AACArithState; + +typedef struct AACArith { + uint16_t low; + uint16_t high; + uint16_t val; +} AACArith; + +#define FF_AAC_AC_ESCAPE 16 + +uint32_t ff_aac_ac_map_process(AACArithState *state, int reset, int len); +uint32_t ff_aac_ac_get_context(AACArithState *state, uint32_t old_c, int idx, int len); +uint32_t ff_aac_ac_get_pk(uint32_t c); + +void ff_aac_ac_update_context(AACArithState *state, int idx, uint16_t a, uint16_t b); +void ff_aac_ac_init(AACArith *ac, GetBitContext *gb); + +uint16_t ff_aac_ac_decode(AACArith *ac, GetBitContext *gb, + const uint16_t *cdf, uint16_t cdf_len); + +void ff_aac_ac_finish(AACArithState *state, int offset, int nb); + +#endif /* AVCODEC_AACDEC_AC_H */ diff --git a/libavcodec/aac/aacdec_dsp_template.c b/libavcodec/aac/aacdec_dsp_template.c index 59a69d88f3..8d31af22f8 100644 --- a/libavcodec/aac/aacdec_dsp_template.c +++ b/libavcodec/aac/aacdec_dsp_template.c @@ -88,8 +88,8 @@ static void AAC_RENAME(apply_mid_side_stereo)(AACDecContext *ac, ChannelElement INTFLOAT *ch1 = cpe->ch[1].AAC_RENAME(coeffs); const uint16_t *offsets = ics->swb_offset; for (int g = 0; g < ics->num_window_groups; g++) { - for (int sfb = 0; sfb < ics->max_sfb; sfb++) { - const int idx = g*ics->max_sfb + sfb; + for (int sfb = 0; sfb < cpe->max_sfb_ste; sfb++) { + const int idx = g*cpe->max_sfb_ste + sfb; if (cpe->ms_mask[idx] && cpe->ch[0].band_type[idx] < NOISE_BT && cpe->ch[1].band_type[idx] < NOISE_BT) { diff --git a/libavcodec/aac/aacdec_latm.h b/libavcodec/aac/aacdec_latm.h index e40a2fe1a7..047c11e0fb 100644 --- a/libavcodec/aac/aacdec_latm.h +++ b/libavcodec/aac/aacdec_latm.h @@ -56,7 +56,8 @@ static int latm_decode_audio_specific_config(struct LATMContext *latmctx, { AACDecContext *ac = &latmctx->aac_ctx; AVCodecContext *avctx = ac->avctx; - MPEG4AudioConfig m4ac = { 0 }; + OutputConfiguration oc = { 0 }; + MPEG4AudioConfig *m4ac = &oc.m4ac; GetBitContext gbc; int config_start_bit = get_bits_count(gb); int sync_extension = 0; @@ -76,7 +77,7 @@ static int latm_decode_audio_specific_config(struct LATMContext *latmctx, if (get_bits_left(gb) <= 0) return AVERROR_INVALIDDATA; - bits_consumed = decode_audio_specific_config_gb(NULL, avctx, &m4ac, + bits_consumed = decode_audio_specific_config_gb(NULL, avctx, &oc, &gbc, config_start_bit, sync_extension); @@ -88,11 +89,12 @@ static int latm_decode_audio_specific_config(struct LATMContext *latmctx, asclen = bits_consumed; if (!latmctx->initialized || - ac->oc[1].m4ac.sample_rate != m4ac.sample_rate || - ac->oc[1].m4ac.chan_config != m4ac.chan_config) { + ac->oc[1].m4ac.sample_rate != m4ac->sample_rate || + ac->oc[1].m4ac.chan_config != m4ac->chan_config) { if (latmctx->initialized) { - av_log(avctx, AV_LOG_INFO, "audio config changed (sample_rate=%d, chan_config=%d)\n", m4ac.sample_rate, m4ac.chan_config); + av_log(avctx, AV_LOG_INFO, "audio config changed (sample_rate=%d, chan_config=%d)\n", + m4ac->sample_rate, m4ac->chan_config); } else { av_log(avctx, AV_LOG_DEBUG, "initializing latmctx\n"); } @@ -280,7 +282,7 @@ static int latm_decode_frame(AVCodecContext *avctx, AVFrame *out, } else { push_output_configuration(&latmctx->aac_ctx); if ((err = decode_audio_specific_config( - &latmctx->aac_ctx, avctx, &latmctx->aac_ctx.oc[1].m4ac, + &latmctx->aac_ctx, avctx, &latmctx->aac_ctx.oc[1], avctx->extradata, avctx->extradata_size*8LL, 1)) < 0) { pop_output_configuration(&latmctx->aac_ctx); return err; diff --git a/libavcodec/aac/aacdec_lpd.c b/libavcodec/aac/aacdec_lpd.c new file mode 100644 index 0000000000..796edd2ab5 --- /dev/null +++ b/libavcodec/aac/aacdec_lpd.c @@ -0,0 +1,198 @@ +/* + * Copyright (c) 2024 Lynne <dev@lynne.ee> + * + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + +#include "aacdec_lpd.h" +#include "aacdec_usac.h" +#include "libavcodec/unary.h" + +const uint8_t ff_aac_lpd_mode_tab[32][4] = { + { 0, 0, 0, 0 }, + { 1, 0, 0, 0 }, + { 0, 1, 0, 0 }, + { 1, 1, 0, 0 }, + { 0, 0, 1, 0 }, + { 1, 0, 1, 0 }, + { 0, 1, 1, 0 }, + { 1, 1, 1, 0 }, + { 0, 0, 0, 1 }, + { 1, 0, 0, 1 }, + { 0, 1, 0, 1 }, + { 1, 1, 0, 1 }, + { 0, 0, 1, 1 }, + { 1, 0, 1, 1 }, + { 0, 1, 1, 1 }, + { 1, 1, 1, 1 }, + { 2, 2, 0, 0 }, + { 2, 2, 1, 0 }, + { 2, 2, 0, 1 }, + { 2, 2, 1, 1 }, + { 0, 0, 2, 2 }, + { 1, 0, 2, 2 }, + { 0, 1, 2, 2 }, + { 1, 1, 2, 2 }, + { 2, 2, 2, 2 }, + { 3, 3, 3, 3 }, + /* Larger values are reserved, but permit them for resilience */ + { 0, 0, 0, 0 }, + { 0, 0, 0, 0 }, + { 0, 0, 0, 0 }, + { 0, 0, 0, 0 }, + { 0, 0, 0, 0 }, + { 0, 0, 0, 0 }, +}; + +static void parse_qn(GetBitContext *gb, int *qn, int nk_mode, int no_qn) +{ + if (nk_mode == 1) { + for (int k = 0; k < no_qn; k++) { + qn[k] = get_unary(gb, 0, INT32_MAX); // TODO: find proper ranges + if (qn[k]) + qn[k]++; + } + return; + } + + for (int k = 0; k < no_qn; k++) + qn[k] = get_bits(gb, 2) + 2; + + if (nk_mode == 2) { + for (int k = 0; k < no_qn; k++) { + if (qn[k] > 4) { + qn[k] = get_unary(gb, 0, INT32_MAX);; + if (qn[k]) + qn[k] += 4; + } + } + return; + } + + for (int k = 0; k < no_qn; k++) { + if (qn[k] > 4) { + int qn_ext = get_unary(gb, 0, INT32_MAX);; + switch (qn_ext) { + case 0: qn[k] = 5; break; + case 1: qn[k] = 6; break; + case 2: qn[k] = 0; break; + default: qn[k] = qn_ext + 4; break; + } + } + } +} + +static int parse_codebook_idx(GetBitContext *gb, uint32_t *kv, + int nk_mode, int no_qn) +{ + int idx, n, nk; + + int qn[2]; + parse_qn(gb, qn, nk_mode, no_qn); + + for (int k = 0; k < no_qn; k++) { + if (qn[k] > 4) { + nk = (qn[k] - 3) / 2; + n = qn[k] - nk*2; + } else { + nk = 0; + n = qn[k]; + } + } + + idx = get_bits(gb, 4*n); + + if (nk > 0) + for (int i = 0; i < 8; i++) + kv[i] = get_bits(gb, nk); + + return 0; +} + +int ff_aac_parse_fac_data(AACUsacElemData *ce, GetBitContext *gb, + int use_gain, int len) +{ + int ret; + if (use_gain) + ce->fac.gain = get_bits(gb, 7); + + for (int i = 0; i < len/8; i++) { + ret = parse_codebook_idx(gb, ce->fac.kv[i], 1, 1); + if (ret < 0) + return ret; + } + + return 0; +} + +int ff_aac_ldp_parse_channel_stream(AACDecContext *ac, AACUSACConfig *usac, + AACUsacElemData *ce, GetBitContext *gb) +{ + int k; + const uint8_t *mod; + int first_ldp_flag; + int first_tcx_flag; + + ce->ldp.acelp_core_mode = get_bits(gb, 3); + ce->ldp.lpd_mode = get_bits(gb, 5); + + ce->ldp.bpf_control_info = get_bits1(gb); + ce->ldp.core_mode_last = get_bits1(gb); + ce->ldp.fac_data_present = get_bits1(gb); + + mod = ff_aac_lpd_mode_tab[ce->ldp.lpd_mode]; + + first_ldp_flag = !ce->ldp.core_mode_last; + first_tcx_flag = 1; + if (first_ldp_flag) + ce->ldp.last_lpd_mode = -1; /* last_ldp_mode is a **STATEFUL** value */ + + k = 0; + while (k < 0) { + if (!k) { + if (ce->ldp.core_mode_last && ce->ldp.fac_data_present) + ff_aac_parse_fac_data(ce, gb, 0, usac->core_frame_len/8); + } else { + if (!ce->ldp.last_lpd_mode && mod[k] > 0 || + ce->ldp.last_lpd_mode && !mod[k]) + ff_aac_parse_fac_data(ce, gb, 0, usac->core_frame_len/8); + } + if (!mod[k]) { +// parse_acelp_coding(); + ce->ldp.last_lpd_mode = 0; + k++; + } else { +// parse_tcx_coding(); + ce->ldp.last_lpd_mode = mod[k]; + k += (1 << (mod[k] - 1)); + first_tcx_flag = 0; + } + } + +// parse_lpc_data(first_lpd_flag); + + if (!ce->ldp.core_mode_last && ce->ldp.fac_data_present) { + uint16_t len_8 = usac->core_frame_len / 8; + uint16_t len_16 = usac->core_frame_len / 16; + uint16_t fac_len = get_bits1(gb) /* short_fac_flag */ ? len_8 : len_16; + int ret = ff_aac_parse_fac_data(ce, gb, 1, fac_len); + if (ret < 0) + return ret; + } + + return 0; +} diff --git a/libavcodec/aac/aacdec_lpd.h b/libavcodec/aac/aacdec_lpd.h new file mode 100644 index 0000000000..924ff75e52 --- /dev/null +++ b/libavcodec/aac/aacdec_lpd.h @@ -0,0 +1,33 @@ +/* + * Copyright (c) 2024 Lynne <dev@lynne.ee> + * + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + +#ifndef AVCODEC_AAC_AACDEC_LPD_H +#define AVCODEC_AAC_AACDEC_LPD_H + +#include "aacdec.h" +#include "libavcodec/get_bits.h" + +int ff_aac_parse_fac_data(AACUsacElemData *ce, GetBitContext *gb, + int use_gain, int len); + +int ff_aac_ldp_parse_channel_stream(AACDecContext *ac, AACUSACConfig *usac, + AACUsacElemData *ce, GetBitContext *gb); + +#endif /* AVCODEC_AAC_AACDEC_LPD_H */ diff --git a/libavcodec/aac/aacdec_usac.c b/libavcodec/aac/aacdec_usac.c new file mode 100644 index 0000000000..c3c9137a2e --- /dev/null +++ b/libavcodec/aac/aacdec_usac.c @@ -0,0 +1,1608 @@ +/* + * Copyright (c) 2024 Lynne <dev@lynne.ee> + * + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + +#include "aacdec_usac.h" +#include "aacdec_tab.h" +#include "aacdec_lpd.h" +#include "aacdec_ac.h" + +#include "libavcodec/aactab.h" +#include "libavutil/mem.h" +#include "libavcodec/mpeg4audio.h" +#include "libavcodec/unary.h" + +/* Number of scalefactor bands per complex prediction band, equal to 2. */ +#define SFB_PER_PRED_BAND 2 + +static inline uint32_t get_escaped_value(GetBitContext *gb, int nb1, int nb2, int nb3) +{ + uint32_t val = get_bits(gb, nb1), val2; + if (val < ((1 << nb1) - 1)) + return val; + + val += val2 = get_bits(gb, nb2); + if (val2 == ((1 << nb2) - 1)) + val += get_bits(gb, nb3); + + return val; +} + +/* ISO/IEC 23003-3, Table 74 — bsOutputChannelPos */ +static const enum AVChannel usac_ch_pos_to_av[64] = { + [0] = AV_CHAN_FRONT_LEFT, + [1] = AV_CHAN_FRONT_RIGHT, + [2] = AV_CHAN_FRONT_CENTER, + [3] = AV_CHAN_LOW_FREQUENCY, + [4] = AV_CHAN_SIDE_LEFT, // +110 degrees, Ls|LS|kAudioChannelLabel_LeftSurround + [5] = AV_CHAN_SIDE_RIGHT, // -110 degrees, Rs|RS|kAudioChannelLabel_RightSurround + [6] = AV_CHAN_FRONT_LEFT_OF_CENTER, + [7] = AV_CHAN_FRONT_RIGHT_OF_CENTER, + [8] = AV_CHAN_BACK_LEFT, // +135 degrees, Lsr|BL|kAudioChannelLabel_RearSurroundLeft + [9] = AV_CHAN_BACK_RIGHT, // -135 degrees, Rsr|BR|kAudioChannelLabel_RearSurroundRight + [10] = AV_CHAN_BACK_CENTER, + [11] = AV_CHAN_SURROUND_DIRECT_LEFT, + [12] = AV_CHAN_SURROUND_DIRECT_RIGHT, + [13] = AV_CHAN_SIDE_SURROUND_LEFT, // +90 degrees, Lss|SL|kAudioChannelLabel_LeftSideSurround + [14] = AV_CHAN_SIDE_SURROUND_RIGHT, // -90 degrees, Rss|SR|kAudioChannelLabel_RightSideSurround + [15] = AV_CHAN_WIDE_LEFT, // +60 degrees, Lw|FLw|kAudioChannelLabel_LeftWide + [16] = AV_CHAN_WIDE_RIGHT, // -60 degrees, Rw|FRw|kAudioChannelLabel_RightWide + [17] = AV_CHAN_TOP_FRONT_LEFT, + [18] = AV_CHAN_TOP_FRONT_RIGHT, + [19] = AV_CHAN_TOP_FRONT_CENTER, + [20] = AV_CHAN_TOP_BACK_LEFT, + [21] = AV_CHAN_TOP_BACK_RIGHT, + [22] = AV_CHAN_TOP_BACK_CENTER, + [23] = AV_CHAN_TOP_SIDE_LEFT, + [24] = AV_CHAN_TOP_SIDE_RIGHT, + [25] = AV_CHAN_TOP_CENTER, + [26] = AV_CHAN_LOW_FREQUENCY_2, + [27] = AV_CHAN_BOTTOM_FRONT_LEFT, + [28] = AV_CHAN_BOTTOM_FRONT_RIGHT, + [29] = AV_CHAN_BOTTOM_FRONT_CENTER, + [30] = AV_CHAN_TOP_SURROUND_LEFT, ///< +110 degrees, Lvs, TpLS + [31] = AV_CHAN_TOP_SURROUND_RIGHT, ///< -110 degrees, Rvs, TpRS +}; + +static int decode_loudness_info(AACDecContext *ac, AACUSACLoudnessInfo *info, + GetBitContext *gb) +{ + info->drc_set_id = get_bits(gb, 6); + info->downmix_id = get_bits(gb, 7); + + if ((info->sample_peak.present = get_bits1(gb))) /* samplePeakLevelPresent */ + info->sample_peak.lvl = get_bits(gb, 12); + + if ((info->true_peak.present = get_bits1(gb))) { /* truePeakLevelPresent */ + info->true_peak.lvl = get_bits(gb, 12); + info->true_peak.measurement = get_bits(gb, 4); + info->true_peak.reliability = get_bits(gb, 2); + } + + info->nb_measurements = get_bits(gb, 4); + for (int i = 0; i < info->nb_measurements; i++) { + info->measurements[i].method_def = get_bits(gb, 4); + info->measurements[i].method_val = get_unary(gb, 0, 8); + info->measurements[i].measurement = get_bits(gb, 4); + info->measurements[i].reliability = get_bits(gb, 2); + } + + return 0; +} + +static int decode_loudness_set(AACDecContext *ac, AACUSACConfig *usac, + GetBitContext *gb) +{ + int ret; + + usac->loudness.nb_album = get_bits(gb, 6); /* loudnessInfoAlbumCount */ + usac->loudness.nb_info = get_bits(gb, 6); /* loudnessInfoCount */ + + for (int i = 0; i < usac->loudness.nb_album; i++) { + ret = decode_loudness_info(ac, &usac->loudness.album_info[i], gb); + if (ret < 0) + return ret; + } + + for (int i = 0; i < usac->loudness.nb_info; i++) { + ret = decode_loudness_info(ac, &usac->loudness.info[i], gb); + if (ret < 0) + return ret; + } + + if (get_bits1(gb)) { /* loudnessInfoSetExtPresent */ + enum AACUSACLoudnessExt type; + while ((type = get_bits(gb, 4)) != UNIDRCLOUDEXT_TERM) { + uint8_t size_bits = get_bits(gb, 4) + 4; + uint8_t bit_size = get_bits(gb, size_bits) + 1; + switch (type) { + case UNIDRCLOUDEXT_EQ: + avpriv_report_missing_feature(ac->avctx, "loudnessInfoV1"); + return AVERROR_PATCHWELCOME; + default: + for (int i = 0; i < bit_size; i++) + skip_bits1(gb); + } + } + } + + return 0; +} + +static void decode_usac_sbr_data(AACUsacElemConfig *e, GetBitContext *gb) +{ + uint8_t header_extra1; + uint8_t header_extra2; + + e->sbr.harmonic_sbr = get_bits1(gb); /* harmonicSBR */ + e->sbr.bs_intertes = get_bits1(gb); /* bs_interTes */ + e->sbr.bs_pvc = get_bits1(gb); /* bs_pvc */ + + e->sbr.dflt.start_freq = get_bits(gb, 4); /* dflt_start_freq */ + e->sbr.dflt.stop_freq = get_bits(gb, 4); /* dflt_stop_freq */ + + header_extra1 = get_bits1(gb); /* dflt_header_extra1 */ + header_extra2 = get_bits1(gb); /* dflt_header_extra2 */ + + e->sbr.dflt.freq_scale = 2; + e->sbr.dflt.alter_scale = 1; + e->sbr.dflt.noise_scale = 2; + if (header_extra1) { + e->sbr.dflt.freq_scale = get_bits(gb, 2); /* dflt_freq_scale */ + e->sbr.dflt.alter_scale = get_bits1(gb); /* dflt_alter_scale */ + e->sbr.dflt.noise_scale = get_bits(gb, 2); /* dflt_noise_scale */ + } + + e->sbr.dflt.limiter_bands = 2; + e->sbr.dflt.limiter_gains = 2; + e->sbr.dflt.interpol_freq = 1; + e->sbr.dflt.smoothing_mode = 1; + if (header_extra2) { + e->sbr.dflt.limiter_bands = get_bits(gb, 2); /* dflt_limiter_bands */ + e->sbr.dflt.limiter_gains = get_bits(gb, 2); /* dflt_limiter_gains */ + e->sbr.dflt.interpol_freq = get_bits1(gb); /* dflt_interpol_freq */ + e->sbr.dflt.smoothing_mode = get_bits1(gb); /* dflt_smoothing_mode */ + } +} + +static void decode_usac_element_core(AACUsacElemConfig *e, + GetBitContext *gb, + int sbr_ratio) +{ + e->tw_mdct = get_bits1(gb); /* tw_mdct */ + e->noise_fill = get_bits1(gb); + e->sbr.ratio = sbr_ratio; +} + +static void decode_usac_element_pair(AACUsacElemConfig *e, GetBitContext *gb) +{ + e->stereo_config_index = 0; + if (e->sbr.ratio) { + decode_usac_sbr_data(e, gb); + e->stereo_config_index = get_bits(gb, 2); + } + if (e->stereo_config_index) { + e->mps.freq_res = get_bits(gb, 3); /* bsFreqRes */ + e->mps.fixed_gain = get_bits(gb, 3); /* bsFixedGainDMX */ + e->mps.temp_shape_config = get_bits(gb, 2); /* bsTempShapeConfig */ + e->mps.decorr_config = get_bits(gb, 2); /* bsDecorrConfig */ + e->mps.high_rate_mode = get_bits1(gb); /* bsHighRateMode */ + e->mps.phase_coding = get_bits1(gb); /* bsPhaseCoding */ + + if (get_bits1(gb)) /* bsOttBandsPhasePresent */ + e->mps.otts_bands_phase = get_bits(gb, 5); /* bsOttBandsPhase */ + + e->mps.residual_coding = e->stereo_config_index >= 2; /* bsResidualCoding */ + if (e->mps.residual_coding) { + e->mps.residual_bands = get_bits(gb, 5); /* bsResidualBands */ + e->mps.pseudo_lr = get_bits1(gb); /* bsPseudoLr */ + } + if (e->mps.temp_shape_config == 2) + e->mps.env_quant_mode = get_bits1(gb); /* bsEnvQuantMode */ + } +} + +static int decode_usac_extension(AACDecContext *ac, AACUsacElemConfig *e, + GetBitContext *gb) +{ + int len = 0, ext_config_len; + + e->ext.type = get_escaped_value(gb, 4, 8, 16); /* usacExtElementType */ + ext_config_len = get_escaped_value(gb, 4, 8, 16); /* usacExtElementConfigLength */ + + if (get_bits1(gb)) /* usacExtElementDefaultLengthPresent */ + len = get_escaped_value(gb, 8, 16, 0) + 1; + + e->ext.default_len = len; + e->ext.payload_frag = get_bits1(gb); /* usacExtElementPayloadFrag */ + + av_log(ac->avctx, AV_LOG_DEBUG, "Extension present: type %i, len %i\n", + e->ext.type, ext_config_len); + + switch (e->ext.type) { +#if 0 /* Skip unsupported values */ + case ID_EXT_ELE_MPEGS: + break; + case ID_EXT_ELE_SAOC: + break; + case ID_EXT_ELE_UNI_DRC: + break; +#endif + case ID_EXT_ELE_FILL: + break; /* This is what the spec does */ + case ID_EXT_ELE_AUDIOPREROLL: + /* No configuration needed - fallthrough (len should be 0) */ + default: + skip_bits(gb, 8*ext_config_len); + break; + }; + + return 0; +} + +int ff_aac_usac_reset_state(AACDecContext *ac, OutputConfiguration *oc) +{ + AACUSACConfig *usac = &oc->usac; + int elem_id[3 /* SCE, CPE, LFE */] = { 0, 0, 0 }; + + ChannelElement *che; + enum RawDataBlockType type; + int id, ch; + + /* Initialize state */ + for (int i = 0; i < usac->nb_elems; i++) { + AACUsacElemConfig *e = &usac->elems[i]; + if (e->type != ID_USAC_SCE && e->type != ID_USAC_CPE) + continue; + + if (e->type == ID_USAC_SCE) { + ch = 1; + type = TYPE_SCE; + id = elem_id[0]++; + } else { + ch = 2; + type = TYPE_CPE; + id = elem_id[1]++; + } + + che = ff_aac_get_che(ac, type, id); + if (che) { + AACUsacStereo *us = &che->us; + memset(us, 0, sizeof(*us)); + + for (int j = 0; j < ch; j++) { + SingleChannelElement *sce = &che->ch[ch]; + AACUsacElemData *ue = &sce->ue; + + memset(ue, 0, sizeof(*ue)); + + if (!ch) + ue->noise.seed = 0x3039; + else + che->ch[1].ue.noise.seed = 0x10932; + } + } + } + + return 0; +} + +/* UsacConfig */ +int ff_aac_usac_config_decode(AACDecContext *ac, AVCodecContext *avctx, + GetBitContext *gb, OutputConfiguration *oc, + int channel_config) +{ + int ret, idx; + uint8_t freq_idx; + uint8_t channel_config_idx; + int nb_elements; + int samplerate; + int sbr_ratio; + MPEG4AudioConfig *m4ac = &oc->m4ac; + AACUSACConfig *usac = &oc->usac; + int elem_id[3 /* SCE, CPE, LFE */]; + + uint8_t layout_map[MAX_ELEM_ID*4][3]; + + memset(usac, 0, sizeof(*usac)); + + freq_idx = get_bits(gb, 5); /* usacSamplingFrequencyIndex */ + if (freq_idx == 0x1f) { + samplerate = get_bits(gb, 24); /* usacSamplingFrequency */ + + /* Try to match up an index for the custom sample rate. + * TODO: not sure if correct */ + for (idx = 0; idx < /* FF_ARRAY_ELEMS(ff_aac_usac_samplerate) */ 32; idx++) { + if (ff_aac_usac_samplerate[idx] >= samplerate) + break; + } + idx = FFMIN(idx, /* FF_ARRAY_ELEMS(ff_aac_usac_samplerate) */ 32 - 1); + usac->rate_idx = idx; + } else { + samplerate = ff_aac_usac_samplerate[freq_idx]; + if (samplerate < 0) + return AVERROR(EINVAL); + usac->rate_idx = freq_idx; + } + + m4ac->sample_rate = avctx->sample_rate = samplerate; + + usac->core_sbr_frame_len_idx = get_bits(gb, 3); /* coreSbrFrameLengthIndex */ + m4ac->frame_length_short = usac->core_sbr_frame_len_idx == 0 || + usac->core_sbr_frame_len_idx == 2; + + usac->core_frame_len = (usac->core_sbr_frame_len_idx == 0 || + usac->core_sbr_frame_len_idx == 2) ? 768 : 1024; + + sbr_ratio = usac->core_sbr_frame_len_idx == 2 ? 2 : + usac->core_sbr_frame_len_idx == 3 ? 3 : + usac->core_sbr_frame_len_idx == 4 ? 1 : + 0; + + channel_config_idx = get_bits(gb, 5); /* channelConfigurationIndex */ + if (!channel_config_idx) { + /* UsacChannelConfig() */ + uint8_t nb_channels = get_escaped_value(gb, 5, 8, 16); /* numOutChannels */ + if (nb_channels >= 64) + return AVERROR(EINVAL); + + av_channel_layout_uninit(&ac->oc[1].ch_layout); + + ret = av_channel_layout_custom_init(&ac->oc[1].ch_layout, nb_channels); + if (ret < 0) + return ret; + + for (int i = 0; i < nb_channels; i++) { + AVChannelCustom *cm = &ac->oc[1].ch_layout.u.map[i]; + cm->id = usac_ch_pos_to_av[get_bits(gb, 5)]; /* bsOutputChannelPos */ + if (cm->id == AV_CHAN_NONE) + cm->id = AV_CHAN_UNKNOWN; + } + + ret = av_channel_layout_retype(&ac->oc[1].ch_layout, + AV_CHANNEL_ORDER_NATIVE, + AV_CHANNEL_LAYOUT_RETYPE_FLAG_CANONICAL); + if (ret < 0) + return ret; + + ret = av_channel_layout_copy(&avctx->ch_layout, &ac->oc[1].ch_layout); + if (ret < 0) + return ret; + } else { + if ((ret = ff_aac_set_default_channel_config(ac, avctx, layout_map, + &nb_elements, channel_config_idx))) + return ret; + } + + /* UsacDecoderConfig */ + elem_id[0] = elem_id[1] = elem_id[2] = 0; + usac->nb_elems = get_escaped_value(gb, 4, 8, 16) + 1; + + for (int i = 0; i < usac->nb_elems; i++) { + AACUsacElemConfig *e = &usac->elems[i]; + memset(e, 0, sizeof(*e)); + + e->type = get_bits(gb, 2); /* usacElementType */ + av_log(ac->avctx, AV_LOG_DEBUG, "Element present: idx %i, type %i\n", + i, e->type); + + switch (e->type) { + case ID_USAC_SCE: /* SCE */ + /* UsacCoreConfig */ + decode_usac_element_core(e, gb, sbr_ratio); + if (e->sbr.ratio > 0) + decode_usac_sbr_data(e, gb); + layout_map[i][0] = TYPE_SCE; + layout_map[i][1] = i; + layout_map[i][2] = AAC_CHANNEL_FRONT; + elem_id[0]++; + + break; + case ID_USAC_CPE: /* UsacChannelPairElementConf */ + /* UsacCoreConfig */ + decode_usac_element_core(e, gb, sbr_ratio); + decode_usac_element_pair(e, gb); + layout_map[i][0] = TYPE_CPE; + layout_map[i][1] = i; + layout_map[i][2] = AAC_CHANNEL_FRONT; + elem_id[1]++; + + break; + case ID_USAC_LFE: /* LFE */ + /* LFE has no need for any configuration */ + e->tw_mdct = 0; + e->noise_fill = 0; + elem_id[2]++; + break; + case ID_USAC_EXT: /* EXT */ + ret = decode_usac_extension(ac, e, gb); + if (ret < 0) + return ret; + break; + }; + } + + ret = ff_aac_output_configure(ac, layout_map, elem_id[0] + elem_id[1] + elem_id[2], OC_GLOBAL_HDR, 0); + if (ret < 0) { + av_log(avctx, AV_LOG_ERROR, "Unable to parse channel config!\n"); + return ret; + } + + if (get_bits1(gb)) { /* usacConfigExtensionPresent */ + int invalid; + int nb_extensions = get_escaped_value(gb, 2, 4, 8) + 1; /* numConfigExtensions */ + for (int i = 0; i < nb_extensions; i++) { + int type = get_escaped_value(gb, 4, 8, 16); + int len = get_escaped_value(gb, 4, 8, 16); + switch (type) { + case ID_CONFIG_EXT_LOUDNESS_INFO: + ret = decode_loudness_set(ac, usac, gb); + if (ret < 0) + return ret; + break; + case ID_CONFIG_EXT_STREAM_ID: + usac->stream_identifier = get_bits(gb, 16); + break; + case ID_CONFIG_EXT_FILL: /* fallthrough */ + invalid = 0; + while (len--) { + if (get_bits(gb, 8) != 0xA5) + invalid++; + } + if (invalid) + av_log(avctx, AV_LOG_WARNING, "Invalid fill bytes: %i\n", + invalid); + break; + default: + while (len--) + skip_bits(gb, 8); + break; + } + } + } + + ret = ff_aac_usac_reset_state(ac, oc); + if (ret < 0) + return ret; + + return 0; +} + +static int decode_usac_scale_factors(AACDecContext *ac, + SingleChannelElement *sce, + GetBitContext *gb, uint8_t global_gain) +{ + IndividualChannelStream *ics = &sce->ics; + + /* Decode all scalefactors. */ + int offset_sf = global_gain; + for (int g = 0; g < ics->num_window_groups; g++) { + for (int sfb = 0; sfb < ics->max_sfb; sfb++) { + /* First coefficient is just the global gain */ + if (!g && !sfb) { + /* The cannonical representation of quantized scalefactors + * in the spec is with 100 subtracted. */ + sce->sfo[0] = offset_sf - 100; + continue; + } + + offset_sf += get_vlc2(gb, ff_vlc_scalefactors, 7, 3) - SCALE_DIFF_ZERO; + if (offset_sf > 255U) { + av_log(ac->avctx, AV_LOG_ERROR, + "Scalefactor (%d) out of range.\n", offset_sf); + return AVERROR_INVALIDDATA; + } + + sce->sfo[g*ics->max_sfb + sfb] = offset_sf - 100; + } + } + + return 0; +} + +/** + * Decode and dequantize arithmetically coded, uniformly quantized value + * + * @param coef array of dequantized, scaled spectral data + * @param sf array of scalefactors or intensity stereo positions + * + * @return Returns error status. 0 - OK, !0 - error + */ +static int decode_spectrum_and_dequant_ac(AACDecContext *s, float coef[1024], + GetBitContext *gb, const float sf[120], + AACArithState *state, int reset, + uint16_t len, uint16_t N) +{ + AACArith ac; + int i, a, b; + uint32_t c; + + int gb_count; + GetBitContext gb2; + + ff_aac_ac_init(&ac, gb); + c = ff_aac_ac_map_process(state, reset, N); + + /* Backup reader for rolling back by 14 bits at the end */ + gb2 = (GetBitContext)*gb; + gb_count = get_bits_count(&gb2); + + for (i = 0; i < len/2; i++) { + /* MSB */ + int lvl, esc_nb, m; + c = ff_aac_ac_get_context(state, c, i, N); + for (lvl=esc_nb=0;;) { + uint32_t pki = ff_aac_ac_get_pk(c + (esc_nb << 17)); + m = ff_aac_ac_decode(&ac, &gb2, ff_aac_ac_msb_cdfs[pki], + FF_ARRAY_ELEMS(ff_aac_ac_msb_cdfs[pki])); + if (m < FF_AAC_AC_ESCAPE) + break; + lvl++; + + /* Cargo-culted value. */ + if (lvl > 23) + return AVERROR(EINVAL); + + if ((esc_nb = lvl) > 7) + esc_nb = 7; + } + + b = m >> 2; + a = m - (b << 2); + + /* ARITH_STOP detection */ + if (!m) { + if (esc_nb) + break; + a = b = 0; + } + + /* LSB */ + for (int l = lvl; l > 0; l--) { + int lsbidx = !a ? 1 : (!b ? 0 : 2); + uint8_t r = ff_aac_ac_decode(&ac, &gb2, ff_aac_ac_lsb_cdfs[lsbidx], + FF_ARRAY_ELEMS(ff_aac_ac_lsb_cdfs[lsbidx])); + a = (a << 1) | (r & 1); + b = (b << 1) | ((r >> 1) & 1); + } + + /* Dequantize coeffs here */ + coef[2*i + 0] = a * cbrt(a); + coef[2*i + 1] = b * cbrt(b); + ff_aac_ac_update_context(state, i, a, b); + } + + if (len > 1) { + /* "Rewind" bitstream back by 14 bits */ + int gb_count2 = get_bits_count(&gb2); + skip_bits(gb, gb_count2 - gb_count - 14); + } else { + *gb = gb2; + } + + ff_aac_ac_finish(state, i, N); + + for (; i < N/2; i++) { + coef[2*i + 0] = 0; + coef[2*i + 1] = 0; + } + + /* Signs */ + for (i = 0; i < len; i++) { + if (coef[i]) { + if (!get_bits1(gb)) /* s */ + coef[i] *= -1; + } + } + + return 0; +} + +static int decode_usac_stereo_cplx(AACDecContext *ac, AACUsacStereo *us, + ChannelElement *cpe, GetBitContext *gb, + int num_window_groups, int indep_flag) +{ + int delta_code_time; + IndividualChannelStream *ics = &cpe->ch[0].ics; + + if (!get_bits1(gb)) { /* cplx_pred_all */ + for (int g = 0; g < num_window_groups; g++) { + for (int sfb = 0; sfb < cpe->max_sfb_ste; sfb += SFB_PER_PRED_BAND) { + const uint8_t val = get_bits1(gb); + us->pred_used[g*cpe->max_sfb_ste + sfb] = val; + if ((sfb + 1) < cpe->max_sfb_ste) + us->pred_used[g*cpe->max_sfb_ste + sfb + 1] = val; + } + } + } else { + for (int g = 0; g < num_window_groups; g++) + for (int sfb = 0; sfb < cpe->max_sfb_ste; sfb++) + us->pred_used[g*cpe->max_sfb_ste + sfb] = 1; + } + + us->pred_dir = get_bits1(gb); + us->complex_coef = get_bits1(gb); + + us->use_prev_frame = 0; + if (us->complex_coef && !indep_flag) + us->use_prev_frame = get_bits1(gb); + + delta_code_time = 0; + if (!indep_flag) + delta_code_time = get_bits1(gb); + + /* TODO: shouldn't be needed */ + for (int g = 0; g < num_window_groups; g++) { + for (int sfb = 0; sfb < cpe->max_sfb_ste; sfb += SFB_PER_PRED_BAND) { + float last_alpha_q_re = 0; + float last_alpha_q_im = 0; + if (delta_code_time) { + if (g) { + last_alpha_q_re = us->prev_alpha_q_re[(g - 1)*cpe->max_sfb_ste + sfb]; + last_alpha_q_im = us->prev_alpha_q_im[(g - 1)*cpe->max_sfb_ste + sfb]; + } else if ((ics->window_sequence[0] == EIGHT_SHORT_SEQUENCE) && + ics->window_sequence[1] == EIGHT_SHORT_SEQUENCE || + ics->window_sequence[1] == EIGHT_SHORT_SEQUENCE) { + /* The spec doesn't explicitly mention this, but it doesn't make + * any other sense otherwise! */ + last_alpha_q_re = us->prev_alpha_q_re[7*cpe->max_sfb_ste + sfb]; + last_alpha_q_im = us->prev_alpha_q_im[7*cpe->max_sfb_ste + sfb]; + } else { + last_alpha_q_re = us->prev_alpha_q_re[g*cpe->max_sfb_ste + sfb]; + last_alpha_q_im = us->prev_alpha_q_im[g*cpe->max_sfb_ste + sfb]; + } + } else { + if (sfb) { + last_alpha_q_re = us->alpha_q_re[g*cpe->max_sfb_ste + sfb - 1]; + last_alpha_q_im = us->alpha_q_im[g*cpe->max_sfb_ste + sfb - 1]; + } + } + + if (us->pred_used[g*cpe->max_sfb_ste + sfb]) { + int val = -get_vlc2(gb, ff_vlc_scalefactors, 7, 3) + 60; + last_alpha_q_re += val * 0.1f; + if (us->complex_coef) { + val = -get_vlc2(gb, ff_vlc_scalefactors, 7, 3) + 60; + last_alpha_q_im += val * 0.1f; + } + us->alpha_q_re[g*cpe->max_sfb_ste + sfb] = last_alpha_q_re; + us->alpha_q_im[g*cpe->max_sfb_ste + sfb] = last_alpha_q_im; + } else { + us->alpha_q_re[g*cpe->max_sfb_ste + sfb] = 0; + us->alpha_q_im[g*cpe->max_sfb_ste + sfb] = 0; + } + + if ((sfb + 1) < cpe->max_sfb_ste) { + us->alpha_q_re[g*cpe->max_sfb_ste + sfb + 1] = + us->alpha_q_re[g*cpe->max_sfb_ste + sfb]; + us->alpha_q_im[g*cpe->max_sfb_ste + sfb + 1] = + us->alpha_q_im[g*cpe->max_sfb_ste + sfb]; + } + } + } + + return 0; +} + +static int setup_sce(AACDecContext *ac, SingleChannelElement *sce, + AACUSACConfig *usac) +{ + AACUsacElemData *ue = &sce->ue; + IndividualChannelStream *ics = &sce->ics; + + /* Setup window parameters */ + if (ics->window_sequence[0] == EIGHT_SHORT_SEQUENCE) { + if (usac->core_frame_len == 768) { + ics->swb_offset = ff_swb_offset_96[usac->rate_idx]; + ics->num_swb = ff_aac_num_swb_96[usac->rate_idx]; + } else { + ics->swb_offset = ff_swb_offset_128[usac->rate_idx]; + ics->num_swb = ff_aac_num_swb_128[usac->rate_idx]; + } + ics->tns_max_bands = ff_tns_max_bands_128[usac->rate_idx]; + + /* Setup scalefactor grouping. 7 bit mask. */ + ics->num_window_groups = 0; + for (int j = 0; j < 7; j++) { + ics->group_len[j] = 1; + if (ue->scale_factor_grouping & (1 << (6 - j))) + ics->group_len[ics->num_window_groups] += 1; + else + ics->num_window_groups++; + } + + ics->group_len[7] = 1; + ics->num_window_groups++; + ics->num_windows = 8; + } else { + if (usac->core_frame_len == 768) { + ics->swb_offset = ff_swb_offset_768[usac->rate_idx]; + ics->num_swb = ff_aac_num_swb_768[usac->rate_idx]; + } else { + ics->swb_offset = ff_swb_offset_1024[usac->rate_idx]; + ics->num_swb = ff_aac_num_swb_1024[usac->rate_idx]; + } + ics->tns_max_bands = ff_tns_max_bands_1024[usac->rate_idx]; + + ics->group_len[0] = 1; + ics->num_window_groups = 1; + ics->num_windows = 1; + } + + if (ics->max_sfb > ics->num_swb) { + av_log(ac->avctx, AV_LOG_ERROR, + "Number of scalefactor bands in group (%d) " + "exceeds limit (%d).\n", + ics->max_sfb, ics->num_swb); + return AVERROR(EINVAL); + } + + /* Just some defaults for the band types */ + for (int i = 0; i < FF_ARRAY_ELEMS(sce->band_type); i++) + sce->band_type[i] = ESC_BT; + + return 0; +} + +static int decode_usac_stereo_info(AACDecContext *ac, AACUSACConfig *usac, + AACUsacElemConfig *ec, ChannelElement *cpe, + GetBitContext *gb, int indep_flag) +{ + int ret, tns_active; + + AACUsacStereo *us = &cpe->us; + SingleChannelElement *sce1 = &cpe->ch[0]; + SingleChannelElement *sce2 = &cpe->ch[1]; + IndividualChannelStream *ics1 = &sce1->ics; + IndividualChannelStream *ics2 = &sce2->ics; + AACUsacElemData *ue1 = &sce1->ue; + AACUsacElemData *ue2 = &sce2->ue; + + us->common_window = 0; + us->common_tw = 0; + + if (!(!ue1->core_mode && !ue2->core_mode)) + return 0; + + tns_active = get_bits1(gb); + us->common_window = get_bits1(gb); + + if (us->common_window) { + /* ics_info() */ + ics1->window_sequence[1] = ics1->window_sequence[0]; + ics2->window_sequence[1] = ics2->window_sequence[0]; + ics1->window_sequence[0] = ics2->window_sequence[0] = get_bits(gb, 2); + + ics1->use_kb_window[1] = ics1->use_kb_window[0]; + ics2->use_kb_window[1] = ics2->use_kb_window[0]; + ics1->use_kb_window[0] = ics2->use_kb_window[0] = get_bits1(gb); + + if (ics1->window_sequence[0] == EIGHT_SHORT_SEQUENCE) { + ics1->max_sfb = ics2->max_sfb = get_bits(gb, 4); + ue1->scale_factor_grouping = ue2->scale_factor_grouping = get_bits(gb, 7); + } else { + ics1->max_sfb = ics2->max_sfb = get_bits(gb, 6); + } + + if (!get_bits1(gb)) { /* common_max_sfb */ + if (ics2->window_sequence[0] == EIGHT_SHORT_SEQUENCE) + ics2->max_sfb = get_bits(gb, 4); + else + ics2->max_sfb = get_bits(gb, 6); + } + + ret = setup_sce(ac, sce1, usac); + if (ret < 0) + return ret; + + ret = setup_sce(ac, sce2, usac); + if (ret < 0) + return ret; + + cpe->max_sfb_ste = FFMAX(ics1->max_sfb, ics2->max_sfb); + + us->ms_mask_mode = get_bits(gb, 2); /* ms_mask_present */ + memset(cpe->ms_mask, 0, sizeof(cpe->ms_mask)); + if (us->ms_mask_mode == 1) { + for (int g = 0; g < ics1->num_window_groups; g++) + for (int sfb = 0; sfb < cpe->max_sfb_ste; sfb++) + cpe->ms_mask[g*cpe->max_sfb_ste + sfb] = get_bits1(gb); + } else if (us->ms_mask_mode == 2) { + memset(cpe->ms_mask, 0xFF, sizeof(cpe->ms_mask)); + } else if ((us->ms_mask_mode == 3) && !ec->stereo_config_index) { + ret = decode_usac_stereo_cplx(ac, us, cpe, gb, + ics1->num_window_groups, indep_flag); + if (ret < 0) + return ret; + } + } + + if (ec->tw_mdct) { + us->common_tw = get_bits1(gb); + avpriv_report_missing_feature(ac->avctx, + "AAC USAC timewarping"); + return AVERROR_PATCHWELCOME; + } + + sce1->tns.present = sce2->tns.present = 0; + if (tns_active) { + av_unused int tns_on_lr; + int common_tns = 0; + if (us->common_window) + common_tns = get_bits1(gb); + + tns_on_lr = get_bits1(gb); + if (common_tns) { + ret = ff_aac_decode_tns(ac, &sce1->tns, gb, ics1); + if (ret < 0) + return ret; + memcpy(&sce2->tns, &sce1->tns, sizeof(sce1->tns)); + sce2->tns.present = 0; + sce1->tns.present = 0; + } else { + if (get_bits1(gb)) { + sce2->tns.present = 1; + sce1->tns.present = 1; + } else { + sce2->tns.present = get_bits1(gb); + sce1->tns.present = !sce2->tns.present; + } + } + } + + return 0; +} + +/* 7.2.4 Generation of random signs for spectral noise filling + * This function is exactly defined, though we've helped the definition + * along with being slightly faster. */ +static inline float noise_random_sign(unsigned int *seed) +{ + unsigned int new_seed = *seed = ((*seed) * 69069) + 5; + if (((new_seed) & 0x10000) > 0) + return -1.f; + return +1.f; +} + +static void apply_noise_fill(AACDecContext *ac, SingleChannelElement *sce, + AACUsacElemData *ue) +{ + float *coef; + IndividualChannelStream *ics = &sce->ics; + + float noise_val = pow(2, (ue->noise.level - 14)/3); + int noise_offset = ue->noise.offset - 16; + int band_off; + + band_off = ff_usac_noise_fill_start_offset[ac->oc[1].m4ac.frame_length_short] + [ics->window_sequence[0] == EIGHT_SHORT_SEQUENCE]; + + coef = sce->coeffs; + for (int g = 0; g < ics->num_window_groups; g++) { + unsigned g_len = ics->group_len[g]; + + for (int sfb = 0; sfb < ics->max_sfb; sfb++) { + float *cb = coef + ics->swb_offset[sfb]; + int cb_len = ics->swb_offset[sfb + 1] - ics->swb_offset[sfb]; + int band_quantized_to_zero = 1; + + if (ics->swb_offset[sfb] < band_off) + continue; + + for (int group = 0; group < (unsigned)g_len; group++, cb += 128) { + for (int z = 0; z < cb_len; z++) { + if (cb[z] == 0) + cb[z] = noise_random_sign(&sce->ue.noise.seed) * noise_val; + else + band_quantized_to_zero = 0; + } + } + + if (band_quantized_to_zero) + sce->sf[g*ics->max_sfb + sfb] += noise_offset; + } + coef += g_len << 7; + } +} + +static void spectrum_scale(AACDecContext *ac, SingleChannelElement *sce, + AACUsacElemData *ue) +{ + IndividualChannelStream *ics = &sce->ics; + float *coef; + + /* Synthesise noise */ + if (ue->noise.level) + apply_noise_fill(ac, sce, ue); + + /* Apply scalefactors */ + coef = sce->coeffs; + for (int g = 0; g < ics->num_window_groups; g++) { + unsigned g_len = ics->group_len[g]; + + for (int sfb = 0; sfb < ics->max_sfb; sfb++) { + float *cb = coef + ics->swb_offset[sfb]; + int cb_len = ics->swb_offset[sfb + 1] - ics->swb_offset[sfb]; + float sf = sce->sf[g*ics->max_sfb + sfb]; + + for (int group = 0; group < (unsigned)g_len; group++, cb += 128) + ac->fdsp->vector_fmul_scalar(cb, cb, sf, cb_len); + } + coef += g_len << 7; + } +} + +static void complex_stereo_downmix_prev(AACDecContext *ac, ChannelElement *cpe, + float *dmix_re) +{ + IndividualChannelStream *ics = &cpe->ch[0].ics; + int sign = !cpe->us.pred_dir ? +1 : -1; + float *coef1 = cpe->ch[0].coeffs; + float *coef2 = cpe->ch[1].coeffs; + + for (int g = 0; g < ics->num_window_groups; g++) { + unsigned g_len = ics->group_len[g]; + for (int sfb = 0; sfb < cpe->max_sfb_ste; sfb++) { + int off = ics->swb_offset[sfb]; + int cb_len = ics->swb_offset[sfb + 1] - off; + + float *c1 = coef1 + off; + float *c2 = coef2 + off; + float *dm = dmix_re + off; + + for (int group = 0; group < (unsigned)g_len; + group++, c1 += 128, c2 += 128, dm += 128) { + for (int z = 0; z < cb_len; z++) + dm[z] = 0.5*(c1[z] + sign*c2[z]); + } + } + + coef1 += g_len << 7; + coef2 += g_len << 7; + dmix_re += g_len << 7; + } +} + +static void complex_stereo_downmix_cur(AACDecContext *ac, ChannelElement *cpe, + float *dmix_re) +{ + AACUsacStereo *us = &cpe->us; + IndividualChannelStream *ics = &cpe->ch[0].ics; + int sign = !cpe->us.pred_dir ? +1 : -1; + float *coef1 = cpe->ch[0].coeffs; + float *coef2 = cpe->ch[1].coeffs; + + for (int g = 0; g < ics->num_window_groups; g++) { + unsigned g_len = ics->group_len[g]; + for (int sfb = 0; sfb < cpe->max_sfb_ste; sfb++) { + int off = ics->swb_offset[sfb]; + int cb_len = ics->swb_offset[sfb + 1] - off; + + float *c1 = coef1 + off; + float *c2 = coef2 + off; + float *dm = dmix_re + off; + + if (us->pred_used[g*cpe->max_sfb_ste + sfb]) { + for (int group = 0; group < (unsigned)g_len; + group++, c1 += 128, c2 += 128, dm += 128) { + for (int z = 0; z < cb_len; z++) + dm[z] = 0.5*(c1[z] + sign*c2[z]); + } + } else { + for (int group = 0; group < (unsigned)g_len; + group++, c1 += 128, c2 += 128, dm += 128) { + for (int z = 0; z < cb_len; z++) + dm[z] = c1[z]; + } + } + } + + coef1 += g_len << 7; + coef2 += g_len << 7; + dmix_re += g_len << 7; + } +} + +static void complex_stereo_interpolate_imag(float *im, float *re, const float f[6], + int len, int factor_even, int factor_odd) +{ + int i = 0; + float s; + + s = f[6]*re[2] + f[5]*re[1] + f[4]*re[0] + + f[3]*re[0] + + f[2]*re[1] + f[1]*re[2] + f[0]*re[3]; + im[i] += s*factor_even; + + i = 1; + s = f[6]*re[1] + f[5]*re[0] + f[4]*re[0] + + f[3]*re[1] + + f[2]*re[2] + f[1]*re[3] + f[0]*re[4]; + im[i] += s*factor_odd; + + i = 2; + s = f[6]*re[0] + f[5]*re[0] + f[4]*re[1] + + f[3]*re[2] + + f[2]*re[3] + f[1]*re[4] + f[0]*re[5]; + + im[i] += s*factor_even; + for (i = 3; i < len - 4; i += 2) { + s = f[6]*re[i-3] + f[5]*re[i-2] + f[4]*re[i-1] + + f[3]*re[i] + + f[2]*re[i+1] + f[1]*re[i+2] + f[0]*re[i+3]; + im[i+0] += s*factor_odd; + + s = f[6]*re[i-2] + f[5]*re[i-1] + f[4]*re[i] + + f[3]*re[i+1] + + f[2]*re[i+2] + f[1]*re[i+3] + f[0]*re[i+4]; + im[i+1] += s*factor_even; + } + + i = len - 3; + s = f[6]*re[i-3] + f[5]*re[i-2] + f[4]*re[i-1] + + f[3]*re[i] + + f[2]*re[i+1] + f[1]*re[i+2] + f[0]*re[i+2]; + im[i] += s*factor_odd; + + i = len - 2; + s = f[6]*re[i-3] + f[5]*re[i-2] + f[4]*re[i-1] + + f[3]*re[i] + + f[2]*re[i+1] + f[1]*re[i+1] + f[0]*re[i]; + im[i] += s*factor_even; + + i = len - 1; + s = f[6]*re[i-3] + f[5]*re[i-2] + f[4]*re[i-1] + + f[3]*re[i] + + f[2]*re[i] + f[1]*re[i-1] + f[0]*re[i-2]; + im[i] += s*factor_odd; +} + +static void apply_complex_stereo(AACDecContext *ac, ChannelElement *cpe) +{ + AACUsacStereo *us = &cpe->us; + IndividualChannelStream *ics = &cpe->ch[0].ics; + float *coef1 = cpe->ch[0].coeffs; + float *coef2 = cpe->ch[1].coeffs; + float *dmix_im = us->dmix_im; + + for (int g = 0; g < ics->num_window_groups; g++) { + unsigned g_len = ics->group_len[g]; + for (int sfb = 0; sfb < cpe->max_sfb_ste; sfb++) { + int off = ics->swb_offset[sfb]; + int cb_len = ics->swb_offset[sfb + 1] - off; + + float *c1 = coef1 + off; + float *c2 = coef2 + off; + float *dm_im = dmix_im + off; + float alpha_re = us->alpha_q_re[g*cpe->max_sfb_ste + sfb]; + float alpha_im = us->alpha_q_im[g*cpe->max_sfb_ste + sfb]; + + if (!us->pred_used[g*cpe->max_sfb_ste + sfb]) + continue; + + if (!cpe->us.pred_dir) { + for (int group = 0; group < (unsigned)g_len; + group++, c1 += 128, c2 += 128, dm_im += 128) { + for (int z = 0; z < cb_len; z++) { + float side; + side = c2[z] - alpha_re*c1[z] - alpha_im*dm_im[z]; + c2[z] = c1[z] - side; + c1[z] = c1[z] + side; + } + } + } else { + for (int group = 0; group < (unsigned)g_len; + group++, c1 += 128, c2 += 128, dm_im += 128) { + for (int z = 0; z < cb_len; z++) { + float mid; + mid = c2[z] - alpha_re*c1[z] - alpha_im*dm_im[z]; + c2[z] = mid - c1[z]; + c1[z] = mid + c1[z]; + } + } + } + } + + coef1 += g_len << 7; + coef2 += g_len << 7; + dmix_im += g_len << 7; + } +} + +static const float *complex_stereo_get_filter(ChannelElement *cpe, int is_prev) +{ + int win, shape; + if (!is_prev) { + switch (cpe->ch[0].ics.window_sequence[0]) { + default: + case ONLY_LONG_SEQUENCE: + case EIGHT_SHORT_SEQUENCE: + win = 0; + break; + case LONG_START_SEQUENCE: + win = 1; + break; + case LONG_STOP_SEQUENCE: + win = 2; + break; + } + + if (cpe->ch[0].ics.use_kb_window[0] == 0 && + cpe->ch[0].ics.use_kb_window[1] == 0) + shape = 0; + else if (cpe->ch[0].ics.use_kb_window[0] == 1 && + cpe->ch[0].ics.use_kb_window[1] == 1) + shape = 1; + else if (cpe->ch[0].ics.use_kb_window[0] == 0 && + cpe->ch[0].ics.use_kb_window[1] == 1) + shape = 2; + else if (cpe->ch[0].ics.use_kb_window[0] == 1 && + cpe->ch[0].ics.use_kb_window[1] == 0) + shape = 3; + else + shape = 3; + } else { + win = cpe->ch[0].ics.window_sequence[0] == LONG_STOP_SEQUENCE; + shape = cpe->ch[0].ics.use_kb_window[1]; + } + + return ff_aac_usac_mdst_filt_cur[win][shape]; +} + +static void spectrum_decode(AACDecContext *ac, AACUSACConfig *usac, + ChannelElement *cpe, int nb_channels) +{ + AACUsacStereo *us = &cpe->us; + + for (int ch = 0; ch < nb_channels; ch++) { + SingleChannelElement *sce = &cpe->ch[ch]; + AACUsacElemData *ue = &sce->ue; + + spectrum_scale(ac, sce, ue); + } + + if (nb_channels > 1 && us->common_window) { + if (us->ms_mask_mode == 3) { + const float *filt; + complex_stereo_downmix_cur(ac, cpe, us->dmix_re); + complex_stereo_downmix_prev(ac, cpe, us->prev_dmix_re); + + filt = complex_stereo_get_filter(cpe, 0); + complex_stereo_interpolate_imag(us->dmix_im, us->dmix_re, filt, + usac->core_frame_len, 1, 1); + if (us->use_prev_frame) { + filt = complex_stereo_get_filter(cpe, 1); + complex_stereo_interpolate_imag(us->dmix_im, us->prev_dmix_re, filt, + usac->core_frame_len, -1, 1); + } + + apply_complex_stereo(ac, cpe); + } else if (us->ms_mask_mode > 0) { + ac->dsp.apply_mid_side_stereo(ac, cpe); + } + } + + /* Save coefficients and alpha values for prediction reasons */ + if (nb_channels > 1) { + AACUsacStereo *us = &cpe->us; + for (int ch = 0; ch < nb_channels; ch++) { + SingleChannelElement *sce = &cpe->ch[ch]; + memcpy(sce->prev_coeffs, sce->coeffs, sizeof(sce->coeffs)); + } + memcpy(us->prev_alpha_q_re, us->alpha_q_re, sizeof(us->alpha_q_re)); + memcpy(us->prev_alpha_q_im, us->alpha_q_im, sizeof(us->alpha_q_im)); + } + + for (int ch = 0; ch < nb_channels; ch++) { + SingleChannelElement *sce = &cpe->ch[ch]; + + /* Apply TNS */ + if (sce->tns.present) + ac->dsp.apply_tns(sce->coeffs, &sce->tns, &sce->ics, 1); + + ac->oc[1].m4ac.frame_length_short ? ac->dsp.imdct_and_windowing_768(ac, sce) : + ac->dsp.imdct_and_windowing(ac, sce); + } +} + +static int decode_usac_core_coder(AACDecContext *ac, AACUSACConfig *usac, + AACUsacElemConfig *ec, ChannelElement *che, + GetBitContext *gb, int indep_flag, int nb_channels) +{ + int ret; + int arith_reset_flag; + AACUsacStereo *us = &che->us; + + /* Local symbols */ + uint8_t global_gain; + + us->common_window = 0; + che->ch[0].tns.present = che->ch[1].tns.present = 0; + + for (int ch = 0; ch < nb_channels; ch++) { + SingleChannelElement *sce = &che->ch[ch]; + AACUsacElemData *ue = &sce->ue; + + ue->core_mode = get_bits1(gb); + } + + if (nb_channels == 2) { + ret = decode_usac_stereo_info(ac, usac, ec, che, gb, indep_flag); + if (ret) + return ret; + } + + for (int ch = 0; ch < nb_channels; ch++) { + SingleChannelElement *sce = &che->ch[ch]; + IndividualChannelStream *ics = &sce->ics; + AACUsacElemData *ue = &sce->ue; + + if (ue->core_mode) { /* lpd_channel_stream */ + ret = ff_aac_ldp_parse_channel_stream(ac, usac, ue, gb); + if (ret < 0) + return ret; + } + + if ((nb_channels == 1) || + (che->ch[0].ue.core_mode != che->ch[1].ue.core_mode)) + sce->tns.present = get_bits1(gb); + + /* fd_channel_stream */ + global_gain = get_bits(gb, 8); + + ue->noise.level = 0; + if (ec->noise_fill) { + ue->noise.level = get_bits(gb, 3); + ue->noise.offset = get_bits(gb, 5); + } + + if (!us->common_window) { + /* ics_info() */ + ics->window_sequence[1] = ics->window_sequence[0]; + ics->window_sequence[0] = get_bits(gb, 2); + ics->use_kb_window[1] = ics->use_kb_window[0]; + ics->use_kb_window[0] = get_bits1(gb); + if (ics->window_sequence[0] == EIGHT_SHORT_SEQUENCE) { + ics->max_sfb = get_bits(gb, 4); + ue->scale_factor_grouping = get_bits(gb, 7); + } else { + ics->max_sfb = get_bits(gb, 6); + } + + ret = setup_sce(ac, sce, usac); + if (ret < 0) + return ret; + } + + if (ec->tw_mdct && !us->common_tw) { + /* tw_data() */ + if (get_bits1(gb)) { /* tw_data_present */ + /* Time warping is not supported in baseline profile streams. */ + avpriv_report_missing_feature(ac->avctx, + "AAC USAC timewarping"); + return AVERROR_PATCHWELCOME; + } + } + + ret = decode_usac_scale_factors(ac, sce, gb, global_gain); + if (ret < 0) + return ret; + + ac->dsp.dequant_scalefactors(sce); + + if (sce->tns.present) { + ret = ff_aac_decode_tns(ac, &sce->tns, gb, ics); + if (ret < 0) + return ret; + } + + /* ac_spectral_data */ + arith_reset_flag = indep_flag; + if (!arith_reset_flag) + arith_reset_flag = get_bits1(gb); + + /* Decode coeffs */ + memset(&sce->coeffs[0], 0, 1024*sizeof(float)); + for (int win = 0; win < ics->num_windows; win++) { + int lg = ics->swb_offset[ics->max_sfb]; + int N; + if (ics->window_sequence[0] == EIGHT_SHORT_SEQUENCE) + N = usac->core_frame_len / 8; + else + N = usac->core_frame_len; + + ret = decode_spectrum_and_dequant_ac(ac, sce->coeffs + win*128, gb, + sce->sf, &ue->ac, + arith_reset_flag && (win == 0), + lg, N); + if (ret < 0) + return ret; + } + + if (get_bits1(gb)) { /* fac_data_present */ + const uint16_t len_8 = usac->core_frame_len / 8; + const uint16_t len_16 = usac->core_frame_len / 16; + const uint16_t fac_len = ics->window_sequence[0] == EIGHT_SHORT_SEQUENCE ? len_8 : len_16; + ret = ff_aac_parse_fac_data(ue, gb, 1, fac_len); + if (ret < 0) + return ret; + } + } + + spectrum_decode(ac, usac, che, nb_channels); + + return 0; +} + +static int parse_audio_preroll(AACDecContext *ac, GetBitContext *gb) +{ + int ret = 0; + GetBitContext gbc; + OutputConfiguration *oc = &ac->oc[1]; + MPEG4AudioConfig *m4ac = &oc->m4ac; + MPEG4AudioConfig m4ac_bak = oc->m4ac; + uint8_t temp_data[512]; + uint8_t *tmp_buf = temp_data; + size_t tmp_buf_size = sizeof(temp_data); + + av_unused int crossfade; + int num_preroll_frames; + + int config_len = get_escaped_value(gb, 4, 4, 8); + + /* Implementations are free to pad the config to any length, so use a + * different reader for this. */ + gbc = *gb; + ret = ff_aac_usac_config_decode(ac, ac->avctx, &gbc, oc, m4ac->chan_config); + if (ret < 0) { + *m4ac = m4ac_bak; + return ret; + } else { + ac->oc[1].m4ac.chan_config = 0; + } + + /* 7.18.3.3 Bitrate adaption + * If configuration didn't change after applying preroll, continue + * without decoding it. */ + if (!memcmp(m4ac, &m4ac_bak, sizeof(m4ac_bak))) + return 0; + + skip_bits_long(gb, config_len*8); + + crossfade = get_bits1(gb); /* applyCrossfade */ + skip_bits1(gb); /* reserved */ + num_preroll_frames = get_escaped_value(gb, 2, 4, 0); /* numPreRollFrames */ + + for (int i = 0; i < num_preroll_frames; i++) { + int got_frame_ptr = 0; + int au_len = get_escaped_value(gb, 16, 16, 0); + + if (au_len*8 > tmp_buf_size) { + uint8_t *tmp2; + tmp_buf = tmp_buf == temp_data ? NULL : tmp_buf; + tmp2 = realloc(tmp_buf, au_len*8); + if (!tmp2) { + if (tmp_buf != temp_data) + av_free(tmp_buf); + return AVERROR(ENOMEM); + } + tmp_buf = tmp2; + } + + /* Byte alignment is not guaranteed. */ + for (int i = 0; i < au_len; i++) + tmp_buf[i] = get_bits(gb, 8); + + ret = init_get_bits8(&gbc, tmp_buf, au_len); + if (ret < 0) + break; + + ret = ff_aac_usac_decode_frame(ac->avctx, ac, &gbc, &got_frame_ptr); + if (ret < 0) + break; + } + + if (tmp_buf != temp_data) + av_free(tmp_buf); + + return 0; +} + +static int parse_ext_ele(AACDecContext *ac, AACUsacElemConfig *e, + GetBitContext *gb) +{ + uint8_t *tmp; + uint8_t pl_frag_start = 1; + uint8_t pl_frag_end = 1; + uint32_t len; + + if (!get_bits1(gb)) /* usacExtElementPresent */ + return 0; + + if (get_bits1(gb)) { /* usacExtElementUseDefaultLength */ + len = e->ext.default_len; + } else { + len = get_bits(gb, 8); /* usacExtElementPayloadLength */ + if (len == 255) + len += get_bits(gb, 16) - 2; + } + + if (!len) + return 0; + + if (e->ext.payload_frag) { + pl_frag_start = get_bits1(gb); /* usacExtElementStart */ + pl_frag_end = get_bits1(gb); /* usacExtElementStop */ + } + + if (pl_frag_start) + e->ext.pl_data_offset = 0; + + /* If an extension starts and ends this packet, we can directly use it */ + if (!(pl_frag_start && pl_frag_end)) { + tmp = av_realloc(e->ext.pl_data, e->ext.pl_data_offset + len); + if (!tmp) { + av_free(e->ext.pl_data); + return AVERROR(ENOMEM); + } + e->ext.pl_data = tmp; + + /* Readout data to a buffer */ + for (int i = 0; i < len; i++) + e->ext.pl_data[e->ext.pl_data_offset + i] = get_bits(gb, 8); + } + + e->ext.pl_data_offset += len; + + if (pl_frag_end) { + int ret = 0; + int start_bits = get_bits_count(gb); + const int pl_len = e->ext.pl_data_offset; + GetBitContext *gb2 = gb; + GetBitContext gbc; + if (!(pl_frag_start && pl_frag_end)) { + ret = init_get_bits8(&gbc, e->ext.pl_data, pl_len); + if (ret < 0) + return ret; + + gb2 = &gbc; + } + + switch (e->ext.type) { + case ID_EXT_ELE_FILL: + /* Filler elements have no usable payload */ + break; + case ID_EXT_ELE_AUDIOPREROLL: + ret = parse_audio_preroll(ac, gb2); + break; + default: + /* This should never happen */ + av_assert0(0); + } + av_freep(&e->ext.pl_data); + if (ret < 0) + return ret; + + skip_bits_long(gb, pl_len*8 - (get_bits_count(gb) - start_bits)); + } + + return 0; +} + +int ff_aac_usac_decode_frame(AVCodecContext *avctx, AACDecContext *ac, + GetBitContext *gb, int *got_frame_ptr) +{ + int ret, nb_ch_el, is_dmono = 0; + int indep_flag, samples = 0; + int audio_found = 0, sce_count = 0; + AVFrame *frame = ac->frame; + + ff_aac_output_configure(ac, ac->oc[1].layout_map, ac->oc[1].layout_map_tags, + ac->oc[1].status, 0); + + indep_flag = get_bits1(gb); + + nb_ch_el = 0; + for (int i = 0; i < ac->oc[1].usac.nb_elems; i++) { + AACUsacElemConfig *e = &ac->oc[1].usac.elems[i]; + ChannelElement *che; + + switch (e->type) { + case ID_USAC_LFE: + /* Fallthrough */ + case ID_USAC_SCE: + che = ff_aac_get_che(ac, TYPE_SCE, nb_ch_el++); + if (!che) { + av_log(ac->avctx, AV_LOG_ERROR, + "channel element %d.%d is not allocated\n", + TYPE_SCE, nb_ch_el - 1); + return AVERROR_INVALIDDATA; + } + + ret = decode_usac_core_coder(ac, &ac->oc[1].usac, e, che, gb, + indep_flag, 1); + if (ret < 0) + return ret; + + sce_count++; + audio_found = 1; + che->present = 1; + samples = ac->oc[1].m4ac.frame_length_short ? 768 : 1024; + break; + case ID_USAC_CPE: + che = ff_aac_get_che(ac, TYPE_CPE, nb_ch_el++); + if (!che) { + av_log(ac->avctx, AV_LOG_ERROR, + "channel element %d.%d is not allocated\n", + TYPE_SCE, nb_ch_el - 1); + return AVERROR_INVALIDDATA; + } + + ret = decode_usac_core_coder(ac, &ac->oc[1].usac, e, che, gb, + indep_flag, 2); + if (ret < 0) + return ret; + + audio_found = 1; + che->present = 1; + samples = ac->oc[1].m4ac.frame_length_short ? 768 : 1024; + break; + case ID_USAC_EXT: + ret = parse_ext_ele(ac, e, gb); + if (ret < 0) + return ret; + break; + } + } + + if (ac->oc[1].status && audio_found) { + avctx->sample_rate = ac->oc[1].m4ac.sample_rate; + avctx->frame_size = samples; + ac->oc[1].status = OC_LOCKED; + } + + if (!frame->data[0] && samples) { + av_log(avctx, AV_LOG_ERROR, "no frame data found\n"); + return AVERROR_INVALIDDATA; + } + + if (samples) { + frame->nb_samples = samples; + frame->sample_rate = avctx->sample_rate; + frame->flags = indep_flag ? AV_FRAME_FLAG_KEY : 0x0; + *got_frame_ptr = 1; + } else { + av_frame_unref(ac->frame); + frame->flags = indep_flag ? AV_FRAME_FLAG_KEY : 0x0; + *got_frame_ptr = 0; + } + + /* for dual-mono audio (SCE + SCE) */ + is_dmono = ac->dmono_mode && sce_count == 2 && + !av_channel_layout_compare(&ac->oc[1].ch_layout, + &(AVChannelLayout)AV_CHANNEL_LAYOUT_STEREO); + if (is_dmono) { + if (ac->dmono_mode == 1) + frame->data[1] = frame->data[0]; + else if (ac->dmono_mode == 2) + frame->data[0] = frame->data[1]; + } + + return 0; +} diff --git a/libavcodec/aac/aacdec_usac.h b/libavcodec/aac/aacdec_usac.h new file mode 100644 index 0000000000..4116a2073a --- /dev/null +++ b/libavcodec/aac/aacdec_usac.h @@ -0,0 +1,37 @@ +/* + * Copyright (c) 2024 Lynne <dev@lynne.ee> + * + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + +#ifndef AVCODEC_AAC_AACDEC_USAC_H +#define AVCODEC_AAC_AACDEC_USAC_H + +#include "aacdec.h" + +#include "libavcodec/get_bits.h" + +int ff_aac_usac_config_decode(AACDecContext *ac, AVCodecContext *avctx, + GetBitContext *gb, OutputConfiguration *oc, + int channel_config); + +int ff_aac_usac_reset_state(AACDecContext *ac, OutputConfiguration *oc); + +int ff_aac_usac_decode_frame(AVCodecContext *avctx, AACDecContext *ac, + GetBitContext *gb, int *got_frame_ptr); + +#endif /* AVCODEC_AAC_AACDEC_USAC_H */ |