diff options
Diffstat (limited to 'demux/demux.c')
-rw-r--r-- | demux/demux.c | 638 |
1 files changed, 375 insertions, 263 deletions
diff --git a/demux/demux.c b/demux/demux.c index 75d74d0beb..5997a96ed6 100644 --- a/demux/demux.c +++ b/demux/demux.c @@ -15,19 +15,18 @@ * License along with mpv. If not, see <http://www.gnu.org/licenses/>. */ -#include <stdio.h> -#include <stdlib.h> -#include <string.h> #include <assert.h> -#include <unistd.h> +#include <float.h> #include <limits.h> -#include <pthread.h> -#include <stdint.h> - #include <math.h> - -#include <sys/types.h> +#include <stdatomic.h> +#include <stdint.h> +#include <stdio.h> +#include <stdlib.h> +#include <string.h> #include <sys/stat.h> +#include <sys/types.h> +#include <unistd.h> #include "cache.h" #include "config.h" @@ -38,9 +37,9 @@ #include "common/msg.h" #include "common/global.h" #include "common/recorder.h" +#include "common/stats.h" #include "misc/charset_conv.h" #include "misc/thread_tools.h" -#include "osdep/atomic.h" #include "osdep/timer.h" #include "osdep/threads.h" @@ -82,59 +81,43 @@ static const demuxer_desc_t *const demuxer_list[] = { NULL }; -struct demux_opts { - int enable_cache; - int disk_cache; - int64_t max_bytes; - int64_t max_bytes_bw; - int donate_fw; - double min_secs; - int force_seekable; - double min_secs_cache; - int access_references; - int seekable_cache; - int create_ccs; - char *record_file; - int video_back_preroll; - int audio_back_preroll; - int back_batch[STREAM_TYPE_COUNT]; - double back_seek_size; - char *meta_cp; -}; - #define OPT_BASE_STRUCT struct demux_opts -#define MAX_BYTES MPMIN(INT64_MAX, SIZE_MAX / 2) - static bool get_demux_sub_opts(int index, const struct m_sub_options **sub); const struct m_sub_options demux_conf = { .opts = (const struct m_option[]){ - OPT_CHOICE("cache", enable_cache, 0, - ({"no", 0}, {"auto", -1}, {"yes", 1})), - OPT_FLAG("cache-on-disk", disk_cache, 0), - OPT_DOUBLE("demuxer-readahead-secs", min_secs, M_OPT_MIN, .min = 0), - // (The MAX_BYTES sizes may not be accurate because the max field is - // of double type.) - OPT_BYTE_SIZE("demuxer-max-bytes", max_bytes, 0, 0, MAX_BYTES), - OPT_BYTE_SIZE("demuxer-max-back-bytes", max_bytes_bw, 0, 0, MAX_BYTES), - OPT_FLAG("demuxer-donate-buffer", donate_fw, 0), - OPT_FLAG("force-seekable", force_seekable, 0), - OPT_DOUBLE("cache-secs", min_secs_cache, M_OPT_MIN, .min = 0), - OPT_FLAG("access-references", access_references, 0), - OPT_CHOICE("demuxer-seekable-cache", seekable_cache, 0, - ({"auto", -1}, {"no", 0}, {"yes", 1})), - OPT_FLAG("sub-create-cc-track", create_ccs, 0), - OPT_STRING("stream-record", record_file, 0), - OPT_CHOICE_OR_INT("video-backward-overlap", video_back_preroll, 0, 0, - 1024, ({"auto", -1})), - OPT_CHOICE_OR_INT("audio-backward-overlap", audio_back_preroll, 0, 0, - 1024, ({"auto", -1})), - OPT_INTRANGE("video-backward-batch", back_batch[STREAM_VIDEO], 0, 0, 1024), - OPT_INTRANGE("audio-backward-batch", back_batch[STREAM_AUDIO], 0, 0, 1024), - OPT_DOUBLE("demuxer-backward-playback-step", back_seek_size, M_OPT_MIN, - .min = 0), - OPT_STRING("metadata-codepage", meta_cp, 0), + {"cache", OPT_CHOICE(enable_cache, + {"no", 0}, {"auto", -1}, {"yes", 1})}, + {"cache-on-disk", OPT_BOOL(disk_cache)}, + {"demuxer-readahead-secs", OPT_DOUBLE(min_secs), M_RANGE(0, DBL_MAX)}, + {"demuxer-hysteresis-secs", OPT_DOUBLE(hyst_secs), M_RANGE(0, DBL_MAX)}, + {"demuxer-max-bytes", OPT_BYTE_SIZE(max_bytes), + M_RANGE(0, M_MAX_MEM_BYTES)}, + {"demuxer-max-back-bytes", OPT_BYTE_SIZE(max_bytes_bw), + M_RANGE(0, M_MAX_MEM_BYTES)}, + {"demuxer-donate-buffer", OPT_BOOL(donate_fw)}, + {"force-seekable", OPT_BOOL(force_seekable)}, + {"cache-secs", OPT_DOUBLE(min_secs_cache), M_RANGE(0, DBL_MAX)}, + {"access-references", OPT_BOOL(access_references)}, + {"demuxer-seekable-cache", OPT_CHOICE(seekable_cache, + {"auto", -1}, {"no", 0}, {"yes", 1})}, + {"index", OPT_CHOICE(index_mode, {"default", 1}, {"recreate", 0})}, + {"mf-fps", OPT_DOUBLE(mf_fps)}, + {"mf-type", OPT_STRING(mf_type)}, + {"sub-create-cc-track", OPT_BOOL(create_ccs)}, + {"stream-record", OPT_STRING(record_file)}, + {"video-backward-overlap", OPT_CHOICE(video_back_preroll, {"auto", -1}), + M_RANGE(0, 1024)}, + {"audio-backward-overlap", OPT_CHOICE(audio_back_preroll, {"auto", -1}), + M_RANGE(0, 1024)}, + {"video-backward-batch", OPT_INT(back_batch[STREAM_VIDEO]), + M_RANGE(0, 1024)}, + {"audio-backward-batch", OPT_INT(back_batch[STREAM_AUDIO]), + M_RANGE(0, 1024)}, + {"demuxer-backward-playback-step", OPT_DOUBLE(back_seek_size), + M_RANGE(0, DBL_MAX)}, + {"metadata-codepage", OPT_STRING(meta_cp)}, {0} }, .size = sizeof(struct demux_opts), @@ -142,11 +125,13 @@ const struct m_sub_options demux_conf = { .enable_cache = -1, // auto .max_bytes = 150 * 1024 * 1024, .max_bytes_bw = 50 * 1024 * 1024, - .donate_fw = 1, + .donate_fw = true, .min_secs = 1.0, - .min_secs_cache = 10.0 * 60 * 60, + .min_secs_cache = 1000.0 * 60 * 60, .seekable_cache = -1, - .access_references = 1, + .index_mode = 1, + .mf_fps = 1.0, + .access_references = true, .video_back_preroll = -1, .audio_back_preroll = -1, .back_seek_size = 60, @@ -154,7 +139,7 @@ const struct m_sub_options demux_conf = { [STREAM_VIDEO] = 1, [STREAM_AUDIO] = 10, }, - .meta_cp = "utf-8", + .meta_cp = "auto", }, .get_sub_options = get_demux_sub_opts, }; @@ -162,6 +147,7 @@ const struct m_sub_options demux_conf = { struct demux_internal { struct mp_log *log; struct mpv_global *global; + struct stats_ctx *stats; bool can_cache; // not a slave demuxer; caching makes sense bool can_record; // stream recording is allowed @@ -173,15 +159,12 @@ struct demux_internal { // The lock protects the packet queues (struct demux_stream), // and the fields below. - pthread_mutex_t lock; - pthread_cond_t wakeup; - pthread_t thread; + mp_mutex lock; + mp_cond wakeup; + mp_thread thread; // -- All the following fields are protected by lock. - struct demux_opts *opts; - struct m_config_cache *opts_cache; - bool thread_terminate; bool threading; bool shutdown_async; @@ -202,18 +185,21 @@ struct demux_internal { struct demux_cache *cache; bool warned_queue_overflow; - bool last_eof; // last actual global EOF status - bool eof; // whether we're in EOF state (reset for retry) - bool idle; + bool eof; // whether we're in EOF state double min_secs; + double hyst_secs; // stop reading till there's hyst_secs remaining + bool hyst_active; size_t max_bytes; size_t max_bytes_bw; bool seekable_cache; bool using_network_cache_opts; char *record_filename; - // At least one decoder actually requested data since init or the last seek. - // Do this to allow the decoder thread to select streams before starting. + // Whether the demuxer thread should prefetch packets. This is set to false + // if EOF was reached or the demuxer cache is full. This is also important + // in the initial state: the decoder thread needs to select streams before + // the first packet is read, so this is set to true by packet reading only. + // Reset to false again on EOF or if prefetching is done. bool reading; // Set if we just performed a seek, without reading packets yet. Used to @@ -266,6 +252,7 @@ struct demux_internal { // Cached state. int64_t stream_size; int64_t last_speed_query; + double speed_query_prev_sample; uint64_t bytes_per_second; int64_t next_cache_update; @@ -380,7 +367,7 @@ struct demux_stream { bool eager; // try to keep at least 1 packet queued // if false, this stream is disabled, or passively // read (like subtitles) - bool still_image; // stream has still video images + bool still_image; // stream consists of multiple sparse still images bool refreshing; // finding old position after track switches bool eof; // end of demuxed stream? (true if no more packets) @@ -400,6 +387,7 @@ struct demux_stream { bool skip_to_keyframe; bool attached_picture_added; bool need_wakeup; // call wakeup_cb on next reader_head state change + double force_read_until;// eager=false streams (subs): force read-ahead // For demux_internal.dumper. Currently, this is used only temporarily // during blocking dumping. @@ -442,7 +430,7 @@ struct demux_stream { static void switch_to_fresh_cache_range(struct demux_internal *in); static void demuxer_sort_chapters(demuxer_t *demuxer); -static void *demux_thread(void *pctx); +static MP_THREAD_VOID demux_thread(void *pctx); static void update_cache(struct demux_internal *in); static void add_packet_locked(struct sh_stream *stream, demux_packet_t *dp); static struct demux_packet *advance_reader_head(struct demux_stream *ds); @@ -458,7 +446,7 @@ static void prune_old_packets(struct demux_internal *in); static void dumper_close(struct demux_internal *in); static void demux_convert_tags_charset(struct demuxer *demuxer); -static uint64_t get_foward_buffered_bytes(struct demux_stream *ds) +static uint64_t get_forward_buffered_bytes(struct demux_stream *ds) { if (!ds->reader_head) return 0; @@ -528,7 +516,7 @@ static void check_queue_consistency(struct demux_internal *in) // ...reader_head and others must be in the queue. assert(is_forward == !!queue->ds->reader_head); assert(kf_found == !!queue->keyframe_latest); - uint64_t fw_bytes2 = get_foward_buffered_bytes(queue->ds); + uint64_t fw_bytes2 = get_forward_buffered_bytes(queue->ds); assert(fw_bytes == fw_bytes2); } @@ -626,7 +614,9 @@ static void update_seek_ranges(struct demux_cached_range *range) range->is_bof &= queue->is_bof; bool empty = queue->is_eof && !queue->head; - if (queue->seek_start >= queue->seek_end && !empty) + if (queue->seek_start >= queue->seek_end && !empty && + !(queue->seek_start == queue->seek_end && + queue->seek_start != MP_NOPTS_VALUE)) goto broken; } } @@ -673,7 +663,7 @@ static void update_seek_ranges(struct demux_cached_range *range) } } - if (range->seek_start >= range->seek_end) + if (range->seek_start >= range->seek_end && !(range->is_bof && range->is_eof)) goto broken; prune_metadata(range); @@ -824,6 +814,7 @@ static void ds_clear_reader_state(struct demux_stream *ds, ds->attached_picture_added = false; ds->last_ret_pos = -1; ds->last_ret_dts = MP_NOPTS_VALUE; + ds->force_read_until = MP_NOPTS_VALUE; if (clear_back_state) { ds->back_restart_pos = -1; @@ -865,7 +856,7 @@ static void wakeup_ds(struct demux_stream *ds) ds->in->wakeup_cb(ds->in->wakeup_cb_ctx); } ds->need_wakeup = false; - pthread_cond_signal(&ds->in->wakeup); + mp_cond_signal(&ds->in->wakeup); } } @@ -928,9 +919,9 @@ static void update_stream_selection_state(struct demux_internal *in, void demux_set_ts_offset(struct demuxer *demuxer, double offset) { struct demux_internal *in = demuxer->in; - pthread_mutex_lock(&in->lock); + mp_mutex_lock(&in->lock); in->ts_offset = offset; - pthread_mutex_unlock(&in->lock); + mp_mutex_unlock(&in->lock); } static void add_missing_streams(struct demux_internal *in, @@ -962,6 +953,7 @@ struct sh_stream *demux_alloc_sh_stream(enum stream_type type) .index = -1, .ff_index = -1, // may be overwritten by demuxer .demuxer_id = -1, // ... same + .program_id = -1, // ... same .codec = talloc_zero(sh, struct mp_codec_params), .tags = talloc_zero(sh, struct mp_tags), }; @@ -1011,7 +1003,7 @@ static void demux_add_sh_stream_locked(struct demux_internal *in, switch (ds->type) { case STREAM_AUDIO: - ds->back_preroll = in->opts->audio_back_preroll; + ds->back_preroll = in->d_user->opts->audio_back_preroll; if (ds->back_preroll < 0) { // auto ds->back_preroll = mp_codec_is_lossless(sh->codec->codec) ? 0 : 1; if (sh->codec->codec && (strcmp(sh->codec->codec, "opus") == 0 || @@ -1021,7 +1013,7 @@ static void demux_add_sh_stream_locked(struct demux_internal *in, } break; case STREAM_VIDEO: - ds->back_preroll = in->opts->video_back_preroll; + ds->back_preroll = in->d_user->opts->video_back_preroll; if (ds->back_preroll < 0) ds->back_preroll = 0; // auto break; @@ -1043,9 +1035,9 @@ void demux_add_sh_stream(struct demuxer *demuxer, struct sh_stream *sh) { struct demux_internal *in = demuxer->in; assert(demuxer == in->d_thread); - pthread_mutex_lock(&in->lock); + mp_mutex_lock(&in->lock); demux_add_sh_stream_locked(in, sh); - pthread_mutex_unlock(&in->lock); + mp_mutex_unlock(&in->lock); } // Return a stream with the given index. Since streams can only be added during @@ -1055,10 +1047,10 @@ void demux_add_sh_stream(struct demuxer *demuxer, struct sh_stream *sh) struct sh_stream *demux_get_stream(struct demuxer *demuxer, int index) { struct demux_internal *in = demuxer->in; - pthread_mutex_lock(&in->lock); + mp_mutex_lock(&in->lock); assert(index >= 0 && index < in->num_streams); struct sh_stream *r = in->streams[index]; - pthread_mutex_unlock(&in->lock); + mp_mutex_unlock(&in->lock); return r; } @@ -1066,9 +1058,9 @@ struct sh_stream *demux_get_stream(struct demuxer *demuxer, int index) int demux_get_num_stream(struct demuxer *demuxer) { struct demux_internal *in = demuxer->in; - pthread_mutex_lock(&in->lock); + mp_mutex_lock(&in->lock); int r = in->num_streams; - pthread_mutex_unlock(&in->lock); + mp_mutex_unlock(&in->lock); return r; } @@ -1107,8 +1099,8 @@ static void demux_dealloc(struct demux_internal *in) { for (int n = 0; n < in->num_streams; n++) talloc_free(in->streams[n]); - pthread_mutex_destroy(&in->lock); - pthread_cond_destroy(&in->wakeup); + mp_mutex_destroy(&in->lock); + mp_cond_destroy(&in->wakeup); talloc_free(in->d_user); } @@ -1139,11 +1131,11 @@ struct demux_free_async_state *demux_free_async(struct demuxer *demuxer) if (!in->threading) return NULL; - pthread_mutex_lock(&in->lock); + mp_mutex_lock(&in->lock); in->thread_terminate = true; in->shutdown_async = true; - pthread_cond_signal(&in->wakeup); - pthread_mutex_unlock(&in->lock); + mp_cond_signal(&in->wakeup); + mp_mutex_unlock(&in->lock); return (struct demux_free_async_state *)demuxer->in; // lies } @@ -1167,9 +1159,9 @@ bool demux_free_async_finish(struct demux_free_async_state *state) { struct demux_internal *in = (struct demux_internal *)state; // reverse lies - pthread_mutex_lock(&in->lock); + mp_mutex_lock(&in->lock); bool busy = in->shutdown_async; - pthread_mutex_unlock(&in->lock); + mp_mutex_unlock(&in->lock); if (busy) return false; @@ -1198,7 +1190,7 @@ void demux_start_thread(struct demuxer *demuxer) if (!in->threading) { in->threading = true; - if (pthread_create(&in->thread, NULL, demux_thread, in)) + if (mp_thread_create(&in->thread, demux_thread, in)) in->threading = false; } } @@ -1209,11 +1201,11 @@ void demux_stop_thread(struct demuxer *demuxer) assert(demuxer == in->d_user); if (in->threading) { - pthread_mutex_lock(&in->lock); + mp_mutex_lock(&in->lock); in->thread_terminate = true; - pthread_cond_signal(&in->wakeup); - pthread_mutex_unlock(&in->lock); - pthread_join(in->thread, NULL); + mp_cond_signal(&in->wakeup); + mp_mutex_unlock(&in->lock); + mp_thread_join(in->thread); in->threading = false; in->thread_terminate = false; } @@ -1223,10 +1215,10 @@ void demux_stop_thread(struct demuxer *demuxer) void demux_set_wakeup_cb(struct demuxer *demuxer, void (*cb)(void *ctx), void *ctx) { struct demux_internal *in = demuxer->in; - pthread_mutex_lock(&in->lock); + mp_mutex_lock(&in->lock); in->wakeup_cb = cb; in->wakeup_cb_ctx = ctx; - pthread_mutex_unlock(&in->lock); + mp_mutex_unlock(&in->lock); } void demux_start_prefetch(struct demuxer *demuxer) @@ -1234,10 +1226,10 @@ void demux_start_prefetch(struct demuxer *demuxer) struct demux_internal *in = demuxer->in; assert(demuxer == in->d_user); - pthread_mutex_lock(&in->lock); + mp_mutex_lock(&in->lock); in->reading = true; - pthread_cond_signal(&in->wakeup); - pthread_mutex_unlock(&in->lock); + mp_cond_signal(&in->wakeup); + mp_mutex_unlock(&in->lock); } const char *stream_type_name(enum stream_type type) @@ -1260,6 +1252,8 @@ static struct sh_stream *demuxer_get_cc_track_locked(struct sh_stream *stream) return NULL; sh->codec->codec = "eia_608"; sh->default_track = true; + sh->hls_bitrate = stream->hls_bitrate; + sh->program_id = stream->program_id; stream->ds->cc = sh; demux_add_sh_stream_locked(stream->ds->in, sh); sh->ds->ignore_eof = true; @@ -1272,10 +1266,10 @@ void demuxer_feed_caption(struct sh_stream *stream, demux_packet_t *dp) { struct demux_internal *in = stream->ds->in; - pthread_mutex_lock(&in->lock); + mp_mutex_lock(&in->lock); struct sh_stream *sh = demuxer_get_cc_track_locked(stream); if (!sh) { - pthread_mutex_unlock(&in->lock); + mp_mutex_unlock(&in->lock); talloc_free(dp); return; } @@ -1285,7 +1279,7 @@ void demuxer_feed_caption(struct sh_stream *stream, demux_packet_t *dp) dp->dts = MP_ADD_PTS(dp->dts, -in->ts_offset); dp->stream = sh->index; add_packet_locked(sh, dp); - pthread_mutex_unlock(&in->lock); + mp_mutex_unlock(&in->lock); } static void error_on_backward_demuxing(struct demux_internal *in) @@ -1322,8 +1316,8 @@ static void perform_backward_seek(struct demux_internal *in) in->reading = true; // Don't starve other threads. - pthread_mutex_unlock(&in->lock); - pthread_mutex_lock(&in->lock); + mp_mutex_unlock(&in->lock); + mp_mutex_lock(&in->lock); } // For incremental backward demuxing search work. @@ -1447,7 +1441,7 @@ static void find_backward_restart_pos(struct demux_stream *ds) // Number of renderable keyframes to return to user. // (Excludes preroll, which is decoded by user, but then discarded.) - int batch = MPMAX(in->opts->back_batch[ds->type], 1); + int batch = MPMAX(in->d_user->opts->back_batch[ds->type], 1); // Number of keyframes to return to the user in total. int total = batch + ds->back_preroll; @@ -1494,7 +1488,7 @@ static void find_backward_restart_pos(struct demux_stream *ds) // Or if preroll is involved, the first preroll packet. while (ds->reader_head != target) { if (!advance_reader_head(ds)) - assert(0); // target must be in list + MP_ASSERT_UNREACHABLE(); // target must be in list } double seek_pts; @@ -1557,8 +1551,9 @@ resume_earlier: ds->reader_head = t; ds->back_need_recheck = true; in->back_any_need_recheck = true; + mp_cond_signal(&in->wakeup); } else { - ds->back_seek_pos -= in->opts->back_seek_size; + ds->back_seek_pos -= in->d_user->opts->back_seek_size; in->need_back_seek = true; } } @@ -1569,7 +1564,7 @@ static void back_demux_see_packets(struct demux_stream *ds) { struct demux_internal *in = ds->in; - if (!ds->selected || !in->back_demuxing) + if (!ds->selected || !in->back_demuxing || !ds->eager) return; assert(!(ds->back_resuming && ds->back_restarting)); @@ -1672,7 +1667,7 @@ static void attempt_range_joining(struct demux_internal *in) // Try to find a join point, where packets obviously overlap. (It would be // better and faster to do this incrementally, but probably too complex.) // The current range can overlap arbitrarily with the next one, not only by - // by the seek overlap, but for arbitrary packet readahead as well. + // the seek overlap, but for arbitrary packet readahead as well. // We also drop the overlapping packets (if joining fails, we discard the // entire next range anyway, so this does no harm). for (int n = 0; n < in->num_streams; n++) { @@ -1844,7 +1839,8 @@ static struct demux_packet *compute_keyframe_times(struct demux_packet *pkt, break; double ts = MP_PTS_OR_DEF(pkt->pts, pkt->dts); - if (pkt->segmented && (ts < pkt->start || ts > pkt->end)) + if (pkt->segmented && ((pkt->start != MP_NOPTS_VALUE && ts < pkt->start) || + (pkt->end != MP_NOPTS_VALUE && ts > pkt->end))) ts = MP_NOPTS_VALUE; min = MP_PTS_MIN(min, ts); @@ -1913,7 +1909,8 @@ static void adjust_seek_range_on_packet(struct demux_stream *ds, queue->keyframe_latest = dp; } - if (update_ranges) { + // Adding a sparse packet never changes the seek range. + if (update_ranges && ds->eager) { update_seek_ranges(queue->range); attempt_range_joining(ds->in); } @@ -1929,9 +1926,18 @@ static struct mp_recorder *recorder_create(struct demux_internal *in, if (stream->ds->selected) MP_TARRAY_APPEND(NULL, streams, num_streams, stream); } + + struct demuxer *demuxer = in->d_thread; + struct demux_attachment **attachments = talloc_array(NULL, struct demux_attachment*, demuxer->num_attachments); + for (int n = 0; n < demuxer->num_attachments; n++) { + attachments[n] = &demuxer->attachments[n]; + } + struct mp_recorder *res = mp_recorder_create(in->d_thread->global, dst, - streams, num_streams); + streams, num_streams, + attachments, demuxer->num_attachments); talloc_free(streams); + talloc_free(attachments); return res; } @@ -1954,13 +1960,13 @@ static void record_packet(struct demux_internal *in, struct demux_packet *dp) { // (should preferably be outside of the lock) if (in->enable_recording && !in->recorder && - in->opts->record_file && in->opts->record_file[0]) + in->d_user->opts->record_file && in->d_user->opts->record_file[0]) { // Later failures shouldn't make it retry and overwrite the previously // recorded file. in->enable_recording = false; - in->recorder = recorder_create(in, in->opts->record_file); + in->recorder = recorder_create(in, in->d_user->opts->record_file); if (!in->recorder) MP_ERR(in, "Disabling recording.\n"); } @@ -1984,6 +1990,7 @@ static void record_packet(struct demux_internal *in, struct demux_packet *dp) static void add_packet_locked(struct sh_stream *stream, demux_packet_t *dp) { struct demux_stream *ds = stream ? stream->ds : NULL; + assert(ds && ds->in); if (!dp->len || demux_cancel_test(ds->in->d_thread)) { talloc_free(dp); return; @@ -2011,7 +2018,8 @@ static void add_packet_locked(struct sh_stream *stream, demux_packet_t *dp) if (!drop) { // If libavformat splits packets, some packets will have pos unset, so // make up one based on the first packet => makes refresh seeks work. - if (dp->pos < 0 && !dp->keyframe && queue->last_pos_fixup >= 0) + if ((dp->pos < 0 || dp->pos == queue->last_pos_fixup) && + !dp->keyframe && queue->last_pos_fixup >= 0) dp->pos = queue->last_pos_fixup + 1; queue->last_pos_fixup = dp->pos; } @@ -2038,7 +2046,7 @@ static void add_packet_locked(struct sh_stream *stream, demux_packet_t *dp) record_packet(in, dp); - if (in->cache && in->opts->disk_cache) { + if (in->cache && in->d_user->opts->disk_cache) { int64_t pos = demux_cache_write(in->cache, dp); if (pos >= 0) { demux_packet_unref_contents(dp); @@ -2078,7 +2086,7 @@ static void add_packet_locked(struct sh_stream *stream, demux_packet_t *dp) if (!ds->ignore_eof) { // obviously not true anymore ds->eof = false; - in->last_eof = in->eof = false; + in->eof = false; } // For video, PTS determination is not trivial, but for other media types @@ -2092,7 +2100,7 @@ static void add_packet_locked(struct sh_stream *stream, demux_packet_t *dp) ds->base_ts = queue->last_ts; const char *num_pkts = queue->head == queue->tail ? "1" : ">1"; - uint64_t fw_bytes = get_foward_buffered_bytes(ds); + uint64_t fw_bytes = get_forward_buffered_bytes(ds); MP_TRACE(in, "append packet to %s: size=%zu pts=%f dts=%f pos=%"PRIi64" " "[num=%s size=%zd]\n", stream_type_name(stream->type), dp->len, dp->pts, dp->dts, dp->pos, num_pkts, (size_t)fw_bytes); @@ -2138,13 +2146,24 @@ static void mark_stream_eof(struct demux_stream *ds) } } +static bool lazy_stream_needs_wait(struct demux_stream *ds) +{ + struct demux_internal *in = ds->in; + // Attempt to read until force_read_until was reached, or reading has + // stopped for some reason (true EOF, queue overflow). + return !ds->eager && !in->back_demuxing && + !in->eof && ds->force_read_until != MP_NOPTS_VALUE && + (in->demux_ts == MP_NOPTS_VALUE || + in->demux_ts <= ds->force_read_until); +} + // Returns true if there was "progress" (lock was released temporarily). static bool read_packet(struct demux_internal *in) { - in->eof = false; - in->idle = true; + bool was_reading = in->reading; + in->reading = false; - if (!in->reading || in->blocked || demux_cancel_test(in->d_thread)) + if (!was_reading || in->blocked || demux_cancel_test(in->d_thread)) return false; // Check if we need to read a new packet. We do this if all queues are below @@ -2158,22 +2177,35 @@ static bool read_packet(struct demux_internal *in) read_more |= !ds->reader_head; if (in->back_demuxing) read_more |= ds->back_restarting || ds->back_resuming; + } else { + if (lazy_stream_needs_wait(ds)) { + read_more = true; + } else { + mark_stream_eof(ds); // let playback continue + } } refresh_more |= ds->refreshing; if (ds->eager && ds->queue->last_ts != MP_NOPTS_VALUE && in->min_secs > 0 && ds->base_ts != MP_NOPTS_VALUE && ds->queue->last_ts >= ds->base_ts && !in->back_demuxing) - prefetch_more |= ds->queue->last_ts - ds->base_ts < in->min_secs; - total_fw_bytes += get_foward_buffered_bytes(ds); + { + if (ds->queue->last_ts - ds->base_ts <= in->hyst_secs) + in->hyst_active = false; + if (!in->hyst_active) + prefetch_more |= ds->queue->last_ts - ds->base_ts < in->min_secs; + } + total_fw_bytes += get_forward_buffered_bytes(ds); } MP_TRACE(in, "bytes=%zd, read_more=%d prefetch_more=%d, refresh_more=%d\n", (size_t)total_fw_bytes, read_more, prefetch_more, refresh_more); if (total_fw_bytes >= in->max_bytes) { // if we hit the limit just by prefetching, simply stop prefetching - if (!read_more) + if (!read_more) { + in->hyst_active = !!in->hyst_secs; return false; + } if (!in->warned_queue_overflow) { in->warned_queue_overflow = true; MP_WARN(in, "Too many packets in the demuxer packet queues:\n"); @@ -2184,7 +2216,7 @@ static bool read_packet(struct demux_internal *in) for (struct demux_packet *dp = ds->reader_head; dp; dp = dp->next) num_pkts++; - uint64_t fw_bytes = get_foward_buffered_bytes(ds); + uint64_t fw_bytes = get_forward_buffered_bytes(ds); MP_WARN(in, " %s/%d: %zd packets, %zd bytes%s%s\n", stream_type_name(ds->type), n, num_pkts, (size_t)fw_bytes, @@ -2203,8 +2235,10 @@ static bool read_packet(struct demux_internal *in) return false; } - if (!read_more && !prefetch_more && !refresh_more) + if (!read_more && !prefetch_more && !refresh_more) { + in->hyst_active = !!in->hyst_secs; return false; + } if (in->after_seek_to_start) { for (int n = 0; n < in->num_streams; n++) { @@ -2216,10 +2250,10 @@ static bool read_packet(struct demux_internal *in) // Actually read a packet. Drop the lock while doing so, because waiting // for disk or network I/O can take time. - in->idle = false; + in->reading = true; in->after_seek = false; in->after_seek_to_start = false; - pthread_mutex_unlock(&in->lock); + mp_mutex_unlock(&in->lock); struct demuxer *demux = in->d_thread; struct demux_packet *pkt = NULL; @@ -2228,7 +2262,7 @@ static bool read_packet(struct demux_internal *in) if (demux->desc->read_packet && !demux_cancel_test(demux)) eof = !demux->desc->read_packet(demux, &pkt); - pthread_mutex_lock(&in->lock); + mp_mutex_lock(&in->lock); update_cache(in); if (pkt) { @@ -2241,14 +2275,15 @@ static bool read_packet(struct demux_internal *in) for (int n = 0; n < in->num_streams; n++) mark_stream_eof(in->streams[n]->ds); // If we had EOF previously, then don't wakeup (avoids wakeup loop) - if (!in->last_eof) { + if (!in->eof) { if (in->wakeup_cb) in->wakeup_cb(in->wakeup_cb_ctx); - pthread_cond_signal(&in->wakeup); + mp_cond_signal(&in->wakeup); MP_VERBOSE(in, "EOF reached.\n"); } } - in->eof = in->last_eof = eof; + in->eof = eof; + in->reading = !eof; } return true; } @@ -2264,12 +2299,12 @@ static void prune_old_packets(struct demux_internal *in) uint64_t fw_bytes = 0; for (int n = 0; n < in->num_streams; n++) { struct demux_stream *ds = in->streams[n]->ds; - fw_bytes += get_foward_buffered_bytes(ds); + fw_bytes += get_forward_buffered_bytes(ds); } uint64_t max_avail = in->max_bytes_bw; // Backward cache (if enabled at all) can use unused forward cache. // Still leave 1 byte free, so the read_packet logic doesn't get stuck. - if (max_avail && in->max_bytes > (fw_bytes + 1) && in->opts->donate_fw) + if (max_avail && in->max_bytes > (fw_bytes + 1) && in->d_user->opts->donate_fw) max_avail += in->max_bytes - (fw_bytes + 1); if (in->total_bytes - fw_bytes <= max_avail) break; @@ -2372,23 +2407,19 @@ static void execute_trackswitch(struct demux_internal *in) { in->tracks_switched = false; - bool any_selected = false; - for (int n = 0; n < in->num_streams; n++) - any_selected |= in->streams[n]->ds->selected; - - pthread_mutex_unlock(&in->lock); + mp_mutex_unlock(&in->lock); if (in->d_thread->desc->switched_tracks) in->d_thread->desc->switched_tracks(in->d_thread); - pthread_mutex_lock(&in->lock); + mp_mutex_lock(&in->lock); } static void execute_seek(struct demux_internal *in) { int flags = in->seek_flags; double pts = in->seek_pts; - in->last_eof = in->eof = false; + in->eof = false; in->seeking = false; in->seeking_in_progress = pts; in->demux_ts = MP_NOPTS_VALUE; @@ -2404,7 +2435,7 @@ static void execute_seek(struct demux_internal *in) if (in->recorder) mp_recorder_mark_discontinuity(in->recorder); - pthread_mutex_unlock(&in->lock); + mp_mutex_unlock(&in->lock); MP_VERBOSE(in, "execute seek (to %f flags %d)\n", pts, flags); @@ -2413,16 +2444,18 @@ static void execute_seek(struct demux_internal *in) MP_VERBOSE(in, "seek done\n"); - pthread_mutex_lock(&in->lock); + mp_mutex_lock(&in->lock); in->seeking_in_progress = MP_NOPTS_VALUE; } -static void update_opts(struct demux_internal *in) +static void update_opts(struct demuxer *demuxer) { - struct demux_opts *opts = in->opts; + struct demux_opts *opts = demuxer->opts; + struct demux_internal *in = demuxer->in; in->min_secs = opts->min_secs; + in->hyst_secs = opts->hyst_secs; in->max_bytes = opts->max_bytes; in->max_bytes_bw = opts->max_bytes_bw; @@ -2485,8 +2518,8 @@ static void update_opts(struct demux_internal *in) // Make demuxing progress. Return whether progress was made. static bool thread_work(struct demux_internal *in) { - if (m_config_cache_update(in->opts_cache)) - update_opts(in); + if (m_config_cache_update(in->d_user->opts_cache)) + update_opts(in->d_user); if (in->tracks_switched) { execute_trackswitch(in); return true; @@ -2503,42 +2536,43 @@ static bool thread_work(struct demux_internal *in) execute_seek(in); return true; } - if (!in->eof) { - if (read_packet(in)) - return true; // read_packet unlocked, so recheck conditions - } - if (mp_time_us() >= in->next_cache_update) { + if (read_packet(in)) + return true; // read_packet unlocked, so recheck conditions + if (mp_time_ns() >= in->next_cache_update) { update_cache(in); return true; } return false; } -static void *demux_thread(void *pctx) +static MP_THREAD_VOID demux_thread(void *pctx) { struct demux_internal *in = pctx; - mpthread_set_name("demux"); - pthread_mutex_lock(&in->lock); + mp_thread_set_name("demux"); + mp_mutex_lock(&in->lock); + + stats_register_thread_cputime(in->stats, "thread"); while (!in->thread_terminate) { if (thread_work(in)) continue; - pthread_cond_signal(&in->wakeup); - struct timespec until = mp_time_us_to_timespec(in->next_cache_update); - pthread_cond_timedwait(&in->wakeup, &in->lock, &until); + mp_cond_signal(&in->wakeup); + mp_cond_timedwait_until(&in->wakeup, &in->lock, in->next_cache_update); } if (in->shutdown_async) { - pthread_mutex_unlock(&in->lock); + mp_mutex_unlock(&in->lock); demux_shutdown(in); - pthread_mutex_lock(&in->lock); + mp_mutex_lock(&in->lock); in->shutdown_async = false; if (in->wakeup_cb) in->wakeup_cb(in->wakeup_cb_ctx); } - pthread_mutex_unlock(&in->lock); - return NULL; + stats_unregister_thread(in->stats, "thre |