123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410 |
- import urllib.parse
- from .common import InfoExtractor
- from ..utils import (
- ExtractorError,
- float_or_none,
- int_or_none,
- parse_qs,
- traverse_obj,
- try_get,
- unescapeHTML,
- update_url_query,
- url_or_none,
- urlencode_postdata,
- )
- class RedditIE(InfoExtractor):
- _NETRC_MACHINE = 'reddit'
- _VALID_URL = r'https?://(?P<host>(?:\w+\.)?reddit(?:media)?\.com)/(?P<slug>(?:(?:r|user)/[^/]+/)?comments/(?P<id>[^/?#&]+))'
- _TESTS = [{
- 'url': 'https://www.reddit.com/r/videos/comments/6rrwyj/that_small_heart_attack/',
- 'info_dict': {
- 'id': 'zv89llsvexdz',
- 'ext': 'mp4',
- 'display_id': '6rrwyj',
- 'title': 'That small heart attack.',
- 'thumbnail': r're:^https?://.*\.(?:jpg|png)',
- 'thumbnails': 'count:4',
- 'timestamp': 1501941939,
- 'upload_date': '20170805',
- 'uploader': 'Antw87',
- 'duration': 12,
- 'like_count': int,
- 'dislike_count': int,
- 'comment_count': int,
- 'age_limit': 0,
- 'channel_id': 'videos',
- },
- 'params': {
- 'skip_download': True,
- },
- }, {
- # 1080p fallback format
- 'url': 'https://www.reddit.com/r/aww/comments/90bu6w/heat_index_was_110_degrees_so_we_offered_him_a/',
- 'md5': '8b5902cfda3006bf90faea7adf765a49',
- 'info_dict': {
- 'id': 'gyh95hiqc0b11',
- 'ext': 'mp4',
- 'display_id': '90bu6w',
- 'title': 'Heat index was 110 degrees so we offered him a cold drink. He went for a full body soak instead',
- 'thumbnail': r're:^https?://.*\.(?:jpg|png)',
- 'thumbnails': 'count:7',
- 'timestamp': 1532051078,
- 'upload_date': '20180720',
- 'uploader': 'FootLoosePickleJuice',
- 'duration': 14,
- 'like_count': int,
- 'dislike_count': int,
- 'comment_count': int,
- 'age_limit': 0,
- 'channel_id': 'aww',
- },
- }, {
- # User post
- 'url': 'https://www.reddit.com/user/creepyt0es/comments/nip71r/i_plan_to_make_more_stickers_and_prints_check/',
- 'info_dict': {
- 'id': 'zasobba6wp071',
- 'ext': 'mp4',
- 'display_id': 'nip71r',
- 'title': 'I plan to make more stickers and prints! Check them out on my Etsy! Or get them through my Patreon. Links below.',
- 'thumbnail': r're:^https?://.*\.(?:jpg|png)',
- 'thumbnails': 'count:5',
- 'timestamp': 1621709093,
- 'upload_date': '20210522',
- 'uploader': 'creepyt0es',
- 'duration': 6,
- 'like_count': int,
- 'dislike_count': int,
- 'comment_count': int,
- 'age_limit': 18,
- 'channel_id': 'u_creepyt0es',
- },
- 'params': {
- 'skip_download': True,
- },
- }, {
- # videos embedded in reddit text post
- 'url': 'https://www.reddit.com/r/KamenRider/comments/wzqkxp/finale_kamen_rider_revice_episode_50_family_to/',
- 'playlist_count': 2,
- 'info_dict': {
- 'id': 'wzqkxp',
- 'title': 'md5:72d3d19402aa11eff5bd32fc96369b37',
- },
- }, {
- # crossposted reddit-hosted media
- 'url': 'https://www.reddit.com/r/dumbfuckers_club/comments/zjjw82/cringe/',
- 'md5': '746180895c7b75a9d6b05341f507699a',
- 'info_dict': {
- 'id': 'a1oneun6pa5a1',
- 'ext': 'mp4',
- 'display_id': 'zjjw82',
- 'title': 'Cringe',
- 'uploader': 'Otaku-senpai69420',
- 'thumbnail': r're:^https?://.*\.(?:jpg|png)',
- 'upload_date': '20221212',
- 'timestamp': 1670812309,
- 'duration': 16,
- 'like_count': int,
- 'dislike_count': int,
- 'comment_count': int,
- 'age_limit': 0,
- 'channel_id': 'dumbfuckers_club',
- },
- }, {
- # post link without subreddit
- 'url': 'https://www.reddit.com/comments/124pp33',
- 'md5': '15eec9d828adcef4468b741a7e45a395',
- 'info_dict': {
- 'id': 'antsenjc2jqa1',
- 'ext': 'mp4',
- 'display_id': '124pp33',
- 'title': 'Harmless prank of some old friends',
- 'uploader': 'Dudezila',
- 'channel_id': 'ContagiousLaughter',
- 'duration': 17,
- 'upload_date': '20230328',
- 'timestamp': 1680012043,
- 'thumbnail': r're:^https?://.*\.(?:jpg|png)',
- 'age_limit': 0,
- 'comment_count': int,
- 'dislike_count': int,
- 'like_count': int,
- },
- }, {
- # quarantined subreddit post
- 'url': 'https://old.reddit.com/r/GenZedong/comments/12fujy3/based_hasan/',
- 'md5': '3156ea69e3c1f1b6259683c5abd36e71',
- 'info_dict': {
- 'id': '8bwtclfggpsa1',
- 'ext': 'mp4',
- 'display_id': '12fujy3',
- 'title': 'Based Hasan?',
- 'uploader': 'KingNigelXLII',
- 'channel_id': 'GenZedong',
- 'duration': 16,
- 'upload_date': '20230408',
- 'timestamp': 1680979138,
- 'age_limit': 0,
- 'comment_count': int,
- 'dislike_count': int,
- 'like_count': int,
- },
- 'skip': 'Requires account that has opted-in to the GenZedong subreddit',
- }, {
- # subtitles in HLS manifest
- 'url': 'https://www.reddit.com/r/Unexpected/comments/1cl9h0u/the_insurance_claim_will_be_interesting/',
- 'info_dict': {
- 'id': 'a2mdj5d57qyc1',
- 'ext': 'mp4',
- 'display_id': '1cl9h0u',
- 'title': 'The insurance claim will be interesting',
- 'uploader': 'darrenpauli',
- 'channel_id': 'Unexpected',
- 'duration': 53,
- 'upload_date': '20240506',
- 'timestamp': 1714966382,
- 'age_limit': 0,
- 'comment_count': int,
- 'dislike_count': int,
- 'like_count': int,
- 'subtitles': {'en': 'mincount:1'},
- },
- 'params': {
- 'skip_download': True,
- },
- }, {
- # subtitles from caption-url
- 'url': 'https://www.reddit.com/r/soccer/comments/1cxwzso/tottenham_1_0_newcastle_united_james_maddison_31/',
- 'info_dict': {
- 'id': 'xbmj4t3igy1d1',
- 'ext': 'mp4',
- 'display_id': '1cxwzso',
- 'title': 'Tottenham [1] - 0 Newcastle United - James Maddison 31\'',
- 'uploader': 'Woodstovia',
- 'channel_id': 'soccer',
- 'duration': 30,
- 'upload_date': '20240522',
- 'timestamp': 1716373798,
- 'age_limit': 0,
- 'comment_count': int,
- 'dislike_count': int,
- 'like_count': int,
- 'subtitles': {'en': 'mincount:1'},
- },
- 'params': {
- 'skip_download': True,
- 'writesubtitles': True,
- },
- }, {
- 'url': 'https://www.reddit.com/r/videos/comments/6rrwyj',
- 'only_matching': True,
- }, {
- # imgur
- 'url': 'https://www.reddit.com/r/MadeMeSmile/comments/6t7wi5/wait_for_it/',
- 'only_matching': True,
- }, {
- # imgur @ old reddit
- 'url': 'https://old.reddit.com/r/MadeMeSmile/comments/6t7wi5/wait_for_it/',
- 'only_matching': True,
- }, {
- # streamable
- 'url': 'https://www.reddit.com/r/videos/comments/6t7sg9/comedians_hilarious_joke_about_the_guam_flag/',
- 'only_matching': True,
- }, {
- # youtube
- 'url': 'https://www.reddit.com/r/videos/comments/6t75wq/southern_man_tries_to_speak_without_an_accent/',
- 'only_matching': True,
- }, {
- # reddit video @ nm reddit
- 'url': 'https://nm.reddit.com/r/Cricket/comments/8idvby/lousy_cameraman_finds_himself_in_cairns_line_of/',
- 'only_matching': True,
- }, {
- 'url': 'https://www.redditmedia.com/r/serbia/comments/pu9wbx/ako_vu%C4%8Di%C4%87_izgubi_izbore_ja_%C4%87u_da_crknem/',
- 'only_matching': True,
- }]
- def _perform_login(self, username, password):
- captcha = self._download_json(
- 'https://www.reddit.com/api/requires_captcha/login.json', None,
- 'Checking login requirement')['required']
- if captcha:
- raise ExtractorError('Reddit is requiring captcha before login', expected=True)
- login = self._download_json(
- f'https://www.reddit.com/api/login/{username}', None, data=urlencode_postdata({
- 'op': 'login-main',
- 'user': username,
- 'passwd': password,
- 'api_type': 'json',
- }), note='Logging in', errnote='Login request failed')
- errors = '; '.join(traverse_obj(login, ('json', 'errors', ..., 1)))
- if errors:
- raise ExtractorError(f'Unable to login, Reddit API says {errors}', expected=True)
- elif not traverse_obj(login, ('json', 'data', 'cookie', {str})):
- raise ExtractorError('Unable to login, no cookie was returned')
- def _get_subtitles(self, video_id):
- # Fallback if there were no subtitles provided by DASH or HLS manifests
- caption_url = f'https://v.redd.it/{video_id}/wh_ben_en.vtt'
- if self._is_valid_url(caption_url, video_id, item='subtitles'):
- return {'en': [{'url': caption_url}]}
- def _real_extract(self, url):
- host, slug, video_id = self._match_valid_url(url).group('host', 'slug', 'id')
- data = self._download_json(
- f'https://{host}/{slug}/.json', video_id, fatal=False, expected_status=403)
- if not data:
- fallback_host = 'old.reddit.com' if host != 'old.reddit.com' else 'www.reddit.com'
- self.to_screen(f'{host} request failed, retrying with {fallback_host}')
- data = self._download_json(
- f'https://{fallback_host}/{slug}/.json', video_id, expected_status=403)
- if traverse_obj(data, 'error') == 403:
- reason = data.get('reason')
- if reason == 'quarantined':
- self.raise_login_required('Quarantined subreddit; an account that has opted in is required')
- elif reason == 'private':
- self.raise_login_required('Private subreddit; an account that has been approved is required')
- else:
- raise ExtractorError(f'HTTP Error 403 Forbidden; reason given: {reason}')
- data = data[0]['data']['children'][0]['data']
- video_url = data['url']
- over_18 = data.get('over_18')
- if over_18 is True:
- age_limit = 18
- elif over_18 is False:
- age_limit = 0
- else:
- age_limit = None
- thumbnails = []
- def add_thumbnail(src):
- if not isinstance(src, dict):
- return
- thumbnail_url = url_or_none(src.get('url'))
- if not thumbnail_url:
- return
- thumbnails.append({
- 'url': unescapeHTML(thumbnail_url),
- 'width': int_or_none(src.get('width')),
- 'height': int_or_none(src.get('height')),
- 'http_headers': {'Accept': '*/*'},
- })
- for image in try_get(data, lambda x: x['preview']['images']) or []:
- if not isinstance(image, dict):
- continue
- add_thumbnail(image.get('source'))
- resolutions = image.get('resolutions')
- if isinstance(resolutions, list):
- for resolution in resolutions:
- add_thumbnail(resolution)
- info = {
- 'title': data.get('title'),
- 'thumbnails': thumbnails,
- 'timestamp': float_or_none(data.get('created_utc')),
- 'uploader': data.get('author'),
- 'channel_id': data.get('subreddit'),
- 'like_count': int_or_none(data.get('ups')),
- 'dislike_count': int_or_none(data.get('downs')),
- 'comment_count': int_or_none(data.get('num_comments')),
- 'age_limit': age_limit,
- }
- parsed_url = urllib.parse.urlparse(video_url)
- # Check for embeds in text posts, or else raise to avoid recursing into the same reddit URL
- if 'reddit.com' in parsed_url.netloc and f'/{video_id}/' in parsed_url.path:
- entries = []
- for media in traverse_obj(data, ('media_metadata', ...), expected_type=dict):
- if not media.get('id') or media.get('e') != 'RedditVideo':
- continue
- formats = []
- if media.get('hlsUrl'):
- formats.extend(self._extract_m3u8_formats(
- unescapeHTML(media['hlsUrl']), video_id, 'mp4', m3u8_id='hls', fatal=False))
- if media.get('dashUrl'):
- formats.extend(self._extract_mpd_formats(
- unescapeHTML(media['dashUrl']), video_id, mpd_id='dash', fatal=False))
- if formats:
- entries.append({
- 'id': media['id'],
- 'display_id': video_id,
- 'formats': formats,
- **info,
- })
- if entries:
- return self.playlist_result(entries, video_id, info.get('title'))
- raise ExtractorError('No media found', expected=True)
- # Check if media is hosted on reddit:
- reddit_video = traverse_obj(data, (
- (None, ('crosspost_parent_list', ...)), ('secure_media', 'media'), 'reddit_video'), get_all=False)
- if reddit_video:
- playlist_urls = [
- try_get(reddit_video, lambda x: unescapeHTML(x[y]))
- for y in ('dash_url', 'hls_url')
- ]
- # Update video_id
- display_id = video_id
- video_id = self._search_regex(
- r'https?://v\.redd\.it/(?P<id>[^/?#&]+)', reddit_video['fallback_url'],
- 'video_id', default=display_id)
- dash_playlist_url = playlist_urls[0] or f'https://v.redd.it/{video_id}/DASHPlaylist.mpd'
- hls_playlist_url = playlist_urls[1] or f'https://v.redd.it/{video_id}/HLSPlaylist.m3u8'
- qs = traverse_obj(parse_qs(hls_playlist_url), {
- 'f': ('f', 0, {lambda x: ','.join([x, 'subsAll']) if x else 'hd,subsAll'}),
- })
- hls_playlist_url = update_url_query(hls_playlist_url, qs)
- formats = [{
- 'url': unescapeHTML(reddit_video['fallback_url']),
- 'height': int_or_none(reddit_video.get('height')),
- 'width': int_or_none(reddit_video.get('width')),
- 'tbr': int_or_none(reddit_video.get('bitrate_kbps')),
- 'acodec': 'none',
- 'vcodec': 'h264',
- 'ext': 'mp4',
- 'format_id': 'fallback',
- 'format_note': 'DASH video, mp4_dash',
- }]
- hls_fmts, subtitles = self._extract_m3u8_formats_and_subtitles(
- hls_playlist_url, display_id, 'mp4', m3u8_id='hls', fatal=False)
- formats.extend(hls_fmts)
- dash_fmts, dash_subs = self._extract_mpd_formats_and_subtitles(
- dash_playlist_url, display_id, mpd_id='dash', fatal=False)
- formats.extend(dash_fmts)
- self._merge_subtitles(dash_subs, target=subtitles)
- return {
- **info,
- 'id': video_id,
- 'display_id': display_id,
- 'formats': formats,
- 'subtitles': subtitles or self.extract_subtitles(video_id),
- 'duration': int_or_none(reddit_video.get('duration')),
- }
- if parsed_url.netloc == 'v.redd.it':
- self.raise_no_formats('This video is processing', expected=True, video_id=video_id)
- return {
- **info,
- 'id': parsed_url.path.split('/')[1],
- 'display_id': video_id,
- }
- # Not hosted on reddit, must continue extraction
- return {
- **info,
- 'display_id': video_id,
- '_type': 'url_transparent',
- 'url': video_url,
- }
|