# coding: utf-8 from __future__ import unicode_literals import re from .common import InfoExtractor from ..compat import compat_str from ..utils import ( int_or_none, parse_iso8601, unescapeHTML, qualities, ) class Revision3EmbedIE(InfoExtractor): IE_NAME = 'revision3:embed' _VALID_URL = r'(?:revision3:(?:(?P[^:]+):)?|https?://(?:(?:(?:www|embed)\.)?(?:revision3|animalist)|(?:(?:api|embed)\.)?seekernetwork)\.com/player/embed\?videoId=)(?P\d+)' _TEST = { 'url': 'http://api.seekernetwork.com/player/embed?videoId=67558', 'md5': '83bcd157cab89ad7318dd7b8c9cf1306', 'info_dict': { 'id': '67558', 'ext': 'mp4', 'title': 'The Pros & Cons Of Zoos', 'description': 'Zoos are often depicted as a terrible place for animals to live, but is there any truth to this?', 'uploader_id': 'dnews', 'uploader': 'DNews', } } _API_KEY = 'ba9c741bce1b9d8e3defcc22193f3651b8867e62' def _real_extract(self, url): mobj = re.match(self._VALID_URL, url) playlist_id = mobj.group('playlist_id') playlist_type = mobj.group('playlist_type') or 'video_id' video_data = self._download_json( 'http://revision3.com/api/getPlaylist.json', playlist_id, query={ 'api_key': self._API_KEY, 'codecs': 'h264,vp8,theora', playlist_type: playlist_id, })['items'][0] formats = [] for vcodec, media in video_data['media'].items(): for quality_id, quality in media.items(): if quality_id == 'hls': formats.extend(self._extract_m3u8_formats( quality['url'], playlist_id, 'mp4', 'm3u8_native', m3u8_id='hls', fatal=False)) else: formats.append({ 'url': quality['url'], 'format_id': '%s-%s' % (vcodec, quality_id), 'tbr': int_or_none(quality.get('bitrate')), 'vcodec': vcodec, }) self._sort_formats(formats) return { 'id': playlist_id, 'title': unescapeHTML(video_data['title']), 'description': unescapeHTML(video_data.get('summary')), 'uploader': video_data.get('show', {}).get('name'), 'uploader_id': video_data.get('show', {}).get('slug'), 'duration': int_or_none(video_data.get('duration')), 'formats': formats, } class Revision3IE(InfoExtractor): IE_NAME = 'revision' _VALID_URL = r'https?://(?:www\.)?(?P(?:revision3|animalist)\.com)/(?P[^/]+(?:/[^/?#]+)?)' _TESTS = [{ 'url': 'http://www.revision3.com/technobuffalo/5-google-predictions-for-2016', 'md5': 'd94a72d85d0a829766de4deb8daaf7df', 'info_dict': { 'id': '71089', 'display_id': 'technobuffalo/5-google-predictions-for-2016', 'ext': 'webm', 'title': '5 Google Predictions for 2016', 'description': 'Google had a great 2015, but it\'s already time to look ahead. Here are our five predictions for 2016.', 'upload_date': '20151228', 'timestamp': 1451325600, 'duration': 187, 'uploader': 'TechnoBuffalo', 'uploader_id': 'technobuffalo', } }, { # Show 'url': 'http://revision3.com/variant', 'only_matching': True, }, { # Tag 'url': 'http://revision3.com/vr', 'only_matching': True, }] _PAGE_DATA_TEMPLATE = 'http://www.%s/apiProxy/ddn/%s?domain=%s' def _real_extract(self, url): domain, display_id = re.match(self._VALID_URL, url).groups() site = domain.split('.')[0] page_info = self._download_json( self._PAGE_DATA_TEMPLATE % (domain, display_id, domain), display_id) page_data = page_info['data'] page_type = page_data['type'] if page_type in ('episode', 'embed'): show_data = page_data['show']['data'] page_id = compat_str(page_data['id']) video_id = compat_str(page_data['video']['data']['id']) preference = qualities(['mini', 'small', 'medium', 'large']) thumbnails = [{ 'url': image_url, 'id': image_id, 'preference': preference(image_id) } for image_id, image_url in page_data.get('images', {}).items()] info = { 'id': page_id, 'display_id': display_id, 'title': unescapeHTML(page_data['name']), 'description': unescapeHTML(page_data.get('summary')), 'timestamp': parse_iso8601(page_data.get('publishTime'), ' '), 'author': page_data.get('author'), 'uploader': show_data.get('name'), 'uploader_id': show_data.get('slug'), 'thumbnails': thumbnails, 'extractor_key': site, } if page_type == 'embed': info.update({ '_type': 'url_transparent', 'url': page_data['video']['data']['embed'], }) return info info.update({ '_type': 'url_transparent', 'url': 'revision3:%s' % video_id, }) return info else: list_data = page_info[page_type]['data'] episodes_data = page_info['episodes']['data'] num_episodes = page_info['meta']['totalEpisodes'] processed_episodes = 0 entries = [] page_num = 1 while True: entries.extend([{ '_type': 'url', 'url': 'http://%s%s' % (domain, episode['path']), 'id': compat_str(episode['id']), 'ie_key': 'Revision3', 'extractor_key': site, } for episode in episodes_data]) processed_episodes += len(episodes_data) if processed_episodes == num_episodes: break page_num += 1 episodes_data = self._download_json(self._PAGE_DATA_TEMPLATE % ( domain, display_id + '/' + compat_str(page_num), domain), display_id)['episodes']['data'] return self.playlist_result( entries, compat_str(list_data['id']), list_data.get('name'), list_data.get('summary'))