+from __future__ import unicode_literals
+
+import random
import re
from .common import InfoExtractor
+from ..compat import (
+ compat_str,
+ compat_urlparse,
+)
+from ..utils import (
+ encode_data_uri,
+ ExtractorError,
+ int_or_none,
+ float_or_none,
+ mimetype2ext,
+ str_or_none,
+)
class UstreamIE(InfoExtractor):
- _VALID_URL = r'https?://www\.ustream\.tv/recorded/(?P<videoID>\d+)'
- IE_NAME = u'ustream'
+ _VALID_URL = r'https?://(?:www\.)?ustream\.tv/(?P<type>recorded|embed|embed/recorded)/(?P<id>\d+)'
+ IE_NAME = 'ustream'
+ _TESTS = [{
+ 'url': 'http://www.ustream.tv/recorded/20274954',
+ 'md5': '088f151799e8f572f84eb62f17d73e5c',
+ 'info_dict': {
+ 'id': '20274954',
+ 'ext': 'flv',
+ 'title': 'Young Americans for Liberty February 7, 2012 2:28 AM',
+ 'description': 'Young Americans for Liberty February 7, 2012 2:28 AM',
+ 'timestamp': 1328577035,
+ 'upload_date': '20120207',
+ 'uploader': 'yaliberty',
+ 'uploader_id': '6780869',
+ },
+ }, {
+ # From http://sportscanada.tv/canadagames/index.php/week2/figure-skating/444
+ # Title and uploader available only from params JSON
+ 'url': 'http://www.ustream.tv/embed/recorded/59307601?ub=ff0000&lc=ff0000&oc=ffffff&uc=ffffff&v=3&wmode=direct',
+ 'md5': '5a2abf40babeac9812ed20ae12d34e10',
+ 'info_dict': {
+ 'id': '59307601',
+ 'ext': 'flv',
+ 'title': '-CG11- Canada Games Figure Skating',
+ 'uploader': 'sportscanadatv',
+ },
+ 'skip': 'This Pro Broadcaster has chosen to remove this video from the ustream.tv site.',
+ }, {
+ 'url': 'http://www.ustream.tv/embed/10299409',
+ 'info_dict': {
+ 'id': '10299409',
+ },
+ 'playlist_count': 3,
+ }, {
+ 'url': 'http://www.ustream.tv/recorded/91343263',
+ 'info_dict': {
+ 'id': '91343263',
+ 'ext': 'mp4',
+ 'title': 'GitHub Universe - General Session - Day 1',
+ 'upload_date': '20160914',
+ 'description': 'GitHub Universe - General Session - Day 1',
+ 'timestamp': 1473872730,
+ 'uploader': 'wa0dnskeqkr',
+ 'uploader_id': '38977840',
+ },
+ 'params': {
+ 'skip_download': True, # m3u8 download
+ },
+ }]
+
+ @staticmethod
+ def _extract_url(webpage):
+ mobj = re.search(
+ r'<iframe[^>]+?src=(["\'])(?P<url>http://www\.ustream\.tv/embed/.+?)\1', webpage)
+ if mobj is not None:
+ return mobj.group('url')
+
+ def _get_stream_info(self, url, video_id, app_id_ver, extra_note=None):
+ def num_to_hex(n):
+ return hex(n)[2:]
+
+ rnd = random.randrange
+
+ if not extra_note:
+ extra_note = ''
+
+ conn_info = self._download_json(
+ 'http://r%d-1-%s-recorded-lp-live.ums.ustream.tv/1/ustream' % (rnd(1e8), video_id),
+ video_id, note='Downloading connection info' + extra_note,
+ query={
+ 'type': 'viewer',
+ 'appId': app_id_ver[0],
+ 'appVersion': app_id_ver[1],
+ 'rsid': '%s:%s' % (num_to_hex(rnd(1e8)), num_to_hex(rnd(1e8))),
+ 'rpin': '_rpin.%d' % rnd(1e15),
+ 'referrer': url,
+ 'media': video_id,
+ 'application': 'recorded',
+ })
+ host = conn_info[0]['args'][0]['host']
+ connection_id = conn_info[0]['args'][0]['connectionId']
+
+ return self._download_json(
+ 'http://%s/1/ustream?connectionId=%s' % (host, connection_id),
+ video_id, note='Downloading stream info' + extra_note)
+
+ def _get_streams(self, url, video_id, app_id_ver):
+ # Sometimes the return dict does not have 'stream'
+ for trial_count in range(3):
+ stream_info = self._get_stream_info(
+ url, video_id, app_id_ver,
+ extra_note=' (try %d)' % (trial_count + 1) if trial_count > 0 else '')
+ if 'stream' in stream_info[0]['args'][0]:
+ return stream_info[0]['args'][0]['stream']
+ return []
+
+ def _parse_segmented_mp4(self, dash_stream_info):
+ def resolve_dash_template(template, idx, chunk_hash):
+ return template.replace('%', compat_str(idx), 1).replace('%', chunk_hash)
+
+ formats = []
+ for stream in dash_stream_info['streams']:
+ # Use only one provider to avoid too many formats
+ provider = dash_stream_info['providers'][0]
+ fragments = [{
+ 'url': resolve_dash_template(
+ provider['url'] + stream['initUrl'], 0, dash_stream_info['hashes']['0'])
+ }]
+ for idx in range(dash_stream_info['videoLength'] // dash_stream_info['chunkTime']):
+ fragments.append({
+ 'url': resolve_dash_template(
+ provider['url'] + stream['segmentUrl'], idx,
+ dash_stream_info['hashes'][compat_str(idx // 10 * 10)])
+ })
+ content_type = stream['contentType']
+ kind = content_type.split('/')[0]
+ f = {
+ 'format_id': '-'.join(filter(None, [
+ 'dash', kind, str_or_none(stream.get('bitrate'))])),
+ 'protocol': 'http_dash_segments',
+ # TODO: generate a MPD doc for external players?
+ 'url': encode_data_uri(b'<MPD/>', 'text/xml'),
+ 'ext': mimetype2ext(content_type),
+ 'height': stream.get('height'),
+ 'width': stream.get('width'),
+ 'fragments': fragments,
+ }
+ if kind == 'video':
+ f.update({
+ 'vcodec': stream.get('codec'),
+ 'acodec': 'none',
+ 'vbr': stream.get('bitrate'),
+ })
+ else:
+ f.update({
+ 'vcodec': 'none',
+ 'acodec': stream.get('codec'),
+ 'abr': stream.get('bitrate'),
+ })
+ formats.append(f)
+ return formats
def _real_extract(self, url):
m = re.match(self._VALID_URL, url)
- video_id = m.group('videoID')
+ video_id = m.group('id')
+
+ # some sites use this embed format (see: https://github.com/rg3/youtube-dl/issues/2990)
+ if m.group('type') == 'embed/recorded':
+ video_id = m.group('id')
+ desktop_url = 'http://www.ustream.tv/recorded/' + video_id
+ return self.url_result(desktop_url, 'Ustream')
+ if m.group('type') == 'embed':
+ video_id = m.group('id')
+ webpage = self._download_webpage(url, video_id)
+ content_video_ids = self._parse_json(self._search_regex(
+ r'ustream\.vars\.offAirContentVideoIds=([^;]+);', webpage,
+ 'content video IDs'), video_id)
+ return self.playlist_result(
+ map(lambda u: self.url_result('http://www.ustream.tv/recorded/' + u, 'Ustream'), content_video_ids),
+ video_id)
+
+ params = self._download_json(
+ 'https://api.ustream.tv/videos/%s.json' % video_id, video_id)
- video_url = u'http://tcdn.ustream.tv/video/%s' % video_id
- webpage = self._download_webpage(url, video_id)
+ error = params.get('error')
+ if error:
+ raise ExtractorError(
+ '%s returned error: %s' % (self.IE_NAME, error), expected=True)
- self.report_extraction(video_id)
+ video = params['video']
- video_title = self._html_search_regex(r'data-title="(?P<title>.+)"',
- webpage, u'title')
+ title = video['title']
+ filesize = float_or_none(video.get('file_size'))
- uploader = self._html_search_regex(r'data-content-type="channel".*?>(?P<uploader>.*?)</a>',
- webpage, u'uploader', fatal=False, flags=re.DOTALL)
+ formats = [{
+ 'id': video_id,
+ 'url': video_url,
+ 'ext': format_id,
+ 'filesize': filesize,
+ } for format_id, video_url in video['media_urls'].items() if video_url]
+
+ if not formats:
+ hls_streams = self._get_streams(url, video_id, app_id_ver=(11, 2))
+ if hls_streams:
+ # m3u8_native leads to intermittent ContentTooShortError
+ formats.extend(self._extract_m3u8_formats(
+ hls_streams[0]['url'], video_id, ext='mp4', m3u8_id='hls'))
+
+ '''
+ # DASH streams handling is incomplete as 'url' is missing
+ dash_streams = self._get_streams(url, video_id, app_id_ver=(3, 1))
+ if dash_streams:
+ formats.extend(self._parse_segmented_mp4(dash_streams))
+ '''
+
+ self._sort_formats(formats)
+
+ description = video.get('description')
+ timestamp = int_or_none(video.get('created_at'))
+ duration = float_or_none(video.get('length'))
+ view_count = int_or_none(video.get('views'))
+
+ uploader = video.get('owner', {}).get('username')
+ uploader_id = video.get('owner', {}).get('id')
+
+ thumbnails = [{
+ 'id': thumbnail_id,
+ 'url': thumbnail_url,
+ } for thumbnail_id, thumbnail_url in video.get('thumbnail', {}).items()]
+
+ return {
+ 'id': video_id,
+ 'title': title,
+ 'description': description,
+ 'thumbnails': thumbnails,
+ 'timestamp': timestamp,
+ 'duration': duration,
+ 'view_count': view_count,
+ 'uploader': uploader,
+ 'uploader_id': uploader_id,
+ 'formats': formats,
+ }
+
+
+class UstreamChannelIE(InfoExtractor):
+ _VALID_URL = r'https?://(?:www\.)?ustream\.tv/channel/(?P<slug>.+)'
+ IE_NAME = 'ustream:channel'
+ _TEST = {
+ 'url': 'http://www.ustream.tv/channel/channeljapan',
+ 'info_dict': {
+ 'id': '10874166',
+ },
+ 'playlist_mincount': 17,
+ }
+
+ def _real_extract(self, url):
+ m = re.match(self._VALID_URL, url)
+ display_id = m.group('slug')
+ webpage = self._download_webpage(url, display_id)
+ channel_id = self._html_search_meta('ustream:channel_id', webpage)
- thumbnail = self._html_search_regex(r'<link rel="image_src" href="(?P<thumb>.*?)"',
- webpage, u'thumbnail', fatal=False)
+ BASE = 'http://www.ustream.tv'
+ next_url = '/ajax/socialstream/videos/%s/1.json' % channel_id
+ video_ids = []
+ while next_url:
+ reply = self._download_json(
+ compat_urlparse.urljoin(BASE, next_url), display_id,
+ note='Downloading video information (next: %d)' % (len(video_ids) + 1))
+ video_ids.extend(re.findall(r'data-content-id="(\d.*)"', reply['data']))
+ next_url = reply['nextUrl']
- info = {
- 'id': video_id,
- 'url': video_url,
- 'ext': 'flv',
- 'title': video_title,
- 'uploader': uploader,
- 'thumbnail': thumbnail,
- }
- return info
+ entries = [
+ self.url_result('http://www.ustream.tv/recorded/' + vid, 'Ustream')
+ for vid in video_ids]
+ return {
+ '_type': 'playlist',
+ 'id': channel_id,
+ 'display_id': display_id,
+ 'entries': entries,
+ }