]> Raphaël G. Git Repositories - youtubedl/blob - youtube-dl
Imported Upstream version 2012.01.05
[youtubedl] / youtube-dl
1 #!/usr/bin/env python
2 # -*- coding: utf-8 -*-
3
4 __authors__ = (
5 'Ricardo Garcia Gonzalez',
6 'Danny Colligan',
7 'Benjamin Johnson',
8 'Vasyl\' Vavrychuk',
9 'Witold Baryluk',
10 'Paweł Paprota',
11 'Gergely Imreh',
12 'Rogério Brito',
13 'Philipp Hagemeister',
14 'Sören Schulze',
15 'Kevin Ngo',
16 'Ori Avtalion',
17 'shizeeg',
18 )
19
20 __license__ = 'Public Domain'
21 __version__ = '2012.01.05'
22
23 UPDATE_URL = 'https://raw.github.com/rg3/youtube-dl/master/youtube-dl'
24
25 import cookielib
26 import datetime
27 import gzip
28 import htmlentitydefs
29 import HTMLParser
30 import httplib
31 import locale
32 import math
33 import netrc
34 import os
35 import os.path
36 import re
37 import socket
38 import string
39 import subprocess
40 import sys
41 import time
42 import urllib
43 import urllib2
44 import warnings
45 import zlib
46
47 if os.name == 'nt':
48 import ctypes
49
50 try:
51 import email.utils
52 except ImportError: # Python 2.4
53 import email.Utils
54 try:
55 import cStringIO as StringIO
56 except ImportError:
57 import StringIO
58
59 # parse_qs was moved from the cgi module to the urlparse module recently.
60 try:
61 from urlparse import parse_qs
62 except ImportError:
63 from cgi import parse_qs
64
65 try:
66 import lxml.etree
67 except ImportError:
68 pass # Handled below
69
70 try:
71 import xml.etree.ElementTree
72 except ImportError: # Python<2.5: Not officially supported, but let it slip
73 warnings.warn('xml.etree.ElementTree support is missing. Consider upgrading to Python >= 2.5 if you get related errors.')
74
75 std_headers = {
76 'User-Agent': 'Mozilla/5.0 (X11; Linux x86_64; rv:5.0.1) Gecko/20100101 Firefox/5.0.1',
77 'Accept-Charset': 'ISO-8859-1,utf-8;q=0.7,*;q=0.7',
78 'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8',
79 'Accept-Encoding': 'gzip, deflate',
80 'Accept-Language': 'en-us,en;q=0.5',
81 }
82
83 try:
84 import json
85 except ImportError: # Python <2.6, use trivialjson (https://github.com/phihag/trivialjson):
86 import re
87 class json(object):
88 @staticmethod
89 def loads(s):
90 s = s.decode('UTF-8')
91 def raiseError(msg, i):
92 raise ValueError(msg + ' at position ' + str(i) + ' of ' + repr(s) + ': ' + repr(s[i:]))
93 def skipSpace(i, expectMore=True):
94 while i < len(s) and s[i] in ' \t\r\n':
95 i += 1
96 if expectMore:
97 if i >= len(s):
98 raiseError('Premature end', i)
99 return i
100 def decodeEscape(match):
101 esc = match.group(1)
102 _STATIC = {
103 '"': '"',
104 '\\': '\\',
105 '/': '/',
106 'b': unichr(0x8),
107 'f': unichr(0xc),
108 'n': '\n',
109 'r': '\r',
110 't': '\t',
111 }
112 if esc in _STATIC:
113 return _STATIC[esc]
114 if esc[0] == 'u':
115 if len(esc) == 1+4:
116 return unichr(int(esc[1:5], 16))
117 if len(esc) == 5+6 and esc[5:7] == '\\u':
118 hi = int(esc[1:5], 16)
119 low = int(esc[7:11], 16)
120 return unichr((hi - 0xd800) * 0x400 + low - 0xdc00 + 0x10000)
121 raise ValueError('Unknown escape ' + str(esc))
122 def parseString(i):
123 i += 1
124 e = i
125 while True:
126 e = s.index('"', e)
127 bslashes = 0
128 while s[e-bslashes-1] == '\\':
129 bslashes += 1
130 if bslashes % 2 == 1:
131 e += 1
132 continue
133 break
134 rexp = re.compile(r'\\(u[dD][89aAbB][0-9a-fA-F]{2}\\u[0-9a-fA-F]{4}|u[0-9a-fA-F]{4}|.|$)')
135 stri = rexp.sub(decodeEscape, s[i:e])
136 return (e+1,stri)
137 def parseObj(i):
138 i += 1
139 res = {}
140 i = skipSpace(i)
141 if s[i] == '}': # Empty dictionary
142 return (i+1,res)
143 while True:
144 if s[i] != '"':
145 raiseError('Expected a string object key', i)
146 i,key = parseString(i)
147 i = skipSpace(i)
148 if i >= len(s) or s[i] != ':':
149 raiseError('Expected a colon', i)
150 i,val = parse(i+1)
151 res[key] = val
152 i = skipSpace(i)
153 if s[i] == '}':
154 return (i+1, res)
155 if s[i] != ',':
156 raiseError('Expected comma or closing curly brace', i)
157 i = skipSpace(i+1)
158 def parseArray(i):
159 res = []
160 i = skipSpace(i+1)
161 if s[i] == ']': # Empty array
162 return (i+1,res)
163 while True:
164 i,val = parse(i)
165 res.append(val)
166 i = skipSpace(i) # Raise exception if premature end
167 if s[i] == ']':
168 return (i+1, res)
169 if s[i] != ',':
170 raiseError('Expected a comma or closing bracket', i)
171 i = skipSpace(i+1)
172 def parseDiscrete(i):
173 for k,v in {'true': True, 'false': False, 'null': None}.items():
174 if s.startswith(k, i):
175 return (i+len(k), v)
176 raiseError('Not a boolean (or null)', i)
177 def parseNumber(i):
178 mobj = re.match('^(-?(0|[1-9][0-9]*)(\.[0-9]*)?([eE][+-]?[0-9]+)?)', s[i:])
179 if mobj is None:
180 raiseError('Not a number', i)
181 nums = mobj.group(1)
182 if '.' in nums or 'e' in nums or 'E' in nums:
183 return (i+len(nums), float(nums))
184 return (i+len(nums), int(nums))
185 CHARMAP = {'{': parseObj, '[': parseArray, '"': parseString, 't': parseDiscrete, 'f': parseDiscrete, 'n': parseDiscrete}
186 def parse(i):
187 i = skipSpace(i)
188 i,res = CHARMAP.get(s[i], parseNumber)(i)
189 i = skipSpace(i, False)
190 return (i,res)
191 i,res = parse(0)
192 if i < len(s):
193 raise ValueError('Extra data at end of input (index ' + str(i) + ' of ' + repr(s) + ': ' + repr(s[i:]) + ')')
194 return res
195
196 def preferredencoding():
197 """Get preferred encoding.
198
199 Returns the best encoding scheme for the system, based on
200 locale.getpreferredencoding() and some further tweaks.
201 """
202 def yield_preferredencoding():
203 try:
204 pref = locale.getpreferredencoding()
205 u'TEST'.encode(pref)
206 except:
207 pref = 'UTF-8'
208 while True:
209 yield pref
210 return yield_preferredencoding().next()
211
212
213 def htmlentity_transform(matchobj):
214 """Transforms an HTML entity to a Unicode character.
215
216 This function receives a match object and is intended to be used with
217 the re.sub() function.
218 """
219 entity = matchobj.group(1)
220
221 # Known non-numeric HTML entity
222 if entity in htmlentitydefs.name2codepoint:
223 return unichr(htmlentitydefs.name2codepoint[entity])
224
225 # Unicode character
226 mobj = re.match(ur'(?u)#(x?\d+)', entity)
227 if mobj is not None:
228 numstr = mobj.group(1)
229 if numstr.startswith(u'x'):
230 base = 16
231 numstr = u'0%s' % numstr
232 else:
233 base = 10
234 return unichr(long(numstr, base))
235
236 # Unknown entity in name, return its literal representation
237 return (u'&%s;' % entity)
238
239
240 def sanitize_title(utitle):
241 """Sanitizes a video title so it could be used as part of a filename."""
242 utitle = re.sub(ur'(?u)&(.+?);', htmlentity_transform, utitle)
243 return utitle.replace(unicode(os.sep), u'%')
244
245
246 def sanitize_open(filename, open_mode):
247 """Try to open the given filename, and slightly tweak it if this fails.
248
249 Attempts to open the given filename. If this fails, it tries to change
250 the filename slightly, step by step, until it's either able to open it
251 or it fails and raises a final exception, like the standard open()
252 function.
253
254 It returns the tuple (stream, definitive_file_name).
255 """
256 try:
257 if filename == u'-':
258 if sys.platform == 'win32':
259 import msvcrt
260 msvcrt.setmode(sys.stdout.fileno(), os.O_BINARY)
261 return (sys.stdout, filename)
262 stream = open(filename, open_mode)
263 return (stream, filename)
264 except (IOError, OSError), err:
265 # In case of error, try to remove win32 forbidden chars
266 filename = re.sub(ur'[/<>:"\|\?\*]', u'#', filename)
267
268 # An exception here should be caught in the caller
269 stream = open(filename, open_mode)
270 return (stream, filename)
271
272
273 def timeconvert(timestr):
274 """Convert RFC 2822 defined time string into system timestamp"""
275 timestamp = None
276 timetuple = email.utils.parsedate_tz(timestr)
277 if timetuple is not None:
278 timestamp = email.utils.mktime_tz(timetuple)
279 return timestamp
280
281 def _simplify_title(title):
282 expr = re.compile(ur'[^\w\d_\-]+', flags=re.UNICODE)
283 return expr.sub(u'_', title).strip(u'_')
284
285 def _orderedSet(iterable):
286 """ Remove all duplicates from the input iterable """
287 res = []
288 for el in iterable:
289 if el not in res:
290 res.append(el)
291 return res
292
293 def _unescapeHTML(s):
294 """
295 @param s a string (of type unicode)
296 """
297 assert type(s) == type(u'')
298
299 htmlParser = HTMLParser.HTMLParser()
300 return htmlParser.unescape(s)
301
302 class DownloadError(Exception):
303 """Download Error exception.
304
305 This exception may be thrown by FileDownloader objects if they are not
306 configured to continue on errors. They will contain the appropriate
307 error message.
308 """
309 pass
310
311
312 class SameFileError(Exception):
313 """Same File exception.
314
315 This exception will be thrown by FileDownloader objects if they detect
316 multiple files would have to be downloaded to the same file on disk.
317 """
318 pass
319
320
321 class PostProcessingError(Exception):
322 """Post Processing exception.
323
324 This exception may be raised by PostProcessor's .run() method to
325 indicate an error in the postprocessing task.
326 """
327 pass
328
329 class MaxDownloadsReached(Exception):
330 """ --max-downloads limit has been reached. """
331 pass
332
333
334 class UnavailableVideoError(Exception):
335 """Unavailable Format exception.
336
337 This exception will be thrown when a video is requested
338 in a format that is not available for that video.
339 """
340 pass
341
342
343 class ContentTooShortError(Exception):
344 """Content Too Short exception.
345
346 This exception may be raised by FileDownloader objects when a file they
347 download is too small for what the server announced first, indicating
348 the connection was probably interrupted.
349 """
350 # Both in bytes
351 downloaded = None
352 expected = None
353
354 def __init__(self, downloaded, expected):
355 self.downloaded = downloaded
356 self.expected = expected
357
358
359 class YoutubeDLHandler(urllib2.HTTPHandler):
360 """Handler for HTTP requests and responses.
361
362 This class, when installed with an OpenerDirector, automatically adds
363 the standard headers to every HTTP request and handles gzipped and
364 deflated responses from web servers. If compression is to be avoided in
365 a particular request, the original request in the program code only has
366 to include the HTTP header "Youtubedl-No-Compression", which will be
367 removed before making the real request.
368
369 Part of this code was copied from:
370
371 http://techknack.net/python-urllib2-handlers/
372
373 Andrew Rowls, the author of that code, agreed to release it to the
374 public domain.
375 """
376
377 @staticmethod
378 def deflate(data):
379 try:
380 return zlib.decompress(data, -zlib.MAX_WBITS)
381 except zlib.error:
382 return zlib.decompress(data)
383
384 @staticmethod
385 def addinfourl_wrapper(stream, headers, url, code):
386 if hasattr(urllib2.addinfourl, 'getcode'):
387 return urllib2.addinfourl(stream, headers, url, code)
388 ret = urllib2.addinfourl(stream, headers, url)
389 ret.code = code
390 return ret
391
392 def http_request(self, req):
393 for h in std_headers:
394 if h in req.headers:
395 del req.headers[h]
396 req.add_header(h, std_headers[h])
397 if 'Youtubedl-no-compression' in req.headers:
398 if 'Accept-encoding' in req.headers:
399 del req.headers['Accept-encoding']
400 del req.headers['Youtubedl-no-compression']
401 return req
402
403 def http_response(self, req, resp):
404 old_resp = resp
405 # gzip
406 if resp.headers.get('Content-encoding', '') == 'gzip':
407 gz = gzip.GzipFile(fileobj=StringIO.StringIO(resp.read()), mode='r')
408 resp = self.addinfourl_wrapper(gz, old_resp.headers, old_resp.url, old_resp.code)
409 resp.msg = old_resp.msg
410 # deflate
411 if resp.headers.get('Content-encoding', '') == 'deflate':
412 gz = StringIO.StringIO(self.deflate(resp.read()))
413 resp = self.addinfourl_wrapper(gz, old_resp.headers, old_resp.url, old_resp.code)
414 resp.msg = old_resp.msg
415 return resp
416
417
418 class FileDownloader(object):
419 """File Downloader class.
420
421 File downloader objects are the ones responsible of downloading the
422 actual video file and writing it to disk if the user has requested
423 it, among some other tasks. In most cases there should be one per
424 program. As, given a video URL, the downloader doesn't know how to
425 extract all the needed information, task that InfoExtractors do, it
426 has to pass the URL to one of them.
427
428 For this, file downloader objects have a method that allows
429 InfoExtractors to be registered in a given order. When it is passed
430 a URL, the file downloader handles it to the first InfoExtractor it
431 finds that reports being able to handle it. The InfoExtractor extracts
432 all the information about the video or videos the URL refers to, and
433 asks the FileDownloader to process the video information, possibly
434 downloading the video.
435
436 File downloaders accept a lot of parameters. In order not to saturate
437 the object constructor with arguments, it receives a dictionary of
438 options instead. These options are available through the params
439 attribute for the InfoExtractors to use. The FileDownloader also
440 registers itself as the downloader in charge for the InfoExtractors
441 that are added to it, so this is a "mutual registration".
442
443 Available options:
444
445 username: Username for authentication purposes.
446 password: Password for authentication purposes.
447 usenetrc: Use netrc for authentication instead.
448 quiet: Do not print messages to stdout.
449 forceurl: Force printing final URL.
450 forcetitle: Force printing title.
451 forcethumbnail: Force printing thumbnail URL.
452 forcedescription: Force printing description.
453 forcefilename: Force printing final filename.
454 simulate: Do not download the video files.
455 format: Video format code.
456 format_limit: Highest quality format to try.
457 outtmpl: Template for output names.
458 ignoreerrors: Do not stop on download errors.
459 ratelimit: Download speed limit, in bytes/sec.
460 nooverwrites: Prevent overwriting files.
461 retries: Number of times to retry for HTTP error 5xx
462 continuedl: Try to continue downloads if possible.
463 noprogress: Do not print the progress bar.
464 playliststart: Playlist item to start at.
465 playlistend: Playlist item to end at.
466 matchtitle: Download only matching titles.
467 rejecttitle: Reject downloads for matching titles.
468 logtostderr: Log messages to stderr instead of stdout.
469 consoletitle: Display progress in console window's titlebar.
470 nopart: Do not use temporary .part files.
471 updatetime: Use the Last-modified header to set output file timestamps.
472 writedescription: Write the video description to a .description file
473 writeinfojson: Write the video description to a .info.json file
474 """
475
476 params = None
477 _ies = []
478 _pps = []
479 _download_retcode = None
480 _num_downloads = None
481 _screen_file = None
482
483 def __init__(self, params):
484 """Create a FileDownloader object with the given options."""
485 self._ies = []
486 self._pps = []
487 self._download_retcode = 0
488 self._num_downloads = 0
489 self._screen_file = [sys.stdout, sys.stderr][params.get('logtostderr', False)]
490 self.params = params
491
492 @staticmethod
493 def format_bytes(bytes):
494 if bytes is None:
495 return 'N/A'
496 if type(bytes) is str:
497 bytes = float(bytes)
498 if bytes == 0.0:
499 exponent = 0
500 else:
501 exponent = long(math.log(bytes, 1024.0))
502 suffix = 'bkMGTPEZY'[exponent]
503 converted = float(bytes) / float(1024 ** exponent)
504 return '%.2f%s' % (converted, suffix)
505
506 @staticmethod
507 def calc_percent(byte_counter, data_len):
508 if data_len is None:
509 return '---.-%'
510 return '%6s' % ('%3.1f%%' % (float(byte_counter) / float(data_len) * 100.0))
511
512 @staticmethod
513 def calc_eta(start, now, total, current):
514 if total is None:
515 return '--:--'
516 dif = now - start
517 if current == 0 or dif < 0.001: # One millisecond
518 return '--:--'
519 rate = float(current) / dif
520 eta = long((float(total) - float(current)) / rate)
521 (eta_mins, eta_secs) = divmod(eta, 60)
522 if eta_mins > 99:
523 return '--:--'
524 return '%02d:%02d' % (eta_mins, eta_secs)
525
526 @staticmethod
527 def calc_speed(start, now, bytes):
528 dif = now - start
529 if bytes == 0 or dif < 0.001: # One millisecond
530 return '%10s' % '---b/s'
531 return '%10s' % ('%s/s' % FileDownloader.format_bytes(float(bytes) / dif))
532
533 @staticmethod
534 def best_block_size(elapsed_time, bytes):
535 new_min = max(bytes / 2.0, 1.0)
536 new_max = min(max(bytes * 2.0, 1.0), 4194304) # Do not surpass 4 MB
537 if elapsed_time < 0.001:
538 return long(new_max)
539 rate = bytes / elapsed_time
540 if rate > new_max:
541 return long(new_max)
542 if rate < new_min:
543 return long(new_min)
544 return long(rate)
545
546 @staticmethod
547 def parse_bytes(bytestr):
548 """Parse a string indicating a byte quantity into a long integer."""
549 matchobj = re.match(r'(?i)^(\d+(?:\.\d+)?)([kMGTPEZY]?)$', bytestr)
550 if matchobj is None:
551 return None
552 number = float(matchobj.group(1))
553 multiplier = 1024.0 ** 'bkmgtpezy'.index(matchobj.group(2).lower())
554 return long(round(number * multiplier))
555
556 def add_info_extractor(self, ie):
557 """Add an InfoExtractor object to the end of the list."""
558 self._ies.append(ie)
559 ie.set_downloader(self)
560
561 def add_post_processor(self, pp):
562 """Add a PostProcessor object to the end of the chain."""
563 self._pps.append(pp)
564 pp.set_downloader(self)
565
566 def to_screen(self, message, skip_eol=False, ignore_encoding_errors=False):
567 """Print message to stdout if not in quiet mode."""
568 try:
569 if not self.params.get('quiet', False):
570 terminator = [u'\n', u''][skip_eol]
571 print >>self._screen_file, (u'%s%s' % (message, terminator)).encode(preferredencoding()),
572 self._screen_file.flush()
573 except (UnicodeEncodeError), err:
574 if not ignore_encoding_errors:
575 raise
576
577 def to_stderr(self, message):
578 """Print message to stderr."""
579 print >>sys.stderr, message.encode(preferredencoding())
580
581 def to_cons_title(self, message):
582 """Set console/terminal window title to message."""
583 if not self.params.get('consoletitle', False):
584 return
585 if os.name == 'nt' and ctypes.windll.kernel32.GetConsoleWindow():
586 # c_wchar_p() might not be necessary if `message` is
587 # already of type unicode()
588 ctypes.windll.kernel32.SetConsoleTitleW(ctypes.c_wchar_p(message))
589 elif 'TERM' in os.environ:
590 sys.stderr.write('\033]0;%s\007' % message.encode(preferredencoding()))
591
592 def fixed_template(self):
593 """Checks if the output template is fixed."""
594 return (re.search(ur'(?u)%\(.+?\)s', self.params['outtmpl']) is None)
595
596 def trouble(self, message=None):
597 """Determine action to take when a download problem appears.
598
599 Depending on if the downloader has been configured to ignore
600 download errors or not, this method may throw an exception or
601 not when errors are found, after printing the message.
602 """
603 if message is not None:
604 self.to_stderr(message)
605 if not self.params.get('ignoreerrors', False):
606 raise DownloadError(message)
607 self._download_retcode = 1
608
609 def slow_down(self, start_time, byte_counter):
610 """Sleep if the download speed is over the rate limit."""
611 rate_limit = self.params.get('ratelimit', None)
612 if rate_limit is None or byte_counter == 0:
613 return
614 now = time.time()
615 elapsed = now - start_time
616 if elapsed <= 0.0:
617 return
618 speed = float(byte_counter) / elapsed
619 if speed > rate_limit:
620 time.sleep((byte_counter - rate_limit * (now - start_time)) / rate_limit)
621
622 def temp_name(self, filename):
623 """Returns a temporary filename for the given filename."""
624 if self.params.get('nopart', False) or filename == u'-' or \
625 (os.path.exists(filename) and not os.path.isfile(filename)):
626 return filename
627 return filename + u'.part'
628
629 def undo_temp_name(self, filename):
630 if filename.endswith(u'.part'):
631 return filename[:-len(u'.part')]
632 return filename
633
634 def try_rename(self, old_filename, new_filename):
635 try:
636 if old_filename == new_filename:
637 return
638 os.rename(old_filename, new_filename)
639 except (IOError, OSError), err:
640 self.trouble(u'ERROR: unable to rename file')
641
642 def try_utime(self, filename, last_modified_hdr):
643 """Try to set the last-modified time of the given file."""
644 if last_modified_hdr is None:
645 return
646 if not os.path.isfile(filename):
647 return
648 timestr = last_modified_hdr
649 if timestr is None:
650 return
651 filetime = timeconvert(timestr)
652 if filetime is None:
653 return filetime
654 try:
655 os.utime(filename, (time.time(), filetime))
656 except:
657 pass
658 return filetime
659
660 def report_writedescription(self, descfn):
661 """ Report that the description file is being written """
662 self.to_screen(u'[info] Writing video description to: %s' % descfn, ignore_encoding_errors=True)
663
664 def report_writeinfojson(self, infofn):
665 """ Report that the metadata file has been written """
666 self.to_screen(u'[info] Video description metadata as JSON to: %s' % infofn, ignore_encoding_errors=True)
667
668 def report_destination(self, filename):
669 """Report destination filename."""
670 self.to_screen(u'[download] Destination: %s' % filename, ignore_encoding_errors=True)
671
672 def report_progress(self, percent_str, data_len_str, speed_str, eta_str):
673 """Report download progress."""
674 if self.params.get('noprogress', False):
675 return
676 self.to_screen(u'\r[download] %s of %s at %s ETA %s' %
677 (percent_str, data_len_str, speed_str, eta_str), skip_eol=True)
678 self.to_cons_title(u'youtube-dl - %s of %s at %s ETA %s' %
679 (percent_str.strip(), data_len_str.strip(), speed_str.strip(), eta_str.strip()))
680
681 def report_resuming_byte(self, resume_len):
682 """Report attempt to resume at given byte."""
683 self.to_screen(u'[download] Resuming download at byte %s' % resume_len)
684
685 def report_retry(self, count, retries):
686 """Report retry in case of HTTP error 5xx"""
687 self.to_screen(u'[download] Got server HTTP error. Retrying (attempt %d of %d)...' % (count, retries))
688
689 def report_file_already_downloaded(self, file_name):
690 """Report file has already been fully downloaded."""
691 try:
692 self.to_screen(u'[download] %s has already been downloaded' % file_name)
693 except (UnicodeEncodeError), err:
694 self.to_screen(u'[download] The file has already been downloaded')
695
696 def report_unable_to_resume(self):
697 """Report it was impossible to resume download."""
698 self.to_screen(u'[download] Unable to resume')
699
700 def report_finish(self):
701 """Report download finished."""
702 if self.params.get('noprogress', False):
703 self.to_screen(u'[download] Download completed')
704 else:
705 self.to_screen(u'')
706
707 def increment_downloads(self):
708 """Increment the ordinal that assigns a number to each file."""
709 self._num_downloads += 1
710
711 def prepare_filename(self, info_dict):
712 """Generate the output filename."""
713 try:
714 template_dict = dict(info_dict)
715 template_dict['epoch'] = unicode(long(time.time()))
716 template_dict['autonumber'] = unicode('%05d' % self._num_downloads)
717 filename = self.params['outtmpl'] % template_dict
718 return filename
719 except (ValueError, KeyError), err:
720 self.trouble(u'ERROR: invalid system charset or erroneous output template')
721 return None
722
723 def _match_entry(self, info_dict):
724 """ Returns None iff the file should be downloaded """
725
726 title = info_dict['title']
727 matchtitle = self.params.get('matchtitle', False)
728 if matchtitle and not re.search(matchtitle, title, re.IGNORECASE):
729 return u'[download] "' + title + '" title did not match pattern "' + matchtitle + '"'
730 rejecttitle = self.params.get('rejecttitle', False)
731 if rejecttitle and re.search(rejecttitle, title, re.IGNORECASE):
732 return u'"' + title + '" title matched reject pattern "' + rejecttitle + '"'
733 return None
734
735 def process_info(self, info_dict):
736 """Process a single dictionary returned by an InfoExtractor."""
737
738 reason = self._match_entry(info_dict)
739 if reason is not None:
740 self.to_screen(u'[download] ' + reason)
741 return
742
743 max_downloads = self.params.get('max_downloads')
744 if max_downloads is not None:
745 if self._num_downloads > int(max_downloads):
746 raise MaxDownloadsReached()
747
748 filename = self.prepare_filename(info_dict)
749
750 # Forced printings
751 if self.params.get('forcetitle', False):
752 print info_dict['title'].encode(preferredencoding(), 'xmlcharrefreplace')
753 if self.params.get('forceurl', False):
754 print info_dict['url'].encode(preferredencoding(), 'xmlcharrefreplace')
755 if self.params.get('forcethumbnail', False) and 'thumbnail' in info_dict:
756 print info_dict['thumbnail'].encode(preferredencoding(), 'xmlcharrefreplace')
757 if self.params.get('forcedescription', False) and 'description' in info_dict:
758 print info_dict['description'].encode(preferredencoding(), 'xmlcharrefreplace')
759 if self.params.get('forcefilename', False) and filename is not None:
760 print filename.encode(preferredencoding(), 'xmlcharrefreplace')
761 if self.params.get('forceformat', False):
762 print info_dict['format'].encode(preferredencoding(), 'xmlcharrefreplace')
763
764 # Do nothing else if in simulate mode
765 if self.params.get('simulate', False):
766 return
767
768 if filename is None:
769 return
770
771 try:
772 dn = os.path.dirname(filename)
773 if dn != '' and not os.path.exists(dn):
774 os.makedirs(dn)
775 except (OSError, IOError), err:
776 self.trouble(u'ERROR: unable to create directory ' + unicode(err))
777 return
778
779 if self.params.get('writedescription', False):
780 try:
781 descfn = filename + '.description'
782 self.report_writedescription(descfn)
783 descfile = open(descfn, 'wb')
784 try:
785 descfile.write(info_dict['description'].encode('utf-8'))
786 finally:
787 descfile.close()
788 except (OSError, IOError):
789 self.trouble(u'ERROR: Cannot write description file ' + descfn)
790 return
791
792 if self.params.get('writeinfojson', False):
793 infofn = filename + '.info.json'
794 self.report_writeinfojson(infofn)
795 try:
796 json.dump
797 except (NameError,AttributeError):
798 self.trouble(u'ERROR: No JSON encoder found. Update to Python 2.6+, setup a json module, or leave out --write-info-json.')
799 return
800 try:
801 infof = open(infofn, 'wb')
802 try:
803 json_info_dict = dict((k,v) for k,v in info_dict.iteritems() if not k in ('urlhandle',))
804 json.dump(json_info_dict, infof)
805 finally:
806 infof.close()
807 except (OSError, IOError):
808 self.trouble(u'ERROR: Cannot write metadata to JSON file ' + infofn)
809 return
810
811 if not self.params.get('skip_download', False):
812 if self.params.get('nooverwrites', False) and os.path.exists(filename):
813 success = True
814 else:
815 try:
816 success = self._do_download(filename, info_dict)
817 except (OSError, IOError), err:
818 raise UnavailableVideoError
819 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
820 self.trouble(u'ERROR: unable to download video data: %s' % str(err))
821 return
822 except (ContentTooShortError, ), err:
823 self.trouble(u'ERROR: content too short (expected %s bytes and served %s)' % (err.expected, err.downloaded))
824 return
825
826 if success:
827 try:
828 self.post_process(filename, info_dict)
829 except (PostProcessingError), err:
830 self.trouble(u'ERROR: postprocessing: %s' % str(err))
831 return
832
833 def download(self, url_list):
834 """Download a given list of URLs."""
835 if len(url_list) > 1 and self.fixed_template():
836 raise SameFileError(self.params['outtmpl'])
837
838 for url in url_list:
839 suitable_found = False
840 for ie in self._ies:
841 # Go to next InfoExtractor if not suitable
842 if not ie.suitable(url):
843 continue
844
845 # Suitable InfoExtractor found
846 suitable_found = True
847
848 # Extract information from URL and process it
849 ie.extract(url)
850
851 # Suitable InfoExtractor had been found; go to next URL
852 break
853
854 if not suitable_found:
855 self.trouble(u'ERROR: no suitable InfoExtractor: %s' % url)
856
857 return self._download_retcode
858
859 def post_process(self, filename, ie_info):
860 """Run the postprocessing chain on the given file."""
861 info = dict(ie_info)
862 info['filepath'] = filename
863 for pp in self._pps:
864 info = pp.run(info)
865 if info is None:
866 break
867
868 def _download_with_rtmpdump(self, filename, url, player_url):
869 self.report_destination(filename)
870 tmpfilename = self.temp_name(filename)
871
872 # Check for rtmpdump first
873 try:
874 subprocess.call(['rtmpdump', '-h'], stdout=(file(os.path.devnull, 'w')), stderr=subprocess.STDOUT)
875 except (OSError, IOError):
876 self.trouble(u'ERROR: RTMP download detected but "rtmpdump" could not be run')
877 return False
878
879 # Download using rtmpdump. rtmpdump returns exit code 2 when
880 # the connection was interrumpted and resuming appears to be
881 # possible. This is part of rtmpdump's normal usage, AFAIK.
882 basic_args = ['rtmpdump', '-q'] + [[], ['-W', player_url]][player_url is not None] + ['-r', url, '-o', tmpfilename]
883 retval = subprocess.call(basic_args + [[], ['-e', '-k', '1']][self.params.get('continuedl', False)])
884 while retval == 2 or retval == 1:
885 prevsize = os.path.getsize(tmpfilename)
886 self.to_screen(u'\r[rtmpdump] %s bytes' % prevsize, skip_eol=True)
887 time.sleep(5.0) # This seems to be needed
888 retval = subprocess.call(basic_args + ['-e'] + [[], ['-k', '1']][retval == 1])
889 cursize = os.path.getsize(tmpfilename)
890 if prevsize == cursize and retval == 1:
891 break
892 # Some rtmp streams seem abort after ~ 99.8%. Don't complain for those
893 if prevsize == cursize and retval == 2 and cursize > 1024:
894 self.to_screen(u'\r[rtmpdump] Could not download the whole video. This can happen for some advertisements.')
895 retval = 0
896 break
897 if retval == 0:
898 self.to_screen(u'\r[rtmpdump] %s bytes' % os.path.getsize(tmpfilename))
899 self.try_rename(tmpfilename, filename)
900 return True
901 else:
902 self.trouble(u'\nERROR: rtmpdump exited with code %d' % retval)
903 return False
904
905 def _do_download(self, filename, info_dict):
906 url = info_dict['url']
907 player_url = info_dict.get('player_url', None)
908
909 # Check file already present
910 if self.params.get('continuedl', False) and os.path.isfile(filename) and not self.params.get('nopart', False):
911 self.report_file_already_downloaded(filename)
912 return True
913
914 # Attempt to download using rtmpdump
915 if url.startswith('rtmp'):
916 return self._download_with_rtmpdump(filename, url, player_url)
917
918 tmpfilename = self.temp_name(filename)
919 stream = None
920
921 # Do not include the Accept-Encoding header
922 headers = {'Youtubedl-no-compression': 'True'}
923 basic_request = urllib2.Request(url, None, headers)
924 request = urllib2.Request(url, None, headers)
925
926 # Establish possible resume length
927 if os.path.isfile(tmpfilename):
928 resume_len = os.path.getsize(tmpfilename)
929 else:
930 resume_len = 0
931
932 open_mode = 'wb'
933 if resume_len != 0:
934 if self.params.get('continuedl', False):
935 self.report_resuming_byte(resume_len)
936 request.add_header('Range','bytes=%d-' % resume_len)
937 open_mode = 'ab'
938 else:
939 resume_len = 0
940
941 count = 0
942 retries = self.params.get('retries', 0)
943 while count <= retries:
944 # Establish connection
945 try:
946 if count == 0 and 'urlhandle' in info_dict:
947 data = info_dict['urlhandle']
948 data = urllib2.urlopen(request)
949 break
950 except (urllib2.HTTPError, ), err:
951 if (err.code < 500 or err.code >= 600) and err.code != 416:
952 # Unexpected HTTP error
953 raise
954 elif err.code == 416:
955 # Unable to resume (requested range not satisfiable)
956 try:
957 # Open the connection again without the range header
958 data = urllib2.urlopen(basic_request)
959 content_length = data.info()['Content-Length']
960 except (urllib2.HTTPError, ), err:
961 if err.code < 500 or err.code >= 600:
962 raise
963 else:
964 # Examine the reported length
965 if (content_length is not None and
966 (resume_len - 100 < long(content_length) < resume_len + 100)):
967 # The file had already been fully downloaded.
968 # Explanation to the above condition: in issue #175 it was revealed that
969 # YouTube sometimes adds or removes a few bytes from the end of the file,
970 # changing the file size slightly and causing problems for some users. So
971 # I decided to implement a suggested change and consider the file
972 # completely downloaded if the file size differs less than 100 bytes from
973 # the one in the hard drive.
974 self.report_file_already_downloaded(filename)
975 self.try_rename(tmpfilename, filename)
976 return True
977 else:
978 # The length does not match, we start the download over
979 self.report_unable_to_resume()
980 open_mode = 'wb'
981 break
982 # Retry
983 count += 1
984 if count <= retries:
985 self.report_retry(count, retries)
986
987 if count > retries:
988 self.trouble(u'ERROR: giving up after %s retries' % retries)
989 return False
990
991 data_len = data.info().get('Content-length', None)
992 if data_len is not None:
993 data_len = long(data_len) + resume_len
994 data_len_str = self.format_bytes(data_len)
995 byte_counter = 0 + resume_len
996 block_size = 1024
997 start = time.time()
998 while True:
999 # Download and write
1000 before = time.time()
1001 data_block = data.read(block_size)
1002 after = time.time()
1003 if len(data_block) == 0:
1004 break
1005 byte_counter += len(data_block)
1006
1007 # Open file just in time
1008 if stream is None:
1009 try:
1010 (stream, tmpfilename) = sanitize_open(tmpfilename, open_mode)
1011 assert stream is not None
1012 filename = self.undo_temp_name(tmpfilename)
1013 self.report_destination(filename)
1014 except (OSError, IOError), err:
1015 self.trouble(u'ERROR: unable to open for writing: %s' % str(err))
1016 return False
1017 try:
1018 stream.write(data_block)
1019 except (IOError, OSError), err:
1020 self.trouble(u'\nERROR: unable to write data: %s' % str(err))
1021 return False
1022 block_size = self.best_block_size(after - before, len(data_block))
1023
1024 # Progress message
1025 speed_str = self.calc_speed(start, time.time(), byte_counter - resume_len)
1026 if data_len is None:
1027 self.report_progress('Unknown %', data_len_str, speed_str, 'Unknown ETA')
1028 else:
1029 percent_str = self.calc_percent(byte_counter, data_len)
1030 eta_str = self.calc_eta(start, time.time(), data_len - resume_len, byte_counter - resume_len)
1031 self.report_progress(percent_str, data_len_str, speed_str, eta_str)
1032
1033 # Apply rate limit
1034 self.slow_down(start, byte_counter - resume_len)
1035
1036 if stream is None:
1037 self.trouble(u'\nERROR: Did not get any data blocks')
1038 return False
1039 stream.close()
1040 self.report_finish()
1041 if data_len is not None and byte_counter != data_len:
1042 raise ContentTooShortError(byte_counter, long(data_len))
1043 self.try_rename(tmpfilename, filename)
1044
1045 # Update file modification time
1046 if self.params.get('updatetime', True):
1047 info_dict['filetime'] = self.try_utime(filename, data.info().get('last-modified', None))
1048
1049 return True
1050
1051
1052 class InfoExtractor(object):
1053 """Information Extractor class.
1054
1055 Information extractors are the classes that, given a URL, extract
1056 information from the video (or videos) the URL refers to. This
1057 information includes the real video URL, the video title and simplified
1058 title, author and others. The information is stored in a dictionary
1059 which is then passed to the FileDownloader. The FileDownloader
1060 processes this information possibly downloading the video to the file
1061 system, among other possible outcomes. The dictionaries must include
1062 the following fields:
1063
1064 id: Video identifier.
1065 url: Final video URL.
1066 uploader: Nickname of the video uploader.
1067 title: Literal title.
1068 stitle: Simplified title.
1069 ext: Video filename extension.
1070 format: Video format.
1071 player_url: SWF Player URL (may be None).
1072
1073 The following fields are optional. Their primary purpose is to allow
1074 youtube-dl to serve as the backend for a video search function, such
1075 as the one in youtube2mp3. They are only used when their respective
1076 forced printing functions are called:
1077
1078 thumbnail: Full URL to a video thumbnail image.
1079 description: One-line video description.
1080
1081 Subclasses of this one should re-define the _real_initialize() and
1082 _real_extract() methods and define a _VALID_URL regexp.
1083 Probably, they should also be added to the list of extractors.
1084 """
1085
1086 _ready = False
1087 _downloader = None
1088
1089 def __init__(self, downloader=None):
1090 """Constructor. Receives an optional downloader."""
1091 self._ready = False
1092 self.set_downloader(downloader)
1093
1094 def suitable(self, url):
1095 """Receives a URL and returns True if suitable for this IE."""
1096 return re.match(self._VALID_URL, url) is not None
1097
1098 def initialize(self):
1099 """Initializes an instance (authentication, etc)."""
1100 if not self._ready:
1101 self._real_initialize()
1102 self._ready = True
1103
1104 def extract(self, url):
1105 """Extracts URL information and returns it in list of dicts."""
1106 self.initialize()
1107 return self._real_extract(url)
1108
1109 def set_downloader(self, downloader):
1110 """Sets the downloader for this IE."""
1111 self._downloader = downloader
1112
1113 def _real_initialize(self):
1114 """Real initialization process. Redefine in subclasses."""
1115 pass
1116
1117 def _real_extract(self, url):
1118 """Real extraction process. Redefine in subclasses."""
1119 pass
1120
1121
1122 class YoutubeIE(InfoExtractor):
1123 """Information extractor for youtube.com."""
1124
1125 _VALID_URL = r'^((?:https?://)?(?:youtu\.be/|(?:\w+\.)?youtube(?:-nocookie)?\.com/)(?!view_play_list|my_playlists|artist|playlist)(?:(?:(?:v|embed|e)/)|(?:(?:watch(?:_popup)?(?:\.php)?)?(?:\?|#!?)(?:.+&)?v=))?)?([0-9A-Za-z_-]+)(?(1).+)?$'
1126 _LANG_URL = r'http://www.youtube.com/?hl=en&persist_hl=1&gl=US&persist_gl=1&opt_out_ackd=1'
1127 _LOGIN_URL = 'https://www.youtube.com/signup?next=/&gl=US&hl=en'
1128 _AGE_URL = 'http://www.youtube.com/verify_age?next_url=/&gl=US&hl=en'
1129 _NETRC_MACHINE = 'youtube'
1130 # Listed in order of quality
1131 _available_formats = ['38', '37', '22', '45', '35', '44', '34', '18', '43', '6', '5', '17', '13']
1132 _available_formats_prefer_free = ['38', '37', '45', '22', '44', '35', '43', '34', '18', '6', '5', '17', '13']
1133 _video_extensions = {
1134 '13': '3gp',
1135 '17': 'mp4',
1136 '18': 'mp4',
1137 '22': 'mp4',
1138 '37': 'mp4',
1139 '38': 'video', # You actually don't know if this will be MOV, AVI or whatever
1140 '43': 'webm',
1141 '44': 'webm',
1142 '45': 'webm',
1143 }
1144 _video_dimensions = {
1145 '5': '240x400',
1146 '6': '???',
1147 '13': '???',
1148 '17': '144x176',
1149 '18': '360x640',
1150 '22': '720x1280',
1151 '34': '360x640',
1152 '35': '480x854',
1153 '37': '1080x1920',
1154 '38': '3072x4096',
1155 '43': '360x640',
1156 '44': '480x854',
1157 '45': '720x1280',
1158 }
1159 IE_NAME = u'youtube'
1160
1161 def report_lang(self):
1162 """Report attempt to set language."""
1163 self._downloader.to_screen(u'[youtube] Setting language')
1164
1165 def report_login(self):
1166 """Report attempt to log in."""
1167 self._downloader.to_screen(u'[youtube] Logging in')
1168
1169 def report_age_confirmation(self):
1170 """Report attempt to confirm age."""
1171 self._downloader.to_screen(u'[youtube] Confirming age')
1172
1173 def report_video_webpage_download(self, video_id):
1174 """Report attempt to download video webpage."""
1175 self._downloader.to_screen(u'[youtube] %s: Downloading video webpage' % video_id)
1176
1177 def report_video_info_webpage_download(self, video_id):
1178 """Report attempt to download video info webpage."""
1179 self._downloader.to_screen(u'[youtube] %s: Downloading video info webpage' % video_id)
1180
1181 def report_information_extraction(self, video_id):
1182 """Report attempt to extract video information."""
1183 self._downloader.to_screen(u'[youtube] %s: Extracting video information' % video_id)
1184
1185 def report_unavailable_format(self, video_id, format):
1186 """Report extracted video URL."""
1187 self._downloader.to_screen(u'[youtube] %s: Format %s not available' % (video_id, format))
1188
1189 def report_rtmp_download(self):
1190 """Indicate the download will use the RTMP protocol."""
1191 self._downloader.to_screen(u'[youtube] RTMP download detected')
1192
1193 def _print_formats(self, formats):
1194 print 'Available formats:'
1195 for x in formats:
1196 print '%s\t:\t%s\t[%s]' %(x, self._video_extensions.get(x, 'flv'), self._video_dimensions.get(x, '???'))
1197
1198 def _real_initialize(self):
1199 if self._downloader is None:
1200 return
1201
1202 username = None
1203 password = None
1204 downloader_params = self._downloader.params
1205
1206 # Attempt to use provided username and password or .netrc data
1207 if downloader_params.get('username', None) is not None:
1208 username = downloader_params['username']
1209 password = downloader_params['password']
1210 elif downloader_params.get('usenetrc', False):
1211 try:
1212 info = netrc.netrc().authenticators(self._NETRC_MACHINE)
1213 if info is not None:
1214 username = info[0]
1215 password = info[2]
1216 else:
1217 raise netrc.NetrcParseError('No authenticators for %s' % self._NETRC_MACHINE)
1218 except (IOError, netrc.NetrcParseError), err:
1219 self._downloader.to_stderr(u'WARNING: parsing .netrc: %s' % str(err))
1220 return
1221
1222 # Set language
1223 request = urllib2.Request(self._LANG_URL)
1224 try:
1225 self.report_lang()
1226 urllib2.urlopen(request).read()
1227 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
1228 self._downloader.to_stderr(u'WARNING: unable to set language: %s' % str(err))
1229 return
1230
1231 # No authentication to be performed
1232 if username is None:
1233 return
1234
1235 # Log in
1236 login_form = {
1237 'current_form': 'loginForm',
1238 'next': '/',
1239 'action_login': 'Log In',
1240 'username': username,
1241 'password': password,
1242 }
1243 request = urllib2.Request(self._LOGIN_URL, urllib.urlencode(login_form))
1244 try:
1245 self.report_login()
1246 login_results = urllib2.urlopen(request).read()
1247 if re.search(r'(?i)<form[^>]* name="loginForm"', login_results) is not None:
1248 self._downloader.to_stderr(u'WARNING: unable to log in: bad username or password')
1249 return
1250 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
1251 self._downloader.to_stderr(u'WARNING: unable to log in: %s' % str(err))
1252 return
1253
1254 # Confirm age
1255 age_form = {
1256 'next_url': '/',
1257 'action_confirm': 'Confirm',
1258 }
1259 request = urllib2.Request(self._AGE_URL, urllib.urlencode(age_form))
1260 try:
1261 self.report_age_confirmation()
1262 age_results = urllib2.urlopen(request).read()
1263 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
1264 self._downloader.trouble(u'ERROR: unable to confirm age: %s' % str(err))
1265 return
1266
1267 def _real_extract(self, url):
1268 # Extract video id from URL
1269 mobj = re.match(self._VALID_URL, url)
1270 if mobj is None:
1271 self._downloader.trouble(u'ERROR: invalid URL: %s' % url)
1272 return
1273 video_id = mobj.group(2)
1274
1275 # Get video webpage
1276 self.report_video_webpage_download(video_id)
1277 request = urllib2.Request('http://www.youtube.com/watch?v=%s&gl=US&hl=en&has_verified=1' % video_id)
1278 try:
1279 video_webpage = urllib2.urlopen(request).read()
1280 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
1281 self._downloader.trouble(u'ERROR: unable to download video webpage: %s' % str(err))
1282 return
1283
1284 # Attempt to extract SWF player URL
1285 mobj = re.search(r'swfConfig.*?"(http:\\/\\/.*?watch.*?-.*?\.swf)"', video_webpage)
1286 if mobj is not None:
1287 player_url = re.sub(r'\\(.)', r'\1', mobj.group(1))
1288 else:
1289 player_url = None
1290
1291 # Get video info
1292 self.report_video_info_webpage_download(video_id)
1293 for el_type in ['&el=embedded', '&el=detailpage', '&el=vevo', '']:
1294 video_info_url = ('http://www.youtube.com/get_video_info?&video_id=%s%s&ps=default&eurl=&gl=US&hl=en'
1295 % (video_id, el_type))
1296 request = urllib2.Request(video_info_url)
1297 try:
1298 video_info_webpage = urllib2.urlopen(request).read()
1299 video_info = parse_qs(video_info_webpage)
1300 if 'token' in video_info:
1301 break
1302 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
1303 self._downloader.trouble(u'ERROR: unable to download video info webpage: %s' % str(err))
1304 return
1305 if 'token' not in video_info:
1306 if 'reason' in video_info:
1307 self._downloader.trouble(u'ERROR: YouTube said: %s' % video_info['reason'][0].decode('utf-8'))
1308 else:
1309 self._downloader.trouble(u'ERROR: "token" parameter not in video info for unknown reason')
1310 return
1311
1312 # Start extracting information
1313 self.report_information_extraction(video_id)
1314
1315 # uploader
1316 if 'author' not in video_info:
1317 self._downloader.trouble(u'ERROR: unable to extract uploader nickname')
1318 return
1319 video_uploader = urllib.unquote_plus(video_info['author'][0])
1320
1321 # title
1322 if 'title' not in video_info:
1323 self._downloader.trouble(u'ERROR: unable to extract video title')
1324 return
1325 video_title = urllib.unquote_plus(video_info['title'][0])
1326 video_title = video_title.decode('utf-8')
1327 video_title = sanitize_title(video_title)
1328
1329 # simplified title
1330 simple_title = _simplify_title(video_title)
1331
1332 # thumbnail image
1333 if 'thumbnail_url' not in video_info:
1334 self._downloader.trouble(u'WARNING: unable to extract video thumbnail')
1335 video_thumbnail = ''
1336 else: # don't panic if we can't find it
1337 video_thumbnail = urllib.unquote_plus(video_info['thumbnail_url'][0])
1338
1339 # upload date
1340 upload_date = u'NA'
1341 mobj = re.search(r'id="eow-date.*?>(.*?)</span>', video_webpage, re.DOTALL)
1342 if mobj is not None:
1343 upload_date = ' '.join(re.sub(r'[/,-]', r' ', mobj.group(1)).split())
1344 format_expressions = ['%d %B %Y', '%B %d %Y', '%b %d %Y']
1345 for expression in format_expressions:
1346 try:
1347 upload_date = datetime.datetime.strptime(upload_date, expression).strftime('%Y%m%d')
1348 except:
1349 pass
1350
1351 # description
1352 try:
1353 lxml.etree
1354 except NameError:
1355 video_description = u'No description available.'
1356 if self._downloader.params.get('forcedescription', False) or self._downloader.params.get('writedescription', False):
1357 mobj = re.search(r'<meta name="description" content="(.*)"(?:\s*/)?>', video_webpage)
1358 if mobj is not None:
1359 video_description = mobj.group(1).decode('utf-8')
1360 else:
1361 html_parser = lxml.etree.HTMLParser(encoding='utf-8')
1362 vwebpage_doc = lxml.etree.parse(StringIO.StringIO(video_webpage), html_parser)
1363 video_description = u''.join(vwebpage_doc.xpath('id("eow-description")//text()'))
1364 # TODO use another parser
1365
1366 # token
1367 video_token = urllib.unquote_plus(video_info['token'][0])
1368
1369 # Decide which formats to download
1370 req_format = self._downloader.params.get('format', None)
1371
1372 if 'conn' in video_info and video_info['conn'][0].startswith('rtmp'):
1373 self.report_rtmp_download()
1374 video_url_list = [(None, video_info['conn'][0])]
1375 elif 'url_encoded_fmt_stream_map' in video_info and len(video_info['url_encoded_fmt_stream_map']) >= 1:
1376 url_data_strs = video_info['url_encoded_fmt_stream_map'][0].split(',')
1377 url_data = [parse_qs(uds) for uds in url_data_strs]
1378 url_data = filter(lambda ud: 'itag' in ud and 'url' in ud, url_data)
1379 url_map = dict((ud['itag'][0], ud['url'][0]) for ud in url_data)
1380
1381 format_limit = self._downloader.params.get('format_limit', None)
1382 available_formats = self._available_formats_prefer_free if self._downloader.params.get('prefer_free_formats', False) else self._available_formats
1383 if format_limit is not None and format_limit in available_formats:
1384 format_list = available_formats[available_formats.index(format_limit):]
1385 else:
1386 format_list = available_formats
1387 existing_formats = [x for x in format_list if x in url_map]
1388 if len(existing_formats) == 0:
1389 self._downloader.trouble(u'ERROR: no known formats available for video')
1390 return
1391 if self._downloader.params.get('listformats', None):
1392 self._print_formats(existing_formats)
1393 return
1394 if req_format is None or req_format == 'best':
1395 video_url_list = [(existing_formats[0], url_map[existing_formats[0]])] # Best quality
1396 elif req_format == 'worst':
1397 video_url_list = [(existing_formats[len(existing_formats)-1], url_map[existing_formats[len(existing_formats)-1]])] # worst quality
1398 elif req_format in ('-1', 'all'):
1399 video_url_list = [(f, url_map[f]) for f in existing_formats] # All formats
1400 else:
1401 # Specific formats. We pick the first in a slash-delimeted sequence.
1402 # For example, if '1/2/3/4' is requested and '2' and '4' are available, we pick '2'.
1403 req_formats = req_format.split('/')
1404 video_url_list = None
1405 for rf in req_formats:
1406 if rf in url_map:
1407 video_url_list = [(rf, url_map[rf])]
1408 break
1409 if video_url_list is None:
1410 self._downloader.trouble(u'ERROR: requested format not available')
1411 return
1412 else:
1413 self._downloader.trouble(u'ERROR: no conn or url_encoded_fmt_stream_map information found in video info')
1414 return
1415
1416 for format_param, video_real_url in video_url_list:
1417 # At this point we have a new video
1418 self._downloader.increment_downloads()
1419
1420 # Extension
1421 video_extension = self._video_extensions.get(format_param, 'flv')
1422
1423 try:
1424 # Process video information
1425 self._downloader.process_info({
1426 'id': video_id.decode('utf-8'),
1427 'url': video_real_url.decode('utf-8'),
1428 'uploader': video_uploader.decode('utf-8'),
1429 'upload_date': upload_date,
1430 'title': video_title,
1431 'stitle': simple_title,
1432 'ext': video_extension.decode('utf-8'),
1433 'format': (format_param is None and u'NA' or format_param.decode('utf-8')),
1434 'thumbnail': video_thumbnail.decode('utf-8'),
1435 'description': video_description,
1436 'player_url': player_url,
1437 })
1438 except UnavailableVideoError, err:
1439 self._downloader.trouble(u'\nERROR: unable to download video')
1440
1441
1442 class MetacafeIE(InfoExtractor):
1443 """Information Extractor for metacafe.com."""
1444
1445 _VALID_URL = r'(?:http://)?(?:www\.)?metacafe\.com/watch/([^/]+)/([^/]+)/.*'
1446 _DISCLAIMER = 'http://www.metacafe.com/family_filter/'
1447 _FILTER_POST = 'http://www.metacafe.com/f/index.php?inputType=filter&controllerGroup=user'
1448 _youtube_ie = None
1449 IE_NAME = u'metacafe'
1450
1451 def __init__(self, youtube_ie, downloader=None):
1452 InfoExtractor.__init__(self, downloader)
1453 self._youtube_ie = youtube_ie
1454
1455 def report_disclaimer(self):
1456 """Report disclaimer retrieval."""
1457 self._downloader.to_screen(u'[metacafe] Retrieving disclaimer')
1458
1459 def report_age_confirmation(self):
1460 """Report attempt to confirm age."""
1461 self._downloader.to_screen(u'[metacafe] Confirming age')
1462
1463 def report_download_webpage(self, video_id):
1464 """Report webpage download."""
1465 self._downloader.to_screen(u'[metacafe] %s: Downloading webpage' % video_id)
1466
1467 def report_extraction(self, video_id):
1468 """Report information extraction."""
1469 self._downloader.to_screen(u'[metacafe] %s: Extracting information' % video_id)
1470
1471 def _real_initialize(self):
1472 # Retrieve disclaimer
1473 request = urllib2.Request(self._DISCLAIMER)
1474 try:
1475 self.report_disclaimer()
1476 disclaimer = urllib2.urlopen(request).read()
1477 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
1478 self._downloader.trouble(u'ERROR: unable to retrieve disclaimer: %s' % str(err))
1479 return
1480
1481 # Confirm age
1482 disclaimer_form = {
1483 'filters': '0',
1484 'submit': "Continue - I'm over 18",
1485 }
1486 request = urllib2.Request(self._FILTER_POST, urllib.urlencode(disclaimer_form))
1487 try:
1488 self.report_age_confirmation()
1489 disclaimer = urllib2.urlopen(request).read()
1490 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
1491 self._downloader.trouble(u'ERROR: unable to confirm age: %s' % str(err))
1492 return
1493
1494 def _real_extract(self, url):
1495 # Extract id and simplified title from URL
1496 mobj = re.match(self._VALID_URL, url)
1497 if mobj is None:
1498 self._downloader.trouble(u'ERROR: invalid URL: %s' % url)
1499 return
1500
1501 video_id = mobj.group(1)
1502
1503 # Check if video comes from YouTube
1504 mobj2 = re.match(r'^yt-(.*)$', video_id)
1505 if mobj2 is not None:
1506 self._youtube_ie.extract('http://www.youtube.com/watch?v=%s' % mobj2.group(1))
1507 return
1508
1509 # At this point we have a new video
1510 self._downloader.increment_downloads()
1511
1512 simple_title = mobj.group(2).decode('utf-8')
1513
1514 # Retrieve video webpage to extract further information
1515 request = urllib2.Request('http://www.metacafe.com/watch/%s/' % video_id)
1516 try:
1517 self.report_download_webpage(video_id)
1518 webpage = urllib2.urlopen(request).read()
1519 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
1520 self._downloader.trouble(u'ERROR: unable retrieve video webpage: %s' % str(err))
1521 return
1522
1523 # Extract URL, uploader and title from webpage
1524 self.report_extraction(video_id)
1525 mobj = re.search(r'(?m)&mediaURL=([^&]+)', webpage)
1526 if mobj is not None:
1527 mediaURL = urllib.unquote(mobj.group(1))
1528 video_extension = mediaURL[-3:]
1529
1530 # Extract gdaKey if available
1531 mobj = re.search(r'(?m)&gdaKey=(.*?)&', webpage)
1532 if mobj is None:
1533 video_url = mediaURL
1534 else:
1535 gdaKey = mobj.group(1)
1536 video_url = '%s?__gda__=%s' % (mediaURL, gdaKey)
1537 else:
1538 mobj = re.search(r' name="flashvars" value="(.*?)"', webpage)
1539 if mobj is None:
1540 self._downloader.trouble(u'ERROR: unable to extract media URL')
1541 return
1542 vardict = parse_qs(mobj.group(1))
1543 if 'mediaData' not in vardict:
1544 self._downloader.trouble(u'ERROR: unable to extract media URL')
1545 return
1546 mobj = re.search(r'"mediaURL":"(http.*?)","key":"(.*?)"', vardict['mediaData'][0])
1547 if mobj is None:
1548 self._downloader.trouble(u'ERROR: unable to extract media URL')
1549 return
1550 mediaURL = mobj.group(1).replace('\\/', '/')
1551 video_extension = mediaURL[-3:]
1552 video_url = '%s?__gda__=%s' % (mediaURL, mobj.group(2))
1553
1554 mobj = re.search(r'(?im)<title>(.*) - Video</title>', webpage)
1555 if mobj is None:
1556 self._downloader.trouble(u'ERROR: unable to extract title')
1557 return
1558 video_title = mobj.group(1).decode('utf-8')
1559 video_title = sanitize_title(video_title)
1560
1561 mobj = re.search(r'(?ms)By:\s*<a .*?>(.+?)<', webpage)
1562 if mobj is None:
1563 self._downloader.trouble(u'ERROR: unable to extract uploader nickname')
1564 return
1565 video_uploader = mobj.group(1)
1566
1567 try:
1568 # Process video information
1569 self._downloader.process_info({
1570 'id': video_id.decode('utf-8'),
1571 'url': video_url.decode('utf-8'),
1572 'uploader': video_uploader.decode('utf-8'),
1573 'upload_date': u'NA',
1574 'title': video_title,
1575 'stitle': simple_title,
1576 'ext': video_extension.decode('utf-8'),
1577 'format': u'NA',
1578 'player_url': None,
1579 })
1580 except UnavailableVideoError:
1581 self._downloader.trouble(u'\nERROR: unable to download video')
1582
1583
1584 class DailymotionIE(InfoExtractor):
1585 """Information Extractor for Dailymotion"""
1586
1587 _VALID_URL = r'(?i)(?:https?://)?(?:www\.)?dailymotion\.[a-z]{2,3}/video/([^_/]+)_([^/]+)'
1588 IE_NAME = u'dailymotion'
1589
1590 def __init__(self, downloader=None):
1591 InfoExtractor.__init__(self, downloader)
1592
1593 def report_download_webpage(self, video_id):
1594 """Report webpage download."""
1595 self._downloader.to_screen(u'[dailymotion] %s: Downloading webpage' % video_id)
1596
1597 def report_extraction(self, video_id):
1598 """Report information extraction."""
1599 self._downloader.to_screen(u'[dailymotion] %s: Extracting information' % video_id)
1600
1601 def _real_extract(self, url):
1602 # Extract id and simplified title from URL
1603 mobj = re.match(self._VALID_URL, url)
1604 if mobj is None:
1605 self._downloader.trouble(u'ERROR: invalid URL: %s' % url)
1606 return
1607
1608 # At this point we have a new video
1609 self._downloader.increment_downloads()
1610 video_id = mobj.group(1)
1611
1612 video_extension = 'flv'
1613
1614 # Retrieve video webpage to extract further information
1615 request = urllib2.Request(url)
1616 request.add_header('Cookie', 'family_filter=off')
1617 try:
1618 self.report_download_webpage(video_id)
1619 webpage = urllib2.urlopen(request).read()
1620 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
1621 self._downloader.trouble(u'ERROR: unable retrieve video webpage: %s' % str(err))
1622 return
1623
1624 # Extract URL, uploader and title from webpage
1625 self.report_extraction(video_id)
1626 mobj = re.search(r'(?i)addVariable\(\"sequence\"\s*,\s*\"([^\"]+?)\"\)', webpage)
1627 if mobj is None:
1628 self._downloader.trouble(u'ERROR: unable to extract media URL')
1629 return
1630 sequence = urllib.unquote(mobj.group(1))
1631 mobj = re.search(r',\"sdURL\"\:\"([^\"]+?)\",', sequence)
1632 if mobj is None:
1633 self._downloader.trouble(u'ERROR: unable to extract media URL')
1634 return
1635 mediaURL = urllib.unquote(mobj.group(1)).replace('\\', '')
1636
1637 # if needed add http://www.dailymotion.com/ if relative URL
1638
1639 video_url = mediaURL
1640
1641 mobj = re.search(r'<meta property="og:title" content="(?P<title>[^"]*)" />', webpage)
1642 if mobj is None:
1643 self._downloader.trouble(u'ERROR: unable to extract title')
1644 return
1645 video_title = _unescapeHTML(mobj.group('title').decode('utf-8'))
1646 video_title = sanitize_title(video_title)
1647 simple_title = _simplify_title(video_title)
1648
1649 mobj = re.search(r'(?im)<span class="owner[^\"]+?">[^<]+?<a [^>]+?>([^<]+?)</a></span>', webpage)
1650 if mobj is None:
1651 self._downloader.trouble(u'ERROR: unable to extract uploader nickname')
1652 return
1653 video_uploader = mobj.group(1)
1654
1655 try:
1656 # Process video information
1657 self._downloader.process_info({
1658 'id': video_id.decode('utf-8'),
1659 'url': video_url.decode('utf-8'),
1660 'uploader': video_uploader.decode('utf-8'),
1661 'upload_date': u'NA',
1662 'title': video_title,
1663 'stitle': simple_title,
1664 'ext': video_extension.decode('utf-8'),
1665 'format': u'NA',
1666 'player_url': None,
1667 })
1668 except UnavailableVideoError:
1669 self._downloader.trouble(u'\nERROR: unable to download video')
1670
1671
1672 class GoogleIE(InfoExtractor):
1673 """Information extractor for video.google.com."""
1674
1675 _VALID_URL = r'(?:http://)?video\.google\.(?:com(?:\.au)?|co\.(?:uk|jp|kr|cr)|ca|de|es|fr|it|nl|pl)/videoplay\?docid=([^\&]+).*'
1676 IE_NAME = u'video.google'
1677
1678 def __init__(self, downloader=None):
1679 InfoExtractor.__init__(self, downloader)
1680
1681 def report_download_webpage(self, video_id):
1682 """Report webpage download."""
1683 self._downloader.to_screen(u'[video.google] %s: Downloading webpage' % video_id)
1684
1685 def report_extraction(self, video_id):
1686 """Report information extraction."""
1687 self._downloader.to_screen(u'[video.google] %s: Extracting information' % video_id)
1688
1689 def _real_extract(self, url):
1690 # Extract id from URL
1691 mobj = re.match(self._VALID_URL, url)
1692 if mobj is None:
1693 self._downloader.trouble(u'ERROR: Invalid URL: %s' % url)
1694 return
1695
1696 # At this point we have a new video
1697 self._downloader.increment_downloads()
1698 video_id = mobj.group(1)
1699
1700 video_extension = 'mp4'
1701
1702 # Retrieve video webpage to extract further information
1703 request = urllib2.Request('http://video.google.com/videoplay?docid=%s&hl=en&oe=utf-8' % video_id)
1704 try:
1705 self.report_download_webpage(video_id)
1706 webpage = urllib2.urlopen(request).read()
1707 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
1708 self._downloader.trouble(u'ERROR: Unable to retrieve video webpage: %s' % str(err))
1709 return
1710
1711 # Extract URL, uploader, and title from webpage
1712 self.report_extraction(video_id)
1713 mobj = re.search(r"download_url:'([^']+)'", webpage)
1714 if mobj is None:
1715 video_extension = 'flv'
1716 mobj = re.search(r"(?i)videoUrl\\x3d(.+?)\\x26", webpage)
1717 if mobj is None:
1718 self._downloader.trouble(u'ERROR: unable to extract media URL')
1719 return
1720 mediaURL = urllib.unquote(mobj.group(1))
1721 mediaURL = mediaURL.replace('\\x3d', '\x3d')
1722 mediaURL = mediaURL.replace('\\x26', '\x26')
1723
1724 video_url = mediaURL
1725
1726 mobj = re.search(r'<title>(.*)</title>', webpage)
1727 if mobj is None:
1728 self._downloader.trouble(u'ERROR: unable to extract title')
1729 return
1730 video_title = mobj.group(1).decode('utf-8')
1731 video_title = sanitize_title(video_title)
1732 simple_title = _simplify_title(video_title)
1733
1734 # Extract video description
1735 mobj = re.search(r'<span id=short-desc-content>([^<]*)</span>', webpage)
1736 if mobj is None:
1737 self._downloader.trouble(u'ERROR: unable to extract video description')
1738 return
1739 video_description = mobj.group(1).decode('utf-8')
1740 if not video_description:
1741 video_description = 'No description available.'
1742
1743 # Extract video thumbnail
1744 if self._downloader.params.get('forcethumbnail', False):
1745 request = urllib2.Request('http://video.google.com/videosearch?q=%s+site:video.google.com&hl=en' % abs(int(video_id)))
1746 try:
1747 webpage = urllib2.urlopen(request).read()
1748 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
1749 self._downloader.trouble(u'ERROR: Unable to retrieve video webpage: %s' % str(err))
1750 return
1751 mobj = re.search(r'<img class=thumbnail-img (?:.* )?src=(http.*)>', webpage)
1752 if mobj is None:
1753 self._downloader.trouble(u'ERROR: unable to extract video thumbnail')
1754 return
1755 video_thumbnail = mobj.group(1)
1756 else: # we need something to pass to process_info
1757 video_thumbnail = ''
1758
1759 try:
1760 # Process video information
1761 self._downloader.process_info({
1762 'id': video_id.decode('utf-8'),
1763 'url': video_url.decode('utf-8'),
1764 'uploader': u'NA',
1765 'upload_date': u'NA',
1766 'title': video_title,
1767 'stitle': simple_title,
1768 'ext': video_extension.decode('utf-8'),
1769 'format': u'NA',
1770 'player_url': None,
1771 })
1772 except UnavailableVideoError:
1773 self._downloader.trouble(u'\nERROR: unable to download video')
1774
1775
1776 class PhotobucketIE(InfoExtractor):
1777 """Information extractor for photobucket.com."""
1778
1779 _VALID_URL = r'(?:http://)?(?:[a-z0-9]+\.)?photobucket\.com/.*[\?\&]current=(.*\.flv)'
1780 IE_NAME = u'photobucket'
1781
1782 def __init__(self, downloader=None):
1783 InfoExtractor.__init__(self, downloader)
1784
1785 def report_download_webpage(self, video_id):
1786 """Report webpage download."""
1787 self._downloader.to_screen(u'[photobucket] %s: Downloading webpage' % video_id)
1788
1789 def report_extraction(self, video_id):
1790 """Report information extraction."""
1791 self._downloader.to_screen(u'[photobucket] %s: Extracting information' % video_id)
1792
1793 def _real_extract(self, url):
1794 # Extract id from URL
1795 mobj = re.match(self._VALID_URL, url)
1796 if mobj is None:
1797 self._downloader.trouble(u'ERROR: Invalid URL: %s' % url)
1798 return
1799
1800 # At this point we have a new video
1801 self._downloader.increment_downloads()
1802 video_id = mobj.group(1)
1803
1804 video_extension = 'flv'
1805
1806 # Retrieve video webpage to extract further information
1807 request = urllib2.Request(url)
1808 try:
1809 self.report_download_webpage(video_id)
1810 webpage = urllib2.urlopen(request).read()
1811 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
1812 self._downloader.trouble(u'ERROR: Unable to retrieve video webpage: %s' % str(err))
1813 return
1814
1815 # Extract URL, uploader, and title from webpage
1816 self.report_extraction(video_id)
1817 mobj = re.search(r'<link rel="video_src" href=".*\?file=([^"]+)" />', webpage)
1818 if mobj is None:
1819 self._downloader.trouble(u'ERROR: unable to extract media URL')
1820 return
1821 mediaURL = urllib.unquote(mobj.group(1))
1822
1823 video_url = mediaURL
1824
1825 mobj = re.search(r'<title>(.*) video by (.*) - Photobucket</title>', webpage)
1826 if mobj is None:
1827 self._downloader.trouble(u'ERROR: unable to extract title')
1828 return
1829 video_title = mobj.group(1).decode('utf-8')
1830 video_title = sanitize_title(video_title)
1831 simple_title = _simplify_title(vide_title)
1832
1833 video_uploader = mobj.group(2).decode('utf-8')
1834
1835 try:
1836 # Process video information
1837 self._downloader.process_info({
1838 'id': video_id.decode('utf-8'),
1839 'url': video_url.decode('utf-8'),
1840 'uploader': video_uploader,
1841 'upload_date': u'NA',
1842 'title': video_title,
1843 'stitle': simple_title,
1844 'ext': video_extension.decode('utf-8'),
1845 'format': u'NA',
1846 'player_url': None,
1847 })
1848 except UnavailableVideoError:
1849 self._downloader.trouble(u'\nERROR: unable to download video')
1850
1851
1852 class YahooIE(InfoExtractor):
1853 """Information extractor for video.yahoo.com."""
1854
1855 # _VALID_URL matches all Yahoo! Video URLs
1856 # _VPAGE_URL matches only the extractable '/watch/' URLs
1857 _VALID_URL = r'(?:http://)?(?:[a-z]+\.)?video\.yahoo\.com/(?:watch|network)/([0-9]+)(?:/|\?v=)([0-9]+)(?:[#\?].*)?'
1858 _VPAGE_URL = r'(?:http://)?video\.yahoo\.com/watch/([0-9]+)/([0-9]+)(?:[#\?].*)?'
1859 IE_NAME = u'video.yahoo'
1860
1861 def __init__(self, downloader=None):
1862 InfoExtractor.__init__(self, downloader)
1863
1864 def report_download_webpage(self, video_id):
1865 """Report webpage download."""
1866 self._downloader.to_screen(u'[video.yahoo] %s: Downloading webpage' % video_id)
1867
1868 def report_extraction(self, video_id):
1869 """Report information extraction."""
1870 self._downloader.to_screen(u'[video.yahoo] %s: Extracting information' % video_id)
1871
1872 def _real_extract(self, url, new_video=True):
1873 # Extract ID from URL
1874 mobj = re.match(self._VALID_URL, url)
1875 if mobj is None:
1876 self._downloader.trouble(u'ERROR: Invalid URL: %s' % url)
1877 return
1878
1879 # At this point we have a new video
1880 self._downloader.increment_downloads()
1881 video_id = mobj.group(2)
1882 video_extension = 'flv'
1883
1884 # Rewrite valid but non-extractable URLs as
1885 # extractable English language /watch/ URLs
1886 if re.match(self._VPAGE_URL, url) is None:
1887 request = urllib2.Request(url)
1888 try:
1889 webpage = urllib2.urlopen(request).read()
1890 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
1891 self._downloader.trouble(u'ERROR: Unable to retrieve video webpage: %s' % str(err))
1892 return
1893
1894 mobj = re.search(r'\("id", "([0-9]+)"\);', webpage)
1895 if mobj is None:
1896 self._downloader.trouble(u'ERROR: Unable to extract id field')
1897 return
1898 yahoo_id = mobj.group(1)
1899
1900 mobj = re.search(r'\("vid", "([0-9]+)"\);', webpage)
1901 if mobj is None:
1902 self._downloader.trouble(u'ERROR: Unable to extract vid field')
1903 return
1904 yahoo_vid = mobj.group(1)
1905
1906 url = 'http://video.yahoo.com/watch/%s/%s' % (yahoo_vid, yahoo_id)
1907 return self._real_extract(url, new_video=False)
1908
1909 # Retrieve video webpage to extract further information
1910 request = urllib2.Request(url)
1911 try:
1912 self.report_download_webpage(video_id)
1913 webpage = urllib2.urlopen(request).read()
1914 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
1915 self._downloader.trouble(u'ERROR: Unable to retrieve video webpage: %s' % str(err))
1916 return
1917
1918 # Extract uploader and title from webpage
1919 self.report_extraction(video_id)
1920 mobj = re.search(r'<meta name="title" content="(.*)" />', webpage)
1921 if mobj is None:
1922 self._downloader.trouble(u'ERROR: unable to extract video title')
1923 return
1924 video_title = mobj.group(1).decode('utf-8')
1925 simple_title = _simplify_title(video_title)
1926
1927 mobj = re.search(r'<h2 class="ti-5"><a href="http://video\.yahoo\.com/(people|profile)/[0-9]+" beacon=".*">(.*)</a></h2>', webpage)
1928 if mobj is None:
1929 self._downloader.trouble(u'ERROR: unable to extract video uploader')
1930 return
1931 video_uploader = mobj.group(1).decode('utf-8')
1932
1933 # Extract video thumbnail
1934 mobj = re.search(r'<link rel="image_src" href="(.*)" />', webpage)
1935 if mobj is None:
1936 self._downloader.trouble(u'ERROR: unable to extract video thumbnail')
1937 return
1938 video_thumbnail = mobj.group(1).decode('utf-8')
1939
1940 # Extract video description
1941 mobj = re.search(r'<meta name="description" content="(.*)" />', webpage)
1942 if mobj is None:
1943 self._downloader.trouble(u'ERROR: unable to extract video description')
1944 return
1945 video_description = mobj.group(1).decode('utf-8')
1946 if not video_description:
1947 video_description = 'No description available.'
1948
1949 # Extract video height and width
1950 mobj = re.search(r'<meta name="video_height" content="([0-9]+)" />', webpage)
1951 if mobj is None:
1952 self._downloader.trouble(u'ERROR: unable to extract video height')
1953 return
1954 yv_video_height = mobj.group(1)
1955
1956 mobj = re.search(r'<meta name="video_width" content="([0-9]+)" />', webpage)
1957 if mobj is None:
1958 self._downloader.trouble(u'ERROR: unable to extract video width')
1959 return
1960 yv_video_width = mobj.group(1)
1961
1962 # Retrieve video playlist to extract media URL
1963 # I'm not completely sure what all these options are, but we
1964 # seem to need most of them, otherwise the server sends a 401.
1965 yv_lg = 'R0xx6idZnW2zlrKP8xxAIR' # not sure what this represents
1966 yv_bitrate = '700' # according to Wikipedia this is hard-coded
1967 request = urllib2.Request('http://cosmos.bcst.yahoo.com/up/yep/process/getPlaylistFOP.php?node_id=' + video_id +
1968 '&tech=flash&mode=playlist&lg=' + yv_lg + '&bitrate=' + yv_bitrate + '&vidH=' + yv_video_height +
1969 '&vidW=' + yv_video_width + '&swf=as3&rd=video.yahoo.com&tk=null&adsupported=v1,v2,&eventid=1301797')
1970 try:
1971 self.report_download_webpage(video_id)
1972 webpage = urllib2.urlopen(request).read()
1973 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
1974 self._downloader.trouble(u'ERROR: Unable to retrieve video webpage: %s' % str(err))
1975 return
1976
1977 # Extract media URL from playlist XML
1978 mobj = re.search(r'<STREAM APP="(http://.*)" FULLPATH="/?(/.*\.flv\?[^"]*)"', webpage)
1979 if mobj is None:
1980 self._downloader.trouble(u'ERROR: Unable to extract media URL')
1981 return
1982 video_url = urllib.unquote(mobj.group(1) + mobj.group(2)).decode('utf-8')
1983 video_url = re.sub(r'(?u)&(.+?);', htmlentity_transform, video_url)
1984
1985 try:
1986 # Process video information
1987 self._downloader.process_info({
1988 'id': video_id.decode('utf-8'),
1989 'url': video_url,
1990 'uploader': video_uploader,
1991 'upload_date': u'NA',
1992 'title': video_title,
1993 'stitle': simple_title,
1994 'ext': video_extension.decode('utf-8'),
1995 'thumbnail': video_thumbnail.decode('utf-8'),
1996 'description': video_description,
1997 'thumbnail': video_thumbnail,
1998 'player_url': None,
1999 })
2000 except UnavailableVideoError:
2001 self._downloader.trouble(u'\nERROR: unable to download video')
2002
2003
2004 class VimeoIE(InfoExtractor):
2005 """Information extractor for vimeo.com."""
2006
2007 # _VALID_URL matches Vimeo URLs
2008 _VALID_URL = r'(?:https?://)?(?:(?:www|player).)?vimeo\.com/(?:groups/[^/]+/)?(?:videos?/)?([0-9]+)'
2009 IE_NAME = u'vimeo'
2010
2011 def __init__(self, downloader=None):
2012 InfoExtractor.__init__(self, downloader)
2013
2014 def report_download_webpage(self, video_id):
2015 """Report webpage download."""
2016 self._downloader.to_screen(u'[vimeo] %s: Downloading webpage' % video_id)
2017
2018 def report_extraction(self, video_id):
2019 """Report information extraction."""
2020 self._downloader.to_screen(u'[vimeo] %s: Extracting information' % video_id)
2021
2022 def _real_extract(self, url, new_video=True):
2023 # Extract ID from URL
2024 mobj = re.match(self._VALID_URL, url)
2025 if mobj is None:
2026 self._downloader.trouble(u'ERROR: Invalid URL: %s' % url)
2027 return
2028
2029 # At this point we have a new video
2030 self._downloader.increment_downloads()
2031 video_id = mobj.group(1)
2032
2033 # Retrieve video webpage to extract further information
2034 request = urllib2.Request("http://vimeo.com/moogaloop/load/clip:%s" % video_id, None, std_headers)
2035 try:
2036 self.report_download_webpage(video_id)
2037 webpage = urllib2.urlopen(request).read()
2038 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
2039 self._downloader.trouble(u'ERROR: Unable to retrieve video webpage: %s' % str(err))
2040 return
2041
2042 # Now we begin extracting as much information as we can from what we
2043 # retrieved. First we extract the information common to all extractors,
2044 # and latter we extract those that are Vimeo specific.
2045 self.report_extraction(video_id)
2046
2047 # Extract title
2048 mobj = re.search(r'<caption>(.*?)</caption>', webpage)
2049 if mobj is None:
2050 self._downloader.trouble(u'ERROR: unable to extract video title')
2051 return
2052 video_title = mobj.group(1).decode('utf-8')
2053 simple_title = _simplify_title(video_title)
2054
2055 # Extract uploader
2056 mobj = re.search(r'<uploader_url>http://vimeo.com/(.*?)</uploader_url>', webpage)
2057 if mobj is None:
2058 self._downloader.trouble(u'ERROR: unable to extract video uploader')
2059 return
2060 video_uploader = mobj.group(1).decode('utf-8')
2061
2062 # Extract video thumbnail
2063 mobj = re.search(r'<thumbnail>(.*?)</thumbnail>', webpage)
2064 if mobj is None:
2065 self._downloader.trouble(u'ERROR: unable to extract video thumbnail')
2066 return
2067 video_thumbnail = mobj.group(1).decode('utf-8')
2068
2069 # # Extract video description
2070 # mobj = re.search(r'<meta property="og:description" content="(.*)" />', webpage)
2071 # if mobj is None:
2072 # self._downloader.trouble(u'ERROR: unable to extract video description')
2073 # return
2074 # video_description = mobj.group(1).decode('utf-8')
2075 # if not video_description: video_description = 'No description available.'
2076 video_description = 'Foo.'
2077
2078 # Vimeo specific: extract request signature
2079 mobj = re.search(r'<request_signature>(.*?)</request_signature>', webpage)
2080 if mobj is None:
2081 self._downloader.trouble(u'ERROR: unable to extract request signature')
2082 return
2083 sig = mobj.group(1).decode('utf-8')
2084
2085 # Vimeo specific: extract video quality information
2086 mobj = re.search(r'<isHD>(\d+)</isHD>', webpage)
2087 if mobj is None:
2088 self._downloader.trouble(u'ERROR: unable to extract video quality information')
2089 return
2090 quality = mobj.group(1).decode('utf-8')
2091
2092 if int(quality) == 1:
2093 quality = 'hd'
2094 else:
2095 quality = 'sd'
2096
2097 # Vimeo specific: Extract request signature expiration
2098 mobj = re.search(r'<request_signature_expires>(.*?)</request_signature_expires>', webpage)
2099 if mobj is None:
2100 self._downloader.trouble(u'ERROR: unable to extract request signature expiration')
2101 return
2102 sig_exp = mobj.group(1).decode('utf-8')
2103
2104 video_url = "http://vimeo.com/moogaloop/play/clip:%s/%s/%s/?q=%s" % (video_id, sig, sig_exp, quality)
2105
2106 try:
2107 # Process video information
2108 self._downloader.process_info({
2109 'id': video_id.decode('utf-8'),
2110 'url': video_url,
2111 'uploader': video_uploader,
2112 'upload_date': u'NA',
2113 'title': video_title,
2114 'stitle': simple_title,
2115 'ext': u'mp4',
2116 'thumbnail': video_thumbnail.decode('utf-8'),
2117 'description': video_description,
2118 'thumbnail': video_thumbnail,
2119 'description': video_description,
2120 'player_url': None,
2121 })
2122 except UnavailableVideoError:
2123 self._downloader.trouble(u'ERROR: unable to download video')
2124
2125
2126 class GenericIE(InfoExtractor):
2127 """Generic last-resort information extractor."""
2128
2129 _VALID_URL = r'.*'
2130 IE_NAME = u'generic'
2131
2132 def __init__(self, downloader=None):
2133 InfoExtractor.__init__(self, downloader)
2134
2135 def report_download_webpage(self, video_id):
2136 """Report webpage download."""
2137 self._downloader.to_screen(u'WARNING: Falling back on generic information extractor.')
2138 self._downloader.to_screen(u'[generic] %s: Downloading webpage' % video_id)
2139
2140 def report_extraction(self, video_id):
2141 """Report information extraction."""
2142 self._downloader.to_screen(u'[generic] %s: Extracting information' % video_id)
2143
2144 def _real_extract(self, url):
2145 # At this point we have a new video
2146 self._downloader.increment_downloads()
2147
2148 video_id = url.split('/')[-1]
2149 request = urllib2.Request(url)
2150 try:
2151 self.report_download_webpage(video_id)
2152 webpage = urllib2.urlopen(request).read()
2153 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
2154 self._downloader.trouble(u'ERROR: Unable to retrieve video webpage: %s' % str(err))
2155 return
2156 except ValueError, err:
2157 # since this is the last-resort InfoExtractor, if
2158 # this error is thrown, it'll be thrown here
2159 self._downloader.trouble(u'ERROR: Invalid URL: %s' % url)
2160 return
2161
2162 self.report_extraction(video_id)
2163 # Start with something easy: JW Player in SWFObject
2164 mobj = re.search(r'flashvars: [\'"](?:.*&)?file=(http[^\'"&]*)', webpage)
2165 if mobj is None:
2166 # Broaden the search a little bit
2167 mobj = re.search(r'[^A-Za-z0-9]?(?:file|source)=(http[^\'"&]*)', webpage)
2168 if mobj is None:
2169 self._downloader.trouble(u'ERROR: Invalid URL: %s' % url)
2170 return
2171
2172 # It's possible that one of the regexes
2173 # matched, but returned an empty group:
2174 if mobj.group(1) is None:
2175 self._downloader.trouble(u'ERROR: Invalid URL: %s' % url)
2176 return
2177
2178 video_url = urllib.unquote(mobj.group(1))
2179 video_id = os.path.basename(video_url)
2180
2181 # here's a fun little line of code for you:
2182 video_extension = os.path.splitext(video_id)[1][1:]
2183 video_id = os.path.splitext(video_id)[0]
2184
2185 # it's tempting to parse this further, but you would
2186 # have to take into account all the variations like
2187 # Video Title - Site Name
2188 # Site Name | Video Title
2189 # Video Title - Tagline | Site Name
2190 # and so on and so forth; it's just not practical
2191 mobj = re.search(r'<title>(.*)</title>', webpage)
2192 if mobj is None:
2193 self._downloader.trouble(u'ERROR: unable to extract title')
2194 return
2195 video_title = mobj.group(1).decode('utf-8')
2196 video_title = sanitize_title(video_title)
2197 simple_title = _simplify_title(video_title)
2198
2199 # video uploader is domain name
2200 mobj = re.match(r'(?:https?://)?([^/]*)/.*', url)
2201 if mobj is None:
2202 self._downloader.trouble(u'ERROR: unable to extract title')
2203 return
2204 video_uploader = mobj.group(1).decode('utf-8')
2205
2206 try:
2207 # Process video information
2208 self._downloader.process_info({
2209 'id': video_id.decode('utf-8'),
2210 'url': video_url.decode('utf-8'),
2211 'uploader': video_uploader,
2212 'upload_date': u'NA',
2213 'title': video_title,
2214 'stitle': simple_title,
2215 'ext': video_extension.decode('utf-8'),
2216 'format': u'NA',
2217 'player_url': None,
2218 })
2219 except UnavailableVideoError, err:
2220 self._downloader.trouble(u'\nERROR: unable to download video')
2221
2222
2223 class YoutubeSearchIE(InfoExtractor):
2224 """Information Extractor for YouTube search queries."""
2225 _VALID_URL = r'ytsearch(\d+|all)?:[\s\S]+'
2226 _TEMPLATE_URL = 'http://www.youtube.com/results?search_query=%s&page=%s&gl=US&hl=en'
2227 _VIDEO_INDICATOR = r'href="/watch\?v=.+?"'
2228 _MORE_PAGES_INDICATOR = r'(?m)>\s*Next\s*</a>'
2229 _youtube_ie = None
2230 _max_youtube_results = 1000
2231 IE_NAME = u'youtube:search'
2232
2233 def __init__(self, youtube_ie, downloader=None):
2234 InfoExtractor.__init__(self, downloader)
2235 self._youtube_ie = youtube_ie
2236
2237 def report_download_page(self, query, pagenum):
2238 """Report attempt to download playlist page with given number."""
2239 query = query.decode(preferredencoding())
2240 self._downloader.to_screen(u'[youtube] query "%s": Downloading page %s' % (query, pagenum))
2241
2242 def _real_initialize(self):
2243 self._youtube_ie.initialize()
2244
2245 def _real_extract(self, query):
2246 mobj = re.match(self._VALID_URL, query)
2247 if mobj is None:
2248 self._downloader.trouble(u'ERROR: invalid search query "%s"' % query)
2249 return
2250
2251 prefix, query = query.split(':')
2252 prefix = prefix[8:]
2253 query = query.encode('utf-8')
2254 if prefix == '':
2255 self._download_n_results(query, 1)
2256 return
2257 elif prefix == 'all':
2258 self._download_n_results(query, self._max_youtube_results)
2259 return
2260 else:
2261 try:
2262 n = long(prefix)
2263 if n <= 0:
2264 self._downloader.trouble(u'ERROR: invalid download number %s for query "%s"' % (n, query))
2265 return
2266 elif n > self._max_youtube_results:
2267 self._downloader.to_stderr(u'WARNING: ytsearch returns max %i results (you requested %i)' % (self._max_youtube_results, n))
2268 n = self._max_youtube_results
2269 self._download_n_results(query, n)
2270 return
2271 except ValueError: # parsing prefix as integer fails
2272 self._download_n_results(query, 1)
2273 return
2274
2275 def _download_n_results(self, query, n):
2276 """Downloads a specified number of results for a query"""
2277
2278 video_ids = []
2279 already_seen = set()
2280 pagenum = 1
2281
2282 while True:
2283 self.report_download_page(query, pagenum)
2284 result_url = self._TEMPLATE_URL % (urllib.quote_plus(query), pagenum)
2285 request = urllib2.Request(result_url)
2286 try:
2287 page = urllib2.urlopen(request).read()
2288 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
2289 self._downloader.trouble(u'ERROR: unable to download webpage: %s' % str(err))
2290 return
2291
2292 # Extract video identifiers
2293 for mobj in re.finditer(self._VIDEO_INDICATOR, page):
2294 video_id = page[mobj.span()[0]:mobj.span()[1]].split('=')[2][:-1]
2295 if video_id not in already_seen:
2296 video_ids.append(video_id)
2297 already_seen.add(video_id)
2298 if len(video_ids) == n:
2299 # Specified n videos reached
2300 for id in video_ids:
2301 self._youtube_ie.extract('http://www.youtube.com/watch?v=%s' % id)
2302 return
2303
2304 if re.search(self._MORE_PAGES_INDICATOR, page) is None:
2305 for id in video_ids:
2306 self._youtube_ie.extract('http://www.youtube.com/watch?v=%s' % id)
2307 return
2308
2309 pagenum = pagenum + 1
2310
2311
2312 class GoogleSearchIE(InfoExtractor):
2313 """Information Extractor for Google Video search queries."""
2314 _VALID_URL = r'gvsearch(\d+|all)?:[\s\S]+'
2315 _TEMPLATE_URL = 'http://video.google.com/videosearch?q=%s+site:video.google.com&start=%s&hl=en'
2316 _VIDEO_INDICATOR = r'videoplay\?docid=([^\&>]+)\&'
2317 _MORE_PAGES_INDICATOR = r'<span>Next</span>'
2318 _google_ie = None
2319 _max_google_results = 1000
2320 IE_NAME = u'video.google:search'
2321
2322 def __init__(self, google_ie, downloader=None):
2323 InfoExtractor.__init__(self, downloader)
2324 self._google_ie = google_ie
2325
2326 def report_download_page(self, query, pagenum):
2327 """Report attempt to download playlist page with given number."""
2328 query = query.decode(preferredencoding())
2329 self._downloader.to_screen(u'[video.google] query "%s": Downloading page %s' % (query, pagenum))
2330
2331 def _real_initialize(self):
2332 self._google_ie.initialize()
2333
2334 def _real_extract(self, query):
2335 mobj = re.match(self._VALID_URL, query)
2336 if mobj is None:
2337 self._downloader.trouble(u'ERROR: invalid search query "%s"' % query)
2338 return
2339
2340 prefix, query = query.split(':')
2341 prefix = prefix[8:]
2342 query = query.encode('utf-8')
2343 if prefix == '':
2344 self._download_n_results(query, 1)
2345 return
2346 elif prefix == 'all':
2347 self._download_n_results(query, self._max_google_results)
2348 return
2349 else:
2350 try:
2351 n = long(prefix)
2352 if n <= 0:
2353 self._downloader.trouble(u'ERROR: invalid download number %s for query "%s"' % (n, query))
2354 return
2355 elif n > self._max_google_results:
2356 self._downloader.to_stderr(u'WARNING: gvsearch returns max %i results (you requested %i)' % (self._max_google_results, n))
2357 n = self._max_google_results
2358 self._download_n_results(query, n)
2359 return
2360 except ValueError: # parsing prefix as integer fails
2361 self._download_n_results(query, 1)
2362 return
2363
2364 def _download_n_results(self, query, n):
2365 """Downloads a specified number of results for a query"""
2366
2367 video_ids = []
2368 already_seen = set()
2369 pagenum = 1
2370
2371 while True:
2372 self.report_download_page(query, pagenum)
2373 result_url = self._TEMPLATE_URL % (urllib.quote_plus(query), pagenum)
2374 request = urllib2.Request(result_url)
2375 try:
2376 page = urllib2.urlopen(request).read()
2377 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
2378 self._downloader.trouble(u'ERROR: unable to download webpage: %s' % str(err))
2379 return
2380
2381 # Extract video identifiers
2382 for mobj in re.finditer(self._VIDEO_INDICATOR, page):
2383 video_id = mobj.group(1)
2384 if video_id not in already_seen:
2385 video_ids.append(video_id)
2386 already_seen.add(video_id)
2387 if len(video_ids) == n:
2388 # Specified n videos reached
2389 for id in video_ids:
2390 self._google_ie.extract('http://video.google.com/videoplay?docid=%s' % id)
2391 return
2392
2393 if re.search(self._MORE_PAGES_INDICATOR, page) is None:
2394 for id in video_ids:
2395 self._google_ie.extract('http://video.google.com/videoplay?docid=%s' % id)
2396 return
2397
2398 pagenum = pagenum + 1
2399
2400
2401 class YahooSearchIE(InfoExtractor):
2402 """Information Extractor for Yahoo! Video search queries."""
2403 _VALID_URL = r'yvsearch(\d+|all)?:[\s\S]+'
2404 _TEMPLATE_URL = 'http://video.yahoo.com/search/?p=%s&o=%s'
2405 _VIDEO_INDICATOR = r'href="http://video\.yahoo\.com/watch/([0-9]+/[0-9]+)"'
2406 _MORE_PAGES_INDICATOR = r'\s*Next'
2407 _yahoo_ie = None
2408 _max_yahoo_results = 1000
2409 IE_NAME = u'video.yahoo:search'
2410
2411 def __init__(self, yahoo_ie, downloader=None):
2412 InfoExtractor.__init__(self, downloader)
2413 self._yahoo_ie = yahoo_ie
2414
2415 def report_download_page(self, query, pagenum):
2416 """Report attempt to download playlist page with given number."""
2417 query = query.decode(preferredencoding())
2418 self._downloader.to_screen(u'[video.yahoo] query "%s": Downloading page %s' % (query, pagenum))
2419
2420 def _real_initialize(self):
2421 self._yahoo_ie.initialize()
2422
2423 def _real_extract(self, query):
2424 mobj = re.match(self._VALID_URL, query)
2425 if mobj is None:
2426 self._downloader.trouble(u'ERROR: invalid search query "%s"' % query)
2427 return
2428
2429 prefix, query = query.split(':')
2430 prefix = prefix[8:]
2431 query = query.encode('utf-8')
2432 if prefix == '':
2433 self._download_n_results(query, 1)
2434 return
2435 elif prefix == 'all':
2436 self._download_n_results(query, self._max_yahoo_results)
2437 return
2438 else:
2439 try:
2440 n = long(prefix)
2441 if n <= 0:
2442 self._downloader.trouble(u'ERROR: invalid download number %s for query "%s"' % (n, query))
2443 return
2444 elif n > self._max_yahoo_results:
2445 self._downloader.to_stderr(u'WARNING: yvsearch returns max %i results (you requested %i)' % (self._max_yahoo_results, n))
2446 n = self._max_yahoo_results
2447 self._download_n_results(query, n)
2448 return
2449 except ValueError: # parsing prefix as integer fails
2450 self._download_n_results(query, 1)
2451 return
2452
2453 def _download_n_results(self, query, n):
2454 """Downloads a specified number of results for a query"""
2455
2456 video_ids = []
2457 already_seen = set()
2458 pagenum = 1
2459
2460 while True:
2461 self.report_download_page(query, pagenum)
2462 result_url = self._TEMPLATE_URL % (urllib.quote_plus(query), pagenum)
2463 request = urllib2.Request(result_url)
2464 try:
2465 page = urllib2.urlopen(request).read()
2466 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
2467 self._downloader.trouble(u'ERROR: unable to download webpage: %s' % str(err))
2468 return
2469
2470 # Extract video identifiers
2471 for mobj in re.finditer(self._VIDEO_INDICATOR, page):
2472 video_id = mobj.group(1)
2473 if video_id not in already_seen:
2474 video_ids.append(video_id)
2475 already_seen.add(video_id)
2476 if len(video_ids) == n:
2477 # Specified n videos reached
2478 for id in video_ids:
2479 self._yahoo_ie.extract('http://video.yahoo.com/watch/%s' % id)
2480 return
2481
2482 if re.search(self._MORE_PAGES_INDICATOR, page) is None:
2483 for id in video_ids:
2484 self._yahoo_ie.extract('http://video.yahoo.com/watch/%s' % id)
2485 return
2486
2487 pagenum = pagenum + 1
2488
2489
2490 class YoutubePlaylistIE(InfoExtractor):
2491 """Information Extractor for YouTube playlists."""
2492
2493 _VALID_URL = r'(?:https?://)?(?:\w+\.)?youtube\.com/(?:(?:course|view_play_list|my_playlists|artist|playlist)\?.*?(p|a|list)=|user/.*?/user/|p/|user/.*?#[pg]/c/)(?:PL)?([0-9A-Za-z-_]+)(?:/.*?/([0-9A-Za-z_-]+))?.*'
2494 _TEMPLATE_URL = 'http://www.youtube.com/%s?%s=%s&page=%s&gl=US&hl=en'
2495 _VIDEO_INDICATOR = r'/watch\?v=(.+?)&'
2496 _MORE_PAGES_INDICATOR = r'(?m)>\s*Next\s*</a>'
2497 _youtube_ie = None
2498 IE_NAME = u'youtube:playlist'
2499
2500 def __init__(self, youtube_ie, downloader=None):
2501 InfoExtractor.__init__(self, downloader)
2502 self._youtube_ie = youtube_ie
2503
2504 def report_download_page(self, playlist_id, pagenum):
2505 """Report attempt to download playlist page with given number."""
2506 self._downloader.to_screen(u'[youtube] PL %s: Downloading page #%s' % (playlist_id, pagenum))
2507
2508 def _real_initialize(self):
2509 self._youtube_ie.initialize()
2510
2511 def _real_extract(self, url):
2512 # Extract playlist id
2513 mobj = re.match(self._VALID_URL, url)
2514 if mobj is None:
2515 self._downloader.trouble(u'ERROR: invalid url: %s' % url)
2516 return
2517
2518 # Single video case
2519 if mobj.group(3) is not None:
2520 self._youtube_ie.extract(mobj.group(3))
2521 return
2522
2523 # Download playlist pages
2524 # prefix is 'p' as default for playlists but there are other types that need extra care
2525 playlist_prefix = mobj.group(1)
2526 if playlist_prefix == 'a':
2527 playlist_access = 'artist'
2528 else:
2529 playlist_prefix = 'p'
2530 playlist_access = 'view_play_list'
2531 playlist_id = mobj.group(2)
2532 video_ids = []
2533 pagenum = 1
2534
2535 while True:
2536 self.report_download_page(playlist_id, pagenum)
2537 url = self._TEMPLATE_URL % (playlist_access, playlist_prefix, playlist_id, pagenum)
2538 request = urllib2.Request(url)
2539 try:
2540 page = urllib2.urlopen(request).read()
2541 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
2542 self._downloader.trouble(u'ERROR: unable to download webpage: %s' % str(err))
2543 return
2544
2545 # Extract video identifiers
2546 ids_in_page = []
2547 for mobj in re.finditer(self._VIDEO_INDICATOR, page):
2548 if mobj.group(1) not in ids_in_page:
2549 ids_in_page.append(mobj.group(1))
2550 video_ids.extend(ids_in_page)
2551
2552 if re.search(self._MORE_PAGES_INDICATOR, page) is None:
2553 break
2554 pagenum = pagenum + 1
2555
2556 playliststart = self._downloader.params.get('playliststart', 1) - 1
2557 playlistend = self._downloader.params.get('playlistend', -1)
2558 video_ids = video_ids[playliststart:playlistend]
2559
2560 for id in video_ids:
2561 self._youtube_ie.extract('http://www.youtube.com/watch?v=%s' % id)
2562 return
2563
2564
2565 class YoutubeUserIE(InfoExtractor):
2566 """Information Extractor for YouTube users."""
2567
2568 _VALID_URL = r'(?:(?:(?:https?://)?(?:\w+\.)?youtube\.com/user/)|ytuser:)([A-Za-z0-9_-]+)'
2569 _TEMPLATE_URL = 'http://gdata.youtube.com/feeds/api/users/%s'
2570 _GDATA_PAGE_SIZE = 50
2571 _GDATA_URL = 'http://gdata.youtube.com/feeds/api/users/%s/uploads?max-results=%d&start-index=%d'
2572 _VIDEO_INDICATOR = r'/watch\?v=(.+?)[\<&]'
2573 _youtube_ie = None
2574 IE_NAME = u'youtube:user'
2575
2576 def __init__(self, youtube_ie, downloader=None):
2577 InfoExtractor.__init__(self, downloader)
2578 self._youtube_ie = youtube_ie
2579
2580 def report_download_page(self, username, start_index):
2581 """Report attempt to download user page."""
2582 self._downloader.to_screen(u'[youtube] user %s: Downloading video ids from %d to %d' %
2583 (username, start_index, start_index + self._GDATA_PAGE_SIZE))
2584
2585 def _real_initialize(self):
2586 self._youtube_ie.initialize()
2587
2588 def _real_extract(self, url):
2589 # Extract username
2590 mobj = re.match(self._VALID_URL, url)
2591 if mobj is None:
2592 self._downloader.trouble(u'ERROR: invalid url: %s' % url)
2593 return
2594
2595 username = mobj.group(1)
2596
2597 # Download video ids using YouTube Data API. Result size per
2598 # query is limited (currently to 50 videos) so we need to query
2599 # page by page until there are no video ids - it means we got
2600 # all of them.
2601
2602 video_ids = []
2603 pagenum = 0
2604
2605 while True:
2606 start_index = pagenum * self._GDATA_PAGE_SIZE + 1
2607 self.report_download_page(username, start_index)
2608
2609 request = urllib2.Request(self._GDATA_URL % (username, self._GDATA_PAGE_SIZE, start_index))
2610
2611 try:
2612 page = urllib2.urlopen(request).read()
2613 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
2614 self._downloader.trouble(u'ERROR: unable to download webpage: %s' % str(err))
2615 return
2616
2617 # Extract video identifiers
2618 ids_in_page = []
2619
2620 for mobj in re.finditer(self._VIDEO_INDICATOR, page):
2621 if mobj.group(1) not in ids_in_page:
2622 ids_in_page.append(mobj.group(1))
2623
2624 video_ids.extend(ids_in_page)
2625
2626 # A little optimization - if current page is not
2627 # "full", ie. does not contain PAGE_SIZE video ids then
2628 # we can assume that this page is the last one - there
2629 # are no more ids on further pages - no need to query
2630 # again.
2631
2632 if len(ids_in_page) < self._GDATA_PAGE_SIZE:
2633 break
2634
2635 pagenum += 1
2636
2637 all_ids_count = len(video_ids)
2638 playliststart = self._downloader.params.get('playliststart', 1) - 1
2639 playlistend = self._downloader.params.get('playlistend', -1)
2640
2641 if playlistend == -1:
2642 video_ids = video_ids[playliststart:]
2643 else:
2644 video_ids = video_ids[playliststart:playlistend]
2645
2646 self._downloader.to_screen("[youtube] user %s: Collected %d video ids (downloading %d of them)" %
2647 (username, all_ids_count, len(video_ids)))
2648
2649 for video_id in video_ids:
2650 self._youtube_ie.extract('http://www.youtube.com/watch?v=%s' % video_id)
2651
2652
2653 class DepositFilesIE(InfoExtractor):
2654 """Information extractor for depositfiles.com"""
2655
2656 _VALID_URL = r'(?:http://)?(?:\w+\.)?depositfiles\.com/(?:../(?#locale))?files/(.+)'
2657 IE_NAME = u'DepositFiles'
2658
2659 def __init__(self, downloader=None):
2660 InfoExtractor.__init__(self, downloader)
2661
2662 def report_download_webpage(self, file_id):
2663 """Report webpage download."""
2664 self._downloader.to_screen(u'[DepositFiles] %s: Downloading webpage' % file_id)
2665
2666 def report_extraction(self, file_id):
2667 """Report information extraction."""
2668 self._downloader.to_screen(u'[DepositFiles] %s: Extracting information' % file_id)
2669
2670 def _real_extract(self, url):
2671 # At this point we have a new file
2672 self._downloader.increment_downloads()
2673
2674 file_id = url.split('/')[-1]
2675 # Rebuild url in english locale
2676 url = 'http://depositfiles.com/en/files/' + file_id
2677
2678 # Retrieve file webpage with 'Free download' button pressed
2679 free_download_indication = { 'gateway_result' : '1' }
2680 request = urllib2.Request(url, urllib.urlencode(free_download_indication))
2681 try:
2682 self.report_download_webpage(file_id)
2683 webpage = urllib2.urlopen(request).read()
2684 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
2685 self._downloader.trouble(u'ERROR: Unable to retrieve file webpage: %s' % str(err))
2686 return
2687
2688 # Search for the real file URL
2689 mobj = re.search(r'<form action="(http://fileshare.+?)"', webpage)
2690 if (mobj is None) or (mobj.group(1) is None):
2691 # Try to figure out reason of the error.
2692 mobj = re.search(r'<strong>(Attention.*?)</strong>', webpage, re.DOTALL)
2693 if (mobj is not None) and (mobj.group(1) is not None):
2694 restriction_message = re.sub('\s+', ' ', mobj.group(1)).strip()
2695 self._downloader.trouble(u'ERROR: %s' % restriction_message)
2696 else:
2697 self._downloader.trouble(u'ERROR: unable to extract download URL from: %s' % url)
2698 return
2699
2700 file_url = mobj.group(1)
2701 file_extension = os.path.splitext(file_url)[1][1:]
2702
2703 # Search for file title
2704 mobj = re.search(r'<b title="(.*?)">', webpage)
2705 if mobj is None:
2706 self._downloader.trouble(u'ERROR: unable to extract title')
2707 return
2708 file_title = mobj.group(1).decode('utf-8')
2709
2710 try:
2711 # Process file information
2712 self._downloader.process_info({
2713 'id': file_id.decode('utf-8'),
2714 'url': file_url.decode('utf-8'),
2715 'uploader': u'NA',
2716 'upload_date': u'NA',
2717 'title': file_title,
2718 'stitle': file_title,
2719 'ext': file_extension.decode('utf-8'),
2720 'format': u'NA',
2721 'player_url': None,
2722 })
2723 except UnavailableVideoError, err:
2724 self._downloader.trouble(u'ERROR: unable to download file')
2725
2726
2727 class FacebookIE(InfoExtractor):
2728 """Information Extractor for Facebook"""
2729
2730 _VALID_URL = r'^(?:https?://)?(?:\w+\.)?facebook\.com/(?:video/video|photo)\.php\?(?:.*?)v=(?P<ID>\d+)(?:.*)'
2731 _LOGIN_URL = 'https://login.facebook.com/login.php?m&next=http%3A%2F%2Fm.facebook.com%2Fhome.php&'
2732 _NETRC_MACHINE = 'facebook'
2733 _available_formats = ['video', 'highqual', 'lowqual']
2734 _video_extensions = {
2735 'video': 'mp4',
2736 'highqual': 'mp4',
2737 'lowqual': 'mp4',
2738 }
2739 IE_NAME = u'facebook'
2740
2741 def __init__(self, downloader=None):
2742 InfoExtractor.__init__(self, downloader)
2743
2744 def _reporter(self, message):
2745 """Add header and report message."""
2746 self._downloader.to_screen(u'[facebook] %s' % message)
2747
2748 def report_login(self):
2749 """Report attempt to log in."""
2750 self._reporter(u'Logging in')
2751
2752 def report_video_webpage_download(self, video_id):
2753 """Report attempt to download video webpage."""
2754 self._reporter(u'%s: Downloading video webpage' % video_id)
2755
2756 def report_information_extraction(self, video_id):
2757 """Report attempt to extract video information."""
2758 self._reporter(u'%s: Extracting video information' % video_id)
2759
2760 def _parse_page(self, video_webpage):
2761 """Extract video information from page"""
2762 # General data
2763 data = {'title': r'\("video_title", "(.*?)"\)',
2764 'description': r'<div class="datawrap">(.*?)</div>',
2765 'owner': r'\("video_owner_name", "(.*?)"\)',
2766 'thumbnail': r'\("thumb_url", "(?P<THUMB>.*?)"\)',
2767 }
2768 video_info = {}
2769 for piece in data.keys():
2770 mobj = re.search(data[piece], video_webpage)
2771 if mobj is not None:
2772 video_info[piece] = urllib.unquote_plus(mobj.group(1).decode("unicode_escape"))
2773
2774 # Video urls
2775 video_urls = {}
2776 for fmt in self._available_formats:
2777 mobj = re.search(r'\("%s_src\", "(.+?)"\)' % fmt, video_webpage)
2778 if mobj is not None:
2779 # URL is in a Javascript segment inside an escaped Unicode format within
2780 # the generally utf-8 page
2781 video_urls[fmt] = urllib.unquote_plus(mobj.group(1).decode("unicode_escape"))
2782 video_info['video_urls'] = video_urls
2783
2784 return video_info
2785
2786 def _real_initialize(self):
2787 if self._downloader is None:
2788 return
2789
2790 useremail = None
2791 password = None
2792 downloader_params = self._downloader.params
2793
2794 # Attempt to use provided username and password or .netrc data
2795 if downloader_params.get('username', None) is not None:
2796 useremail = downloader_params['username']
2797 password = downloader_params['password']
2798 elif downloader_params.get('usenetrc', False):
2799 try:
2800 info = netrc.netrc().authenticators(self._NETRC_MACHINE)
2801 if info is not None:
2802 useremail = info[0]
2803 password = info[2]
2804 else:
2805 raise netrc.NetrcParseError('No authenticators for %s' % self._NETRC_MACHINE)
2806 except (IOError, netrc.NetrcParseError), err:
2807 self._downloader.to_stderr(u'WARNING: parsing .netrc: %s' % str(err))
2808 return
2809
2810 if useremail is None:
2811 return
2812
2813 # Log in
2814 login_form = {
2815 'email': useremail,
2816 'pass': password,
2817 'login': 'Log+In'
2818 }
2819 request = urllib2.Request(self._LOGIN_URL, urllib.urlencode(login_form))
2820 try:
2821 self.report_login()
2822 login_results = urllib2.urlopen(request).read()
2823 if re.search(r'<form(.*)name="login"(.*)</form>', login_results) is not None:
2824 self._downloader.to_stderr(u'WARNING: unable to log in: bad username/password, or exceded login rate limit (~3/min). Check credentials or wait.')
2825 return
2826 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
2827 self._downloader.to_stderr(u'WARNING: unable to log in: %s' % str(err))
2828 return
2829
2830 def _real_extract(self, url):
2831 mobj = re.match(self._VALID_URL, url)
2832 if mobj is None:
2833 self._downloader.trouble(u'ERROR: invalid URL: %s' % url)
2834 return
2835 video_id = mobj.group('ID')
2836
2837 # Get video webpage
2838 self.report_video_webpage_download(video_id)
2839 request = urllib2.Request('https://www.facebook.com/video/video.php?v=%s' % video_id)
2840 try:
2841 page = urllib2.urlopen(request)
2842 video_webpage = page.read()
2843 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
2844 self._downloader.trouble(u'ERROR: unable to download video webpage: %s' % str(err))
2845 return
2846
2847 # Start extracting information
2848 self.report_information_extraction(video_id)
2849
2850 # Extract information
2851 video_info = self._parse_page(video_webpage)
2852
2853 # uploader
2854 if 'owner' not in video_info:
2855 self._downloader.trouble(u'ERROR: unable to extract uploader nickname')
2856 return
2857 video_uploader = video_info['owner']
2858
2859 # title
2860 if 'title' not in video_info:
2861 self._downloader.trouble(u'ERROR: unable to extract video title')
2862 return
2863 video_title = video_info['title']
2864 video_title = video_title.decode('utf-8')
2865 video_title = sanitize_title(video_title)
2866
2867 simple_title = _simplify_title(video_title)
2868
2869 # thumbnail image
2870 if 'thumbnail' not in video_info:
2871 self._downloader.trouble(u'WARNING: unable to extract video thumbnail')
2872 video_thumbnail = ''
2873 else:
2874 video_thumbnail = video_info['thumbnail']
2875
2876 # upload date
2877 upload_date = u'NA'
2878 if 'upload_date' in video_info:
2879 upload_time = video_info['upload_date']
2880 timetuple = email.utils.parsedate_tz(upload_time)
2881 if timetuple is not None:
2882 try:
2883 upload_date = time.strftime('%Y%m%d', timetuple[0:9])
2884 except:
2885 pass
2886
2887 # description
2888 video_description = video_info.get('description', 'No description available.')
2889
2890 url_map = video_info['video_urls']
2891 if len(url_map.keys()) > 0:
2892 # Decide which formats to download
2893 req_format = self._downloader.params.get('format', None)
2894 format_limit = self._downloader.params.get('format_limit', None)
2895
2896 if format_limit is not None and format_limit in self._available_formats:
2897 format_list = self._available_formats[self._available_formats.index(format_limit):]
2898 else:
2899 format_list = self._available_formats
2900 existing_formats = [x for x in format_list if x in url_map]
2901 if len(existing_formats) == 0:
2902 self._downloader.trouble(u'ERROR: no known formats available for video')
2903 return
2904 if req_format is None:
2905 video_url_list = [(existing_formats[0], url_map[existing_formats[0]])] # Best quality
2906 elif req_format == 'worst':
2907 video_url_list = [(existing_formats[len(existing_formats)-1], url_map[existing_formats[len(existing_formats)-1]])] # worst quality
2908 elif req_format == '-1':
2909 video_url_list = [(f, url_map[f]) for f in existing_formats] # All formats
2910 else:
2911 # Specific format
2912 if req_format not in url_map:
2913 self._downloader.trouble(u'ERROR: requested format not available')
2914 return
2915 video_url_list = [(req_format, url_map[req_format])] # Specific format
2916
2917 for format_param, video_real_url in video_url_list:
2918
2919 # At this point we have a new video
2920 self._downloader.increment_downloads()
2921
2922 # Extension
2923 video_extension = self._video_extensions.get(format_param, 'mp4')
2924
2925 try:
2926 # Process video information
2927 self._downloader.process_info({
2928 'id': video_id.decode('utf-8'),
2929 'url': video_real_url.decode('utf-8'),
2930 'uploader': video_uploader.decode('utf-8'),
2931 'upload_date': upload_date,
2932 'title': video_title,
2933 'stitle': simple_title,
2934 'ext': video_extension.decode('utf-8'),
2935 'format': (format_param is None and u'NA' or format_param.decode('utf-8')),
2936 'thumbnail': video_thumbnail.decode('utf-8'),
2937 'description': video_description.decode('utf-8'),
2938 'player_url': None,
2939 })
2940 except UnavailableVideoError, err:
2941 self._downloader.trouble(u'\nERROR: unable to download video')
2942
2943 class BlipTVIE(InfoExtractor):
2944 """Information extractor for blip.tv"""
2945
2946 _VALID_URL = r'^(?:https?://)?(?:\w+\.)?blip\.tv(/.+)$'
2947 _URL_EXT = r'^.*\.([a-z0-9]+)$'
2948 IE_NAME = u'blip.tv'
2949
2950 def report_extraction(self, file_id):
2951 """Report information extraction."""
2952 self._downloader.to_screen(u'[%s] %s: Extracting information' % (self.IE_NAME, file_id))
2953
2954 def report_direct_download(self, title):
2955 """Report information extraction."""
2956 self._downloader.to_screen(u'[%s] %s: Direct download detected' % (self.IE_NAME, title))
2957
2958 def _real_extract(self, url):
2959 mobj = re.match(self._VALID_URL, url)
2960 if mobj is None:
2961 self._downloader.trouble(u'ERROR: invalid URL: %s' % url)
2962 return
2963
2964 if '?' in url:
2965 cchar = '&'
2966 else:
2967 cchar = '?'
2968 json_url = url + cchar + 'skin=json&version=2&no_wrap=1'
2969 request = urllib2.Request(json_url)
2970 self.report_extraction(mobj.group(1))
2971 info = None
2972 try:
2973 urlh = urllib2.urlopen(request)
2974 if urlh.headers.get('Content-Type', '').startswith('video/'): # Direct download
2975 basename = url.split('/')[-1]
2976 title,ext = os.path.splitext(basename)
2977 title = title.decode('UTF-8')
2978 ext = ext.replace('.', '')
2979 self.report_direct_download(title)
2980 info = {
2981 'id': title,
2982 'url': url,
2983 'title': title,
2984 'stitle': _simplify_title(title),
2985 'ext': ext,
2986 'urlhandle': urlh
2987 }
2988 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
2989 self._downloader.trouble(u'ERROR: unable to download video info webpage: %s' % str(err))
2990 return
2991 if info is None: # Regular URL
2992 try:
2993 json_code = urlh.read()
2994 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
2995 self._downloader.trouble(u'ERROR: unable to read video info webpage: %s' % str(err))
2996 return
2997
2998 try:
2999 json_data = json.loads(json_code)
3000 if 'Post' in json_data:
3001 data = json_data['Post']
3002 else:
3003 data = json_data
3004
3005 upload_date = datetime.datetime.strptime(data['datestamp'], '%m-%d-%y %H:%M%p').strftime('%Y%m%d')
3006 video_url = data['media']['url']
3007 umobj = re.match(self._URL_EXT, video_url)
3008 if umobj is None:
3009 raise ValueError('Can not determine filename extension')
3010 ext = umobj.group(1)
3011
3012 info = {
3013 'id': data['item_id'],
3014 'url': video_url,
3015 'uploader': data['display_name'],
3016 'upload_date': upload_date,
3017 'title': data['title'],
3018 'stitle': _simplify_title(data['title']),
3019 'ext': ext,
3020 'format': data['media']['mimeType'],
3021 'thumbnail': data['thumbnailUrl'],
3022 'description': data['description'],
3023 'player_url': data['embedUrl']
3024 }
3025 except (ValueError,KeyError), err:
3026 self._downloader.trouble(u'ERROR: unable to parse video information: %s' % repr(err))
3027 return
3028
3029 self._downloader.increment_downloads()
3030
3031 try:
3032 self._downloader.process_info(info)
3033 except UnavailableVideoError, err:
3034 self._downloader.trouble(u'\nERROR: unable to download video')
3035
3036
3037 class MyVideoIE(InfoExtractor):
3038 """Information Extractor for myvideo.de."""
3039
3040 _VALID_URL = r'(?:http://)?(?:www\.)?myvideo\.de/watch/([0-9]+)/([^?/]+).*'
3041 IE_NAME = u'myvideo'
3042
3043 def __init__(self, downloader=None):
3044 InfoExtractor.__init__(self, downloader)
3045
3046 def report_download_webpage(self, video_id):
3047 """Report webpage download."""
3048 self._downloader.to_screen(u'[myvideo] %s: Downloading webpage' % video_id)
3049
3050 def report_extraction(self, video_id):
3051 """Report information extraction."""
3052 self._downloader.to_screen(u'[myvideo] %s: Extracting information' % video_id)
3053
3054 def _real_extract(self,url):
3055 mobj = re.match(self._VALID_URL, url)
3056 if mobj is None:
3057 self._download.trouble(u'ERROR: invalid URL: %s' % url)
3058 return
3059
3060 video_id = mobj.group(1)
3061
3062 # Get video webpage
3063 request = urllib2.Request('http://www.myvideo.de/watch/%s' % video_id)
3064 try:
3065 self.report_download_webpage(video_id)
3066 webpage = urllib2.urlopen(request).read()
3067 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3068 self._downloader.trouble(u'ERROR: Unable to retrieve video webpage: %s' % str(err))
3069 return
3070
3071 self.report_extraction(video_id)
3072 mobj = re.search(r'<link rel=\'image_src\' href=\'(http://is[0-9].myvideo\.de/de/movie[0-9]+/[a-f0-9]+)/thumbs/[^.]+\.jpg\' />',
3073 webpage)
3074 if mobj is None:
3075 self._downloader.trouble(u'ERROR: unable to extract media URL')
3076 return
3077 video_url = mobj.group(1) + ('/%s.flv' % video_id)
3078
3079 mobj = re.search('<title>([^<]+)</title>', webpage)
3080 if mobj is None:
3081 self._downloader.trouble(u'ERROR: unable to extract title')
3082 return
3083
3084 video_title = mobj.group(1)
3085 video_title = sanitize_title(video_title)
3086
3087 simple_title = _simplify_title(video_title)
3088
3089 try:
3090 self._downloader.process_info({
3091 'id': video_id,
3092 'url': video_url,
3093 'uploader': u'NA',
3094 'upload_date': u'NA',
3095 'title': video_title,
3096 'stitle': simple_title,
3097 'ext': u'flv',
3098 'format': u'NA',
3099 'player_url': None,
3100 })
3101 except UnavailableVideoError:
3102 self._downloader.trouble(u'\nERROR: Unable to download video')
3103
3104 class ComedyCentralIE(InfoExtractor):
3105 """Information extractor for The Daily Show and Colbert Report """
3106
3107 _VALID_URL = r'^(:(?P<shortname>tds|thedailyshow|cr|colbert|colbertnation|colbertreport))|(https?://)?(www\.)?(?P<showname>thedailyshow|colbertnation)\.com/full-episodes/(?P<episode>.*)$'
3108 IE_NAME = u'comedycentral'
3109
3110 def report_extraction(self, episode_id):
3111 self._downloader.to_screen(u'[comedycentral] %s: Extracting information' % episode_id)
3112
3113 def report_config_download(self, episode_id):
3114 self._downloader.to_screen(u'[comedycentral] %s: Downloading configuration' % episode_id)
3115
3116 def report_index_download(self, episode_id):
3117 self._downloader.to_screen(u'[comedycentral] %s: Downloading show index' % episode_id)
3118
3119 def report_player_url(self, episode_id):
3120 self._downloader.to_screen(u'[comedycentral] %s: Determining player URL' % episode_id)
3121
3122 def _real_extract(self, url):
3123 mobj = re.match(self._VALID_URL, url)
3124 if mobj is None:
3125 self._downloader.trouble(u'ERROR: invalid URL: %s' % url)
3126 return
3127
3128 if mobj.group('shortname'):
3129 if mobj.group('shortname') in ('tds', 'thedailyshow'):
3130 url = u'http://www.thedailyshow.com/full-episodes/'
3131 else:
3132 url = u'http://www.colbertnation.com/full-episodes/'
3133 mobj = re.match(self._VALID_URL, url)
3134 assert mobj is not None
3135
3136 dlNewest = not mobj.group('episode')
3137 if dlNewest:
3138 epTitle = mobj.group('showname')
3139 else:
3140 epTitle = mobj.group('episode')
3141
3142 req = urllib2.Request(url)
3143 self.report_extraction(epTitle)
3144 try:
3145 htmlHandle = urllib2.urlopen(req)
3146 html = htmlHandle.read()
3147 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3148 self._downloader.trouble(u'ERROR: unable to download webpage: %s' % unicode(err))
3149 return
3150 if dlNewest:
3151 url = htmlHandle.geturl()
3152 mobj = re.match(self._VALID_URL, url)
3153 if mobj is None:
3154 self._downloader.trouble(u'ERROR: Invalid redirected URL: ' + url)
3155 return
3156 if mobj.group('episode') == '':
3157 self._downloader.trouble(u'ERROR: Redirected URL is still not specific: ' + url)
3158 return
3159 epTitle = mobj.group('episode')
3160
3161 mMovieParams = re.findall('(?:<param name="movie" value="|var url = ")(http://media.mtvnservices.com/([^"]*episode.*?:.*?))"', html)
3162 if len(mMovieParams) == 0:
3163 self._downloader.trouble(u'ERROR: unable to find Flash URL in webpage ' + url)
3164 return
3165
3166 playerUrl_raw = mMovieParams[0][0]
3167 self.report_player_url(epTitle)
3168 try:
3169 urlHandle = urllib2.urlopen(playerUrl_raw)
3170 playerUrl = urlHandle.geturl()
3171 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3172 self._downloader.trouble(u'ERROR: unable to find out player URL: ' + unicode(err))
3173 return
3174
3175 uri = mMovieParams[0][1]
3176 indexUrl = 'http://shadow.comedycentral.com/feeds/video_player/mrss/?' + urllib.urlencode({'uri': uri})
3177 self.report_index_download(epTitle)
3178 try:
3179 indexXml = urllib2.urlopen(indexUrl).read()
3180 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3181 self._downloader.trouble(u'ERROR: unable to download episode index: ' + unicode(err))
3182 return
3183
3184 idoc = xml.etree.ElementTree.fromstring(indexXml)
3185 itemEls = idoc.findall('.//item')
3186 for itemEl in itemEls:
3187 mediaId = itemEl.findall('./guid')[0].text
3188 shortMediaId = mediaId.split(':')[-1]
3189 showId = mediaId.split(':')[-2].replace('.com', '')
3190 officialTitle = itemEl.findall('./title')[0].text
3191 officialDate = itemEl.findall('./pubDate')[0].text
3192
3193 configUrl = ('http://www.comedycentral.com/global/feeds/entertainment/media/mediaGenEntertainment.jhtml?' +
3194 urllib.urlencode({'uri': mediaId}))
3195 configReq = urllib2.Request(configUrl)
3196 self.report_config_download(epTitle)
3197 try:
3198 configXml = urllib2.urlopen(configReq).read()
3199 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3200 self._downloader.trouble(u'ERROR: unable to download webpage: %s' % unicode(err))
3201 return
3202
3203 cdoc = xml.etree.ElementTree.fromstring(configXml)
3204 turls = []
3205 for rendition in cdoc.findall('.//rendition'):
3206 finfo = (rendition.attrib['bitrate'], rendition.findall('./src')[0].text)
3207 turls.append(finfo)
3208
3209 if len(turls) == 0:
3210 self._downloader.trouble(u'\nERROR: unable to download ' + mediaId + ': No videos found')
3211 continue
3212
3213 # For now, just pick the highest bitrate
3214 format,video_url = turls[-1]
3215
3216 self._downloader.increment_downloads()
3217
3218 effTitle = showId + u'-' + epTitle
3219 info = {
3220 'id': shortMediaId,
3221 'url': video_url,
3222 'uploader': showId,
3223 'upload_date': officialDate,
3224 'title': effTitle,
3225 'stitle': _simplify_title(effTitle),
3226 'ext': 'mp4',
3227 'format': format,
3228 'thumbnail': None,
3229 'description': officialTitle,
3230 'player_url': playerUrl
3231 }
3232
3233 try:
3234 self._downloader.process_info(info)
3235 except UnavailableVideoError, err:
3236 self._downloader.trouble(u'\nERROR: unable to download ' + mediaId)
3237 continue
3238
3239
3240 class EscapistIE(InfoExtractor):
3241 """Information extractor for The Escapist """
3242
3243 _VALID_URL = r'^(https?://)?(www\.)?escapistmagazine\.com/videos/view/(?P<showname>[^/]+)/(?P<episode>[^/?]+)[/?]?.*$'
3244 IE_NAME = u'escapist'
3245
3246 def report_extraction(self, showName):
3247 self._downloader.to_screen(u'[escapist] %s: Extracting information' % showName)
3248
3249 def report_config_download(self, showName):
3250 self._downloader.to_screen(u'[escapist] %s: Downloading configuration' % showName)
3251
3252 def _real_extract(self, url):
3253 htmlParser = HTMLParser.HTMLParser()
3254
3255 mobj = re.match(self._VALID_URL, url)
3256 if mobj is None:
3257 self._downloader.trouble(u'ERROR: invalid URL: %s' % url)
3258 return
3259 showName = mobj.group('showname')
3260 videoId = mobj.group('episode')
3261
3262 self.report_extraction(showName)
3263 try:
3264 webPage = urllib2.urlopen(url).read()
3265 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3266 self._downloader.trouble(u'ERROR: unable to download webpage: ' + unicode(err))
3267 return
3268
3269 descMatch = re.search('<meta name="description" content="([^"]*)"', webPage)
3270 description = htmlParser.unescape(descMatch.group(1))
3271 imgMatch = re.search('<meta property="og:image" content="([^"]*)"', webPage)
3272 imgUrl = htmlParser.unescape(imgMatch.group(1))
3273 playerUrlMatch = re.search('<meta property="og:video" content="([^"]*)"', webPage)
3274 playerUrl = htmlParser.unescape(playerUrlMatch.group(1))
3275 configUrlMatch = re.search('config=(.*)$', playerUrl)
3276 configUrl = urllib2.unquote(configUrlMatch.group(1))
3277
3278 self.report_config_download(showName)
3279 try:
3280 configJSON = urllib2.urlopen(configUrl).read()
3281 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3282 self._downloader.trouble(u'ERROR: unable to download configuration: ' + unicode(err))
3283 return
3284
3285 # Technically, it's JavaScript, not JSON
3286 configJSON = configJSON.replace("'", '"')
3287
3288 try:
3289 config = json.loads(configJSON)
3290 except (ValueError,), err:
3291 self._downloader.trouble(u'ERROR: Invalid JSON in configuration file: ' + unicode(err))
3292 return
3293
3294 playlist = config['playlist']
3295 videoUrl = playlist[1]['url']
3296
3297 self._downloader.increment_downloads()
3298 info = {
3299 'id': videoId,
3300 'url': videoUrl,
3301 'uploader': showName,
3302 'upload_date': None,
3303 'title': showName,
3304 'stitle': _simplify_title(showName),
3305 'ext': 'flv',
3306 'format': 'flv',
3307 'thumbnail': imgUrl,
3308 'description': description,
3309 'player_url': playerUrl,
3310 }
3311
3312 try:
3313 self._downloader.process_info(info)
3314 except UnavailableVideoError, err:
3315 self._downloader.trouble(u'\nERROR: unable to download ' + videoId)
3316
3317
3318 class CollegeHumorIE(InfoExtractor):
3319 """Information extractor for collegehumor.com"""
3320
3321 _VALID_URL = r'^(?:https?://)?(?:www\.)?collegehumor\.com/video/(?P<videoid>[0-9]+)/(?P<shorttitle>.*)$'
3322 IE_NAME = u'collegehumor'
3323
3324 def report_webpage(self, video_id):
3325 """Report information extraction."""
3326 self._downloader.to_screen(u'[%s] %s: Downloading webpage' % (self.IE_NAME, video_id))
3327
3328 def report_extraction(self, video_id):
3329 """Report information extraction."""
3330 self._downloader.to_screen(u'[%s] %s: Extracting information' % (self.IE_NAME, video_id))
3331
3332 def _real_extract(self, url):
3333 htmlParser = HTMLParser.HTMLParser()
3334
3335 mobj = re.match(self._VALID_URL, url)
3336 if mobj is None:
3337 self._downloader.trouble(u'ERROR: invalid URL: %s' % url)
3338 return
3339 video_id = mobj.group('videoid')
3340
3341 self.report_webpage(video_id)
3342 request = urllib2.Request(url)
3343 try:
3344 webpage = urllib2.urlopen(request).read()
3345 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3346 self._downloader.trouble(u'ERROR: unable to download video webpage: %s' % str(err))
3347 return
3348
3349 m = re.search(r'id="video:(?P<internalvideoid>[0-9]+)"', webpage)
3350 if m is None:
3351 self._downloader.trouble(u'ERROR: Cannot extract internal video ID')
3352 return
3353 internal_video_id = m.group('internalvideoid')
3354
3355 info = {
3356 'id': video_id,
3357 'internal_id': internal_video_id,
3358 }
3359
3360 self.report_extraction(video_id)
3361 xmlUrl = 'http://www.collegehumor.com/moogaloop/video:' + internal_video_id
3362 try:
3363 metaXml = urllib2.urlopen(xmlUrl).read()
3364 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3365 self._downloader.trouble(u'ERROR: unable to download video info XML: %s' % str(err))
3366 return
3367
3368 mdoc = xml.etree.ElementTree.fromstring(metaXml)
3369 try:
3370 videoNode = mdoc.findall('./video')[0]
3371 info['description'] = videoNode.findall('./description')[0].text
3372 info['title'] = videoNode.findall('./caption')[0].text
3373 info['stitle'] = _simplify_title(info['title'])
3374 info['url'] = videoNode.findall('./file')[0].text
3375 info['thumbnail'] = videoNode.findall('./thumbnail')[0].text
3376 info['ext'] = info['url'].rpartition('.')[2]
3377 info['format'] = info['ext']
3378 except IndexError:
3379 self._downloader.trouble(u'\nERROR: Invalid metadata XML file')
3380 return
3381
3382 self._downloader.increment_downloads()
3383
3384 try:
3385 self._downloader.process_info(info)
3386 except UnavailableVideoError, err:
3387 self._downloader.trouble(u'\nERROR: unable to download video')
3388
3389
3390 class XVideosIE(InfoExtractor):
3391 """Information extractor for xvideos.com"""
3392
3393 _VALID_URL = r'^(?:https?://)?(?:www\.)?xvideos\.com/video([0-9]+)(?:.*)'
3394 IE_NAME = u'xvideos'
3395
3396 def report_webpage(self, video_id):
3397 """Report information extraction."""
3398 self._downloader.to_screen(u'[%s] %s: Downloading webpage' % (self.IE_NAME, video_id))
3399
3400 def report_extraction(self, video_id):
3401 """Report information extraction."""
3402 self._downloader.to_screen(u'[%s] %s: Extracting information' % (self.IE_NAME, video_id))
3403
3404 def _real_extract(self, url):
3405 htmlParser = HTMLParser.HTMLParser()
3406
3407 mobj = re.match(self._VALID_URL, url)
3408 if mobj is None:
3409 self._downloader.trouble(u'ERROR: invalid URL: %s' % url)
3410 return
3411 video_id = mobj.group(1).decode('utf-8')
3412
3413 self.report_webpage(video_id)
3414
3415 request = urllib2.Request(r'http://www.xvideos.com/video' + video_id)
3416 try:
3417 webpage = urllib2.urlopen(request).read()
3418 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3419 self._downloader.trouble(u'ERROR: unable to download video webpage: %s' % str(err))
3420 return
3421
3422 self.report_extraction(video_id)
3423
3424
3425 # Extract video URL
3426 mobj = re.search(r'flv_url=(.+?)&', webpage)
3427 if mobj is None:
3428 self._downloader.trouble(u'ERROR: unable to extract video url')
3429 return
3430 video_url = urllib2.unquote(mobj.group(1).decode('utf-8'))
3431
3432
3433 # Extract title
3434 mobj = re.search(r'<title>(.*?)\s+-\s+XVID', webpage)
3435 if mobj is None:
3436 self._downloader.trouble(u'ERROR: unable to extract video title')
3437 return
3438 video_title = mobj.group(1).decode('utf-8')
3439
3440
3441 # Extract video thumbnail
3442 mobj = re.search(r'http://(?:img.*?\.)xvideos.com/videos/thumbs/[a-fA-F0-9]/[a-fA-F0-9]/[a-fA-F0-9]/([a-fA-F0-9.]+jpg)', webpage)
3443 if mobj is None:
3444 self._downloader.trouble(u'ERROR: unable to extract video thumbnail')
3445 return
3446 video_thumbnail = mobj.group(1).decode('utf-8')
3447
3448
3449
3450 self._downloader.increment_downloads()
3451 info = {
3452 'id': video_id,
3453 'url': video_url,
3454 'uploader': None,
3455 'upload_date': None,
3456 'title': video_title,
3457 'stitle': _simplify_title(video_title),
3458 'ext': 'flv',
3459 'format': 'flv',
3460 'thumbnail': video_thumbnail,
3461 'description': None,
3462 'player_url': None,
3463 }
3464
3465 try:
3466 self._downloader.process_info(info)
3467 except UnavailableVideoError, err:
3468 self._downloader.trouble(u'\nERROR: unable to download ' + video_id)
3469
3470
3471 class SoundcloudIE(InfoExtractor):
3472 """Information extractor for soundcloud.com
3473 To access the media, the uid of the song and a stream token
3474 must be extracted from the page source and the script must make
3475 a request to media.soundcloud.com/crossdomain.xml. Then
3476 the media can be grabbed by requesting from an url composed
3477 of the stream token and uid
3478 """
3479
3480 _VALID_URL = r'^(?:https?://)?(?:www\.)?soundcloud\.com/([\w\d-]+)/([\w\d-]+)'
3481 IE_NAME = u'soundcloud'
3482
3483 def __init__(self, downloader=None):
3484 InfoExtractor.__init__(self, downloader)
3485
3486 def report_webpage(self, video_id):
3487 """Report information extraction."""
3488 self._downloader.to_screen(u'[%s] %s: Downloading webpage' % (self.IE_NAME, video_id))
3489
3490 def report_extraction(self, video_id):
3491 """Report information extraction."""
3492 self._downloader.to_screen(u'[%s] %s: Extracting information' % (self.IE_NAME, video_id))
3493
3494 def _real_extract(self, url):
3495 htmlParser = HTMLParser.HTMLParser()
3496
3497 mobj = re.match(self._VALID_URL, url)
3498 if mobj is None:
3499 self._downloader.trouble(u'ERROR: invalid URL: %s' % url)
3500 return
3501
3502 # extract uploader (which is in the url)
3503 uploader = mobj.group(1).decode('utf-8')
3504 # extract simple title (uploader + slug of song title)
3505 slug_title = mobj.group(2).decode('utf-8')
3506 simple_title = uploader + '-' + slug_title
3507
3508 self.report_webpage('%s/%s' % (uploader, slug_title))
3509
3510 request = urllib2.Request('http://soundcloud.com/%s/%s' % (uploader, slug_title))
3511 try:
3512 webpage = urllib2.urlopen(request).read()
3513 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3514 self._downloader.trouble(u'ERROR: unable to download video webpage: %s' % str(err))
3515 return
3516
3517 self.report_extraction('%s/%s' % (uploader, slug_title))
3518
3519 # extract uid and stream token that soundcloud hands out for access
3520 mobj = re.search('"uid":"([\w\d]+?)".*?stream_token=([\w\d]+)', webpage)
3521 if mobj:
3522 video_id = mobj.group(1)
3523 stream_token = mobj.group(2)
3524
3525 # extract unsimplified title
3526 mobj = re.search('"title":"(.*?)",', webpage)
3527 if mobj:
3528 title = mobj.group(1)
3529
3530 # construct media url (with uid/token)
3531 mediaURL = "http://media.soundcloud.com/stream/%s?stream_token=%s"
3532 mediaURL = mediaURL % (video_id, stream_token)
3533
3534 # description
3535 description = u'No description available'
3536 mobj = re.search('track-description-value"><p>(.*?)</p>', webpage)
3537 if mobj:
3538 description = mobj.group(1)
3539
3540 # upload date
3541 upload_date = None
3542 mobj = re.search("pretty-date'>on ([\w]+ [\d]+, [\d]+ \d+:\d+)</abbr></h2>", webpage)
3543 if mobj:
3544 try:
3545 upload_date = datetime.datetime.strptime(mobj.group(1), '%B %d, %Y %H:%M').strftime('%Y%m%d')
3546 except Exception, e:
3547 print str(e)
3548
3549 # for soundcloud, a request to a cross domain is required for cookies
3550 request = urllib2.Request('http://media.soundcloud.com/crossdomain.xml', std_headers)
3551
3552 try:
3553 self._downloader.process_info({
3554 'id': video_id.decode('utf-8'),
3555 'url': mediaURL,
3556 'uploader': uploader.decode('utf-8'),
3557 'upload_date': upload_date,
3558 'title': simple_title.decode('utf-8'),
3559 'stitle': simple_title.decode('utf-8'),
3560 'ext': u'mp3',
3561 'format': u'NA',
3562 'player_url': None,
3563 'description': description.decode('utf-8')
3564 })
3565 except UnavailableVideoError:
3566 self._downloader.trouble(u'\nERROR: unable to download video')
3567
3568
3569 class InfoQIE(InfoExtractor):
3570 """Information extractor for infoq.com"""
3571
3572 _VALID_URL = r'^(?:https?://)?(?:www\.)?infoq\.com/[^/]+/[^/]+$'
3573 IE_NAME = u'infoq'
3574
3575 def report_webpage(self, video_id):
3576 """Report information extraction."""
3577 self._downloader.to_screen(u'[%s] %s: Downloading webpage' % (self.IE_NAME, video_id))
3578
3579 def report_extraction(self, video_id):
3580 """Report information extraction."""
3581 self._downloader.to_screen(u'[%s] %s: Extracting information' % (self.IE_NAME, video_id))
3582
3583 def _real_extract(self, url):
3584 htmlParser = HTMLParser.HTMLParser()
3585
3586 mobj = re.match(self._VALID_URL, url)
3587 if mobj is None:
3588 self._downloader.trouble(u'ERROR: invalid URL: %s' % url)
3589 return
3590
3591 self.report_webpage(url)
3592
3593 request = urllib2.Request(url)
3594 try:
3595 webpage = urllib2.urlopen(request).read()
3596 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3597 self._downloader.trouble(u'ERROR: unable to download video webpage: %s' % str(err))
3598 return
3599
3600 self.report_extraction(url)
3601
3602
3603 # Extract video URL
3604 mobj = re.search(r"jsclassref='([^']*)'", webpage)
3605 if mobj is None:
3606 self._downloader.trouble(u'ERROR: unable to extract video url')
3607 return
3608 video_url = 'rtmpe://video.infoq.com/cfx/st/' + urllib2.unquote(mobj.group(1).decode('base64'))
3609
3610
3611 # Extract title
3612 mobj = re.search(r'contentTitle = "(.*?)";', webpage)
3613 if mobj is None:
3614 self._downloader.trouble(u'ERROR: unable to extract video title')
3615 return
3616 video_title = mobj.group(1).decode('utf-8')
3617
3618 # Extract description
3619 video_description = u'No description available.'
3620 mobj = re.search(r'<meta name="description" content="(.*)"(?:\s*/)?>', webpage)
3621 if mobj is not None:
3622 video_description = mobj.group(1).decode('utf-8')
3623
3624 video_filename = video_url.split('/')[-1]
3625 video_id, extension = video_filename.split('.')
3626
3627 self._downloader.increment_downloads()
3628 info = {
3629 'id': video_id,
3630 'url': video_url,
3631 'uploader': None,
3632 'upload_date': None,
3633 'title': video_title,
3634 'stitle': _simplify_title(video_title),
3635 'ext': extension,
3636 'format': extension, # Extension is always(?) mp4, but seems to be flv
3637 'thumbnail': None,
3638 'description': video_description,
3639 'player_url': None,
3640 }
3641
3642 try:
3643 self._downloader.process_info(info)
3644 except UnavailableVideoError, err:
3645 self._downloader.trouble(u'\nERROR: unable to download ' + video_url)
3646
3647 class MixcloudIE(InfoExtractor):
3648 """Information extractor for www.mixcloud.com"""
3649 _VALID_URL = r'^(?:https?://)?(?:www\.)?mixcloud\.com/([\w\d-]+)/([\w\d-]+)'
3650 IE_NAME = u'mixcloud'
3651
3652 def __init__(self, downloader=None):
3653 InfoExtractor.__init__(self, downloader)
3654
3655 def report_download_json(self, file_id):
3656 """Report JSON download."""
3657 self._downloader.to_screen(u'[%s] Downloading json' % self.IE_NAME)
3658
3659 def report_extraction(self, file_id):
3660 """Report information extraction."""
3661 self._downloader.to_screen(u'[%s] %s: Extracting information' % (self.IE_NAME, file_id))
3662
3663 def get_urls(self, jsonData, fmt, bitrate='best'):
3664 """Get urls from 'audio_formats' section in json"""
3665 file_url = None
3666 try:
3667 bitrate_list = jsonData[fmt]
3668 if bitrate is None or bitrate == 'best' or bitrate not in bitrate_list:
3669 bitrate = max(bitrate_list) # select highest
3670
3671 url_list = jsonData[fmt][bitrate]
3672 except TypeError: # we have no bitrate info.
3673 url_list = jsonData[fmt]
3674
3675 return url_list
3676
3677 def check_urls(self, url_list):
3678 """Returns 1st active url from list"""
3679 for url in url_list:
3680 try:
3681 urllib2.urlopen(url)
3682 return url
3683 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3684 url = None
3685
3686 return None
3687
3688 def _print_formats(self, formats):
3689 print 'Available formats:'
3690 for fmt in formats.keys():
3691 for b in formats[fmt]:
3692 try:
3693 ext = formats[fmt][b][0]
3694 print '%s\t%s\t[%s]' % (fmt, b, ext.split('.')[-1])
3695 except TypeError: # we have no bitrate info
3696 ext = formats[fmt][0]
3697 print '%s\t%s\t[%s]' % (fmt, '??', ext.split('.')[-1])
3698 break
3699
3700 def _real_extract(self, url):
3701 mobj = re.match(self._VALID_URL, url)
3702 if mobj is None:
3703 self._downloader.trouble(u'ERROR: invalid URL: %s' % url)
3704 return
3705 # extract uploader & filename from url
3706 uploader = mobj.group(1).decode('utf-8')
3707 file_id = uploader + "-" + mobj.group(2).decode('utf-8')
3708
3709 # construct API request
3710 file_url = 'http://www.mixcloud.com/api/1/cloudcast/' + '/'.join(url.split('/')[-3:-1]) + '.json'
3711 # retrieve .json file with links to files
3712 request = urllib2.Request(file_url)
3713 try:
3714 self.report_download_json(file_url)
3715 jsonData = urllib2.urlopen(request).read()
3716 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3717 self._downloader.trouble(u'ERROR: Unable to retrieve file: %s' % str(err))
3718 return
3719
3720 # parse JSON
3721 json_data = json.loads(jsonData)
3722 player_url = json_data['player_swf_url']
3723 formats = dict(json_data['audio_formats'])
3724
3725 req_format = self._downloader.params.get('format', None)
3726 bitrate = None
3727
3728 if self._downloader.params.get('listformats', None):
3729 self._print_formats(formats)
3730 return
3731
3732 if req_format is None or req_format == 'best':
3733 for format_param in formats.keys():
3734 url_list = self.get_urls(formats, format_param)
3735 # check urls
3736 file_url = self.check_urls(url_list)
3737 if file_url is not None:
3738 break # got it!
3739 else:
3740 if req_format not in formats.keys():
3741 self._downloader.trouble(u'ERROR: format is not available')
3742 return
3743
3744 url_list = self.get_urls(formats, req_format)
3745 file_url = self.check_urls(url_list)
3746 format_param = req_format
3747
3748 # We have audio
3749 self._downloader.increment_downloads()
3750 try:
3751 # Process file information
3752 self._downloader.process_info({
3753 'id': file_id.decode('utf-8'),
3754 'url': file_url.decode('utf-8'),
3755 'uploader': uploader.decode('utf-8'),
3756 'upload_date': u'NA',
3757 'title': json_data['name'],
3758 'stitle': _simplify_title(json_data['name']),
3759 'ext': file_url.split('.')[-1].decode('utf-8'),
3760 'format': (format_param is None and u'NA' or format_param.decode('utf-8')),
3761 'thumbnail': json_data['thumbnail_url'],
3762 'description': json_data['description'],
3763 'player_url': player_url.decode('utf-8'),
3764 })
3765 except UnavailableVideoError, err:
3766 self._downloader.trouble(u'ERROR: unable to download file')
3767
3768 class StanfordOpenClassroomIE(InfoExtractor):
3769 """Information extractor for Stanford's Open ClassRoom"""
3770
3771 _VALID_URL = r'^(?:https?://)?openclassroom.stanford.edu(?P<path>/?|(/MainFolder/(?:HomePage|CoursePage|VideoPage)\.php([?]course=(?P<course>[^&]+)(&video=(?P<video>[^&]+))?(&.*)?)?))$'
3772 IE_NAME = u'stanfordoc'
3773
3774 def report_download_webpage(self, objid):
3775 """Report information extraction."""
3776 self._downloader.to_screen(u'[%s] %s: Downloading webpage' % (self.IE_NAME, objid))
3777
3778 def report_extraction(self, video_id):
3779 """Report information extraction."""
3780 self._downloader.to_screen(u'[%s] %s: Extracting information' % (self.IE_NAME, video_id))
3781
3782 def _real_extract(self, url):
3783 mobj = re.match(self._VALID_URL, url)
3784 if mobj is None:
3785 self._downloader.trouble(u'ERROR: invalid URL: %s' % url)
3786 return
3787
3788 if mobj.group('course') and mobj.group('video'): # A specific video
3789 course = mobj.group('course')
3790 video = mobj.group('video')
3791 info = {
3792 'id': _simplify_title(course + '_' + video),
3793 }
3794
3795 self.report_extraction(info['id'])
3796 baseUrl = 'http://openclassroom.stanford.edu/MainFolder/courses/' + course + '/videos/'
3797 xmlUrl = baseUrl + video + '.xml'
3798 try:
3799 metaXml = urllib2.urlopen(xmlUrl).read()
3800 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3801 self._downloader.trouble(u'ERROR: unable to download video info XML: %s' % unicode(err))
3802 return
3803 mdoc = xml.etree.ElementTree.fromstring(metaXml)
3804 try:
3805 info['title'] = mdoc.findall('./title')[0].text
3806 info['url'] = baseUrl + mdoc.findall('./videoFile')[0].text
3807 except IndexError:
3808 self._downloader.trouble(u'\nERROR: Invalid metadata XML file')
3809 return
3810 info['stitle'] = _simplify_title(info['title'])
3811 info['ext'] = info['url'].rpartition('.')[2]
3812 info['format'] = info['ext']
3813 self._downloader.increment_downloads()
3814 try:
3815 self._downloader.process_info(info)
3816 except UnavailableVideoError, err:
3817 self._downloader.trouble(u'\nERROR: unable to download video')
3818 elif mobj.group('course'): # A course page
3819 unescapeHTML = HTMLParser.HTMLParser().unescape
3820
3821 course = mobj.group('course')
3822 info = {
3823 'id': _simplify_title(course),
3824 'type': 'playlist',
3825 }
3826
3827 self.report_download_webpage(info['id'])
3828 try:
3829 coursepage = urllib2.urlopen(url).read()
3830 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3831 self._downloader.trouble(u'ERROR: unable to download course info page: ' + unicode(err))
3832 return
3833
3834 m = re.search('<h1>([^<]+)</h1>', coursepage)
3835 if m:
3836 info['title'] = unescapeHTML(m.group(1))
3837 else:
3838 info['title'] = info['id']
3839 info['stitle'] = _simplify_title(info['title'])
3840
3841 m = re.search('<description>([^<]+)</description>', coursepage)
3842 if m:
3843 info['description'] = unescapeHTML(m.group(1))
3844
3845 links = _orderedSet(re.findall('<a href="(VideoPage.php\?[^"]+)">', coursepage))
3846 info['list'] = [
3847 {
3848 'type': 'reference',
3849 'url': 'http://openclassroom.stanford.edu/MainFolder/' + unescapeHTML(vpage),
3850 }
3851 for vpage in links]
3852
3853 for entry in info['list']:
3854 assert entry['type'] == 'reference'
3855 self.extract(entry['url'])
3856 else: # Root page
3857 unescapeHTML = HTMLParser.HTMLParser().unescape
3858
3859 info = {
3860 'id': 'Stanford OpenClassroom',
3861 'type': 'playlist',
3862 }
3863
3864 self.report_download_webpage(info['id'])
3865 rootURL = 'http://openclassroom.stanford.edu/MainFolder/HomePage.php'
3866 try:
3867 rootpage = urllib2.urlopen(rootURL).read()
3868 except (urllib2.URLError, httplib.HTTPException, socket.error), err:
3869 self._downloader.trouble(u'ERROR: unable to download course info page: ' + unicode(err))
3870 return
3871
3872 info['title'] = info['id']
3873 info['stitle'] = _simplify_title(info['title'])
3874
3875 links = _orderedSet(re.findall('<a href="(CoursePage.php\?[^"]+)">', rootpage))
3876 info['list'] = [
3877 {
3878 'type': 'reference',
3879 'url': 'http://openclassroom.stanford.edu/MainFolder/' + unescapeHTML(cpage),
3880 }
3881 for cpage in links]
3882
3883 for entry in info['list']:
3884 assert entry['type'] == 'reference'
3885 self.extract(entry['url'])
3886
3887
3888 class PostProcessor(object):
3889 """Post Processor class.
3890
3891 PostProcessor objects can be added to downloaders with their
3892 add_post_processor() method. When the downloader has finished a
3893 successful download, it will take its internal chain of PostProcessors
3894 and start calling the run() method on each one of them, first with
3895 an initial argument and then with the returned value of the previous
3896 PostProcessor.
3897
3898 The chain will be stopped if one of them ever returns None or the end
3899 of the chain is reached.
3900
3901 PostProcessor objects follow a "mutual registration" process similar
3902 to InfoExtractor objects.
3903 """
3904
3905 _downloader = None
3906
3907 def __init__(self, downloader=None):
3908 self._downloader = downloader
3909
3910 def set_downloader(self, downloader):
3911 """Sets the downloader for this PP."""
3912 self._downloader = downloader
3913
3914 def run(self, information):
3915 """Run the PostProcessor.
3916
3917 The "information" argument is a dictionary like the ones
3918 composed by InfoExtractors. The only difference is that this
3919 one has an extra field called "filepath" that points to the
3920 downloaded file.
3921
3922 When this method returns None, the postprocessing chain is
3923 stopped. However, this method may return an information
3924 dictionary that will be passed to the next postprocessing
3925 object in the chain. It can be the one it received after
3926 changing some fields.
3927
3928 In addition, this method may raise a PostProcessingError
3929 exception that will be taken into account by the downloader
3930 it was called from.
3931 """
3932 return information # by default, do nothing
3933
3934 class AudioConversionError(BaseException):
3935 def __init__(self, message):
3936 self.message = message
3937
3938 class FFmpegExtractAudioPP(PostProcessor):
3939
3940 def __init__(self, downloader=None, preferredcodec=None, preferredquality=None, keepvideo=False):
3941 PostProcessor.__init__(self, downloader)
3942 if preferredcodec is None:
3943 preferredcodec = 'best'
3944 self._preferredcodec = preferredcodec
3945 self._preferredquality = preferredquality
3946 self._keepvideo = keepvideo
3947
3948 @staticmethod
3949 def get_audio_codec(path):
3950 try:
3951 cmd = ['ffprobe', '-show_streams', '--', path]
3952 handle = subprocess.Popen(cmd, stderr=file(os.path.devnull, 'w'), stdout=subprocess.PIPE)
3953 output = handle.communicate()[0]
3954 if handle.wait() != 0:
3955 return None
3956 except (IOError, OSError):
3957 return None
3958 audio_codec = None
3959 for line in output.split('\n'):
3960 if line.startswith('codec_name='):
3961 audio_codec = line.split('=')[1].strip()
3962 elif line.strip() == 'codec_type=audio' and audio_codec is not None:
3963 return audio_codec
3964 return None
3965
3966 @staticmethod
3967 def run_ffmpeg(path, out_path, codec, more_opts):
3968 if codec is None:
3969 acodec_opts = []
3970 else:
3971 acodec_opts = ['-acodec', codec]
3972 cmd = ['ffmpeg', '-y', '-i', path, '-vn'] + acodec_opts + more_opts + ['--', out_path]
3973 try:
3974 p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
3975 stdout,stderr = p.communicate()
3976 except (IOError, OSError):
3977 e = sys.exc_info()[1]
3978 if isinstance(e, OSError) and e.errno == 2:
3979 raise AudioConversionError('ffmpeg not found. Please install ffmpeg.')
3980 else:
3981 raise e
3982 if p.returncode != 0:
3983 msg = stderr.strip().split('\n')[-1]
3984 raise AudioConversionError(msg)
3985
3986 def run(self, information):
3987 path = information['filepath']
3988
3989 filecodec = self.get_audio_codec(path)
3990 if filecodec is None:
3991 self._downloader.to_stderr(u'WARNING: unable to obtain file audio codec with ffprobe')
3992 return None
3993
3994 more_opts = []
3995 if self._preferredcodec == 'best' or self._preferredcodec == filecodec or (self._preferredcodec == 'm4a' and filecodec == 'aac'):
3996 if self._preferredcodec == 'm4a' and filecodec == 'aac':
3997 # Lossless, but in another container
3998 acodec = 'copy'
3999 extension = self._preferredcodec
4000 more_opts = ['-absf', 'aac_adtstoasc']
4001 elif filecodec in ['aac', 'mp3', 'vorbis']:
4002 # Lossless if possible
4003 acodec = 'copy'
4004 extension = filecodec
4005 if filecodec == 'aac':
4006 more_opts = ['-f', 'adts']
4007 if filecodec == 'vorbis':
4008 extension = 'ogg'
4009 else:
4010 # MP3 otherwise.
4011 acodec = 'libmp3lame'
4012 extension = 'mp3'
4013 more_opts = []
4014 if self._preferredquality is not None:
4015 more_opts += ['-ab', self._preferredquality]
4016 else:
4017 # We convert the audio (lossy)
4018 acodec = {'mp3': 'libmp3lame', 'aac': 'aac', 'm4a': 'aac', 'vorbis': 'libvorbis', 'wav': None}[self._preferredcodec]
4019 extension = self._preferredcodec
4020 more_opts = []
4021 if self._preferredquality is not None:
4022 more_opts += ['-ab', self._preferredquality]
4023 if self._preferredcodec == 'aac':
4024 more_opts += ['-f', 'adts']
4025 if self._preferredcodec == 'm4a':
4026 more_opts += ['-absf', 'aac_adtstoasc']
4027 if self._preferredcodec == 'vorbis':
4028 extension = 'ogg'
4029 if self._preferredcodec == 'wav':
4030 extension = 'wav'
4031 more_opts += ['-f', 'wav']
4032
4033 (prefix, ext) = os.path.splitext(path)
4034 new_path = prefix + '.' + extension
4035 self._downloader.to_screen(u'[ffmpeg] Destination: %s' % new_path)
4036 try:
4037 self.run_ffmpeg(path, new_path, acodec, more_opts)
4038 except:
4039 etype,e,tb = sys.exc_info()
4040 if isinstance(e, AudioConversionError):
4041 self._downloader.to_stderr(u'ERROR: audio conversion failed: ' + e.message)
4042 else:
4043 self._downloader.to_stderr(u'ERROR: error running ffmpeg')
4044 return None
4045
4046 # Try to update the date time for extracted audio file.
4047 if information.get('filetime') is not None:
4048 try:
4049 os.utime(new_path, (time.time(), information['filetime']))
4050 except:
4051 self._downloader.to_stderr(u'WARNING: Cannot update utime of audio file')
4052
4053 if not self._keepvideo:
4054 try:
4055 os.remove(path)
4056 except (IOError, OSError):
4057 self._downloader.to_stderr(u'WARNING: Unable to remove downloaded video file')
4058 return None
4059
4060 information['filepath'] = new_path
4061 return information
4062
4063
4064 def updateSelf(downloader, filename):
4065 ''' Update the program file with the latest version from the repository '''
4066 # Note: downloader only used for options
4067 if not os.access(filename, os.W_OK):
4068 sys.exit('ERROR: no write permissions on %s' % filename)
4069
4070 downloader.to_screen('Updating to latest version...')
4071
4072 try:
4073 try:
4074 urlh = urllib.urlopen(UPDATE_URL)
4075 newcontent = urlh.read()
4076
4077 vmatch = re.search("__version__ = '([^']+)'", newcontent)
4078 if vmatch is not None and vmatch.group(1) == __version__:
4079 downloader.to_screen('youtube-dl is up-to-date (' + __version__ + ')')
4080 return
4081 finally:
4082 urlh.close()
4083 except (IOError, OSError), err:
4084 sys.exit('ERROR: unable to download latest version')
4085
4086 try:
4087 outf = open(filename, 'wb')
4088 try:
4089 outf.write(newcontent)
4090 finally:
4091 outf.close()
4092 except (IOError, OSError), err:
4093 sys.exit('ERROR: unable to overwrite current version')
4094
4095 downloader.to_screen('Updated youtube-dl. Restart youtube-dl to use the new version.')
4096
4097 def parseOpts():
4098 # Deferred imports
4099 import getpass
4100 import optparse
4101 import shlex
4102
4103 def _readOptions(filename):
4104 try:
4105 optionf = open(filename)
4106 except IOError:
4107 return [] # silently skip if file is not present
4108 try:
4109 res = []
4110 for l in optionf:
4111 res += shlex.split(l, comments=True)
4112 finally:
4113 optionf.close()
4114 return res
4115
4116 def _format_option_string(option):
4117 ''' ('-o', '--option') -> -o, --format METAVAR'''
4118
4119 opts = []
4120
4121 if option._short_opts: opts.append(option._short_opts[0])
4122 if option._long_opts: opts.append(option._long_opts[0])
4123 if len(opts) > 1: opts.insert(1, ', ')
4124
4125 if option.takes_value(): opts.append(' %s' % option.metavar)
4126
4127 return "".join(opts)
4128
4129 def _find_term_columns():
4130 columns = os.environ.get('COLUMNS', None)
4131 if columns:
4132 return int(columns)
4133
4134 try:
4135 sp = subprocess.Popen(['stty', 'size'], stdout=subprocess.PIPE, stderr=subprocess.PIPE)
4136 out,err = sp.communicate()
4137 return int(out.split()[1])
4138 except:
4139 pass
4140 return None
4141
4142 max_width = 80
4143 max_help_position = 80
4144
4145 # No need to wrap help messages if we're on a wide console
4146 columns = _find_term_columns()
4147 if columns: max_width = columns
4148
4149 fmt = optparse.IndentedHelpFormatter(width=max_width, max_help_position=max_help_position)
4150 fmt.format_option_strings = _format_option_string
4151
4152 kw = {
4153 'version' : __version__,
4154 'formatter' : fmt,
4155 'usage' : '%prog [options] url [url...]',
4156 'conflict_handler' : 'resolve',
4157 }
4158
4159 parser = optparse.OptionParser(**kw)
4160
4161 # option groups
4162 general = optparse.OptionGroup(parser, 'General Options')
4163 selection = optparse.OptionGroup(parser, 'Video Selection')
4164 authentication = optparse.OptionGroup(parser, 'Authentication Options')
4165 video_format = optparse.OptionGroup(parser, 'Video Format Options')
4166 postproc = optparse.OptionGroup(parser, 'Post-processing Options')
4167 filesystem = optparse.OptionGroup(parser, 'Filesystem Options')
4168 verbosity = optparse.OptionGroup(parser, 'Verbosity / Simulation Options')
4169
4170 general.add_option('-h', '--help',
4171 action='help', help='print this help text and exit')
4172 general.add_option('-v', '--version',
4173 action='version', help='print program version and exit')
4174 general.add_option('-U', '--update',
4175 action='store_true', dest='update_self', help='update this program to latest version')
4176 general.add_option('-i', '--ignore-errors',
4177 action='store_true', dest='ignoreerrors', help='continue on download errors', default=False)
4178 general.add_option('-r', '--rate-limit',
4179 dest='ratelimit', metavar='LIMIT', help='download rate limit (e.g. 50k or 44.6m)')
4180 general.add_option('-R', '--retries',
4181 dest='retries', metavar='RETRIES', help='number of retries (default is 10)', default=10)
4182 general.add_option('--dump-user-agent',
4183 action='store_true', dest='dump_user_agent',
4184 help='display the current browser identification', default=False)
4185 general.add_option('--list-extractors',
4186 action='store_true', dest='list_extractors',
4187 help='List all supported extractors and the URLs they would handle', default=False)
4188
4189 selection.add_option('--playlist-start',
4190 dest='playliststart', metavar='NUMBER', help='playlist video to start at (default is 1)', default=1)
4191 selection.add_option('--playlist-end',
4192 dest='playlistend', metavar='NUMBER', help='playlist video to end at (default is last)', default=-1)
4193 selection.add_option('--match-title', dest='matchtitle', metavar='REGEX',help='download only matching titles (regex or caseless sub-string)')
4194 selection.add_option('--reject-title', dest='rejecttitle', metavar='REGEX',help='skip download for matching titles (regex or caseless sub-string)')
4195 selection.add_option('--max-downloads', metavar='NUMBER', dest='max_downloads', help='Abort after downloading NUMBER files', default=None)
4196
4197 authentication.add_option('-u', '--username',
4198 dest='username', metavar='USERNAME', help='account username')
4199 authentication.add_option('-p', '--password',
4200 dest='password', metavar='PASSWORD', help='account password')
4201 authentication.add_option('-n', '--netrc',
4202 action='store_true', dest='usenetrc', help='use .netrc authentication data', default=False)
4203
4204
4205 video_format.add_option('-f', '--format',
4206 action='store', dest='format', metavar='FORMAT', help='video format code')
4207 video_format.add_option('--all-formats',
4208 action='store_const', dest='format', help='download all available video formats', const='all')
4209 video_format.add_option('--prefer-free-formats',
4210 action='store_true', dest='prefer_free_formats', default=False, help='prefer free video formats unless a specific one is requested')
4211 video_format.add_option('--max-quality',
4212 action='store', dest='format_limit', metavar='FORMAT', help='highest quality format to download')
4213 video_format.add_option('-F', '--list-formats',
4214 action='store_true', dest='listformats', help='list all available formats (currently youtube only)')
4215
4216
4217 verbosity.add_option('-q', '--quiet',
4218 action='store_true', dest='quiet', help='activates quiet mode', default=False)
4219 verbosity.add_option('-s', '--simulate',
4220 action='store_true', dest='simulate', help='do not download the video and do not write anything to disk', default=False)
4221 verbosity.add_option('--skip-download',
4222 action='store_true', dest='skip_download', help='do not download the video', default=False)
4223 verbosity.add_option('-g', '--get-url',
4224 action='store_true', dest='geturl', help='simulate, quiet but print URL', default=False)
4225 verbosity.add_option('-e', '--get-title',
4226 action='store_true', dest='gettitle', help='simulate, quiet but print title', default=False)
4227 verbosity.add_option('--get-thumbnail',
4228 action='store_true', dest='getthumbnail',
4229 help='simulate, quiet but print thumbnail URL', default=False)
4230 verbosity.add_option('--get-description',
4231 action='store_true', dest='getdescription',
4232 help='simulate, quiet but print video description', default=False)
4233 verbosity.add_option('--get-filename',
4234 action='store_true', dest='getfilename',
4235 help='simulate, quiet but print output filename', default=False)
4236 verbosity.add_option('--get-format',
4237 action='store_true', dest='getformat',
4238 help='simulate, quiet but print output format', default=False)
4239 verbosity.add_option('--no-progress',
4240 action='store_true', dest='noprogress', help='do not print progress bar', default=False)
4241 verbosity.add_option('--console-title',
4242 action='store_true', dest='consoletitle',
4243 help='display progress in console titlebar', default=False)
4244
4245
4246 filesystem.add_option('-t', '--title',
4247 action='store_true', dest='usetitle', help='use title in file name', default=False)
4248 filesystem.add_option('-l', '--literal',
4249 action='store_true', dest='useliteral', help='use literal title in file name', default=False)
4250 filesystem.add_option('-A', '--auto-number',
4251 action='store_true', dest='autonumber',
4252 help='number downloaded files starting from 00000', default=False)
4253 filesystem.add_option('-o', '--output',
4254 dest='outtmpl', metavar='TEMPLATE', help='output filename template. Use %(stitle)s to get the title, %(uploader)s for the uploader name, %(autonumber)s to get an automatically incremented number, %(ext)s for the filename extension, %(upload_date)s for the upload date (YYYYMMDD), and %% for a literal percent. Use - to output to stdout.')
4255 filesystem.add_option('-a', '--batch-file',
4256 dest='batchfile', metavar='FILE', help='file containing URLs to download (\'-\' for stdin)')
4257 filesystem.add_option('-w', '--no-overwrites',
4258 action='store_true', dest='nooverwrites', help='do not overwrite files', default=False)
4259 filesystem.add_option('-c', '--continue',
4260 action='store_true', dest='continue_dl', help='resume partially downloaded files', default=False)
4261 filesystem.add_option('--no-continue',
4262 action='store_false', dest='continue_dl',
4263 help='do not resume partially downloaded files (restart from beginning)')
4264 filesystem.add_option('--cookies',
4265 dest='cookiefile', metavar='FILE', help='file to read cookies from and dump cookie jar in')
4266 filesystem.add_option('--no-part',
4267 action='store_true', dest='nopart', help='do not use .part files', default=False)
4268 filesystem.add_option('--no-mtime',
4269 action='store_false', dest='updatetime',
4270 help='do not use the Last-modified header to set the file modification time', default=True)
4271 filesystem.add_option('--write-description',
4272 action='store_true', dest='writedescription',
4273 help='write video description to a .description file', default=False)
4274 filesystem.add_option('--write-info-json',
4275 action='store_true', dest='writeinfojson',
4276 help='write video metadata to a .info.json file', default=False)
4277
4278
4279 postproc.add_option('--extract-audio', action='store_true', dest='extractaudio', default=False,
4280 help='convert video files to audio-only files (requires ffmpeg and ffprobe)')
4281 postproc.add_option('--audio-format', metavar='FORMAT', dest='audioformat', default='best',
4282 help='"best", "aac", "vorbis", "mp3", "m4a", or "wav"; best by default')
4283 postproc.add_option('--audio-quality', metavar='QUALITY', dest='audioquality', default='128K',
4284 help='ffmpeg audio bitrate specification, 128k by default')
4285 postproc.add_option('-k', '--keep-video', action='store_true', dest='keepvideo', default=False,
4286 help='keeps the video file on disk after the post-processing; the video is erased by default')
4287
4288
4289 parser.add_option_group(general)
4290 parser.add_option_group(selection)
4291 parser.add_option_group(filesystem)
4292 parser.add_option_group(verbosity)
4293 parser.add_option_group(video_format)
4294 parser.add_option_group(authentication)
4295 parser.add_option_group(postproc)
4296
4297 xdg_config_home = os.environ.get('XDG_CONFIG_HOME')
4298 if xdg_config_home:
4299 userConf = os.path.join(xdg_config_home, 'youtube-dl.conf')
4300 else:
4301 userConf = os.path.join(os.path.expanduser('~'), '.config', 'youtube-dl.conf')
4302 argv = _readOptions('/etc/youtube-dl.conf') + _readOptions(userConf) + sys.argv[1:]
4303 opts, args = parser.parse_args(argv)
4304
4305 return parser, opts, args
4306
4307 def gen_extractors():
4308 """ Return a list of an instance of every supported extractor.
4309 The order does matter; the first extractor matched is the one handling the URL.
4310 """
4311 youtube_ie = YoutubeIE()
4312 google_ie = GoogleIE()
4313 yahoo_ie = YahooIE()
4314 return [
4315 YoutubePlaylistIE(youtube_ie),
4316 YoutubeUserIE(youtube_ie),
4317 YoutubeSearchIE(youtube_ie),
4318 youtube_ie,
4319 MetacafeIE(youtube_ie),
4320 DailymotionIE(),
4321 google_ie,
4322 GoogleSearchIE(google_ie),
4323 PhotobucketIE(),
4324 yahoo_ie,
4325 YahooSearchIE(yahoo_ie),
4326 DepositFilesIE(),
4327 FacebookIE(),
4328 BlipTVIE(),
4329 VimeoIE(),
4330 MyVideoIE(),
4331 ComedyCentralIE(),
4332 EscapistIE(),
4333 CollegeHumorIE(),
4334 XVideosIE(),
4335 SoundcloudIE(),
4336 InfoQIE(),
4337 MixcloudIE(),
4338 StanfordOpenClassroomIE(),
4339
4340 GenericIE()
4341 ]
4342
4343 def _real_main():
4344 parser, opts, args = parseOpts()
4345
4346 # Open appropriate CookieJar
4347 if opts.cookiefile is None:
4348 jar = cookielib.CookieJar()
4349 else:
4350 try:
4351 jar = cookielib.MozillaCookieJar(opts.cookiefile)
4352 if os.path.isfile(opts.cookiefile) and os.access(opts.cookiefile, os.R_OK):
4353 jar.load()
4354 except (IOError, OSError), err:
4355 sys.exit(u'ERROR: unable to open cookie file')
4356
4357 # Dump user agent
4358 if opts.dump_user_agent:
4359 print std_headers['User-Agent']
4360 sys.exit(0)
4361
4362 # Batch file verification
4363 batchurls = []
4364 if opts.batchfile is not None:
4365 try:
4366 if opts.batchfile == '-':
4367 batchfd = sys.stdin
4368 else:
4369 batchfd = open(opts.batchfile, 'r')
4370 batchurls = batchfd.readlines()
4371 batchurls = [x.strip() for x in batchurls]
4372 batchurls = [x for x in batchurls if len(x) > 0 and not re.search(r'^[#/;]', x)]
4373 except IOError:
4374 sys.exit(u'ERROR: batch file could not be read')
4375 all_urls = batchurls + args
4376
4377 # General configuration
4378 cookie_processor = urllib2.HTTPCookieProcessor(jar)
4379 opener = urllib2.build_opener(urllib2.ProxyHandler(), cookie_processor, YoutubeDLHandler())
4380 urllib2.install_opener(opener)
4381 socket.setdefaulttimeout(300) # 5 minutes should be enough (famous last words)
4382
4383 extractors = gen_extractors()
4384
4385 if opts.list_extractors:
4386 for ie in extractors:
4387 print(ie.IE_NAME)
4388 matchedUrls = filter(lambda url: ie.suitable(url), all_urls)
4389 all_urls = filter(lambda url: url not in matchedUrls, all_urls)
4390 for mu in matchedUrls:
4391 print(u' ' + mu)
4392 sys.exit(0)
4393
4394 # Conflicting, missing and erroneous options
4395 if opts.usenetrc and (opts.username is not None or opts.password is not None):
4396 parser.error(u'using .netrc conflicts with giving username/password')
4397 if opts.password is not None and opts.username is None:
4398 parser.error(u'account username missing')
4399 if opts.outtmpl is not None and (opts.useliteral or opts.usetitle or opts.autonumber):
4400 parser.error(u'using output template conflicts with using title, literal title or auto number')
4401 if opts.usetitle and opts.useliteral:
4402 parser.error(u'using title conflicts with using literal title')
4403 if opts.username is not None and opts.password is None:
4404 opts.password = getpass.getpass(u'Type account password and press return:')
4405 if opts.ratelimit is not None:
4406 numeric_limit = FileDownloader.parse_bytes(opts.ratelimit)
4407 if numeric_limit is None:
4408 parser.error(u'invalid rate limit specified')
4409 opts.ratelimit = numeric_limit
4410 if opts.retries is not None:
4411 try:
4412 opts.retries = long(opts.retries)
4413 except (TypeError, ValueError), err:
4414 parser.error(u'invalid retry count specified')
4415 try:
4416 opts.playliststart = int(opts.playliststart)
4417 if opts.playliststart <= 0:
4418 raise ValueError(u'Playlist start must be positive')
4419 except (TypeError, ValueError), err:
4420 parser.error(u'invalid playlist start number specified')
4421 try:
4422 opts.playlistend = int(opts.playlistend)
4423 if opts.playlistend != -1 and (opts.playlistend <= 0 or opts.playlistend < opts.playliststart):
4424 raise ValueError(u'Playlist end must be greater than playlist start')
4425 except (TypeError, ValueError), err:
4426 parser.error(u'invalid playlist end number specified')
4427 if opts.extractaudio:
4428 if opts.audioformat not in ['best', 'aac', 'mp3', 'vorbis', 'm4a', 'wav']:
4429 parser.error(u'invalid audio format specified')
4430
4431 # File downloader
4432 fd = FileDownloader({
4433 'usenetrc': opts.usenetrc,
4434 'username': opts.username,
4435 'password': opts.password,
4436 'quiet': (opts.quiet or opts.geturl or opts.gettitle or opts.getthumbnail or opts.getdescription or opts.getfilename or opts.getformat),
4437 'forceurl': opts.geturl,
4438 'forcetitle': opts.gettitle,
4439 'forcethumbnail': opts.getthumbnail,
4440 'forcedescription': opts.getdescription,
4441 'forcefilename': opts.getfilename,
4442 'forceformat': opts.getformat,
4443 'simulate': opts.simulate,
4444 'skip_download': (opts.skip_download or opts.simulate or opts.geturl or opts.gettitle or opts.getthumbnail or opts.getdescription or opts.getfilename or opts.getformat),
4445 'format': opts.format,
4446 'format_limit': opts.format_limit,
4447 'listformats': opts.listformats,
4448 'outtmpl': ((opts.outtmpl is not None and opts.outtmpl.decode(preferredencoding()))
4449 or (opts.format == '-1' and opts.usetitle and u'%(stitle)s-%(id)s-%(format)s.%(ext)s')
4450 or (opts.format == '-1' and opts.useliteral and u'%(title)s-%(id)s-%(format)s.%(ext)s')
4451 or (opts.format == '-1' and u'%(id)s-%(format)s.%(ext)s')
4452 or (opts.usetitle and opts.autonumber and u'%(autonumber)s-%(stitle)s-%(id)s.%(ext)s')
4453 or (opts.useliteral and opts.autonumber and u'%(autonumber)s-%(title)s-%(id)s.%(ext)s')
4454 or (opts.usetitle and u'%(stitle)s-%(id)s.%(ext)s')
4455 or (opts.useliteral and u'%(title)s-%(id)s.%(ext)s')
4456 or (opts.autonumber and u'%(autonumber)s-%(id)s.%(ext)s')
4457 or u'%(id)s.%(ext)s'),
4458 'ignoreerrors': opts.ignoreerrors,
4459 'ratelimit': opts.ratelimit,
4460 'nooverwrites': opts.nooverwrites,
4461 'retries': opts.retries,
4462 'continuedl': opts.continue_dl,
4463 'noprogress': opts.noprogress,
4464 'playliststart': opts.playliststart,
4465 'playlistend': opts.playlistend,
4466 'logtostderr': opts.outtmpl == '-',
4467 'consoletitle': opts.consoletitle,
4468 'nopart': opts.nopart,
4469 'updatetime': opts.updatetime,
4470 'writedescription': opts.writedescription,
4471 'writeinfojson': opts.writeinfojson,
4472 'matchtitle': opts.matchtitle,
4473 'rejecttitle': opts.rejecttitle,
4474 'max_downloads': opts.max_downloads,
4475 'prefer_free_formats': opts.prefer_free_formats,
4476 })
4477 for extractor in extractors:
4478 fd.add_info_extractor(extractor)
4479
4480 # PostProcessors
4481 if opts.extractaudio:
4482 fd.add_post_processor(FFmpegExtractAudioPP(preferredcodec=opts.audioformat, preferredquality=opts.audioquality, keepvideo=opts.keepvideo))
4483
4484 # Update version
4485 if opts.update_self:
4486 updateSelf(fd, sys.argv[0])
4487
4488 # Maybe do nothing
4489 if len(all_urls) < 1:
4490 if not opts.update_self:
4491 parser.error(u'you must provide at least one URL')
4492 else:
4493 sys.exit()
4494
4495 try:
4496 retcode = fd.download(all_urls)
4497 except MaxDownloadsReached:
4498 fd.to_screen(u'--max-download limit reached, aborting.')
4499 retcode = 101
4500
4501 # Dump cookie jar if requested
4502 if opts.cookiefile is not None:
4503 try:
4504 jar.save()
4505 except (IOError, OSError), err:
4506 sys.exit(u'ERROR: unable to save cookie jar')
4507
4508 sys.exit(retcode)
4509
4510 def main():
4511 try:
4512 _real_main()
4513 except DownloadError:
4514 sys.exit(1)
4515 except SameFileError:
4516 sys.exit(u'ERROR: fixed output name but more than one file to download')
4517 except KeyboardInterrupt:
4518 sys.exit(u'\nERROR: Interrupted by user')
4519
4520 if __name__ == '__main__':
4521 main()
4522
4523 # vim: set ts=4 sw=4 sts=4 noet ai si filetype=python: