X-Git-Url: http://git.cielonegro.org/gitweb.cgi?a=blobdiff_plain;ds=sidebyside;f=youtube-dl;h=20e26f99c06d7b70fda144c6a95a9b5655caa3a9;hb=b0eddb2eb4b8cad325ba3224c8550a21f67e4315;hp=46d45d37d95630d5fd742ee16d182530a3dcb34f;hpb=1c5e23021ed23a1d143d7b0a9554574d1e7050fb;p=youtube-dl.git diff --git a/youtube-dl b/youtube-dl index 46d45d37d..20e26f99c 100755 --- a/youtube-dl +++ b/youtube-dl @@ -1,9 +1,11 @@ #!/usr/bin/env python # -*- coding: utf-8 -*- # Author: Ricardo Garcia Gonzalez +# Author: Danny Colligan # License: Public domain code import htmlentitydefs import httplib +import locale import math import netrc import os @@ -16,8 +18,8 @@ import time import urllib import urllib2 -std_headers = { - 'User-Agent': 'Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.9.0.1) Gecko/2008070208 Firefox/3.0.1', +std_headers = { + 'User-Agent': 'Mozilla/5.0 (Windows; U; Windows NT 6.0; en-US; rv:1.9.0.8) Gecko/2009032609 Firefox/3.0.8', 'Accept-Charset': 'ISO-8859-1,utf-8;q=0.7,*;q=0.7', 'Accept': 'text/xml,application/xml,application/xhtml+xml,text/html;q=0.9,text/plain;q=0.8,image/png,*/*;q=0.5', 'Accept-Language': 'en-us,en;q=0.5', @@ -25,6 +27,31 @@ std_headers = { simple_title_chars = string.ascii_letters.decode('ascii') + string.digits.decode('ascii') +class DownloadError(Exception): + """Download Error exception. + + This exception may be thrown by FileDownloader objects if they are not + configured to continue on errors. They will contain the appropriate + error message. + """ + pass + +class SameFileError(Exception): + """Same File exception. + + This exception will be thrown by FileDownloader objects if they detect + multiple files would have to be downloaded to the same file on disk. + """ + pass + +class PostProcessingError(Exception): + """Post Processing exception. + + This exception may be raised by PostProcessor's .run() method to + indicate an error in the postprocessing task. + """ + pass + class FileDownloader(object): """File Downloader class. @@ -44,10 +71,10 @@ class FileDownloader(object): File downloaders accept a lot of parameters. In order not to saturate the object constructor with arguments, it receives a dictionary of - options instead. These options are available through the get_params() - method for the InfoExtractors to use. The FileDownloader also registers - itself as the downloader in charge for the InfoExtractors that are - added to it, so this is a "mutual registration". + options instead. These options are available through the params + attribute for the InfoExtractors to use. The FileDownloader also + registers itself as the downloader in charge for the InfoExtractors + that are added to it, so this is a "mutual registration". Available options: @@ -61,21 +88,26 @@ class FileDownloader(object): format: Video format code. outtmpl: Template for output names. ignoreerrors: Do not stop on download errors. + ratelimit: Download speed limit, in bytes/sec. + nooverwrites: Prevent overwriting files. """ - _params = None + params = None _ies = [] + _pps = [] def __init__(self, params): """Create a FileDownloader object with the given options.""" self._ies = [] - self.set_params(params) + self._pps = [] + self.params = params @staticmethod def pmkdir(filename): """Create directory components in filename. Similar to Unix "mkdir -p".""" components = filename.split(os.sep) aggregate = [os.sep.join(components[0:x]) for x in xrange(1, len(components))] + aggregate = ['%s%s' % (x, os.sep) for x in aggregate] # Finish names with separator for dir in aggregate: if not os.path.exists(dir): os.mkdir(dir) @@ -112,7 +144,7 @@ class FileDownloader(object): return '--:--' return '%02d:%02d' % (eta_mins, eta_secs) - @staticmethod + @staticmethod def calc_speed(start, now, bytes): dif = now - start if bytes == 0 or dif < 0.001: # One millisecond @@ -132,112 +164,175 @@ class FileDownloader(object): return int(new_min) return int(rate) - def set_params(self, params): - """Sets parameters.""" - if type(params) != dict: - raise ValueError('params: dictionary expected') - self._params = params - - def get_params(self): - """Get parameters.""" - return self._params + @staticmethod + def parse_bytes(bytestr): + """Parse a string indicating a byte quantity into a long integer.""" + matchobj = re.match(r'(?i)^(\d+(?:\.\d+)?)([kMGTPEZY]?)$', bytestr) + if matchobj is None: + return None + number = float(matchobj.group(1)) + multiplier = 1024.0 ** 'bkmgtpezy'.index(matchobj.group(2).lower()) + return long(round(number * multiplier)) def add_info_extractor(self, ie): """Add an InfoExtractor object to the end of the list.""" self._ies.append(ie) ie.set_downloader(self) + def add_post_processor(self, pp): + """Add a PostProcessor object to the end of the chain.""" + self._pps.append(pp) + pp.set_downloader(self) + def to_stdout(self, message, skip_eol=False): """Print message to stdout if not in quiet mode.""" - if not self._params.get('quiet', False): - sys.stdout.write('%s%s' % (message, ['\n', ''][skip_eol])) + if not self.params.get('quiet', False): + print u'%s%s' % (message, [u'\n', u''][skip_eol]), sys.stdout.flush() def to_stderr(self, message): """Print message to stderr.""" - sys.stderr.write('%s\n' % message) + print >>sys.stderr, message def fixed_template(self): """Checks if the output template is fixed.""" - return (re.search(ur'(?u)%\(.+?\)s', self._params['outtmpl']) is None) + return (re.search(ur'(?u)%\(.+?\)s', self.params['outtmpl']) is None) def trouble(self, message=None): """Determine action to take when a download problem appears. Depending on if the downloader has been configured to ignore - download errors or not, this method may exit the program or + download errors or not, this method may throw an exception or not when errors are found, after printing the message. If it - doesn't exit, it returns an error code suitable to be returned + doesn't raise, it returns an error code suitable to be returned later as a program exit code to indicate error. """ if message is not None: self.to_stderr(message) - if not self._params.get('ignoreerrors', False): - sys.exit(1) + if not self.params.get('ignoreerrors', False): + raise DownloadError(message) return 1 + def slow_down(self, start_time, byte_counter): + """Sleep if the download speed is over the rate limit.""" + rate_limit = self.params.get('ratelimit', None) + if rate_limit is None or byte_counter == 0: + return + now = time.time() + elapsed = now - start_time + if elapsed <= 0.0: + return + speed = float(byte_counter) / elapsed + if speed > rate_limit: + time.sleep((byte_counter - rate_limit * (now - start_time)) / rate_limit) + + def report_destination(self, filename): + """Report destination filename.""" + self.to_stdout(u'[download] Destination: %s' % filename) + + def report_progress(self, percent_str, data_len_str, speed_str, eta_str): + """Report download progress.""" + self.to_stdout(u'\r[download] %s of %s at %s ETA %s' % + (percent_str, data_len_str, speed_str, eta_str), skip_eol=True) + + def report_finish(self): + """Report download finished.""" + self.to_stdout(u'') + + def process_info(self, info_dict): + """Process a single dictionary returned by an InfoExtractor.""" + # Forced printings + if self.params.get('forcetitle', False): + print info_dict['title'] + if self.params.get('forceurl', False): + print info_dict['url'] + + # Do nothing else if in simulate mode + if self.params.get('simulate', False): + return 0 + + try: + filename = self.params['outtmpl'] % info_dict + self.report_destination(filename) + except (ValueError, KeyError), err: + return self.trouble('ERROR: invalid output template or system charset: %s' % str(err)) + if self.params['nooverwrites'] and os.path.exists(filename): + self.to_stderr('WARNING: file exists: %s; skipping' % filename) + return 0 + try: + self.pmkdir(filename) + except (OSError, IOError), err: + return self.trouble('ERROR: unable to create directories: %s' % str(err)) + try: + outstream = open(filename, 'wb') + except (OSError, IOError), err: + return self.trouble('ERROR: unable to open for writing: %s' % str(err)) + try: + self._do_download(outstream, info_dict['url']) + outstream.close() + except (OSError, IOError), err: + return self.trouble('ERROR: unable to write video data: %s' % str(err)) + except (urllib2.URLError, httplib.HTTPException, socket.error), err: + return self.trouble('ERROR: unable to download video data: %s' % str(err)) + try: + self.post_process(filename, info_dict) + except (PostProcessingError), err: + return self.trouble('ERROR: postprocessing: %s' % str(err)) + + return 0 + def download(self, url_list): """Download a given list of URLs.""" retcode = 0 if len(url_list) > 1 and self.fixed_template(): - sys.exit('ERROR: fixed output name but more than one file to download') + raise SameFileError(self.params['outtmpl']) for url in url_list: suitable_found = False for ie in self._ies: + # Go to next InfoExtractor if not suitable if not ie.suitable(url): continue + # Suitable InfoExtractor found suitable_found = True + + # Extract information from URL all_results = ie.extract(url) results = [x for x in all_results if x is not None] + + # See if there were problems extracting any information if len(results) != len(all_results): retcode = self.trouble() + # Two results could go to the same file if len(results) > 1 and self.fixed_template(): - sys.exit('ERROR: fixed output name but more than one file to download') + raise SameFileError(self.params['outtmpl']) + # Process each result for result in results: + result = self.process_info(result) + + # Do not overwrite an error code with a success code + if result != 0: + retcode = result - # Forced printings - if self._params.get('forcetitle', False): - print result['title'] - if self._params.get('forceurl', False): - print result['url'] - - # Do nothing else if in simulate mode - if self._params.get('simulate', False): - continue - - try: - filename = self._params['outtmpl'] % result - except (ValueError, KeyError), err: - retcode = self.trouble('ERROR: invalid output template: %s' % str(err)) - continue - try: - self.pmkdir(filename) - except (OSError, IOError), err: - retcode = self.trouble('ERROR: unable to create directories: %s' % str(err)) - continue - try: - outstream = open(filename, 'wb') - except (OSError, IOError), err: - retcode = self.trouble('ERROR: unable to open for writing: %s' % str(err)) - continue - try: - self._do_download(outstream, result['url']) - outstream.close() - except (OSError, IOError), err: - retcode = self.trouble('ERROR: unable to write video data: %s' % str(err)) - continue - except (urllib2.URLError, httplib.HTTPException, socket.error), err: - retcode = self.trouble('ERROR: unable to download video data: %s' % str(err)) - continue + # Suitable InfoExtractor had been found; go to next URL break + if not suitable_found: retcode = self.trouble('ERROR: no suitable InfoExtractor: %s' % url) return retcode + + def post_process(self, filename, ie_info): + """Run the postprocessing chain on the given file.""" + info = dict(ie_info) + info['filepath'] = filename + for pp in self._pps: + info = pp.run(info) + if info is None: + break def _do_download(self, stream, url): request = urllib2.Request(url, None, std_headers) @@ -248,12 +343,13 @@ class FileDownloader(object): block_size = 1024 start = time.time() while True: + # Progress message percent_str = self.calc_percent(byte_counter, data_len) eta_str = self.calc_eta(start, time.time(), data_len, byte_counter) speed_str = self.calc_speed(start, time.time(), byte_counter) - self.to_stdout('\r[download] %s of %s at %s ETA %s' % - (percent_str, data_len_str, speed_str, eta_str), skip_eol=True) + self.report_progress(percent_str, data_len_str, speed_str, eta_str) + # Download and write before = time.time() data_block = data.read(block_size) after = time.time() @@ -264,7 +360,10 @@ class FileDownloader(object): stream.write(data_block) block_size = self.best_block_size(after - before, data_block_len) - self.to_stdout('') + # Apply rate limit + self.slow_down(start, byte_counter) + + self.report_finish() if data_len is not None and str(byte_counter) != data_len: raise ValueError('Content too short: %s/%s bytes' % (byte_counter, data_len)) @@ -303,7 +402,7 @@ class InfoExtractor(object): @staticmethod def suitable(url): """Receives a URL and returns True if suitable for this IE.""" - return True + return False def initialize(self): """Initializes an instance (authentication, etc).""" @@ -322,12 +421,12 @@ class InfoExtractor(object): def to_stdout(self, message): """Print message to stdout if downloader is not in quiet mode.""" - if self._downloader is None or not self._downloader.get_params().get('quiet', False): + if self._downloader is None or not self._downloader.params.get('quiet', False): print message def to_stderr(self, message): """Print message to stderr.""" - sys.stderr.write('%s\n' % message) + print >>sys.stderr, message def _real_initialize(self): """Real initialization process. Redefine in subclasses.""" @@ -340,17 +439,70 @@ class InfoExtractor(object): class YoutubeIE(InfoExtractor): """Information extractor for youtube.com.""" - _LOGIN_URL = 'http://www.youtube.com/login?next=/' - _AGE_URL = 'http://www.youtube.com/verify_age?next_url=/' + _VALID_URL = r'^((?:http://)?(?:\w+\.)?youtube\.com/(?:(?:v/)|(?:(?:watch(?:\.php)?)?\?(?:.+&)?v=)))?([0-9A-Za-z_-]+)(?(1).+)?$' + _LANG_URL = r'http://uk.youtube.com/?hl=en&persist_hl=1&gl=US&persist_gl=1&opt_out_ackd=1' + _LOGIN_URL = 'http://www.youtube.com/signup?next=/&gl=US&hl=en' + _AGE_URL = 'http://www.youtube.com/verify_age?next_url=/&gl=US&hl=en' _NETRC_MACHINE = 'youtube' + @staticmethod + def suitable(url): + return (re.match(YoutubeIE._VALID_URL, url) is not None) + + @staticmethod + def htmlentity_transform(matchobj): + """Transforms an HTML entity to a Unicode character.""" + entity = matchobj.group(1) + + # Known non-numeric HTML entity + if entity in htmlentitydefs.name2codepoint: + return unichr(htmlentitydefs.name2codepoint[entity]) + + # Unicode character + mobj = re.match(ur'(?u)#(x?\d+)', entity) + if mobj is not None: + numstr = mobj.group(1) + if numstr.startswith(u'x'): + base = 16 + numstr = u'0%s' % numstr + else: + base = 10 + return unichr(long(numstr, base)) + + # Unknown entity in name, return its literal representation + return (u'&%s;' % entity) + + def report_lang(self): + """Report attempt to set language.""" + self.to_stdout(u'[youtube] Setting language') + + def report_login(self): + """Report attempt to log in.""" + self.to_stdout(u'[youtube] Logging in') + + def report_age_confirmation(self): + """Report attempt to confirm age.""" + self.to_stdout(u'[youtube] Confirming age') + + def report_webpage_download(self, video_id): + """Report attempt to download webpage.""" + self.to_stdout(u'[youtube] %s: Downloading video webpage' % video_id) + + def report_information_extraction(self, video_id): + """Report attempt to extract video information.""" + self.to_stdout(u'[youtube] %s: Extracting video information' % video_id) + + def report_video_url(self, video_id, video_real_url): + """Report extracted video URL.""" + self.to_stdout(u'[youtube] %s: URL: %s' % (video_id, video_real_url)) + def _real_initialize(self): if self._downloader is None: return username = None password = None - downloader_params = self._downloader.get_params() + downloader_params = self._downloader.params # Attempt to use provided username and password or .netrc data if downloader_params.get('username', None) is not None: @@ -365,9 +517,19 @@ class YoutubeIE(InfoExtractor): else: raise netrc.NetrcParseError('No authenticators for %s' % self._NETRC_MACHINE) except (IOError, netrc.NetrcParseError), err: - self.to_stderr('WARNING: parsing .netrc: %s' % str(err)) + self.to_stderr(u'WARNING: parsing .netrc: %s' % str(err)) return + # Set language + request = urllib2.Request(self._LANG_URL, None, std_headers) + try: + self.report_lang() + urllib2.urlopen(request).read() + except (urllib2.URLError, httplib.HTTPException, socket.error), err: + self.to_stderr(u'WARNING: unable to set language: %s' % str(err)) + return + + # No authentication to be performed if username is None: return @@ -381,13 +543,13 @@ class YoutubeIE(InfoExtractor): } request = urllib2.Request(self._LOGIN_URL, urllib.urlencode(login_form), std_headers) try: - self.to_stdout('[youtube] Logging in') + self.report_login() login_results = urllib2.urlopen(request).read() if re.search(r'(?i)