Fix GoogleSearchIE (Fixes #822)
This commit is contained in:
		@@ -1494,82 +1494,53 @@ class YoutubeSearchIE(InfoExtractor):
 | 
			
		||||
 | 
			
		||||
class GoogleSearchIE(InfoExtractor):
 | 
			
		||||
    """Information Extractor for Google Video search queries."""
 | 
			
		||||
    _VALID_URL = r'gvsearch(\d+|all)?:[\s\S]+'
 | 
			
		||||
    _TEMPLATE_URL = 'http://video.google.com/videosearch?q=%s+site:video.google.com&start=%s&hl=en'
 | 
			
		||||
    _VIDEO_INDICATOR = r'<a href="http://video\.google\.com/videoplay\?docid=([^"\&]+)'
 | 
			
		||||
    _VALID_URL = r'gvsearch(?P<prefix>|\d+|all):(?P<query>[\s\S]+)'
 | 
			
		||||
    _MORE_PAGES_INDICATOR = r'class="pn" id="pnnext"'
 | 
			
		||||
    _max_google_results = 1000
 | 
			
		||||
    IE_NAME = u'video.google:search'
 | 
			
		||||
 | 
			
		||||
    def report_download_page(self, query, pagenum):
 | 
			
		||||
        """Report attempt to download playlist page with given number."""
 | 
			
		||||
        query = query.decode(preferredencoding())
 | 
			
		||||
        self.to_screen(u'query "%s": Downloading page %s' % (query, pagenum))
 | 
			
		||||
 | 
			
		||||
    def _real_extract(self, query):
 | 
			
		||||
        mobj = re.match(self._VALID_URL, query)
 | 
			
		||||
        if mobj is None:
 | 
			
		||||
            self._downloader.report_error(u'invalid search query "%s"' % query)
 | 
			
		||||
            return
 | 
			
		||||
 | 
			
		||||
        prefix, query = query.split(':')
 | 
			
		||||
        prefix = prefix[8:]
 | 
			
		||||
        query = query.encode('utf-8')
 | 
			
		||||
        prefix = mobj.group('prefix')
 | 
			
		||||
        query = mobj.group('query')
 | 
			
		||||
        if prefix == '':
 | 
			
		||||
            self._download_n_results(query, 1)
 | 
			
		||||
            return
 | 
			
		||||
            return self._download_n_results(query, 1)
 | 
			
		||||
        elif prefix == 'all':
 | 
			
		||||
            self._download_n_results(query, self._max_google_results)
 | 
			
		||||
            return
 | 
			
		||||
            return self._download_n_results(query, self._max_google_results)
 | 
			
		||||
        else:
 | 
			
		||||
            try:
 | 
			
		||||
                n = int(prefix)
 | 
			
		||||
                if n <= 0:
 | 
			
		||||
                    self._downloader.report_error(u'invalid download number %s for query "%s"' % (n, query))
 | 
			
		||||
                    return
 | 
			
		||||
                elif n > self._max_google_results:
 | 
			
		||||
                    self._downloader.report_warning(u'gvsearch returns max %i results (you requested %i)' % (self._max_google_results, n))
 | 
			
		||||
                    n = self._max_google_results
 | 
			
		||||
                self._download_n_results(query, n)
 | 
			
		||||
                return
 | 
			
		||||
            except ValueError: # parsing prefix as integer fails
 | 
			
		||||
                self._download_n_results(query, 1)
 | 
			
		||||
                return
 | 
			
		||||
            n = int(prefix)
 | 
			
		||||
            if n <= 0:
 | 
			
		||||
                raise ExtractorError(u'invalid download number %s for query "%s"' % (n, query))
 | 
			
		||||
            elif n > self._max_google_results:
 | 
			
		||||
                self._downloader.report_warning(u'gvsearch returns max %i results (you requested %i)' % (self._max_google_results, n))
 | 
			
		||||
                n = self._max_google_results
 | 
			
		||||
            return self._download_n_results(query, n)
 | 
			
		||||
 | 
			
		||||
    def _download_n_results(self, query, n):
 | 
			
		||||
        """Downloads a specified number of results for a query"""
 | 
			
		||||
 | 
			
		||||
        video_ids = []
 | 
			
		||||
        pagenum = 0
 | 
			
		||||
        res = {
 | 
			
		||||
            '_type': 'playlist',
 | 
			
		||||
            'id': query,
 | 
			
		||||
            'entries': []
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
        while True:
 | 
			
		||||
            self.report_download_page(query, pagenum)
 | 
			
		||||
            result_url = self._TEMPLATE_URL % (compat_urllib_parse.quote_plus(query), pagenum*10)
 | 
			
		||||
            request = compat_urllib_request.Request(result_url)
 | 
			
		||||
            try:
 | 
			
		||||
                page = compat_urllib_request.urlopen(request).read()
 | 
			
		||||
            except (compat_urllib_error.URLError, compat_http_client.HTTPException, socket.error) as err:
 | 
			
		||||
                self._downloader.report_error(u'unable to download webpage: %s' % compat_str(err))
 | 
			
		||||
                return
 | 
			
		||||
        for pagenum in itertools.count(1):
 | 
			
		||||
            result_url = u'http://video.google.com/videosearch?q=%s&start=%s&hl=en' % (compat_urllib_parse.quote_plus(query), pagenum*10)
 | 
			
		||||
            webpage = self._download_webpage(result_url, u'gvsearch:' + query,
 | 
			
		||||
                                             note='Downloading result page ' + str(pagenum))
 | 
			
		||||
 | 
			
		||||
            # Extract video identifiers
 | 
			
		||||
            for mobj in re.finditer(self._VIDEO_INDICATOR, page):
 | 
			
		||||
                video_id = mobj.group(1)
 | 
			
		||||
                if video_id not in video_ids:
 | 
			
		||||
                    video_ids.append(video_id)
 | 
			
		||||
                    if len(video_ids) == n:
 | 
			
		||||
                        # Specified n videos reached
 | 
			
		||||
                        for id in video_ids:
 | 
			
		||||
                            self._downloader.download(['http://video.google.com/videoplay?docid=%s' % id])
 | 
			
		||||
                        return
 | 
			
		||||
 | 
			
		||||
            if re.search(self._MORE_PAGES_INDICATOR, page) is None:
 | 
			
		||||
                for id in video_ids:
 | 
			
		||||
                    self._downloader.download(['http://video.google.com/videoplay?docid=%s' % id])
 | 
			
		||||
                return
 | 
			
		||||
 | 
			
		||||
            pagenum = pagenum + 1
 | 
			
		||||
            for mobj in re.finditer(r'<h3 class="r"><a href="([^"]+)"', webpage):
 | 
			
		||||
                e = {
 | 
			
		||||
                    '_type': 'url',
 | 
			
		||||
                    'url': mobj.group(1)
 | 
			
		||||
                }
 | 
			
		||||
                res['entries'].append(e)
 | 
			
		||||
 | 
			
		||||
            if (pagenum * 10 > n) or not re.search(self._MORE_PAGES_INDICATOR, webpage):
 | 
			
		||||
                return res
 | 
			
		||||
 | 
			
		||||
class YahooSearchIE(InfoExtractor):
 | 
			
		||||
    """Information Extractor for Yahoo! Video search queries."""
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user