Move GoogleSearchIE into its own file
This commit is contained in:
		
							parent
							
								
									9b122384e9
								
							
						
					
					
						commit
						3fc03845a1
					
				| @ -49,36 +49,6 @@ from .extractor.zdf import ZDFIE | |||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| class GoogleSearchIE(SearchInfoExtractor): |  | ||||||
|     """Information Extractor for Google Video search queries.""" |  | ||||||
|     _MORE_PAGES_INDICATOR = r'id="pnnext" class="pn"' |  | ||||||
|     _MAX_RESULTS = 1000 |  | ||||||
|     IE_NAME = u'video.google:search' |  | ||||||
|     _SEARCH_KEY = 'gvsearch' |  | ||||||
| 
 |  | ||||||
|     def _get_n_results(self, query, n): |  | ||||||
|         """Get a specified number of results for a query""" |  | ||||||
| 
 |  | ||||||
|         res = { |  | ||||||
|             '_type': 'playlist', |  | ||||||
|             'id': query, |  | ||||||
|             'entries': [] |  | ||||||
|         } |  | ||||||
| 
 |  | ||||||
|         for pagenum in itertools.count(1): |  | ||||||
|             result_url = u'http://www.google.com/search?tbm=vid&q=%s&start=%s&hl=en' % (compat_urllib_parse.quote_plus(query), pagenum*10) |  | ||||||
|             webpage = self._download_webpage(result_url, u'gvsearch:' + query, |  | ||||||
|                                              note='Downloading result page ' + str(pagenum)) |  | ||||||
| 
 |  | ||||||
|             for mobj in re.finditer(r'<h3 class="r"><a href="([^"]+)"', webpage): |  | ||||||
|                 e = { |  | ||||||
|                     '_type': 'url', |  | ||||||
|                     'url': mobj.group(1) |  | ||||||
|                 } |  | ||||||
|                 res['entries'].append(e) |  | ||||||
| 
 |  | ||||||
|             if (pagenum * 10 > n) or not re.search(self._MORE_PAGES_INDICATOR, webpage): |  | ||||||
|                 return res |  | ||||||
| 
 | 
 | ||||||
| class YahooSearchIE(SearchInfoExtractor): | class YahooSearchIE(SearchInfoExtractor): | ||||||
|     """Information Extractor for Yahoo! Video search queries.""" |     """Information Extractor for Yahoo! Video search queries.""" | ||||||
|  | |||||||
							
								
								
									
										39
									
								
								youtube_dl/extractor/google.py
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										39
									
								
								youtube_dl/extractor/google.py
									
									
									
									
									
										Normal file
									
								
							| @ -0,0 +1,39 @@ | |||||||
|  | import itertools | ||||||
|  | import re | ||||||
|  | 
 | ||||||
|  | from .common import SearchInfoExtractor | ||||||
|  | from ..utils import ( | ||||||
|  |     compat_urllib_parse, | ||||||
|  | ) | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | class GoogleSearchIE(SearchInfoExtractor): | ||||||
|  |     """Information Extractor for Google Video search queries.""" | ||||||
|  |     _MORE_PAGES_INDICATOR = r'id="pnnext" class="pn"' | ||||||
|  |     _MAX_RESULTS = 1000 | ||||||
|  |     IE_NAME = u'video.google:search' | ||||||
|  |     _SEARCH_KEY = 'gvsearch' | ||||||
|  | 
 | ||||||
|  |     def _get_n_results(self, query, n): | ||||||
|  |         """Get a specified number of results for a query""" | ||||||
|  | 
 | ||||||
|  |         res = { | ||||||
|  |             '_type': 'playlist', | ||||||
|  |             'id': query, | ||||||
|  |             'entries': [] | ||||||
|  |         } | ||||||
|  | 
 | ||||||
|  |         for pagenum in itertools.count(1): | ||||||
|  |             result_url = u'http://www.google.com/search?tbm=vid&q=%s&start=%s&hl=en' % (compat_urllib_parse.quote_plus(query), pagenum*10) | ||||||
|  |             webpage = self._download_webpage(result_url, u'gvsearch:' + query, | ||||||
|  |                                              note='Downloading result page ' + str(pagenum)) | ||||||
|  | 
 | ||||||
|  |             for mobj in re.finditer(r'<h3 class="r"><a href="([^"]+)"', webpage): | ||||||
|  |                 e = { | ||||||
|  |                     '_type': 'url', | ||||||
|  |                     'url': mobj.group(1) | ||||||
|  |                 } | ||||||
|  |                 res['entries'].append(e) | ||||||
|  | 
 | ||||||
|  |             if (pagenum * 10 > n) or not re.search(self._MORE_PAGES_INDICATOR, webpage): | ||||||
|  |                 return res | ||||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user