[airmozilla] Be more tolerant when nonessential items are missing (#5030)

This commit is contained in:
Philipp Hagemeister 2015-02-26 01:25:00 +01:00
parent cd5b4b0bc2
commit 3e675fabe0
3 changed files with 30 additions and 23 deletions

View File

@ -246,6 +246,7 @@ def test_parse_duration(self):
self.assertEqual(parse_duration('2.5 hours'), 9000) self.assertEqual(parse_duration('2.5 hours'), 9000)
self.assertEqual(parse_duration('02:03:04'), 7384) self.assertEqual(parse_duration('02:03:04'), 7384)
self.assertEqual(parse_duration('01:02:03:04'), 93784) self.assertEqual(parse_duration('01:02:03:04'), 93784)
self.assertEqual(parse_duration('1 hour 3 minutes'), 3780)
def test_fix_xml_ampersands(self): def test_fix_xml_ampersands(self):
self.assertEqual( self.assertEqual(

View File

@ -4,7 +4,11 @@
import re import re
from .common import InfoExtractor from .common import InfoExtractor
from ..utils import parse_iso8601 from ..utils import (
int_or_none,
parse_duration,
parse_iso8601,
)
class AirMozillaIE(InfoExtractor): class AirMozillaIE(InfoExtractor):
@ -27,13 +31,6 @@ class AirMozillaIE(InfoExtractor):
} }
} }
_QUALITY_MAP = {
'360p': 0,
'576p': 1,
'640p': 2,
'720p': 3,
}
def _real_extract(self, url): def _real_extract(self, url):
display_id = self._match_id(url) display_id = self._match_id(url)
webpage = self._download_webpage(url, display_id) webpage = self._download_webpage(url, display_id)
@ -43,19 +40,23 @@ def _real_extract(self, url):
jwconfig = self._search_regex(r'\svar jwconfig = (\{.*?\});\s', embed_script, 'metadata') jwconfig = self._search_regex(r'\svar jwconfig = (\{.*?\});\s', embed_script, 'metadata')
metadata = self._parse_json(jwconfig, video_id) metadata = self._parse_json(jwconfig, video_id)
formats = [] formats = [{
for source in metadata['playlist'][0]['sources']: 'url': source['file'],
fmt = { 'ext': source['type'],
'url': source['file'], 'format_id': self._search_regex(r'&format=(.*)$', source['file'], 'video format'),
'ext': source['type'], 'format': source['label'],
'format_id': self._search_regex(r'&format=(.*)$', source['file'], 'video format'), 'height': int(source['label'].rstrip('p')),
'resolution': source['label'], } for source in metadata['playlist'][0]['sources']]
'quality': self._QUALITY_MAP.get(source['label'], -1),
}
formats.append(fmt)
self._sort_formats(formats) self._sort_formats(formats)
duration_match = re.search(r'Duration:(?: (?P<H>\d+) hours?)?(?: (?P<M>\d+) minutes?)?', webpage) view_count = int_or_none(self._html_search_regex(
r'Views since archived: ([0-9]+)',
webpage, 'view count', fatal=False))
timestamp = parse_iso8601(self._html_search_regex(
r'<time datetime="(.*?)"', webpage, 'timestamp', fatal=False))
duration = parse_duration(self._search_regex(
r'Duration:\s*(\d+\s*hours?\s*\d+\s*minutes?)',
webpage, 'duration', fatal=False))
return { return {
'id': video_id, 'id': video_id,
@ -63,11 +64,11 @@ def _real_extract(self, url):
'formats': formats, 'formats': formats,
'url': self._og_search_url(webpage), 'url': self._og_search_url(webpage),
'display_id': display_id, 'display_id': display_id,
'thumbnail': metadata['playlist'][0]['image'], 'thumbnail': metadata['playlist'][0].get('image'),
'description': self._og_search_description(webpage), 'description': self._og_search_description(webpage),
'timestamp': parse_iso8601(self._html_search_regex(r'<time datetime="(.*?)"', webpage, 'timestamp')), 'timestamp': timestamp,
'location': self._html_search_regex(r'Location: (.*)', webpage, 'location', default=None), 'location': self._html_search_regex(r'Location: (.*)', webpage, 'location', default=None),
'duration': int(duration_match.groupdict()['H'] or 0) * 3600 + int(duration_match.groupdict()['M'] or 0) * 60, 'duration': duration,
'view_count': int(self._html_search_regex(r'Views since archived: ([0-9]+)', webpage, 'view count')), 'view_count': view_count,
'categories': re.findall(r'<a href=".*?" class="channel">(.*?)</a>', webpage), 'categories': re.findall(r'<a href=".*?" class="channel">(.*?)</a>', webpage),
} }

View File

@ -1290,6 +1290,7 @@ def parse_duration(s):
(?P<only_mins>[0-9.]+)\s*(?:mins?|minutes?)\s*| (?P<only_mins>[0-9.]+)\s*(?:mins?|minutes?)\s*|
(?P<only_hours>[0-9.]+)\s*(?:hours?)| (?P<only_hours>[0-9.]+)\s*(?:hours?)|
\s*(?P<hours_reversed>[0-9]+)\s*(?:[:h]|hours?)\s*(?P<mins_reversed>[0-9]+)\s*(?:[:m]|mins?|minutes?)\s*|
(?: (?:
(?: (?:
(?:(?P<days>[0-9]+)\s*(?:[:d]|days?)\s*)? (?:(?P<days>[0-9]+)\s*(?:[:d]|days?)\s*)?
@ -1308,10 +1309,14 @@ def parse_duration(s):
return float_or_none(m.group('only_hours'), invscale=60 * 60) return float_or_none(m.group('only_hours'), invscale=60 * 60)
if m.group('secs'): if m.group('secs'):
res += int(m.group('secs')) res += int(m.group('secs'))
if m.group('mins_reversed'):
res += int(m.group('mins_reversed')) * 60
if m.group('mins'): if m.group('mins'):
res += int(m.group('mins')) * 60 res += int(m.group('mins')) * 60
if m.group('hours'): if m.group('hours'):
res += int(m.group('hours')) * 60 * 60 res += int(m.group('hours')) * 60 * 60
if m.group('hours_reversed'):
res += int(m.group('hours_reversed')) * 60 * 60
if m.group('days'): if m.group('days'):
res += int(m.group('days')) * 24 * 60 * 60 res += int(m.group('days')) * 24 * 60 * 60
if m.group('ms'): if m.group('ms'):