mirror of
https://codeberg.org/polarisfm/youtube-dl
synced 2024-11-05 01:44:33 +01:00
116 lines
4.3 KiB
Python
116 lines
4.3 KiB
Python
# coding: utf-8
|
|
from __future__ import unicode_literals
|
|
|
|
import re
|
|
import time
|
|
import hashlib
|
|
|
|
from .common import InfoExtractor
|
|
from ..utils import (
|
|
ExtractorError,
|
|
unified_strdate,
|
|
)
|
|
|
|
|
|
class WatIE(InfoExtractor):
|
|
_VALID_URL = r'http://www\.wat\.tv/video/(?P<display_id>.*)-(?P<short_id>.*?)_.*?\.html'
|
|
IE_NAME = 'wat.tv'
|
|
_TEST = {
|
|
'url': 'http://www.wat.tv/video/soupe-figues-l-orange-aux-epices-6z1uz_2hvf7_.html',
|
|
'md5': 'ce70e9223945ed26a8056d413ca55dc9',
|
|
'info_dict': {
|
|
'id': '11713067',
|
|
'display_id': 'soupe-figues-l-orange-aux-epices',
|
|
'ext': 'mp4',
|
|
'title': 'Soupe de figues à l\'orange et aux épices',
|
|
'description': 'Retrouvez l\'émission "Petits plats en équilibre", diffusée le 18 août 2014.',
|
|
'upload_date': '20140819',
|
|
'duration': 120,
|
|
},
|
|
}
|
|
|
|
def download_video_info(self, real_id):
|
|
# 'contentv4' is used in the website, but it also returns the related
|
|
# videos, we don't need them
|
|
info = self._download_json('http://www.wat.tv/interface/contentv3/' + real_id, real_id)
|
|
return info['media']
|
|
|
|
def _real_extract(self, url):
|
|
def real_id_for_chapter(chapter):
|
|
return chapter['tc_start'].split('-')[0]
|
|
mobj = re.match(self._VALID_URL, url)
|
|
short_id = mobj.group('short_id')
|
|
display_id = mobj.group('display_id')
|
|
webpage = self._download_webpage(url, display_id or short_id)
|
|
real_id = self._search_regex(r'xtpage = ".*-(.*?)";', webpage, 'real id')
|
|
|
|
video_info = self.download_video_info(real_id)
|
|
|
|
if video_info.get('geolock'):
|
|
raise ExtractorError('This content is not available in your area', expected=True)
|
|
|
|
chapters = video_info['chapters']
|
|
first_chapter = chapters[0]
|
|
files = video_info['files']
|
|
first_file = files[0]
|
|
|
|
if real_id_for_chapter(first_chapter) != real_id:
|
|
self.to_screen('Multipart video detected')
|
|
chapter_urls = []
|
|
for chapter in chapters:
|
|
chapter_id = real_id_for_chapter(chapter)
|
|
# Yes, when we this chapter is processed by WatIE,
|
|
# it will download the info again
|
|
chapter_info = self.download_video_info(chapter_id)
|
|
chapter_urls.append(chapter_info['url'])
|
|
entries = [self.url_result(chapter_url) for chapter_url in chapter_urls]
|
|
return self.playlist_result(entries, real_id, video_info['title'])
|
|
|
|
upload_date = None
|
|
if 'date_diffusion' in first_chapter:
|
|
upload_date = unified_strdate(first_chapter['date_diffusion'])
|
|
# Otherwise we can continue and extract just one part, we have to use
|
|
# the short id for getting the video url
|
|
|
|
formats = [{
|
|
'url': 'http://wat.tv/get/android5/%s.mp4' % real_id,
|
|
'format_id': 'Mobile',
|
|
}]
|
|
|
|
fmts = [('SD', 'web')]
|
|
if first_file.get('hasHD'):
|
|
fmts.append(('HD', 'webhd'))
|
|
|
|
def compute_token(param):
|
|
timestamp = '%08x' % int(time.time())
|
|
magic = '9b673b13fa4682ed14c3cfa5af5310274b514c4133e9b3a81e6e3aba009l2564'
|
|
return '%s/%s' % (hashlib.md5((magic + param + timestamp).encode('ascii')).hexdigest(), timestamp)
|
|
|
|
for fmt in fmts:
|
|
webid = '/%s/%s' % (fmt[1], real_id)
|
|
video_url = self._download_webpage(
|
|
'http://www.wat.tv/get%s?token=%s&getURL=1' % (webid, compute_token(webid)),
|
|
real_id,
|
|
'Downloding %s video URL' % fmt[0],
|
|
'Failed to download %s video URL' % fmt[0],
|
|
False)
|
|
if not video_url:
|
|
continue
|
|
formats.append({
|
|
'url': video_url,
|
|
'ext': 'mp4',
|
|
'format_id': fmt[0],
|
|
})
|
|
|
|
return {
|
|
'id': real_id,
|
|
'display_id': display_id,
|
|
'title': first_chapter['title'],
|
|
'thumbnail': first_chapter['preview'],
|
|
'description': first_chapter['description'],
|
|
'view_count': video_info['views'],
|
|
'upload_date': upload_date,
|
|
'duration': first_file['duration'],
|
|
'formats': formats,
|
|
}
|