Merge branch 'dramafever' of https://github.com/ping/youtube-dl into ping-dramafever
commit
ad49fe7c8f
|
@ -112,6 +112,10 @@ from .dfb import DFBIE
|
||||||
from .dhm import DHMIE
|
from .dhm import DHMIE
|
||||||
from .dotsub import DotsubIE
|
from .dotsub import DotsubIE
|
||||||
from .douyutv import DouyuTVIE
|
from .douyutv import DouyuTVIE
|
||||||
|
from .dramafever import (
|
||||||
|
DramaFeverIE,
|
||||||
|
DramaFeverSeriesIE,
|
||||||
|
)
|
||||||
from .dreisat import DreiSatIE
|
from .dreisat import DreiSatIE
|
||||||
from .drbonanza import DRBonanzaIE
|
from .drbonanza import DRBonanzaIE
|
||||||
from .drtuber import DrTuberIE
|
from .drtuber import DrTuberIE
|
||||||
|
|
|
@ -0,0 +1,130 @@
|
||||||
|
# encoding: utf-8
|
||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
import re
|
||||||
|
|
||||||
|
from .common import InfoExtractor
|
||||||
|
|
||||||
|
|
||||||
|
class DramaFeverIE(InfoExtractor):
|
||||||
|
IE_NAME = 'dramafever'
|
||||||
|
_VALID_URL = r'^https?://(?:www\.)?dramafever\.com/drama/(?P<id>[0-9]+/[0-9]+)/'
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'http://www.dramafever.com/drama/4512/1/Cooking_with_Shin/',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '4512.1',
|
||||||
|
'ext': 'flv',
|
||||||
|
'title': 'Cooking with Shin 4512.1',
|
||||||
|
'upload_date': '20140702',
|
||||||
|
'description': 'Served at all special occasions and featured in the hit drama Heirs, Shin cooks Red Bean Rice.',
|
||||||
|
}
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
video_id = self._match_id(url).replace("/", ".")
|
||||||
|
|
||||||
|
consumer_secret = self._get_consumer_secret(video_id)
|
||||||
|
|
||||||
|
ep_json = self._download_json(
|
||||||
|
"http://www.dramafever.com/amp/episode/feed.json?guid=%s" % video_id,
|
||||||
|
video_id, note='Downloading episode metadata',
|
||||||
|
errnote="Video may not be available for your location")["channel"]["item"]
|
||||||
|
|
||||||
|
title = ep_json["media-group"]["media-title"]
|
||||||
|
description = ep_json["media-group"]["media-description"]
|
||||||
|
thumbnail = ep_json["media-group"]["media-thumbnail"]["@attributes"]["url"]
|
||||||
|
duration = int(ep_json["media-group"]["media-content"][0]["@attributes"]["duration"])
|
||||||
|
mobj = re.match(r"([0-9]{4})-([0-9]{2})-([0-9]{2})", ep_json["pubDate"])
|
||||||
|
upload_date = mobj.group(1) + mobj.group(2) + mobj.group(3) if mobj is not None else None
|
||||||
|
|
||||||
|
formats = []
|
||||||
|
for vid_format in ep_json["media-group"]["media-content"]:
|
||||||
|
src = vid_format["@attributes"]["url"]
|
||||||
|
if '.f4m' in src:
|
||||||
|
formats.extend(self._extract_f4m_formats(src, video_id))
|
||||||
|
|
||||||
|
self._sort_formats(formats)
|
||||||
|
video_subtitles = self.extract_subtitles(video_id, consumer_secret)
|
||||||
|
|
||||||
|
return {
|
||||||
|
'id': video_id,
|
||||||
|
'title': title,
|
||||||
|
'description': description,
|
||||||
|
'thumbnail': thumbnail,
|
||||||
|
'upload_date': upload_date,
|
||||||
|
'duration': duration,
|
||||||
|
'formats': formats,
|
||||||
|
'subtitles': video_subtitles,
|
||||||
|
}
|
||||||
|
|
||||||
|
def _get_consumer_secret(self, video_id):
|
||||||
|
df_js = self._download_webpage(
|
||||||
|
"http://www.dramafever.com/static/126960d/v2/js/plugins/jquery.threadedcomments.js", video_id)
|
||||||
|
return self._search_regex(r"'cs': '([0-9a-zA-Z]+)'", df_js, "cs")
|
||||||
|
|
||||||
|
def _get_episodes(self, series_id, consumer_secret, episode_filter=None):
|
||||||
|
_PAGE_SIZE = 60
|
||||||
|
|
||||||
|
curr_page = 1
|
||||||
|
max_pages = curr_page + 1
|
||||||
|
results = []
|
||||||
|
while max_pages >= curr_page:
|
||||||
|
page_url = "http://www.dramafever.com/api/4/episode/series/?cs=%s&series_id=%s&page_size=%d&page_number=%d" % \
|
||||||
|
(consumer_secret, series_id, _PAGE_SIZE, curr_page)
|
||||||
|
series = self._download_json(
|
||||||
|
page_url, series_id, note="Downloading series json page #%d" % curr_page)
|
||||||
|
max_pages = series['num_pages']
|
||||||
|
results.extend([ep for ep in series['value'] if episode_filter is None or episode_filter(ep)])
|
||||||
|
curr_page += 1
|
||||||
|
return results
|
||||||
|
|
||||||
|
def _get_subtitles(self, video_id, consumer_secret):
|
||||||
|
|
||||||
|
res = None
|
||||||
|
info = self._get_episodes(
|
||||||
|
video_id.split(".")[0], consumer_secret,
|
||||||
|
episode_filter=lambda x: x['guid'] == video_id)
|
||||||
|
|
||||||
|
if len(info) == 1 and info[0]['subfile'] != '':
|
||||||
|
res = {'en': [{'url': info[0]['subfile'], 'ext': 'srt'}]}
|
||||||
|
return res
|
||||||
|
|
||||||
|
|
||||||
|
class DramaFeverSeriesIE(DramaFeverIE):
|
||||||
|
IE_NAME = 'dramafever:series'
|
||||||
|
_VALID_URL = r'^https?://(?:www\.)?dramafever\.com/drama/(?P<id>[0-9]+)/\d*[a-zA-Z_][a-zA-Z0-9_]*/'
|
||||||
|
_TESTS = [{
|
||||||
|
'url': 'http://www.dramafever.com/drama/4512/Cooking_with_Shin/',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '4512',
|
||||||
|
'title': 'Cooking with Shin',
|
||||||
|
'description': 'Professional chef and cooking instructor Shin Kim takes some of the delicious dishes featured in your favorite dramas and shows you how to make them right at home.',
|
||||||
|
},
|
||||||
|
'playlist_count': 4,
|
||||||
|
}, {
|
||||||
|
'url': 'http://www.dramafever.com/drama/124/IRIS/',
|
||||||
|
'info_dict': {
|
||||||
|
'id': '124',
|
||||||
|
'title': 'IRIS',
|
||||||
|
'description': 'Lee Byung Hun and Kim Tae Hee star in this powerhouse drama and ratings megahit of action, intrigue and romance.',
|
||||||
|
},
|
||||||
|
'playlist_count': 20,
|
||||||
|
}]
|
||||||
|
|
||||||
|
def _real_extract(self, url):
|
||||||
|
series_id = self._match_id(url)
|
||||||
|
consumer_secret = self._get_consumer_secret(series_id)
|
||||||
|
|
||||||
|
series_json = self._download_json(
|
||||||
|
"http://www.dramafever.com/api/4/series/query/?cs=%s&series_id=%s" % (consumer_secret, series_id),
|
||||||
|
series_id, note='Downloading series metadata')["series"][series_id]
|
||||||
|
|
||||||
|
title = series_json["name"]
|
||||||
|
description = series_json["description_short"]
|
||||||
|
|
||||||
|
episodes = self._get_episodes(series_id, consumer_secret)
|
||||||
|
entries = []
|
||||||
|
for ep in episodes:
|
||||||
|
entries.append(self.url_result(
|
||||||
|
'http://www.dramafever.com%s' % ep['episode_url'], 'DramaFever', ep['guid']))
|
||||||
|
return self.playlist_result(entries, series_id, title, description)
|
Loading…
Reference in New Issue