1
0
mirror of https://github.com/ytdl-org/youtube-dl.git synced 2024-12-21 23:46:48 +00:00
This commit is contained in:
dirkf 2024-12-20 17:46:27 +01:00 committed by GitHub
commit cb168bac1e
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
2 changed files with 238 additions and 0 deletions

View File

@ -494,6 +494,11 @@ from .hotstar import (
)
from .howcast import HowcastIE
from .howstuffworks import HowStuffWorksIE
from .hqporner import (
HQPornerIE,
HQPornerListIE,
HQPornerSearchIE,
)
from .hrfernsehen import HRFernsehenIE
from .hrti import (
HRTiIE,

View File

@ -0,0 +1,233 @@
# coding: utf-8
from __future__ import unicode_literals
import itertools
import re
from string import capwords
from .common import InfoExtractor
from ..utils import (
clean_html,
date_from_str,
get_element_by_class,
get_element_by_id,
int_or_none,
join_nonempty,
parse_duration,
strip_or_none,
unified_strdate,
urljoin,
)
class HQPornerIE(InfoExtractor):
_VALID_URL = r'https?://(?:www\.)?hqporner\.com/hdporn/(?P<id>[\d]+)-'
_TESTS = [{
'url': 'https://hqporner.com/hdporn/110374-looking_for_a_change_of_pace.html',
'md5': '7eb7b791a1ce8a619bde603b2dc334b5',
'info_dict': {
'id': '110374',
'ext': 'mp4',
'title': 'Looking For A Change Of Pace',
'description': 'featuring Myra',
'upload_date': '20230227',
'age_limit': 18,
'tags': list,
'categories': list,
'duration': 3271,
'thumbnail': r're:https?://.*\.jpg',
},
}, {
'url': 'https://hqporner.com/hdporn/86482-all_night_rager.html',
'only_matching': True,
}]
def _real_extract(self, url):
video_id = self._match_id(url)
# some pages need a `referer` to avoid 404
webpage = self._download_webpage(url, video_id, headers={'referer': 'https://hqporner.com/?q=porn'})
# details below video are in a <header> element
heading = self._search_regex(r'''(?s)(<header>.+?</section>)\s*</div>''', webpage, 'heading', default='')
title = (
capwords(clean_html(get_element_by_class('main-h1', heading) or ''))
or self._html_search_regex(
r'<title\b[^>]*>\s*([^<]+)(?:\s+-\s+HQporner\.com)?\s*</title>',
webpage, 'title'))
# video details are in a page loaded by this <iframe>
player = get_element_by_id('playerWrapper', webpage)
player = self._search_regex(
r'''<iframe [^>]*\bsrc\s*=\s*('|")(?P<url>(?:(?!\1)\S)+)''',
player, 'player iframe', group='url')
player = self._download_webpage(urljoin(url, player), video_id, note='Downloading player iframe')
# select the complete set of videos
player = self._search_regex(
r'''\belse\s*\{\s*\$\s*\(\s*('|")#jw\1\s*\)\s*\.\s*html\s*\(\s*(?P<vstr>("|')\s*<video [^>]+>[^)]+</video>\s*\3)''',
player, 'video element', group='vstr')
# it's a string containing HTML5 video
info = self._parse_html5_media_entries(
url, self._parse_json(player, video_id), video_id)[0] # not used now: m3u8_id='hls', m3u8_entry_protocol='m3u8_native', mpd_id='dash')
# site offers no timestamp, but may have "however-many date-units ago"
upload_date = get_element_by_class('fa-calendar', heading) or ''
if upload_date.endswith(' ago'):
upload_date = date_from_str('now-' + upload_date[:-len(' ago')].replace(' ', '').replace('one', '1'))
if upload_date:
upload_date = upload_date.isoformat().replace('-', '')
else:
# probably ineffective
upload_date = unified_strdate(upload_date)
# utils candidate
def tag_list(s, delim=','):
return [t for t in map(strip_or_none, s.split(delim)) if t] or None
desc = self._html_search_meta('description', webpage, default='')
info.update({
'id': video_id,
'title': title,
'age_limit': 18,
'upload_date': upload_date,
'description': clean_html(get_element_by_class('icon fa-star-o', heading)),
'duration': parse_duration(
get_element_by_class('fa-clock-o', heading)
or self._search_regex(r'Video duration is\s+([^.]+)', desc, 'duration', default='')),
'categories': tag_list(
clean_html(self._search_regex(r'(?s)</h3>\s*<p>(.+?)</p>', heading, 'categories', default='').replace('</a>', ','))
or self._html_search_meta('keywords', webpage, default='')),
'tags': tag_list(self._search_regex(r'Tags [\w\s-]+:\s+([^.]+)', desc, 'tags', default='')),
})
return info
class HQPornerListBaseIE(InfoExtractor):
# yt-dlp shim
@classmethod
def _match_valid_url(cls, url):
return re.match(cls._VALID_URL, url)
def _real_extract(self, url):
pl_id, pg = self._match_valid_url(url).group('id', 'pg')
pg = int_or_none(pg)
def entries():
"""
Generate the playlist
If `pg` is a page number, get the list for that page
Otherwise continue from `url` until exhausted
"""
next = url
for pnum in itertools.count(start=pg or 1):
p_url = urljoin(url, next)
first = (pg or 1) == pnum
if not p_url and not first:
break
page = self._download_webpage(
p_url, pl_id,
note='Downloading page' + ('' if first else (' %d' % pnum)),
fatal=first)
if not page:
break
for m in re.finditer(
# each entry in a playlist page has a hyperlinked <img> followed by the caption:
# <h3> containing a hyperlinked title, followed by a <span> containing duration
r'''(?sx)
# <img> with thumbnail (extracted), then closing containing <a><div>
(?:<img\b[^>]+\bsrc\s*=\s*("|')(?P<thm>(?:(?!\1).)+)\1[^>]*>.*?)?
# caption: extract href and title
<h3\b[^>]+\bclass\s*=\s*("|')(?:(?:(?!\3).)+\s)*?meta-data-title(?:\s(?:(?!\3).)+)*\3[^>]*>\s*
<a\b[^>]+\bhref\b=("|')(?P<url>(?:(?!\4).)+)\4[^>]*>(?P<ttl>[^<]*)</a>\s*
</h3>\s*
# extract duration
(?:<span\b[^>]+\bclass\s*=\s*("|')(?:(?:(?!\7).)+\s)*?fa-clock-o(?:\s(?:(?!\7).)+)*\7[^>]*>(?P<dur>[^<]+)</span>)?
''', page):
a_url = self._proto_relative_url(urljoin(url, m.group('url')))
if a_url:
res = self.url_result(a_url, video_title=capwords(clean_html(m.group('ttl')) or '') or None)
res.update({
'duration': parse_duration(m.group('dur')),
'thumbnail': self._proto_relative_url(urljoin(a_url, m.group('thm'))),
})
yield res
if pg is not None:
break
# next is the last link in the pagination block, unless that's the current URL
maybe_next = urljoin(url, self._search_regex(
r'''(?s)<a\b[^>]+\bhref\s*=\s*("|')(?P<url>(?:(?!\1).)+)\1[^>]*>[^>]*</a>\s*</li>\s*$''',
get_element_by_class('pagination', page) or '', 'next page',
group='url', default=None))
next = maybe_next if maybe_next != next else None
return self.playlist_result(entries(), playlist_id=join_nonempty(pl_id, pg, delim=':'))
@staticmethod
def _set_title(info):
pl_id, _, pg = info['id'].rpartition(':')
if not pl_id:
pl_id = pg
pg = None
info['title'] = pl_id.replace('/', ': ')
if pg:
info['title'] += ' [%s]' % (pg, )
class HQPornerListIE(HQPornerListBaseIE):
_VALID_URL = r'https?://(?:www\.)?hqporner\.com/(?P<id>(?:top|(?:category|actress)/[^/]+))(?:/(?P<pg>\d+))?'
_TESTS = [{
'url': 'https://hqporner.com/category/beach-porn',
'info_dict': {
'id': 'category/beach-porn',
'title': 'Category: beach-porn',
},
'playlist_mincount': 250,
}, {
'url': 'https://hqporner.com/category/beach-porn/2',
'info_dict': {
'id': 'category/beach-porn:2',
'title': 'Category: beach-porn [2]',
},
'playlist_count': 50,
}, {
'url': 'https://hqporner.com/actress/mary/1',
'info_dict': {
'id': 'actress/mary:1',
'title': 'Actress: mary [1]',
},
'playlist_count': 50,
},
]
def _real_extract(self, url):
res = super(HQPornerListIE, self)._real_extract(url)
self._set_title(res)
res['title'] = res['title'].capitalize()
return res
class HQPornerSearchIE(HQPornerListBaseIE):
_VALID_URL = r'https?://(?:www\.)?hqporner\.com/\?q=(?P<id>[^&]+)(?:&p=(?P<pg>\d+))?'
_TESTS = [{
'url': 'https://hqporner.com/?q=french',
'info_dict': {
'id': 'french',
'title': 'Searching: french',
},
'playlist_mincount': 250,
}, {
'url': 'https://hqporner.com/?q=french&p=2',
'info_dict': {
'id': 'french:2',
'title': 'Searching: french [2]',
},
'playlist_count': 50,
},
]
def _real_extract(self, url):
res = super(HQPornerSearchIE, self)._real_extract(url)
self._set_title(res)
res['title'] = 'Searching: ' + res['title']
return res