mirror of
https://github.com/simon987/chan_feed.git
synced 2025-04-18 01:36:44 +00:00
add plus4chan
This commit is contained in:
parent
82578daecd
commit
ee666496e0
13
chan/chan.py
13
chan/chan.py
@ -14,6 +14,7 @@ from chan.kev4_php import Kev4PhpHelper
|
|||||||
from chan.lolnada_html import LolNadaHtmlChanHelper
|
from chan.lolnada_html import LolNadaHtmlChanHelper
|
||||||
from chan.mayuri import MayuriChanHelper
|
from chan.mayuri import MayuriChanHelper
|
||||||
from chan.nowere_html import NowereHtmlChanHelper
|
from chan.nowere_html import NowereHtmlChanHelper
|
||||||
|
from chan.plus4chan_html import Plus4ChanHelper
|
||||||
from chan.russian_json import RussianJsonChanHelper
|
from chan.russian_json import RussianJsonChanHelper
|
||||||
from chan.synch_json import SynchJsonChanHelper
|
from chan.synch_json import SynchJsonChanHelper
|
||||||
from chan.tgchan_html import TgChanHtmlChanHelper
|
from chan.tgchan_html import TgChanHtmlChanHelper
|
||||||
@ -486,4 +487,16 @@ CHANS = {
|
|||||||
),
|
),
|
||||||
rps=1 / 20
|
rps=1 / 20
|
||||||
),
|
),
|
||||||
|
"plus4chan": Plus4ChanHelper(
|
||||||
|
33,
|
||||||
|
"https://boards.plus4chan.org/",
|
||||||
|
"https://boards.plus4chan.org/",
|
||||||
|
"",
|
||||||
|
"",
|
||||||
|
(
|
||||||
|
"baw", "co", "cog", "jam", "mtv",
|
||||||
|
"coc", "draw", "pco", "coq", "cod", "a"
|
||||||
|
),
|
||||||
|
rps=1 / 15
|
||||||
|
),
|
||||||
}
|
}
|
||||||
|
@ -41,7 +41,7 @@ class Kev4PhpHelper(ChanHelper):
|
|||||||
|
|
||||||
for threadEl in soup.find_all("p", class_="info"):
|
for threadEl in soup.find_all("p", class_="info"):
|
||||||
threads.append({
|
threads.append({
|
||||||
"id": int(threadEl.find("a").get("id")[len("expandButtun"):]),
|
"id": int(threadEl.find("a").get("id")[len("expandButton"):]),
|
||||||
"omit": int(threadEl.text.split(" ")[1])
|
"omit": int(threadEl.text.split(" ")[1])
|
||||||
})
|
})
|
||||||
|
|
||||||
|
81
chan/plus4chan_html.py
Normal file
81
chan/plus4chan_html.py
Normal file
@ -0,0 +1,81 @@
|
|||||||
|
import datetime
|
||||||
|
from urllib.parse import urljoin
|
||||||
|
|
||||||
|
from bs4 import BeautifulSoup
|
||||||
|
|
||||||
|
from chan.helper import ChanHelper
|
||||||
|
from post_process import get_links_from_html_body
|
||||||
|
|
||||||
|
|
||||||
|
class Plus4ChanHelper(ChanHelper):
|
||||||
|
|
||||||
|
def threads_url(self, board):
|
||||||
|
return "%s%s/" % (self._base_url, board)
|
||||||
|
|
||||||
|
def posts_url(self, board, thread):
|
||||||
|
return "%s%s/t%d.html" % (self._base_url, board, self.item_id(thread))
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def item_id(item):
|
||||||
|
return item["id"]
|
||||||
|
|
||||||
|
def item_urls(self, item, board):
|
||||||
|
return list(set(get_links_from_html_body(item["html"], self._base_url)))
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def item_type(item):
|
||||||
|
return item["type"]
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def thread_mtime(thread):
|
||||||
|
return thread["omit"]
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def item_mtime(item):
|
||||||
|
return item["time"]
|
||||||
|
|
||||||
|
def parse_threads_list(self, r):
|
||||||
|
soup = BeautifulSoup(r.content.decode('utf-8', 'ignore'), "html.parser")
|
||||||
|
|
||||||
|
threads = []
|
||||||
|
|
||||||
|
for threadEl in soup.find_all("section", class_="t", id=lambda x: x and x[1:].isnumeric()):
|
||||||
|
|
||||||
|
omit = threadEl.find("a", class_="omittedbreakdown")
|
||||||
|
|
||||||
|
threads.append({
|
||||||
|
"id": int(threadEl.get("id")[1:]),
|
||||||
|
"omit": int(omit.text.split(" ")[1]) if omit else 0
|
||||||
|
})
|
||||||
|
|
||||||
|
for link in soup.find_all("a", href=lambda x: x):
|
||||||
|
if link.text == "next":
|
||||||
|
return threads, urljoin(r.url, link.get("href"))
|
||||||
|
return threads, None
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def parse_thread(r):
|
||||||
|
soup = BeautifulSoup(r.content.decode('utf-8', 'ignore'), "html.parser")
|
||||||
|
|
||||||
|
thread_el = soup.find("section", class_="t")
|
||||||
|
tid = int(thread_el.get("id")[1:])
|
||||||
|
|
||||||
|
for post_el in soup.find_all("div", class_="p", id=lambda x: x and x[1:].isnumeric()):
|
||||||
|
pid = int(post_el.get("id")[1:])
|
||||||
|
if pid == tid:
|
||||||
|
yield {
|
||||||
|
"id": tid,
|
||||||
|
"type": "thread",
|
||||||
|
"html": str(post_el),
|
||||||
|
"time": int(datetime.datetime.strptime(post_el.find("time", class_="date").text,
|
||||||
|
"%Y/%m/%d %H:%M:%S").timestamp())
|
||||||
|
}
|
||||||
|
else:
|
||||||
|
yield {
|
||||||
|
"id": pid,
|
||||||
|
"type": "post",
|
||||||
|
"html": str(post_el),
|
||||||
|
"time": int(datetime.datetime.strptime(post_el.find("time", class_="date").text,
|
||||||
|
"%Y/%m/%d %H:%M:%S").timestamp()),
|
||||||
|
"parent": tid
|
||||||
|
}
|
Loading…
x
Reference in New Issue
Block a user