Dolda2000 GitWeb
/
automanga.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Added kakalot library.
[automanga.git]
/
manga
/
batoto.py
diff --git
a/manga/batoto.py
b/manga/batoto.py
index
b626dc1
..
42edfac
100644
(file)
--- a/
manga/batoto.py
+++ b/
manga/batoto.py
@@
-38,7
+38,9
@@
def nextel(el):
return el
def fetchreader(lib, readerid, page):
return el
def fetchreader(lib, readerid, page):
- pg = soupify(lib.sess.fetch(lib.base + "areader?" + urllib.parse.urlencode({"id": readerid, "p": str(page)}),
+ pg = soupify(lib.sess.fetch(lib.base + "areader?" + urllib.parse.urlencode({"id": readerid,
+ "p": str(page),
+ "supress_webtoon": "t"}),
headers={"Referer": "http://bato.to/reader"}))
return pg
headers={"Referer": "http://bato.to/reader"}))
return pg
@@
-274,6
+276,7
@@
class session(object):
else:
return
else:
else:
return
else:
+ pass
form = page.find("form", id="login")
if not form and pre:
form = page.find("form", id="login")
if not form and pre:
@@
-286,6
+289,7
@@
class session(object):
values["rememberMe"] = "1"
values["anonymous"] = "1"
req = urllib.request.Request(form["action"], urllib.parse.urlencode(values).encode("ascii"))
values["rememberMe"] = "1"
values["anonymous"] = "1"
req = urllib.request.Request(form["action"], urllib.parse.urlencode(values).encode("ascii"))
+ req.add_header("User-Agent", self.useragent)
with self.web.open(req) as hs:
page = soupify(hs.read())
for resp in page.findAll("p", attrs={"class": "message"}):
with self.web.open(req) as hs:
page = soupify(hs.read())
for resp in page.findAll("p", attrs={"class": "message"}):
@@
-298,8
+302,10
@@
class session(object):
def open(self, url):
return self.web.open(url)
def open(self, url):
return self.web.open(url)
+ useragent = "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.22 (KHTML, like Gecko) Chrome/25.0.1364.160 Safari/537.22"
def fetch(self, url, headers=None):
req = urllib.request.Request(url)
def fetch(self, url, headers=None):
req = urllib.request.Request(url)
+ req.add_header("User-Agent", self.useragent)
if headers is not None:
for k, v in headers.items():
req.add_header(k, v)
if headers is not None:
for k, v in headers.items():
req.add_header(k, v)
@@
-336,7
+342,9
@@
class library(lib.library):
while True:
_pars = dict(pars)
_pars["p"] = str(p)
while True:
_pars = dict(pars)
_pars["p"] = str(p)
- resp = urllib.request.urlopen(self.base + "search?" + urllib.parse.urlencode(_pars))
+ req = urllib.request.Request(self.base + "search?" + urllib.parse.urlencode(_pars))
+ req.add_header("User-Agent", session.useragent)
+ resp = urllib.request.urlopen(req)
try:
page = soupify(resp.read())
finally:
try:
page = soupify(resp.read())
finally: