你好,打擾了,最近dmzj好像某些常有問題 原始碼如下: Grabbing: http://manhua.dmzj.com/mkxdcdcp/ Traceback (most recent call last): File "C:\TooL\ComicCrawler-master\comiccrawler.py", line 477, in worker self.analyze(self.mission) File "C:\TooL\ComicCrawler-master\comiccrawler.py", line 512, in analyze epList = self.waitChild(downloader.getepisodelist, html, url=mission.url) File "C:\TooL\ComicCrawler-master\worker.py", line 356, in waitChild raise child.error.get_nowait() File "C:\TooL\ComicCrawler-master\worker.py", line 242, in _worker self._ret = self.worker(*self._args, **self._kwargs) File "C:\TooL\ComicCrawler-master\cc_dmzj.py", line 41, in getepisodelist comicurl = re.search("comic_url = \"(.+?)\"", html).group(1) AttributeError: 'NoneType' object has no attribute 'group'
Grabbing: http://manhua.dmzj.com/yaoguaihezhu/ Traceback (most recent call last): File "C:\TooL\ComicCrawler-master\comiccrawler.py", line 477, in worker self.analyze(self.mission) File "C:\TooL\ComicCrawler-master\comiccrawler.py", line 507, in analyze html = self.waitChild(grabhtml, mission.url, hd=downloader.header) File "C:\TooL\ComicCrawler-master\worker.py", line 356, in waitChild raise child.error.get_nowait() File "C:\TooL\ComicCrawler-master\worker.py", line 242, in _worker self._ret = self.worker(*self._args, **self._kwargs) File "C:\TooL\ComicCrawler-master\comiccrawler.py", line 177, in grabhtml return grabber(url, hd, encode) File "C:\TooL\ComicCrawler-master\comiccrawler.py", line 155, in grabber rs = urllib.request.urlopen(req, timeout=20) File "C:\Python34\lib\urllib\request.py", line 153, in urlopen return opener.open(url, data, timeout) File "C:\Python34\lib\urllib\request.py", line 455, in open response = self._open(req, data) File "C:\Python34\lib\urllib\request.py", line 473, in _open '_open', req) File "C:\Python34\lib\urllib\request.py", line 433, in _call_chain result = func(*args) File "C:\Python34\lib\urllib\request.py", line 1202, in http_open return self.do_open(http.client.HTTPConnection, req) File "C:\Python34\lib\urllib\request.py", line 1177, in do_open r = h.getresponse() File "C:\Python34\lib\http\client.py", line 1172, in getresponse response.begin() File "C:\Python34\lib\http\client.py", line 351, in begin version, status, reason = self._read_status() File "C:\Python34\lib\http\client.py", line 313, in _read_status line = str(self.fp.readline(_MAXLINE + 1), "iso-8859-1") File "C:\Python34\lib\socket.py", line 371, in readinto return self._sock.recv_into(b) socket.timeout: timed out
Something went wrong in Traceback (most recent call last): File "G:\ComicCrawler-master\worker.py", line 242, in _worker self._ret = self.worker(*self._args, **self._kwargs) File "G:\ComicCrawler-master\comiccrawler.py", line 763, in worker self.conf() File "G:\ComicCrawler-master\comiccrawler.py", line 780, in conf section("DEFAULT", default) File "G:\ComicCrawler-master\config.py", line 36, in section config[name][key] = default[key] File "C:\Python34\lib\configparser.py", line 1207, in __setitem__ self._parser._validate_value_types(option=key, value=value) File "C:\Python34\lib\configparser.py", line 1155, in _validate_value_types raise TypeError("option values must be strings") TypeError: option values must be strings Start analyzing http://manhua.dmzj.com/slxcxuxll/ Analyzing success! Start analyzing http://manhua.dmzj.com/ysty/ Analyzing success!
Python 3.4.2 (v3.4.2:ab2c023a9432, Oct 6 2014, 22:16:31) [MSC v.1600 64 bit (AM D64)] on win32 Type "help", "copyright", "credits" or "license" for more information. >>>
Microsoft Windows [版本 6.1.7601] Copyright (c) 2009 Microsoft Corporation. All rights reserved.
C:\>cd comiccrawler
C:\ComicCrawler>comiccrawlergui.py Something went wrong in <__main__.MainWindow object at 0x0000000002AAB208> Traceback (most recent call last): File "C:\ComicCrawler\worker.py", line 242, in _worker self._ret = self.worker(*self._args, **self._kwargs) File "C:\ComicCrawler\comiccrawler.py", line 1001, in worker self.loadClasses() File "C:\ComicCrawler\comiccrawler.py", line 1008, in loadClasses self.moduleManager = ModuleManager() File "C:\ComicCrawler\comiccrawler.py", line 941, in __init__ self.loadMods() File "C:\ComicCrawler\comiccrawler.py", line 953, in loadMods self.mods.append(importlib.import_module(mod)) File "C:\Python34\lib\importlib\__init__.py", line 109, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "", line 2254, in _gcd_import File "", line 2237, in _find_and_load File "", line 2226, in _find_and_load_unlocked File "", line 1200, in _load_unlocked File "", line 1129, in _exec File "", line 1471, in exec_module File "", line 321, in _call_with_frames_removed File "C:\ComicCrawler\cc_acgn.py", line 10, in import re, execjs ImportError: No module named 'execjs'
Downloading ep Vol_1 Get imgurls failed: SyntaxError: Invalid character Traceback (most recent call last): File "D:\Python Scripts\ComicCrawler-master\comiccrawler.py", line 351, in crawlpage imgurls = self.waitChild(downloader.getimgurls, html, url=ep.firstpageurl) File "D:\Python Scripts\ComicCrawler-master\worker.py", line 356, in waitChild raise child.error.get_nowait() File "D:\Python Scripts\ComicCrawler-master\worker.py", line 242, in _worker self._ret = self.worker(*self._args, **self._kwargs) File "D:\Python Scripts\ComicCrawler-master\cc_iibq.py", line 67, in getimgurls arrFiles = unsuan(sFiles).split("|") File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 280, in call return self.eval("{identifier}.apply(this, {args})".format(identifier=identifier, args=args)) File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 259, in eval return self.exec_(code) File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 276, in exec_ return self._extract_result(output.split("\n")[-2]) File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 315, in _extract_result raise RuntimeError(value) execjs.RuntimeError: SyntaxError: Invalid character Get imgurls failed: SyntaxError: Invalid character
Downloading ep Vol_1 Get imgurls failed: SyntaxError: Invalid character Traceback (most recent call last): File "D:\Python Scripts\ComicCrawler-master\comiccrawler.py", line 351, in crawlpage imgurls = self.waitChild(downloader.getimgurls, html, url=ep.firstpageurl) File "D:\Python Scripts\ComicCrawler-master\worker.py", line 356, in waitChild raise child.error.get_nowait() File "D:\Python Scripts\ComicCrawler-master\worker.py", line 242, in _worker self._ret = self.worker(*self._args, **self._kwargs) File "D:\Python Scripts\ComicCrawler-master\cc_iibq.py", line 67, in getimgurls arrFiles = unsuan(sFiles).split("|") File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 280, in call return self.eval("{identifier}.apply(this, {args})".format(identifier=identifier, args=args)) File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 259, in eval return self.exec_(code) File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 276, in exec_ return self._extract_result(output.split("\n")[-2]) File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 315, in _extract_result raise RuntimeError(value) execjs.RuntimeError: SyntaxError: Invalid character
您好,回報SF的漫畫下載時都會出現 [error_loop] Process error: name 'comiccrawler' is not defined [grabber] http://comic.sfacg.com/HTML/ZHEM/117/
Thread crashed in : Traceback (most recent call last): File "C:\Python34\lib\site-packages\worker\__init__.py", line 281, in worker returned_value = self.target(*args, **kwargs) File "C:\Python34\lib\site-packages\comiccrawler\mods\sfacg.py", line 37, in g etimgurls htmljs = comiccrawler.grabhtml(base + js) NameError: name 'comiccrawler' is not defined
Thread crashed in : Traceback (most recent call last): File "C:\Python34\lib\site-packages\worker\__init__.py", line 281, in worker returned_value = self.target(*args, **kwargs) File "C:\Python34\lib\site-packages\comiccrawler\mods\seemh.py", line 53, in g etimgurls html AttributeError: 'NoneType' object has no attribute 'group'
[error_loop] Process error: 'NoneType' object has no attribute 'group'
就會自動打包壓縮成 zip 檔了 * 記得安裝 7z︰http://www.developershome.com/7-zip/"
就出現下列訊息
Mission complete! Failed to run process: ('zip.bat', 'C:\\TooL\\ComicCrawler-master\\download\\rin ') Traceback (most recent call last): File "C:\Python34\lib\site-packages\worker\__init__.py", line 247, in wait message = self.message_cache.get_nowait() File "C:\Python34\lib\queue.py", line 195, in get_nowait return self.get(block=False) File "C:\Python34\lib\queue.py", line 164, in get raise Empty queue.Empty
During handling of the above exception, another exception occurred:
Traceback (most recent call last): File "C:\Python34\lib\site-packages\comiccrawler\download_manager.py", line 52 , in dummy subprocess.call(command) File "C:\Python34\lib\subprocess.py", line 537, in call with Popen(*popenargs, **kwargs) as p: File "C:\Python34\lib\subprocess.py", line 858, in __init__ restore_signals, start_new_session) File "C:\Python34\lib\subprocess.py", line 1111, in _execute_child startupinfo) FileNotFoundError: [WinError 2] 系統找不到指定的檔案。
C:\WINDOWS\system32>pip install comiccrawler Downloading/unpacking comiccrawler Running setup.py (path:c:\users\wjm\appdata\local\temp\pip_build_wjm\comiccrawler\setup.py) egg_info for package comiccrawler Traceback (most recent call last): File "", line 17, in File "c:\users\wjm\appdata\local\temp\pip_build_wjm\comiccrawler\setup.py", line 17, in with open(path.join(here, 'README.rst'), encoding='utf-8') as f: TypeError: 'encoding' is an invalid keyword argument for this function Complete output from command python setup.py egg_info: Traceback (most recent call last):
File "", line 17, in
File "c:\users\wjm\appdata\local\temp\pip_build_wjm\comiccrawler\setup.py", line 17, in
with open(path.join(here, 'README.rst'), encoding='utf-8') as f:
TypeError: 'encoding' is an invalid keyword argument for this function
---------------------------------------- Cleaning up... Command python setup.py egg_info failed with error code 1 in c:\users\wjm\appdata\local\temp\pip_build_wjm\comiccrawler Storing debug log for failure in C:\Users\wjm\pip\pip.log 新的问题。。。
Thread crashed in : Traceback (most recent call last): File "D:\Python34\lib\site-packages\worker\__init__.py", line 281, in worker returned_value = self.target(*args, **kwargs) File "D:\Python34\lib\site-packages\comiccrawler\mods\seemh.py", line 54, in g etimgurls ).group() AttributeError: 'NoneType' object has no attribute 'group'
[error_loop] Process error: 'NoneType' object has no attribute 'group'
Thread crashed in : Traceback (most recent call last): File "D:\Python34\lib\site-packages\worker\__init__.py", line 281, in worker returned_value = self.target(*args, **kwargs) File "D:\Python34\lib\site-packages\comiccrawler\mods\seemh.py", line 54, in g etimgurls ).group() AttributeError: 'NoneType' object has no attribute 'group'
[error_loop] Process error: 'NoneType' object has no attribute 'group'
我setting弄好了 ge站可以下 但ex站不能下 會出現 解析錯誤 http error 302 :the http server returned a redirect error that would lead to an infinite loop the last 30x error message was: found
Downloading ep 第133話 [grabber] http://www.comicvip.com/show/cool-9554.html?ch=133
Thread crashed in : Traceback (most recent call last): File "C:\Python34\lib\site-packages\worker\__init__.py", line 281, in worker returned_value = self.target(*args, **kwargs) File "C:\Python34\lib\site-packages\comiccrawler\mods\eight.py", line 97, in g etimgurls pages = int(ss(code[7:10])) ValueError: invalid literal for int() with base 10: ''
[error_loop] Process error: invalid literal for int() with base 10: ''
Downloading ep 第98話 [grabber] http://www.comicvip.com/show/cool-10660.html?ch=98
Thread crashed in : Traceback (most recent call last): File "C:\Python34\lib\site-packages\worker\__init__.py", line 281, in worker returned_value = self.target(*args, **kwargs) File "C:\Python34\lib\site-packages\comiccrawler\core.py", line 223, in grabim g return grabber(*args, **kwargs) File "C:\Python34\lib\site-packages\comiccrawler\core.py", line 168, in grabbe r response = urlopen(request, timeout=20) File "C:\Python34\lib\urllib\request.py", line 153, in urlopen return opener.open(url, data, timeout) File "C:\Python34\lib\urllib\request.py", line 461, in open response = meth(req, response) File "C:\Python34\lib\urllib\request.py", line 571, in http_response 'http', request, response, code, msg, hdrs) File "C:\Python34\lib\urllib\request.py", line 499, in error return self._call_chain(*args) File "C:\Python34\lib\urllib\request.py", line 433, in _call_chain result = func(*args) File "C:\Python34\lib\urllib\request.py", line 579, in http_error_default raise HTTPError(req.full_url, code, msg, hdrs, fp) urllib.error.HTTPError: HTTP Error 404: Not Found
[error_loop] Process error: HTTP Error 404: Not Found
Thread crashed in : Traceback (most recent call last): File "C:\Python34\lib\site-packages\worker\__init__.py", line 281, in worker returned_value = self.target(*args, **kwargs) File "C:\Python34\lib\site-packages\comiccrawler\core.py", line 223, in grabim g return grabber(*args, **kwargs) File "C:\Python34\lib\site-packages\comiccrawler\core.py", line 168, in grabbe r response = urlopen(request, timeout=20) File "C:\Python34\lib\urllib\request.py", line 153, in urlopen return opener.open(url, data, timeout) File "C:\Python34\lib\urllib\request.py", line 461, in open response = meth(req, response) File "C:\Python34\lib\urllib\request.py", line 571, in http_response 'http', request, response, code, msg, hdrs) File "C:\Python34\lib\urllib\request.py", line 499, in error return self._call_chain(*args) File "C:\Python34\lib\urllib\request.py", line 433, in _call_chain result = func(*args) File "C:\Python34\lib\urllib\request.py", line 579, in http_error_default raise HTTPError(req.full_url, code, msg, hdrs, fp) urllib.error.HTTPError: HTTP Error 404: Not Found
[error_loop] Process error: HTTP Error 404: Not Found
Microsoft Windows [版本 6.1.7601] Copyright (c) 2009 Microsoft Corporation. All rights reserved.
C:\Users\user>comiccrawler gui Start download trash.黑街杀手 Start downloading trash.黑街杀手 total 6 episode. Downloading ep 第05卷 Traceback (most recent call last): File "C:\Python34\lib\site-packages\comiccrawler\core\__init__.py", line 407, in error_loop process() File "C:\Python34\lib\site-packages\comiccrawler\core\__init__.py", line 367, in download crawler.init() File "C:\Python34\lib\site-packages\comiccrawler\core\__init__.py", line 192, in init self.init_images(self.ep.current_page - 1) File "C:\Python34\lib\site-packages\comiccrawler\core\__init__.py", line 199, in init_images self.get_images() File "C:\Python34\lib\site-packages\comiccrawler\core\__init__.py", line 330, in get_images self.ep.current_url File "C:\Python34\lib\site-packages\comiccrawler\mods\seemh.py", line 97, in g et_images return ctx.call("getFiles", path, files, host) File "C:\Python34\lib\site-packages\execjs\_abstract_runtime_context.py", line 37, in call return self._call(name, *args) File "C:\Python34\lib\site-packages\execjs\_external_runtime.py", line 87, in _call return self._eval("{identifier}.apply(this, {args})".format(identifier=ident ifier, args=args)) File "C:\Python34\lib\site-packages\execjs\_external_runtime.py", line 73, in _eval return self.exec_(code) File "C:\Python34\lib\site-packages\execjs\_abstract_runtime_context.py", line 18, in exec_ return self._exec_(source) File "C:\Python34\lib\site-packages\execjs\_external_runtime.py", line 83, in _exec_ return self._extract_result(output) File "C:\Python34\lib\site-packages\execjs\_external_runtime.py", line 165, in _extract_result raise exceptions.ProgramError(value) execjs._exceptions.ProgramError: TypeError: 物件不支援此屬性或方法
Failed to load session! Traceback (most recent call last): File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\runpy.py", line 170, in _run_module_as_main "__main__", mod_spec) File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\runpy.py", line 85, in _run_code exec(code, run_globals) File "C:\Users\Windows 7\AppData\Local\Programs\Python\Python35-32\Scripts\comiccrawler.exe\__main__.py", line 9, in File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\__init__.py", line 48, in console_init from .gui import MainWindow File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\gui.py", line 22, in from .download_manager import download_manager File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\download_manager.py", line 16, in from .mission_manager import mission_manager, init_episode, uninit_episode File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\mission_manager.py", line 249, in mission_manager = MissionManager() File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\mission_manager.py", line 86, in __init__ self.load() File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\mission_manager.py", line 124, in load self._load() File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\mission_manager.py", line 135, in _load library = load("~/comiccrawler/library.json") or [] File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\mission_manager.py", line 54, in load return json.load(fp) File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\json\__init__.py", line 268, in load parse_constant=parse_constant, object_pairs_hook=object_pairs_hook, **kw) File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\json\__init__.py", line 319, in loads return _default_decoder.decode(s) File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\json\decoder.py", line 339, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\json\decoder.py", line 357, in raw_decode raise JSONDecodeError("Expecting value", s, err.value) from None json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
錯誤訊息是 - Traceback (most recent call last): File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\core\__init__.py", line 428, in error_loop process() File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\core\__init__.py", line 378, in download crawler.init() File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\core\__init__.py", line 201, in init self.init_images(self.ep.current_page - 1) File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\core\__init__.py", line 208, in init_images self.get_images() File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\core\__init__.py", line 340, in get_images self.ep.current_url File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\mods\seemh.py", line 97, in get_images return ctx.call("getFiles", path, files, host) File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\execjs\_abstract_runtime_context.py", line 37, in call return self._call(name, *args) File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\execjs\_external_runtime.py", line 87, in _call return self._eval("{identifier}.apply(this, {args})".format(identifier=identifier, args=args)) File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\execjs\_external_runtime.py", line 73, in _eval return self.exec_(code) File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\execjs\_abstract_runtime_context.py", line 18, in exec_ return self._exec_(source) File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\execjs\_external_runtime.py", line 83, in _exec_ return self._extract_result(output) File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\execjs\_external_runtime.py", line 165, in _extract_result raise exceptions.ProgramError(value) execjs._exceptions.ProgramError: TypeError: 物件不支援此屬性或方法
C:\Users\rrt>comiccrawler gui Start analyzing http://www.hhssee.com/manhua31715.html Analyzing success! Start download 测不准的阿波连同学 Start downloading 测不准的阿波连同学 total 7 episode. Downloading ep 测不准的阿波连同学 001集 Traceback (most recent call last): File "c:\python34\lib\site-packages\comiccrawler\core\__init__.py", line 567, in error_loop process() File "c:\python34\lib\site-packages\comiccrawler\core\__init__.py", line 517, in download crawler.init() File "c:\python34\lib\site-packages\comiccrawler\core\__init__.py", line 358, in init self.init_images(self.ep.current_page - 1) File "c:\python34\lib\site-packages\comiccrawler\core\__init__.py", line 365, in init_images self.get_images() File "c:\python34\lib\site-packages\comiccrawler\core\__init__.py", line 484, in get_images self.ep.current_url File "c:\python34\lib\site-packages\comiccrawler\mods\hhxiee.py", line 86, in get_images return ctx.call("getImages", url, img_name, hd_domain) File "c:\python34\lib\site-packages\node_vm2\__init__.py", line 166, in call "args": args File "c:\python34\lib\site-packages\node_vm2\__init__.py", line 110, in commun icate raise VMError(data["error"]) node_vm2.VMError: Index doesn't exist: null
剛測試了 http://www.dm5.com/manhua-qqtszdsy/ 正常分析下載。
回覆刪除應該也不是版本的問題,因為 cc_dm5.py 的上次更新日期是 6/7。
我想看看網頁原始碼,先進到載不下來的漫畫網頁,然後按 Ctrl + U,
把程式碼複製後貼到 http://codepad.org/ 或是寄 email 也行。
另外為了保險起見,試試更新到最新版問題會不會解決。
說起來難以啟齒,其實睡個兩天三再開起來後又一切正常了...(遮
回覆刪除感謝作者的熱心,下次出現時會記得存下原始碼回報!
DM5下載會出現你盜鏈的圖片
回覆刪除忘記還要更新這裡了。
刪除麻煩到 https://github.com/eight04/ComicCrawler 下載最新版本再試試。
可以了目前
刪除監獄學園第99話
刪除Downloading ep ???? 第99?
Grabbing: http://www.dm5.com/m133958/
Get imgurls failed: SyntaxError: 必須要有 ')'
已修正
刪除執行comiccrawlergui後 無法開啟命令介面... 可以教下我嗎?
回覆刪除命令介面指的是?
刪除您好
回覆刪除我想抓http://www.999comic.com/comic/3300/92142052371f5a982c9d99df13604b03.html?p=1
可是貼上LINK後 按下[加入連結]卻沒有動作
可否麻煩您教一下
謝謝
http://www.999comic.com/comic/3300/ 貼這個llink也不行 可否麻煩您教一下 謝謝
刪除因為程式沒有支援「www.999comic.com」
刪除可以看看其它的網站有沒有相同的漫畫
如果其它的網站可以取代的話,就不加入支援了
謝謝版大您熱心分享 祝福您
刪除如果版大允許 看能不能增加 http://www.999comic.com/comic/3300/ 再謝謝您
檢查後發現此站的圖都導向漫畫島︰
刪除http://www.manhuadao.com/book/wudaokuangzhishi/
已將漫畫島加入
很感謝您 ^^ 祝福
刪除請問 主程式 要更新嗎
可是漫畫島 沒有這部漫
刪除> 主程式 要更新嗎
刪除要。這個程式沒有自動更新功能。
> 可是漫畫島 沒有這部漫
上面的連結就是了
謝謝您 下載程式更新後 貼上面的link 終於OK了 感謝您的幫忙
刪除祝福您 ^^
實在很不好意思 讓您百忙中抽出時間
再次謝謝您 ^^
你好,打擾了,最近dmzj好像某些常有問題 原始碼如下:
回覆刪除Grabbing: http://manhua.dmzj.com/mkxdcdcp/
Traceback (most recent call last):
File "C:\TooL\ComicCrawler-master\comiccrawler.py", line 477, in worker
self.analyze(self.mission)
File "C:\TooL\ComicCrawler-master\comiccrawler.py", line 512, in analyze
epList = self.waitChild(downloader.getepisodelist, html, url=mission.url)
File "C:\TooL\ComicCrawler-master\worker.py", line 356, in waitChild
raise child.error.get_nowait()
File "C:\TooL\ComicCrawler-master\worker.py", line 242, in _worker
self._ret = self.worker(*self._args, **self._kwargs)
File "C:\TooL\ComicCrawler-master\cc_dmzj.py", line 41, in getepisodelist
comicurl = re.search("comic_url = \"(.+?)\"", html).group(1)
AttributeError: 'NoneType' object has no attribute 'group'
Grabbing: http://manhua.dmzj.com/yaoguaihezhu/
Traceback (most recent call last):
File "C:\TooL\ComicCrawler-master\comiccrawler.py", line 477, in worker
self.analyze(self.mission)
File "C:\TooL\ComicCrawler-master\comiccrawler.py", line 507, in analyze
html = self.waitChild(grabhtml, mission.url, hd=downloader.header)
File "C:\TooL\ComicCrawler-master\worker.py", line 356, in waitChild
raise child.error.get_nowait()
File "C:\TooL\ComicCrawler-master\worker.py", line 242, in _worker
self._ret = self.worker(*self._args, **self._kwargs)
File "C:\TooL\ComicCrawler-master\comiccrawler.py", line 177, in grabhtml
return grabber(url, hd, encode)
File "C:\TooL\ComicCrawler-master\comiccrawler.py", line 155, in grabber
rs = urllib.request.urlopen(req, timeout=20)
File "C:\Python34\lib\urllib\request.py", line 153, in urlopen
return opener.open(url, data, timeout)
File "C:\Python34\lib\urllib\request.py", line 455, in open
response = self._open(req, data)
File "C:\Python34\lib\urllib\request.py", line 473, in _open
'_open', req)
File "C:\Python34\lib\urllib\request.py", line 433, in _call_chain
result = func(*args)
File "C:\Python34\lib\urllib\request.py", line 1202, in http_open
return self.do_open(http.client.HTTPConnection, req)
File "C:\Python34\lib\urllib\request.py", line 1177, in do_open
r = h.getresponse()
File "C:\Python34\lib\http\client.py", line 1172, in getresponse
response.begin()
File "C:\Python34\lib\http\client.py", line 351, in begin
version, status, reason = self._read_status()
File "C:\Python34\lib\http\client.py", line 313, in _read_status
line = str(self.fp.readline(_MAXLINE + 1), "iso-8859-1")
File "C:\Python34\lib\socket.py", line 371, in readinto
return self._sock.recv_into(b)
socket.timeout: timed out
您好,動漫之家的漫畫似乎有改版無法下載
回覆刪除可以幫忙看看嗎? 謝謝
已修正 & 可以貼一下網址嗎?
刪除您好,我的右下方有顯示成功解讀,但是沒有跳出選擇集數的視窗
回覆刪除我用幾個連結來試似乎都如此: http://manhua.dmzj.com/ysty/
http://manhua.dmzj.com/slxcxuxll/
這是他顯示的:
Something went wrong in
Traceback (most recent call last):
File "G:\ComicCrawler-master\worker.py", line 242, in _worker
self._ret = self.worker(*self._args, **self._kwargs)
File "G:\ComicCrawler-master\comiccrawler.py", line 763, in worker
self.conf()
File "G:\ComicCrawler-master\comiccrawler.py", line 780, in conf
section("DEFAULT", default)
File "G:\ComicCrawler-master\config.py", line 36, in section
config[name][key] = default[key]
File "C:\Python34\lib\configparser.py", line 1207, in __setitem__
self._parser._validate_value_types(option=key, value=value)
File "C:\Python34\lib\configparser.py", line 1155, in _validate_value_types
raise TypeError("option values must be strings")
TypeError: option values must be strings
Start analyzing http://manhua.dmzj.com/slxcxuxll/
Analyzing success!
Start analyzing http://manhua.dmzj.com/ysty/
Analyzing success!
拜託您了,謝謝
神秘的 bug,看不出來哪裡出錯
回覆刪除我需要 python 版本︰Win + R,輸入 python;按 Enter 後跳出來的視窗,第一行會顯示相關訊息
Python 3.4.2 (v3.4.2:ab2c023a9432, Oct 6 2014, 22:16:31) [MSC v.1600 64 bit (AM
刪除D64)] on win32
Type "help", "copyright", "credits" or "license" for more information.
>>>
已修正
回覆刪除已經成功且順利的下載,真是太謝謝您了!
刪除這算是新版Python才有的bug嗎?
不是,只是我沒注意到錯在哪 :P
刪除抱歉
回覆刪除我無法執行
因為我本身有安裝Blender
Blender需要Python26與27
因此我已經安裝過Python了
接著又照著你的指示
我安裝Python343
是有衝突還是怎樣?
下面是CMD的執行訊息
Microsoft Windows [版本 6.1.7601]
Copyright (c) 2009 Microsoft Corporation. All rights reserved.
C:\>cd comiccrawler
C:\ComicCrawler>comiccrawlergui.py
Something went wrong in <__main__.MainWindow object at 0x0000000002AAB208>
Traceback (most recent call last):
File "C:\ComicCrawler\worker.py", line 242, in _worker
self._ret = self.worker(*self._args, **self._kwargs)
File "C:\ComicCrawler\comiccrawler.py", line 1001, in worker
self.loadClasses()
File "C:\ComicCrawler\comiccrawler.py", line 1008, in loadClasses
self.moduleManager = ModuleManager()
File "C:\ComicCrawler\comiccrawler.py", line 941, in __init__
self.loadMods()
File "C:\ComicCrawler\comiccrawler.py", line 953, in loadMods
self.mods.append(importlib.import_module(mod))
File "C:\Python34\lib\importlib\__init__.py", line 109, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "", line 2254, in _gcd_import
File "", line 2237, in _find_and_load
File "", line 2226, in _find_and_load_unlocked
File "", line 1200, in _load_unlocked
File "", line 1129, in _exec
File "", line 1471, in exec_module
File "", line 321, in _call_with_frames_removed
File "C:\ComicCrawler\cc_acgn.py", line 10, in
import re, execjs
ImportError: No module named 'execjs'
C:\ComicCrawler>
缺少 execjs 這個模組。用以下指令安裝
刪除pip install pyexecjs
抱歉~
回覆刪除沒問題了
我從PPT看到了
https://www.ptt.cc/bbs/EZsoft/M.1405606960.A.1FB.html
並下載Google上的版本
內有中文說明
使用需求:
==========================
Python 3.4 up
下載網站︰
https://www.python.org/
記得要選「Add python.exe to path」,否則安裝 PyExecJS 時會找不到 pip 命令。
PyExecJS
1. 先安裝完 Python
2. 按 Windows + R,輸入 cmd 後按 Enter
3. 輸入 pip install pyexecjs 後按 Enter。安裝時應該會跑出類似︰
====
Downloading/unpacking pyexecjs
Downloading PyExecJS-1.0.4.tar.gz
Running setup.py (path:C:\User.......
Installing collected packages: pyexecjs
Running setup.py install for pyexecjs
Successfully installed pyexecjs
Cleaning up...
====
4. 安裝完畢
您好,我安裝好 Python 3以後,打開comiccrawlergui.py,出現黑色視窗跑了一串文字就自動關閉了,請問是什麼問題呢@@"
回覆刪除可能是缺少 execjs 這個模組。用以下指令安裝
刪除pip install pyexecjs
另外,可以的話麻煩把那串文字貼上來。
可以支援 http://www.iibq.com 嗎?
回覆刪除謝謝
已加入
刪除出現了以下的錯誤訊息: 謝謝
刪除Downloading ep Vol_1
Get imgurls failed: SyntaxError: Invalid character
Traceback (most recent call last):
File "D:\Python Scripts\ComicCrawler-master\comiccrawler.py", line 351, in crawlpage
imgurls = self.waitChild(downloader.getimgurls, html, url=ep.firstpageurl)
File "D:\Python Scripts\ComicCrawler-master\worker.py", line 356, in waitChild
raise child.error.get_nowait()
File "D:\Python Scripts\ComicCrawler-master\worker.py", line 242, in _worker
self._ret = self.worker(*self._args, **self._kwargs)
File "D:\Python Scripts\ComicCrawler-master\cc_iibq.py", line 67, in getimgurls
arrFiles = unsuan(sFiles).split("|")
File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 280, in call
return self.eval("{identifier}.apply(this, {args})".format(identifier=identifier, args=args))
File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 259, in eval
return self.exec_(code)
File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 276, in exec_
return self._extract_result(output.split("\n")[-2])
File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 315, in _extract_result
raise RuntimeError(value)
execjs.RuntimeError: SyntaxError: Invalid character
Get imgurls failed: SyntaxError: Invalid character
出現的錯誤訊息: 謝謝
刪除Downloading ep Vol_1
Get imgurls failed: SyntaxError: Invalid character
Traceback (most recent call last):
File "D:\Python Scripts\ComicCrawler-master\comiccrawler.py", line 351, in crawlpage
imgurls = self.waitChild(downloader.getimgurls, html, url=ep.firstpageurl)
File "D:\Python Scripts\ComicCrawler-master\worker.py", line 356, in waitChild
raise child.error.get_nowait()
File "D:\Python Scripts\ComicCrawler-master\worker.py", line 242, in _worker
self._ret = self.worker(*self._args, **self._kwargs)
File "D:\Python Scripts\ComicCrawler-master\cc_iibq.py", line 67, in getimgurls
arrFiles = unsuan(sFiles).split("|")
File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 280, in call
return self.eval("{identifier}.apply(this, {args})".format(identifier=identifier, args=args))
File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 259, in eval
return self.exec_(code)
File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 276, in exec_
return self._extract_result(output.split("\n")[-2])
File "d:\apps\utilities\python34\lib\site-packages\execjs\__init__.py", line 315, in _extract_result
raise RuntimeError(value)
execjs.RuntimeError: SyntaxError: Invalid character
可以附上網址嗎?
刪除http://www.iibq.com/comic/82012137709/
回覆刪除我想是 grabhtml("http://www.iibq.com/script/viewhtm.js") 的 UTF-8 encoding 出錯了 ...
很感謝您的幫忙
我需要更詳細的 log 檔,這邊正常。
刪除在 setting.ini 內 default 段 加入︰
errorlog = true
重開後在發生錯誤時暫停,資料夾內應該會產生 grabber.xx.log,貼上來
若是檔案太大,用寄信的也可以
寄了 grabber.file.log 和 grabber.header.log 给你 ...
刪除謝謝
跟 viewhtm.js 無關
刪除或許是 javascrtip runtime 的問題
先試試最新版能不能解決
另外請檢查下 runtime,開啟 python 後輸入以下程式碼︰
import execjs
execjs.get().name
就會顯示名稱了
最新版同样的错误 ...
刪除execjs.get().name 顯示 'JScript'
你是用 Windows XP + IE6 嗎?
刪除我猜這是舊版 IE 的 bug(精明眼不支援舊 IE)
暫時的解決方法,安裝 Node.js: https://nodejs.org/
不是, 我是用 Windows 7 Ultimate + IE11 + Firefox 38.0.5.
刪除安裝 Node.js 後有解決了嗎?
刪除安裝 Node.js 後, 成功了! 请问Node.js 和 JScript 有什么分别?
刪除为什么 Node.js 没有問題呢? 很感謝您!
Node.js 是用 Google 開發的 V8 引擎
刪除JScript 則是由微軟隨著 IE 開發的
兩者都是 JavaScript 引擎。不過 JScript 在 IE8 之前的版本,bug 出了名得多,並且只支援到 ECMA3。而 V8 現在還持續在開發,並且支援 ECMA5(目前通用的 JavaScript)
不過我也不清楚要怎麼看 JScript 的版本
最近找到一個不錯的網站:
回覆刪除http://www.seemh.com/
甚至有連大站dm5都沒有的資源,而且圖片不會亂添加浮水印,僅供您參考。
已加入
刪除你好!謝謝你的程式~
回覆刪除想請問,如果要設定D槽的CommicDownload為下載資料夾,該如何設定呢?
在設定檔中把 savepath 改成 D:\ComicDownload。設定檔位於 %userprofile%/comiccrawler/setting.ini。其它的設定請參考 https://github.com/eight04/ComicCrawler#%E8%A8%AD%E5%AE%9A%E6%AA%94
回覆刪除謝謝 更改成功! 一直以為要加引號~
刪除您好,回報SF的漫畫下載時都會出現
回覆刪除[error_loop] Process error: name 'comiccrawler' is not defined
[grabber] http://comic.sfacg.com/HTML/ZHEM/117/
Thread crashed in :
Traceback (most recent call last):
File "C:\Python34\lib\site-packages\worker\__init__.py", line 281, in worker
returned_value = self.target(*args, **kwargs)
File "C:\Python34\lib\site-packages\comiccrawler\mods\sfacg.py", line 37, in g
etimgurls
htmljs = comiccrawler.grabhtml(base + js)
NameError: name 'comiccrawler' is not defined
已修正於 2015.8.7
刪除你好,回報http://www.seemh.com/的漫畫無法下載
回覆刪除例如複製http://www.seemh.com/comic/6320/加入連結及選擇集數沒問題
但開始下載失敗,會出現
[grabber] http://www.seemh.com/comic/6320/55088.html
Thread crashed in :
Traceback (most recent call last):
File "C:\Python34\lib\site-packages\worker\__init__.py", line 281, in worker
returned_value = self.target(*args, **kwargs)
File "C:\Python34\lib\site-packages\comiccrawler\mods\seemh.py", line 53, in g
etimgurls
html
AttributeError: 'NoneType' object has no attribute 'group'
[error_loop] Process error: 'NoneType' object has no attribute 'group'
僅供參考,謝謝!
2015/8/30 測試可正常使用
刪除http://www.seemh.com/comic/5135/
回覆刪除下載時出現
[grabber] http://www.seemh.com/comic/5135/48891.html
Thread crashed in :
Traceback (most recent call last):
File "E:\software\Python\lib\site-packages\worker\__init__.py", line 28
orker
returned_value = self.target(*args, **kwargs)
File "E:\software\Python\lib\site-packages\comiccrawler\mods\seemh.py",
3, in getimgurls
html
AttributeError: 'NoneType' object has no attribute 'group'
[error_loop] Process error: 'NoneType' object has no attribute 'group'
已解決 多試幾次就好了
刪除請教一下下載完後壓縮哪部分有錯?
回覆刪除我參考這段文章後照做
"# 2014/4/17
新增了執行外部程式的功能
在 setting.ini 中,[DEFAULT] 底下,新增了設定 runafterdownload
我順便寫了 7z 壓縮指令,zip.bat。只要設定︰
runafterdownload = zip.bat
就會自動打包壓縮成 zip 檔了
* 記得安裝 7z︰http://www.developershome.com/7-zip/"
就出現下列訊息
Mission complete!
Failed to run process: ('zip.bat', 'C:\\TooL\\ComicCrawler-master\\download\\rin
')
Traceback (most recent call last):
File "C:\Python34\lib\site-packages\worker\__init__.py", line 247, in wait
message = self.message_cache.get_nowait()
File "C:\Python34\lib\queue.py", line 195, in get_nowait
return self.get(block=False)
File "C:\Python34\lib\queue.py", line 164, in get
raise Empty
queue.Empty
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\Python34\lib\site-packages\comiccrawler\download_manager.py", line 52
, in dummy
subprocess.call(command)
File "C:\Python34\lib\subprocess.py", line 537, in call
with Popen(*popenargs, **kwargs) as p:
File "C:\Python34\lib\subprocess.py", line 858, in __init__
restore_signals, start_new_session)
File "C:\Python34\lib\subprocess.py", line 1111, in _execute_child
startupinfo)
FileNotFoundError: [WinError 2] 系統找不到指定的檔案。
zip.bat 這個檔案在新版中被拿掉了,原因有二
刪除1. bat 批次檔只在 Windows 上可用
2. 不是所有人的電腦中都有裝 7zip
我會老慮把「利用 7zip 和 runafterdownload 將資料夾打包成壓縮檔」的教學文章加入說明頁面,或是直接把「製作成壓縮檔」的功能,加進主程式內。
在那之前,可以先參考舊版檔案︰http://is.gd/RhpHZI
把它存進電腦,例如 C:\zip.bat
再把 runafterdownload 設成 C:\zip.bat 就能用了
作者已經移除這則留言。
刪除作者已經移除這則留言。
回覆刪除感謝你的回覆,可以使用了~我後來改了一部分,方便自己習慣,也po上來給大家參考,
回覆刪除我改成每一話或每一集就壓縮一個檔案,而不是欸特的添加壓縮的方式!
大家參考參考 也感謝欸特的幫忙
REM 7z path
set zip="C:\Program Files\7-Zip\7z.exe"
REM move to mission directory
cd /D %1
REM run 7z command to compress to zip file
FOR /d %%i in (*) do (
%zip% a -tzip "%%i.zip" "%%i")
REM exit if failed to compress
if ERRORLEVEL 1 goto exit
exit:
REM pause
感謝你的軟件, 不知道日後可不可以問有關COMIC CRAWLER 支援網站以外的技術問題嗎?
回覆刪除謝謝
當然可以
刪除謝謝欸特, 想問一下. http://www.ebookjapan.jp/ebj/index.asp 這網站可不可用SEEDMH 那種方法, 用EXECJS 來處理圖像? 希望想請教一下.
刪除execjs 這個模組是用來執行 JS (JavaScript) 的,和圖像無關。
刪除Comic Crawler 是這樣運作的︰
1. 分析漫畫頁,找出各集的連結
2. 分析單集的連結,找出每頁上的圖片網址
3. 從圖片網址把圖片下載下來
1, 2 在分析時,圖片網址不一定是直接顯示在網頁上,可能會藏在 JS 內。如果 JS 沒被加密過,可以直接搜尋文字去找;但如果有被加密過,就得執行 JS 再取得結果。這和圖像處理是無關的。
我看了一下 ebookjapan 的原始碼,它是用 Ajax 搭配 timestamp, session 去取得圖像,並且圖像分成兩部份,取得時是 base64 encoded 的 data url。只要找出它的 request 送了什麼過去,應該很容易就能拿到圖片。
其實我的問題是能不能用那網站本身的JAVASCRIPT在PYTHON 處理加密的圖片. 因為本身我對JAVASCRIPT 不太懂, 所以想問欸特能不能找出處理的解密的方法或思路. 謝謝
刪除如果你說的解密是指 base64 轉換,答案是可以,但 python 有內建的 base64 模組可以處理。
刪除如果是指用該網站的 JavaScript 取得圖片,答案也是可以,只要能看懂它的原始碼就行了。但不懂 JavaScript 和網頁運作原理的話,要直接去分析原始碼可能會有點困難。如果不要用該網站的 JavaScript,也可以用 python 自行把整個解密過程重寫一遍(但要明白解密過程,依然要去看 JavaScript 原始碼)。
以下是我找尋圖片的步驟,用 Firefox 的除錯工具︰
1. 觀察網頁的 request,找出和圖片有關的部份(http://i.imgur.com/ZBvGPvK.png)
2. 觀察載入的 JavaScript,找出和網站主要用到的檔案(http://i.imgur.com/XCuL9FZ.png)
3. Beautify 後看原始碼,找出和圖片有關的部份。可以配合除錯工具一起用(http://i.imgur.com/B9gUOlB.png)
4. 把需要的部份丟給 python 的 execjs 去跑。
其實我用CHROME 看到JS 抓取資料後生成BASE64 資料, 得出來的結果是 (http://i.imgur.com/BKpwOHi.jpg), 我想知道得出這個之後, 圖片轉換原圖的JAVASCRIPT 跑法. 謝謝
刪除DRM 的演算法貌似在 BR_pageAction.drawCanvas() 裡面
刪除那個DRM 算法的跑法我還是看不懂....
刪除沒有辦法下,用PYTHON 寫了一個自動截圖的工具把電子書的圖片自動生成圖片檔, 1秒鐘自動截一張圖, 效率並不太高...(圖片檔是用來支援漢化組織) http://i.imgur.com/xi7FdB4.jpg
作者已經移除這則留言。
刪除作者已經移除這則留言。
回覆刪除Downloading/unpacking pyexecjs
回覆刪除Downloading PyExecJS-1.1.0.tar.gz
Running setup.py (path:c:\users\wjm\appdata\local\temp\pip_build_wjm\pyexecjs\setup.py) egg_info for package pyexecjs
Installing collected packages: pyexecjs
Running setup.py install for pyexecjs
Successfully installed pyexecjs
Cleaning up...
C:\WINDOWS\system32>pip install comiccrawler
Downloading/unpacking comiccrawler
Running setup.py (path:c:\users\wjm\appdata\local\temp\pip_build_wjm\comiccrawler\setup.py) egg_info for package comiccrawler
Traceback (most recent call last):
File "", line 17, in
File "c:\users\wjm\appdata\local\temp\pip_build_wjm\comiccrawler\setup.py", line 12, in
from build import find_version, read
ImportError: No module named build
Complete output from command python setup.py egg_info:
Traceback (most recent call last):
File "", line 17, in
File "c:\users\wjm\appdata\local\temp\pip_build_wjm\comiccrawler\setup.py", line 12, in
from build import find_version, read
ImportError: No module named build
----------------------------------------
Cleaning up...
Command python setup.py egg_info failed with error code 1 in c:\users\wjm\appdata\local\temp\pip_build_wjm\comiccrawler
Storing debug log for failure in C:\Users\wjm\pip\pip.log
这个是什么原因
看到這個我才發現 build process 有很大的問題,會找時間修正
刪除https://github.com/eight04/ComicCrawler/issues/6
試試最新版有沒有解決
刪除C:\WINDOWS\system32>pip install comiccrawler
刪除Downloading/unpacking comiccrawler
Running setup.py (path:c:\users\wjm\appdata\local\temp\pip_build_wjm\comiccrawler\setup.py) egg_info for package comiccrawler
Traceback (most recent call last):
File "", line 17, in
File "c:\users\wjm\appdata\local\temp\pip_build_wjm\comiccrawler\setup.py", line 17, in
with open(path.join(here, 'README.rst'), encoding='utf-8') as f:
TypeError: 'encoding' is an invalid keyword argument for this function
Complete output from command python setup.py egg_info:
Traceback (most recent call last):
File "", line 17, in
File "c:\users\wjm\appdata\local\temp\pip_build_wjm\comiccrawler\setup.py", line 17, in
with open(path.join(here, 'README.rst'), encoding='utf-8') as f:
TypeError: 'encoding' is an invalid keyword argument for this function
----------------------------------------
Cleaning up...
Command python setup.py egg_info failed with error code 1 in c:\users\wjm\appdata\local\temp\pip_build_wjm\comiccrawler
Storing debug log for failure in C:\Users\wjm\pip\pip.log
新的问题。。。
檢查一下 python 的版本是否是 3.4 以上,Comic Crawler 需要 3.4 以上才能執行
刪除你好
回覆刪除http://www.seemh.com/的漫畫無法下載
加入連結 選擇話數沒有問題
但是不能下載
[grabber] http://www.seemh.com/comic/282/80031.html
Thread crashed in :
Traceback (most recent call last):
File "D:\Python34\lib\site-packages\worker\__init__.py", line 281, in worker
returned_value = self.target(*args, **kwargs)
File "D:\Python34\lib\site-packages\comiccrawler\mods\seemh.py", line 54, in g
etimgurls
).group()
AttributeError: 'NoneType' object has no attribute 'group'
[error_loop] Process error: 'NoneType' object has no attribute 'group'
你好
回覆刪除http://www.seemh.com/的漫畫無法下載
加入連結 選擇話數沒有問題
但是不能下載
[grabber] http://www.seemh.com/comic/282/80031.html
Thread crashed in :
Traceback (most recent call last):
File "D:\Python34\lib\site-packages\worker\__init__.py", line 281, in worker
returned_value = self.target(*args, **kwargs)
File "D:\Python34\lib\site-packages\comiccrawler\mods\seemh.py", line 54, in g
etimgurls
).group()
AttributeError: 'NoneType' object has no attribute 'group'
[error_loop] Process error: 'NoneType' object has no attribute 'group'
於 2016.1.3 修正
刪除請問怎麼在setting筆記本裡
回覆刪除設定e紳士的帳號密碼
比如說
帳號:a1234
密碼:a5678
請問示範一下嗎?
設定檔要輸入的不是帳號密碼,而是特定的 Cookie,可以在開發工具裡找到(按 F12)。
刪除Firefox: http://i.imgur.com/J3hxeL7.png
Chrome: http://i.imgur.com/ETGSVpL.png
你好 設定好了 還是一樣ge可以 但ex不行
刪除謝謝你的教學
刪除我知道哪邊弄錯了
[e紳士]
ipb_member_id = ipb_member_id:xxxxxx
ipb_pass_hash =
ipb_pass_hash:xxxxxxxxxxxx
我沒有把中間的
ipb_member_id:
和ipb_pass_hash:
加進去
謝謝你
這個解決辦法還滿神秘的,因為程式中並沒有要求把「XXX:」加進去。
刪除經你這樣講,我覺得可能是改完設定後,忘記存檔或忘記「重載設定檔」而導致 Comic Crawler 還在用舊的設定。
應該是 總之問題解決了 謝謝^^
刪除剛試了一下 如你所說 是没按 "重載設定檔"
刪除我setting弄好了
回覆刪除ge站可以下
但ex站不能下
會出現 解析錯誤
http error 302 :the http server returned a redirect error that
would lead to an infinite loop
the last 30x error message was:
found
1. 檢查用瀏覽器是否能上 ExH
刪除2. 檢查 Cookie 有沒有複製完整。注意是 ExH 的 Cookie,EH 是不用登入就能看的
不好意思想请问一下,如果某一话的某一页下载有问题,我删除那一页后,要让ComicCrawler仅仅重新下载被删除的那一页,是要怎样设定才可以做到呢?
回覆刪除你可以把整個任務刪除,再重加一次任務並勾選要下載的集數。之前載過的頁數 Comic Crawler 會自動跳過。
刪除尝试过了,它会整个集数重新下载,对那些一话的漫画还好,对那些一卷就比较麻烦了。
刪除Comic Crawler 會檢查下載資料夾內的頁數,如果已經有那頁的話會略過,不下載圖片︰
刪除https://github.com/eight04/ComicCrawler/blob/master/comiccrawler/core.py#L558
但即使如此,它還是會去抓下一頁的 HTML 內容(視情況而定,有些支援 AJAX 的網站可能不用)。
不然就得手動改存檔(先關閉 Comic Crawler)︰
1. 用文字編輯器開啟 ~comiccrawler/pool.json
2. 搜尋任務的 URL
3. 將 "state": "FINISHED" 改成 "state": "ERROR"
4. 搜尋集數的 URL,通常是第一頁的 URL
5. 將 "current_page": xx 改成你要的頁數
6. 將 "current_url": xx 改成你要的那頁的 URL
7. 將 "complete": true 改成 "complete": false
重開後再下載,就會從指定的那頁開始。
手動改存檔有個風險—就是把存檔給改壞,造成 Comic Crawler 讀不進來。原則上 Comic Crawler 遇到讀不了的存檔會自動把存檔備份,不用擔心整個存檔遺失的情形。但若有疑慮的話還是建議在改之前先把存檔 Copy 一份(包括 pool.json, view.json, library.json)。
好的,了解了,我会再去试试看你建议的方法。:)
刪除不过最主要只要它会略过就好,不然对上dmzj这类型常常短线的网站会是个问题。
谢谢!
你好,我也遇到了这个问题,这个问题是因为默认的路径是~/comiccrawler/ 而不是/home/xxx/comiccrawler/ 引起的。解决的方法是在 core.py文件的292行处插入一句
刪除efd = os.path.expanduser(efd)
我瞭解了,是 os.listdir() 並不會展開 "~" 的問題,這幾天會修復。
刪除已於 2016.1.26 修正
刪除無限的網站好像修改了造成無法下載,我把讀取錯誤地方擷取下來,
回覆刪除Downloading ep 第133話
[grabber] http://www.comicvip.com/show/cool-9554.html?ch=133
Thread crashed in :
Traceback (most recent call last):
File "C:\Python34\lib\site-packages\worker\__init__.py", line 281, in worker
returned_value = self.target(*args, **kwargs)
File "C:\Python34\lib\site-packages\comiccrawler\mods\eight.py", line 97, in g
etimgurls
pages = int(ss(code[7:10]))
ValueError: invalid literal for int() with base 10: ''
[error_loop] Process error: invalid literal for int() with base 10: ''
Downloading ep 第98話
[grabber] http://www.comicvip.com/show/cool-10660.html?ch=98
Downloading 第98話 page 1: http://img.8comic.com/r/10660/98/001_49v.jpg
[grabber] http://img.8comic.com/r/10660/98/001_49v.jpg
Thread crashed in :
Traceback (most recent call last):
File "C:\Python34\lib\site-packages\worker\__init__.py", line 281, in worker
returned_value = self.target(*args, **kwargs)
File "C:\Python34\lib\site-packages\comiccrawler\core.py", line 223, in grabim
g
return grabber(*args, **kwargs)
File "C:\Python34\lib\site-packages\comiccrawler\core.py", line 168, in grabbe
r
response = urlopen(request, timeout=20)
File "C:\Python34\lib\urllib\request.py", line 153, in urlopen
return opener.open(url, data, timeout)
File "C:\Python34\lib\urllib\request.py", line 461, in open
response = meth(req, response)
File "C:\Python34\lib\urllib\request.py", line 571, in http_response
'http', request, response, code, msg, hdrs)
File "C:\Python34\lib\urllib\request.py", line 499, in error
return self._call_chain(*args)
File "C:\Python34\lib\urllib\request.py", line 433, in _call_chain
result = func(*args)
File "C:\Python34\lib\urllib\request.py", line 579, in http_error_default
raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 404: Not Found
[error_loop] Process error: HTTP Error 404: Not Found
無限把網址改成 "www.comicbus.com" 了,請把存檔內的 "www.comicvip.com" 全部取代成 "www.comicbus.com"
刪除改完後 出現的錯誤~
刪除Downloading ep 第84話
[grabber] http://www.comicbus.com/show/best-manga-12324.html?ch=84
Downloading 第84話 page 1: http://img.8comic.com/3/12324/84/001_u7q.jpg
[grabber] http://img.8comic.com/3/12324/84/001_u7q.jpg
Thread crashed in :
Traceback (most recent call last):
File "C:\Python34\lib\site-packages\worker\__init__.py", line 281, in worker
returned_value = self.target(*args, **kwargs)
File "C:\Python34\lib\site-packages\comiccrawler\core.py", line 223, in grabim
g
return grabber(*args, **kwargs)
File "C:\Python34\lib\site-packages\comiccrawler\core.py", line 168, in grabbe
r
response = urlopen(request, timeout=20)
File "C:\Python34\lib\urllib\request.py", line 153, in urlopen
return opener.open(url, data, timeout)
File "C:\Python34\lib\urllib\request.py", line 461, in open
response = meth(req, response)
File "C:\Python34\lib\urllib\request.py", line 571, in http_response
'http', request, response, code, msg, hdrs)
File "C:\Python34\lib\urllib\request.py", line 499, in error
return self._call_chain(*args)
File "C:\Python34\lib\urllib\request.py", line 433, in _call_chain
result = func(*args)
File "C:\Python34\lib\urllib\request.py", line 579, in http_error_default
raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 404: Not Found
[error_loop] Process error: HTTP Error 404: Not Found
這邊有對這次改版的詳細討論︰ https://github.com/eight04/ComicCrawler/issues/7#issuecomment-189981285 看看裡面的方法能不能解決
刪除www.seemh.com的网址已经被改成http://www.ikanman.com
回覆刪除已於 2016.4.27 修正,見 https://github.com/eight04/ComicCrawler/issues/11
刪除你好 我是這樣安裝
回覆刪除1.安裝Python 3.4 以上
2.安裝時記得要選「Add python.exe to path」,才能使用 pip 指令
3.在 cmd 底下輸入以下指令︰ pip install comiccrawler
4.再輸入 comiccrawler gui
結果顯示comiccrawler' 不是內部或外部命令、可執行的程式或批次檔。
請問我漏了哪個步驟嗎?
先確認 comiccrawler 是否已安裝成功。
刪除執行︰
python -m comiccrawler --version
若是能正確的列出版本號就表示安裝成功。可能是捷徑的環境變數還沒更新,重開機就能解決。
順帶一提,執行 comiccrawler gui 和 python -m comiccrawler gui 有相同效果。
若是顯示「No module named comiccrawler」則表示安裝失敗,python 找不到 comiccrawler 這個模組。
話說這篇又被當作垃圾留言了,why??
刪除網誌管理員已經移除這則留言。
回覆刪除你好 的確是出現「No module named comiccrawler」
回覆刪除輸入pip install comiccrawler
出現以下文字
http://imgur.com/a/dbNdO
你好 我已經可以使用了
回覆刪除重新安裝 Python 3.4
就可以 成功安裝comiccrawler
打擾你了 謝謝
「~=」好像是錯誤的格式,只能用「>, >=, <, <=」
刪除已於 2016.5.15 修正
刪除http://www.ikanman.com/comic/7708/
回覆刪除這網站變得不能下了
其餘有支援的網站 都正常
v2016.5.24 測試正常。先試著更新到最新版看看,若還是失敗再貼出錯誤訊息
刪除謝謝 忘了先更新看看 打擾你了
刪除已更新到最新
回覆刪除請問下載速度有異常要怎麼處理
以前載一頁最久不超過20秒
現在載了一整天還載不完一部
平均每一頁載3~5分鐘
還是是因為站點是看漫畫的問題
http://imgur.com/LGekrrY
網址是?檢查看看用瀏覽器會不會比較快
刪除用瀏覽器 圖都是馬上好的
刪除例如這一個 http://tw.ikanman.com/comic/4949/
更新到 2016.6.14,修復了一個嚴重的 bug
刪除感謝回報,我根本沒察覺有這麼嚴重的 bug 存在 Orz
感謝修復 已恢復正常 m( _ _ )m
刪除你好,我應該已經更新到最新版本,以下網址無法下載:
回覆刪除http://www.ikanman.com/comic/6804/
cmd方面顯示:
Microsoft Windows [版本 6.1.7601]
Copyright (c) 2009 Microsoft Corporation. All rights reserved.
C:\Users\user>comiccrawler gui
Start download trash.黑街杀手
Start downloading trash.黑街杀手
total 6 episode.
Downloading ep 第05卷
Traceback (most recent call last):
File "C:\Python34\lib\site-packages\comiccrawler\core\__init__.py", line 407,
in error_loop
process()
File "C:\Python34\lib\site-packages\comiccrawler\core\__init__.py", line 367,
in download
crawler.init()
File "C:\Python34\lib\site-packages\comiccrawler\core\__init__.py", line 192,
in init
self.init_images(self.ep.current_page - 1)
File "C:\Python34\lib\site-packages\comiccrawler\core\__init__.py", line 199,
in init_images
self.get_images()
File "C:\Python34\lib\site-packages\comiccrawler\core\__init__.py", line 330,
in get_images
self.ep.current_url
File "C:\Python34\lib\site-packages\comiccrawler\mods\seemh.py", line 97, in g
et_images
return ctx.call("getFiles", path, files, host)
File "C:\Python34\lib\site-packages\execjs\_abstract_runtime_context.py", line
37, in call
return self._call(name, *args)
File "C:\Python34\lib\site-packages\execjs\_external_runtime.py", line 87, in
_call
return self._eval("{identifier}.apply(this, {args})".format(identifier=ident
ifier, args=args))
File "C:\Python34\lib\site-packages\execjs\_external_runtime.py", line 73, in
_eval
return self.exec_(code)
File "C:\Python34\lib\site-packages\execjs\_abstract_runtime_context.py", line
18, in exec_
return self._exec_(source)
File "C:\Python34\lib\site-packages\execjs\_external_runtime.py", line 83, in
_exec_
return self._extract_result(output)
File "C:\Python34\lib\site-packages\execjs\_external_runtime.py", line 165, in
_extract_result
raise exceptions.ProgramError(value)
execjs._exceptions.ProgramError: TypeError: 物件不支援此屬性或方法
使用 2016.7.2 可以正常下載。
刪除你有安裝 Node.js 嗎?
https://github.com/eight04/ComicCrawler#install-nodejs
感謝,安裝完Node.js再試果然OK,勞煩了
回覆刪除电脑卡机重启后,就无法开启comic crawler了,以下是cmd的显示,请问要如何解决这个问题?
回覆刪除Failed to load session!
Traceback (most recent call last):
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\runpy.py", line 170, in _run_module_as_main
"__main__", mod_spec)
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\runpy.py", line 85, in _run_code
exec(code, run_globals)
File "C:\Users\Windows 7\AppData\Local\Programs\Python\Python35-32\Scripts\comiccrawler.exe\__main__.py", line 9, in
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\__init__.py", line 48, in console_init
from .gui import MainWindow
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\gui.py", line 22, in
from .download_manager import download_manager
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\download_manager.py", line 16, in
from .mission_manager import mission_manager, init_episode, uninit_episode
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\mission_manager.py", line 249, in
mission_manager = MissionManager()
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\mission_manager.py", line 86, in __init__
self.load()
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\mission_manager.py", line 124, in load
self._load()
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\mission_manager.py", line 135, in _load
library = load("~/comiccrawler/library.json") or []
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\mission_manager.py", line 54, in load
return json.load(fp)
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\json\__init__.py", line 268, in load
parse_constant=parse_constant, object_pairs_hook=object_pairs_hook, **kw)
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\json\__init__.py", line 319, in loads
return _default_decoder.decode(s)
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\json\decoder.py", line 339, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\json\decoder.py", line 357, in raw_decode
raise JSONDecodeError("Expecting value", s, err.value) from None
json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
之前的留言不见了,所以我重说一次。
回覆刪除情况是因为我的电脑卡机,所以我暖重启,然后要打开comiccrawler的时候就出现"Failed to load session!"以及一堆的cmd显示。
想请问一下这样的情况要怎样解决呢?
存檔損壞了。
刪除1. 檢查 "~/comiccrawler" 資料夾裡的 .json 檔案是否正常,如果不在意隱私問題的話可以把存檔寄給我。
2. 如果運氣好,在資料夾內可能會有之前的備份,檔名是 "XXX.json@[時間戳記]"。如果有看到的話可以試著用備份檔替換。
3. 建議執行一下磁碟修復工具,漫畫存檔損壞了事小,要是系統檔案也有損壞就麻煩大了
1. 我把library.json档拿掉就正常了,这个.json档案会很重要吗?因为我开始下载漫画后它又自己出现了,不过我有保留损毁的library.json档。
回覆刪除2. 有出现,不过都是在损毁的时候后才出现的,所以替换后依然开启不了。
3. 已经执行,目前看来没有其它问题,谢谢提醒!:)
library.json 儲存著「圖書館」的任務列表。打開來應該會看到很多網址用雙引號括起來,以逗點分隔,最外側再加上中括號。如︰
刪除[
"http://xxx",
"http://yyy",
"http://zzz"
]
哦,那我想应该没有太大的问题了,因为最近喜欢的网站有书架的功能了,所以我都直接用那个来收集漫画。
刪除谢谢你的帮助!:D
你好今天要下載dm5時
回覆刪除http://www.dm5.com/
出現以下資訊
http://imgur.com/a/V7V60
問題已解決 謝謝
刪除阿...抱歉 似乎還是不行...
刪除已於 2016.8.19 修正
刪除站长, 目前有失效的站包括 comic.ck101.com 99comic www.manhuadao.com,建议加入 以下几个站
回覆刪除www.hhxiee.com www.cococomic.com 99770comic 以上几个站漫画资源丰富些。考虑到大陆漫画站的删减力度,希望加大tw站点的支持
我這邊看只有 hhxiee 是正常的,99、coco、99770 全都死圖。
刪除www.ikanman.com 的前身好像就是漫畫島?我不太確定。
是的,99comic和99770服务器好像坏了。。补一个 http://dmeden.net/
回覆刪除汗汗与dmeden的漫画比其它站多不少,ikanman只有2万部 而hh和dmeden是3万
99comic, 99770, dmeden, coco都是連到1112223333.com拿圖片,如 http://eden.1112223333.com/dm02//ok-comic02/G/27966/act_001/z_0001_12851.JPG。
刪除查詢 1112223333.com,註冊訊息中email看起來是汗汗的 http://i.imgur.com/MAYxnya.png
也許其它站點是汗汗的分身站?
--
話說 http://eden.1112223333.com/ 的目錄是開的,這樣沒問題嗎
目前来看是这样 99comic和99770卡ip 大陆和港台ip都上不去。。。dmeden.net 港台以外ip倒能上
回覆刪除应该不是一家,汗汗这边图片的地址不同
如果都指向一个的话,这样可以写一个解析头,下面用其它可上站的图片地址去代替,
回覆刪除已於 2016.8.22 支援 hhxiee。
刪除布卡漫画的目前无法下载,会在完全没有下载的情况下显示下载完毕。
回覆刪除已於2016.9.27修正
刪除請問能把http://www.aacomic.com/
回覆刪除這個網站
加入支援的網域嗎?
想加入它的原因是?有漫畫只在這個網站有嗎?
刪除已不能在tw.ikanman.com下載, 請查看一下. 謝謝
回覆刪除怎麼個不能下載法?網址?錯誤訊息?
刪除網址是http://tw.ikanman.com/comic/20013/
刪除錯誤訊息是 -
Traceback (most recent call last):
File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\core\__init__.py", line 428, in error_loop
process()
File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\core\__init__.py", line 378, in download
crawler.init()
File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\core\__init__.py", line 201, in init
self.init_images(self.ep.current_page - 1)
File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\core\__init__.py", line 208, in init_images
self.get_images()
File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\core\__init__.py", line 340, in get_images
self.ep.current_url
File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\mods\seemh.py", line 97, in get_images
return ctx.call("getFiles", path, files, host)
File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\execjs\_abstract_runtime_context.py", line 37, in call
return self._call(name, *args)
File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\execjs\_external_runtime.py", line 87, in _call
return self._eval("{identifier}.apply(this, {args})".format(identifier=identifier, args=args))
File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\execjs\_external_runtime.py", line 73, in _eval
return self.exec_(code)
File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\execjs\_abstract_runtime_context.py", line 18, in exec_
return self._exec_(source)
File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\execjs\_external_runtime.py", line 83, in _exec_
return self._extract_result(output)
File "c:\users\aabc-3\appdata\local\programs\python\python35-32\lib\site-packages\execjs\_external_runtime.py", line 165, in _extract_result
raise exceptions.ProgramError(value)
execjs._exceptions.ProgramError: TypeError: 物件不支援此屬性或方法
請查看一下. 謝謝
有裝 Node.js 嗎?
刪除https://github.com/eight04/ComicCrawler#install-nodejs
裝完 Node.js後可以了. 謝謝
刪除這個www.cartoonmad.com
回覆刪除能加入支援嗎?
五月妹殺目前似乎這裡才有更新
沒學過程式的人有辦法爬圖嗎...
已於 2016.11.25 加入
刪除> 沒學過程式的人有辦法爬圖嗎
對圖片點右鍵 -> 另存圖片 -> 換下一頁
程式只是把這一系列的動作自動化了而已
感謝
刪除話說有沒有辦法直接執行啊?
比如.exe之類的
可以點兩下就開啟主視窗的
想用.bat結果失敗
不會寫執行指令...
把這行指令存成.bat檔就可以
刪除comiccrawler gui
......看來是我的電腦有問題
刪除同一個執行檔
之前不行
現在可以...
感謝
後面的CMD不能關的樣子...
如果不喜歡的話可以用些工具把它縮到系統列,如
刪除http://rbtray.sourceforge.net/
了解
刪除感謝
刚刚发现ikanman.com下载时会出现错误。
回覆刪除错误讯息如下:
Traceback (most recent call last):
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\core\__init__.py", line 428, in error_loop
process()
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\core\__init__.py", line 404, in download
crawler.save_image()
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\core\__init__.py", line 282, in save_image
content_write(self.get_full_filename(), self.image_bin)
File "c:\users\windows 7\appdata\local\programs\python\python35-32\lib\site-packages\comiccrawler\core\__init__.py", line 258, in get_full_filename
raise Exception("Bad file type: " + self.image_ext)
Exception: Bad file type: .a
重新安装Note.js后依然无法下载。
網址?
刪除这一个:http://www.ikanman.com/comic/2817/
刪除是某一卷的的某一頁嗎?目前我跑到04卷都正常
刪除已於 2016.12.1 修正
刪除好快!!
刪除不过已经成功了,真的很感谢你!:)
作者已經移除這則留言。
回覆刪除http://tw.ikanman.com/comic/7506/
回覆刪除下載下來的檔案副檔名會變成.webp
不能看
Note.js有裝
.webp 是一種較新的影像格式︰https://zh.wikipedia.org/zh-tw/WebP
刪除ikanman 在最近的更新後,預設採用 .webp 格式︰https://github.com/eight04/ComicCrawler/issues/27
如果不能看,請先確定看圖軟體是否有支援 .webp 檔。若是不支援,建議換另一種看圖軟體,或者將 .webp 檔轉成 .png 檔。
原來webp是圖片檔...
刪除是只有ikanman才採用webp?
看來只能轉檔了...
看不太懂問題
刪除## 如果是問︰ikanman 只有提供 webp 格式的圖嗎?
NO,ikanman 會檢查瀏覽器是否支援 webp 來決定要載入哪種格式的圖。Chrome 有支援 webp 就會得到 webp 格式的圖,Firefox 則是 jpg 或 gif。
## 如果是問︰除了 ikanman 還有其它網站提供 webp 圖嗎?
當然有。不過對 Comic Crawler 來說,目前有人回報下載出問題的就只有 ikanman,Comic Crawler 也是在那以後才加入 webp 支援的。但也說不定有其它網站也有只是沒人回報而已。
順帶一提,我目前用 HoneyView 看圖,有支援 webp 檔。
以為webp不是圖檔
刪除因為我的看圖軟體不支援webp
可是好像又沒辦法接下載jpg或gif檔
所以才問的
感谢作者制作如此便利的抓图程序!
回覆刪除关于E绅士,有些抓图的设计是否能有些改进,提高抓图的技巧!
例如:https://exhentai.org/g/563634/e943e48bc0/
Up主的贴图顺序是错误的,https://exhentai.org/s/2fcc0d00f4/563634-5
而E绅士给作品按原作品排序就变成了----pag5:046jpg
即:原作品的46页贴在了第5页的位置,从而comiccrawler抓图也就把046jpg抓成pag5[image_005.jpg]。
一句话,在E绅士抓图是按贴图顺序抓而不是按作品原书页顺序抓,造成抓图下来的阅读错误。
问:作者能否提供些技巧或设计,让comiccrawler按E绅士就原作品的书页顺序抓图[E绅士好像对每部本本都有正 确的排序,单图打开就看到图片的详细信息,如https://exhentai.org/s/2fcc0d00f4/563634-5 046.jpg :: 1109 x 1600 :: 482.4 KB],而不是up主的贴图顺序抓图。避免人为的错序阅读,也给众绅士良好的阅读体验。
谢谢此程序作者!
那不是順序而是上傳時的原始檔名吧。
刪除我覺得應該可行,不過要確保原始檔名不會重覆,否則重覆的就不會被下載了。
汉化本应该不会有重复的档名,UP主拿到汉化本不会乱改档名,除非盗源的不经过同意乱发布[UP]的。贴图顺序乱了我猜是同时上传时,部分文件完成的快慢造成的!如果能直接就原始档案抓图那就最好了,可以避免多数抓图后的阅读错误!
刪除多谢指教!
对于就原始档名抓图,有个例子!
刪除https://exhentai.org/g/434036/6a2793be69/
https://exhentai.org/g/598770/7eabe03dce/
可以看出大多数UP主不会乱改原始档名,就原黑白图进行的彩图修改部分也会跟黑白部分原始档名相同,便于替换,有够佛心。
其實並不是貼圖順序亂了,而是 003~043 的頁數缺失。
刪除已於 2016.12.20 加入。
如果要指定存檔時使用原檔名,在設定檔中加上 originalfilename = true
程序更新后,在设定档摸索一下后[originalfilename 原来要在cook的上一行才能解析],终于能就原始档名抓图了。简单粗暴,好用到不行!
刪除再次感谢作者!:)
設定檔中同個區段內的項目順序是不影響的
刪除作者已經移除這則留言。
回覆刪除下載得到的檔名與路徑有簡體字的話
回覆刪除有些不支援 unicode 檔名的壓縮/解壓縮軟體或圖片瀏覽器會出錯
希望 ComicCrawler 能內建檔名簡繁轉換機制
這功能有點微妙,我傾向於不做這個功能,原因如下︰
刪除1. 繁簡轉換不對等。簡轉繁是一對多,繁轉簡是多對一。
2. 若是包含非簡體字的 Unicode 字元(如︰日文),簡繁轉換後這些字元還是會留在檔名裡
3. 治標不治本。就算 Comic Crawler 可以簡繁轉換,下次用其它軟體無法簡繁轉換怎麼辦?
我建議直接將無法支援 Unicode 的軟體淘汰。目前我用 Bandizip 和 Honeyview,都是同一家公司出品。
原來早就完整評估過這個問題了
刪除以下載漫畫的情況來說,遇到日文的情況比較沒那麼多
而簡轉繁一對多在沒有詞庫輔助的情況下也確實容易出錯
對問題有如此嚴謹精確的考慮態度,真是令人信賴!
可是在這裡要求的僅是漫畫檔名上的用途
而非學術或文學、公文等正式文件
相信鄉民們在這娛樂用途上不會去計較
就是圖個方便而已 (不用再多一個第三方軟體來處理檔名的動作)
一定要留日文的話,當然會採用支援 unicode 的軟體 (或者 rename 為中文譯名)
即便除去 unicode 問題,看到簡體字仍然令人頭痛啊 (雖然有些漢化組採用簡體)
在這個層次上,治標治本的方式也許是把簡體字消滅 XD
您建議的軟體 Honeyview 我也有用,他最大的缺點就是 made in korea (誤)
自己改變軟體習慣容易,把檔案分享給朋友要同時改變他們的習慣困難...
無論如何,感謝分享與維護 ComicCrawler
http://www.dm5.com/manhua-womendejiji/
回覆刪除這個無法下載
---------------------------
動漫屋
---------------------------
解析錯誤!
substring not found
---------------------------
確定
---------------------------
跑出這個
已經更新到最新
我點進去長這樣︰
刪除http://i.imgur.com/9O8XeWx.png
所以這種無法下載囉?
刪除那真不懂他更新做什麼......
您好,不好意思打擾了,請問您若有空閒,能不能增加"http://www.aacomic.com",這個網站的下載辨識呢?謝謝您。
回覆刪除原因是?有漫畫是其它地方沒有的嗎?
刪除的確,您會讀心術嗎,哈哈,有一些中斷沒中文,在那裡可以找到日文版本全集;不急啦,只是一種收集的想法而已,謝謝。
刪除檢查後發現 www.aacomic.com 和 www.iibq.com 共用同一個伺服器。你能檢查你找的漫畫有沒有在 iibq.com 上嗎?如果沒有,則把網址貼上來。
刪除我用火狐到iibq常常點進去"某些"漫畫的頁數就出現以下訊息:
刪除『已知的有害網頁!
這個在 www.iibq.com 的網頁已被回報為有害網頁,依據您所選擇的安全設定予以封鎖。
有害網頁會嘗試安裝能竊取隱私資訊、利用您的電腦攻擊他人或破壞作業系統等的惡意軟體到您的電腦上。
某些有害網頁會故意安裝有害軟體到電腦上,但更多網頁是在連網頁擁有者都不知情的情況下,成為有害軟體散佈的溫床。』
e.g.:http://www.iibq.com/comic/82012133193/
小泉麻將傳說
所以我都不上iibq,雖然它好像跟您說的一樣是同一個伺服器,但我進aa就都正常顯示圖片。
於 2017.4.3 加入。
刪除iibq 大概是評論欄被植入了惡意程式,才會有的頁面有,有的頁面沒有。
了解,謝謝
刪除汗汗好像挂了。下不了 == www.hhssee.com
回覆刪除試試更新主程式
刪除還有https://nodejs.org/en/
最少六版以上
试过了 能分析出漫画 但是下载就出错,,不知为啥
回覆刪除C:\Users\rrt>comiccrawler gui
回覆刪除Start analyzing http://www.hhssee.com/manhua31715.html
Analyzing success!
Start download 测不准的阿波连同学
Start downloading 测不准的阿波连同学
total 7 episode.
Downloading ep 测不准的阿波连同学 001集
Traceback (most recent call last):
File "c:\python34\lib\site-packages\comiccrawler\core\__init__.py", line 567,
in error_loop
process()
File "c:\python34\lib\site-packages\comiccrawler\core\__init__.py", line 517,
in download
crawler.init()
File "c:\python34\lib\site-packages\comiccrawler\core\__init__.py", line 358,
in init
self.init_images(self.ep.current_page - 1)
File "c:\python34\lib\site-packages\comiccrawler\core\__init__.py", line 365,
in init_images
self.get_images()
File "c:\python34\lib\site-packages\comiccrawler\core\__init__.py", line 484,
in get_images
self.ep.current_url
File "c:\python34\lib\site-packages\comiccrawler\mods\hhxiee.py", line 86, in
get_images
return ctx.call("getImages", url, img_name, hd_domain)
File "c:\python34\lib\site-packages\node_vm2\__init__.py", line 166, in call
"args": args
File "c:\python34\lib\site-packages\node_vm2\__init__.py", line 110, in commun
icate
raise VMError(data["error"])
node_vm2.VMError: Index doesn't exist: null
於 2017.4.3 修正。
刪除http://mangafox.me/
回覆刪除請問可以支援這個網站嗎?
有時候要找到無修正圖源(被後製噴霧、笑臉、聖光)不容易,這時候看英文化的漫畫也是一種選擇
也有少數較冷門的漫畫,英文站台釋出速度會比中文站台快
這個站會在圖片下方加上大幅的 banner。有其它選擇嗎?
刪除比較常用的還有
刪除http://mangahere.co
http://mangaseeonline.us
備用的
http://mangapark.me
http://mangatan.com
http://zingbox.me
另外,只能 google site:yoedge.com 搜尋漫畫名,這個沒有首頁的低調站台
也有一些別人沒有的無修正漫畫,希望也有機會能加入支援
在你列的英文網站中,隨便點了幾本,圖都來自 mangastream.com。www.mangahere.co 則是被報為惡意網站。
刪除對於 yoedge.com
> 有一些別人沒有的無修正漫畫
可以舉個例子嗎
www.mangahere.co 在回覆於此之前,仍然是可以正常使用 chrome 瀏覽的。
回覆刪除(沒注意到還有 mangastream.com,立馬 bookmark)
yoedge.com 的話,例如 http://smp.yoedge.com/view/omnibus/1000591 是有更新其他站台所沒有的進度
其他漫畫倒是看過就忘了...
yoedge.com 雖然漫畫總量可能較少,可是不會像一些大站台因畫面尺度刻意修正或停更