1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
|
#!/usr/bin/env python3
import os
import time
import sys
import requests
import json
import subprocess
from lxml import etree
def usage():
print("readms.py command")
print("commands:")
print(" download")
print(" list")
exit(1)
def usage_list():
print("readms.py list <url> [latest]")
print("examples:")
print(" readms.py list \"https://readms.net/manga/a_trail_of_blood\"")
print(" readms.py list \"https://readms.net/manga/a_trail_of_blood\" \"48 - Blood oath\"")
exit(1)
def usage_download():
print("readms.py download <url> <download_dir>")
print("examples:")
print(" readms.py download \"https://readms.net/manga/a_trail_of_blood\" /home/adam/Manga/MangaName")
print("")
print("Note: The manga directory has to exist.")
exit(1)
if len(sys.argv) < 2:
usage()
def download_file(url, save_path):
process = subprocess.Popen(["wget", "-q", "-o", "/dev/null", "-O", save_path, url], stderr=subprocess.PIPE)
_, stderr = process.communicate()
if process.returncode != 0:
print("Failed to download file: {}, error: {}".format(url, stderr.decode('utf-8')))
return False
return True
def list_chapters(url, latest):
response = requests.get(url)
if response.status_code != 200:
print("Failed to list chapters, server responded with status code %d" % response.status_code)
exit(2)
tree = etree.HTML(response.text)
chapters = []
for element in tree.xpath('//table//tr//a'):
if latest and element.text == latest:
break
chapters.append({ "name": element.text.strip(), "url": "https://readms.net" + element.attrib.get("href").strip() })
print(json.dumps({ "items": chapters }))
def download_chapter(url, download_dir):
in_progress_filepath = os.path.join(download_dir, "in_progress")
with open(in_progress_filepath, "w") as file:
file.write(url)
img_number = 1
while True:
response = requests.get(url)
if response.status_code != 200:
print("Failed to list chapters, server responded with status code %d" % response.status_code)
exit(2)
tree = etree.HTML(response.text)
image_sources = tree.xpath('//img[@id="manga-page"]/@src')
if len(image_sources) != 1:
break
image_source = "https:" + image_sources[0]
ext = image_source[image_source.rfind("."):]
image_name = str(img_number) + ext
image_path = os.path.join(download_dir, image_name)
print("Downloading {} to {}".format(image_source, image_path))
if not download_file(image_source, image_path):
exit(1)
next_pages = tree.xpath('//div[@class="page"]//a/@href')
if len(next_pages) != 1:
break
url = "https://readms.net" + next_pages[0]
img_number += 1
with open(os.path.join(download_dir, "finished"), "w") as file:
file.write("1")
os.remove(in_progress_filepath)
command = sys.argv[1]
if command == "list":
if len(sys.argv) < 3:
usage_list()
url = sys.argv[2]
latest = ""
if len(sys.argv) >= 4:
latest = sys.argv[3]
list_chapters(url, latest)
elif command == "download":
if len(sys.argv) < 4:
usage_download()
url = sys.argv[2]
download_dir = sys.argv[3]
download_chapter(url, download_dir)
else:
usage()
|