Dnomd343
2 years ago
9 changed files with 0 additions and 293 deletions
@ -1,68 +0,0 @@ |
|||
#!/usr/bin/env python3 |
|||
# -*- coding: utf-8 -*- |
|||
|
|||
import re |
|||
import json |
|||
from logger import logger |
|||
from bs4 import BeautifulSoup |
|||
|
|||
|
|||
def splitPage(rawHtml: str) -> dict: |
|||
body = BeautifulSoup(rawHtml, 'lxml').body |
|||
script = body.select('script')[5].text |
|||
info = { |
|||
'title': body.select('div[class="size18 w100 text-center lh100 pt30 pb15"]')[0].text.strip(), |
|||
'contents': [x.text.strip() for x in body.select('p[class="content_detail"]')], |
|||
'prePage': body.select('div[class="pt-prechapter"]')[0].a.attrs['href'], |
|||
'nextPage': body.select('div[class="pt-nextchapter"]')[0].a.attrs['href'], |
|||
'preId': re.search(r'window\.__PREVPAGE = "(\d*)"', script)[1], |
|||
'nextId': re.search(r'window\.__NEXTPAGE = "(\d*)"', script)[1], |
|||
'myId': re.search(r'window\.chapterNum = (\d+)', script)[1], |
|||
'caption': re.search(r'window\.chapterName = \'(.+)\'', script)[1], |
|||
} |
|||
if not info['title'].startswith(info['caption']): |
|||
logger.error('Title error -> %s' % info['caption']) |
|||
info['index'] = info['title'].replace(info['caption'], '') |
|||
info.pop('title') |
|||
return info |
|||
|
|||
|
|||
def combinePage(id: str) -> dict: |
|||
page_1 = splitPage(open('./html/%s-1.html' % id).read()) |
|||
page_2 = splitPage(open('./html/%s-2.html' % id).read()) |
|||
|
|||
if not page_1['index'] == '[1/2页]' or not page_2['index'] == '[2/2页]': |
|||
logger.error('Sub page error -> `%s` <-> `%s`' % (page_1['index'], page_2['index'])) |
|||
if not page_1['caption'] == page_2['caption']: |
|||
logger.error('Caption error -> `%s` <-> `%s`' % (page_1['caption'], page_2['caption'])) |
|||
if not page_1['myId'] == page_2['myId']: |
|||
logger.error('Page ID error -> `%s` <-> `%s`' % (page_1['myId'], page_2['myId'])) |
|||
if not page_1['preId'] == page_2['preId']: |
|||
logger.error('Pre page ID error -> `%s` <-> `%s`' % (page_1['preId'], page_2['preId'])) |
|||
if not page_1['nextId'] == page_2['nextId']: |
|||
logger.error('Next page ID error -> `%s` <-> `%s`' % (page_1['nextId'], page_2['nextId'])) |
|||
|
|||
if not page_1['prePage'] == '/novel/57104/read_%s.html' % page_1['preId']: |
|||
logger.warning('Page-1 pre url -> `%s` (ID = %s)' % (page_1['prePage'], id)) |
|||
if not page_1['nextPage'] == '/novel/57104/read_%s/2.html' % page_1['myId']: |
|||
logger.warning('Page-1 next url -> `%s` (ID = %s)' % (page_1['nextPage'], id)) |
|||
if not page_2['prePage'] == '/novel/57104/read_%s.html' % page_2['myId']: |
|||
logger.warning('Page-2 pre url -> `%s` (ID = %s)' % (page_2['prePage'], id)) |
|||
if not page_2['nextPage'] == '/novel/57104/read_%s.html' % page_2['nextId']: |
|||
logger.warning('Page-2 next url -> `%s` (ID = %s)' % (page_2['nextPage'], id)) |
|||
|
|||
return { |
|||
'title': page_1['caption'], |
|||
'preId': page_1['preId'], |
|||
'myId': page_1['myId'], |
|||
'nextId': page_1['nextId'], |
|||
'contents': page_1['contents'] + page_2['contents'] |
|||
} |
|||
|
|||
|
|||
catalog = json.loads(open('./catalog/catalog.json').read()) |
|||
|
|||
for _, pageId in catalog.items(): |
|||
logger.info('Analyse page `%s`' % pageId) |
|||
with open('./json/%s.json' % pageId, 'w') as fileObj: |
|||
fileObj.write(json.dumps(combinePage(pageId))) |
File diff suppressed because one or more lines are too long
@ -1,51 +0,0 @@ |
|||
#!/usr/bin/env python3 |
|||
# -*- coding: utf-8 -*- |
|||
|
|||
import sys |
|||
import json |
|||
import time |
|||
import requests |
|||
from bs4 import BeautifulSoup |
|||
|
|||
basicUrl = 'https://m.wxsy.net/novel/57104/all.html' |
|||
|
|||
userAgent = ( # default user-agent |
|||
'Mozilla/5.0 (Linux; Android 10; moto g(7) play) ' |
|||
'AppleWebKit/537.36 (KHTML, like Gecko) ' |
|||
'Chrome/100.0.4896.79 Mobile Safari/537.36' |
|||
) |
|||
|
|||
|
|||
def httpRequest(url: str) -> str: |
|||
request = requests.get(url, headers = { |
|||
'user-agent': userAgent, # with fake user-agent |
|||
'accept-encoding': 'gzip, deflate', # allow content compress |
|||
}) |
|||
if request.status_code not in range(200, 300): # http status code 2xx |
|||
raise RuntimeError('Http request failed') |
|||
return request.text |
|||
|
|||
|
|||
def analysePage(rawHtml: str) -> list: |
|||
analyseRet = [] |
|||
soup = BeautifulSoup(rawHtml, 'lxml') |
|||
div = soup.select('div[class="border-b"]')[0] |
|||
for row in div.select('a[class="w100 flex-wrp flex-align-center flex-between pt10 pb10"]'): |
|||
analyseRet.append({ |
|||
'name': row.attrs['title'], |
|||
'url': row.attrs['href'], |
|||
}) |
|||
return analyseRet |
|||
|
|||
|
|||
def fetchCatalog(pageNum: int) -> list: |
|||
catalog = [] |
|||
for pageIndex in range(1, pageNum + 1): |
|||
print('Page: %d' % pageIndex, file = sys.stderr) |
|||
pageUrl = '%s?sort=1&page=%d' % (basicUrl, pageIndex) |
|||
catalog.append(analysePage(httpRequest(pageUrl))) |
|||
time.sleep(3) |
|||
return catalog |
|||
|
|||
|
|||
print(json.dumps(fetchCatalog(18))) |
File diff suppressed because one or more lines are too long
@ -1,16 +0,0 @@ |
|||
#!/usr/bin/env python3 |
|||
# -*- coding: utf-8 -*- |
|||
|
|||
import re |
|||
import json |
|||
|
|||
catalog = {} |
|||
for catalogPage in json.loads(open('raw.json').read()): |
|||
for pageInfo in catalogPage: |
|||
pageId = re.search(r'^/novel/57104/read_(\d+)\.html$', pageInfo['url'])[1] |
|||
catalog[pageInfo['name']] = pageId |
|||
|
|||
catalog = sorted(catalog.items(), key = lambda d: int(re.search(r'^第(\d+)章', d[0])[1])) |
|||
catalog = {x[0]: x[1] for x in catalog} |
|||
|
|||
print(json.dumps(catalog)) |
@ -1,58 +0,0 @@ |
|||
#!/usr/bin/env python3 |
|||
# -*- coding: utf-8 -*- |
|||
|
|||
import json |
|||
from logger import logger |
|||
|
|||
|
|||
def loadData(catalog: dict) -> dict: |
|||
data = {} |
|||
for _, pageId in catalog.items(): |
|||
data[pageId] = json.loads( |
|||
open('./json/%s.json' % pageId).read() |
|||
) |
|||
return data |
|||
|
|||
|
|||
def listDiff(list_1: list, list_2: list) -> bool: |
|||
diffFlag = False |
|||
if len(list_1) != len(list_2): |
|||
diffFlag = True |
|||
logger.error('List with different length') |
|||
for i in range(0, len(list_1)): |
|||
if list_1[i] != list_2[i]: |
|||
diffFlag = True |
|||
logger.error('List diff: `%s` <-> `%s`' % (list_1[i], list_2[i])) |
|||
return diffFlag |
|||
|
|||
|
|||
def check(catalog: dict, data: dict) -> None: |
|||
titles = [x['title'] for _, x in data.items()] |
|||
ids = [x['myId'] for _, x in data.items()] |
|||
preIds = [x['preId'] for _, x in data.items()] |
|||
nextIds = [x['nextId'] for _, x in data.items()] |
|||
nextIds.pop(-1) |
|||
preIds.pop(0) |
|||
|
|||
# if listDiff(ids, preIds + [ids[-1]]): |
|||
# logger.warning('Pre IDs mismatch') |
|||
# if listDiff(ids, [ids[0]] + nextIds): |
|||
# logger.warning('Next IDs mismatch') |
|||
if listDiff(ids, [x for _, x in catalog.items()]): |
|||
logger.warning('IDs mismatch') |
|||
if listDiff(titles, [x for x in catalog]): |
|||
logger.warning('Titles mismatch') |
|||
|
|||
|
|||
def combine() -> dict: |
|||
catalog = json.loads(open('./catalog/catalog.json').read()) |
|||
data = loadData(catalog) |
|||
check(catalog, data) |
|||
|
|||
result = {} |
|||
for _, info in data.items(): |
|||
result[info['title']] = info['contents'] |
|||
return result |
|||
|
|||
|
|||
print(json.dumps(combine())) |
@ -1,46 +0,0 @@ |
|||
#!/usr/bin/env python3 |
|||
# -*- coding: utf-8 -*- |
|||
|
|||
import time |
|||
import json |
|||
import requests |
|||
from logger import logger |
|||
|
|||
userAgent = ( # default user-agent |
|||
'Mozilla/5.0 (Linux; Android 10; moto g(7) play) ' |
|||
'AppleWebKit/537.36 (KHTML, like Gecko) ' |
|||
'Chrome/100.0.4896.79 Mobile Safari/537.36' |
|||
) |
|||
|
|||
|
|||
def httpRequest(url: str, fileName: str) -> bool: |
|||
try: |
|||
logger.debug('Http request `%s` -> %s' % (url, fileName)) |
|||
request = requests.get(url, timeout = 30, |
|||
headers = { |
|||
'user-agent': userAgent, # with fake user-agent |
|||
} |
|||
) |
|||
if request.status_code not in range(200, 300): # http status code 2xx |
|||
logger.warning('Http request failed -> %s' % url) |
|||
return False |
|||
logger.debug('Http request success -> %s' % url) |
|||
with open(fileName, 'w') as fileObj: # save html content |
|||
fileObj.write(request.text) |
|||
logger.debug('File save success -> %s' % fileName) |
|||
except: |
|||
return False |
|||
return True |
|||
|
|||
|
|||
catalog = json.loads(open('./catalog/catalog.json').read()) |
|||
|
|||
for _, pageId in catalog.items(): |
|||
for subPage in [1, 2]: |
|||
pageUrl = 'https://m.wxsy.net/novel/57104/read_%s/%d.html' % (pageId, subPage) |
|||
pageFile = './html/%s-%d.html' % (pageId, subPage) |
|||
if httpRequest(pageUrl, pageFile): |
|||
logger.info('Page request success -> %s' % pageUrl) |
|||
else: |
|||
logger.error('Page request failed -> %s' % pageUrl) |
|||
time.sleep(1) |
@ -1,41 +0,0 @@ |
|||
#!/usr/bin/env python3 |
|||
# -*- coding: utf-8 -*- |
|||
|
|||
import sys |
|||
import logging |
|||
import colorlog |
|||
|
|||
logColor = { # log color |
|||
'DEBUG': 'white', |
|||
'INFO': 'green', |
|||
'WARNING': 'yellow', |
|||
'ERROR': 'red', |
|||
'CRITICAL': 'bold_red', |
|||
} |
|||
|
|||
dateFormat = '%Y-%m-%d %H:%M:%S' |
|||
timeFormat = '%(asctime)s.%(msecs)03d' |
|||
logFormat = '[%(levelname)s] %(message)s (%(module)s.%(funcName)s:%(lineno)d)' |
|||
|
|||
# load fileHandler -> log file |
|||
fileHandler = logging.FileHandler('runtime.log', encoding = 'utf-8') |
|||
fileHandler.setFormatter(logging.Formatter( |
|||
'[' + timeFormat + '] ' + logFormat, |
|||
datefmt = dateFormat |
|||
)) |
|||
fileHandler.setLevel(logging.DEBUG) # debug level for log file |
|||
|
|||
# load stdHandler -> stderr |
|||
stdHandler = colorlog.StreamHandler() |
|||
stdHandler.setFormatter(colorlog.ColoredFormatter( |
|||
'%(light_black)s' + timeFormat + '%(log_color)s ' + logFormat, |
|||
datefmt = dateFormat, |
|||
log_colors = logColor, |
|||
stream = sys.stderr |
|||
)) |
|||
stdHandler.setLevel(logging.INFO) # info level for stderr |
|||
|
|||
logger = logging.getLogger() |
|||
logger.addHandler(stdHandler) |
|||
logger.addHandler(fileHandler) |
|||
logger.setLevel(logging.DEBUG) # set log level in handler |
@ -1,11 +0,0 @@ |
|||
#!/usr/bin/env python3 |
|||
# -*- coding: utf-8 -*- |
|||
|
|||
import json |
|||
|
|||
strData = '' |
|||
data = json.loads(open('xxrs.json').read()) |
|||
for title, content in data.items(): |
|||
strData += '%s\n\n' % title |
|||
strData += '%s\n\n\n' % '\n\n'.join(content) |
|||
print(strData.strip()) |
Loading…
Reference in new issue