Run processing in thread

This commit is contained in:
Mark Veidemanis 2022-09-04 21:29:00 +01:00
parent 22cef33342
commit db46fea550
Signed by: m
GPG Key ID: 5ACFCEED46C0904F
7 changed files with 152 additions and 85 deletions

21
db.py
View File

@ -1,12 +1,13 @@
import json
from pprint import pprint
import manticoresearch import manticoresearch
from manticoresearch.rest import ApiException from manticoresearch.rest import ApiException
from redis import StrictRedis from redis import StrictRedis
import util import util
from schemas.mc_s import schema from schemas.mc_s import schema
import ujson
from numpy import array_split
from math import ceil
configuration = manticoresearch.Configuration(host="http://monolith-db-1:9308") configuration = manticoresearch.Configuration(host="http://monolith-db-1:9308")
api_client = manticoresearch.ApiClient(configuration) api_client = manticoresearch.ApiClient(configuration)
api_instance = manticoresearch.IndexApi(api_client) api_instance = manticoresearch.IndexApi(api_client)
@ -39,7 +40,7 @@ def store_message(msg):
] ]
body_post = "" body_post = ""
for item in body: for item in body:
body_post += json.dumps(item) body_post += ujson.dumps(item)
body_post += "\n" body_post += "\n"
#print(body_post) #print(body_post)
@ -50,12 +51,15 @@ def store_message(msg):
except ApiException as e: except ApiException as e:
print("Exception when calling IndexApi->bulk: %s\n" % e) print("Exception when calling IndexApi->bulk: %s\n" % e)
async def store_message_bulk(messages): def store_message_bulk(data):
""" """
Store a message into Manticore Store a message into Manticore
:param msg: dict :param msg: dict
""" """
print("BULK", len(messages)) print("BULK", len(data))
split_posts = array_split(data, ceil(len(data) / 10000))
for messages in split_posts:
print("PROCESSING SPLIT OF", len(messages), "MESSAGES")
total = [] total = []
for msg in messages: for msg in messages:
# normalise fields # normalise fields
@ -77,7 +81,7 @@ async def store_message_bulk(messages):
body_post = "" body_post = ""
for item in total: for item in total:
body_post += json.dumps(item) body_post += ujson.dumps(item)
body_post += "\n" body_post += "\n"
#print(body_post) #print(body_post)
@ -87,6 +91,9 @@ async def store_message_bulk(messages):
#print(api_response) #print(api_response)
except ApiException as e: except ApiException as e:
print("Exception when calling IndexApi->bulk: %s\n" % e) print("Exception when calling IndexApi->bulk: %s\n" % e)
print("FINISHED PROCESSING SPLIT")
print("BULK FINISH")
def update_schema(): def update_schema():
pass pass

View File

@ -17,7 +17,7 @@ services:
db: db:
image: manticoresearch/manticore image: manticoresearch/manticore
restart: always restart: always
expose: ports:
- 9308 - 9308
- 9312 - 9312
ulimits: ulimits:

View File

@ -10,6 +10,27 @@ services:
- .env - .env
volumes_from: volumes_from:
- tmp - tmp
depends_on:
- db
db:
image: manticoresearch/manticore
restart: always
ports:
- 9308
- 9312
ulimits:
nproc: 65535
nofile:
soft: 65535
hard: 65535
memlock:
soft: -1
hard: -1
environment:
- MCL=1
volumes:
- ./docker/data:/var/lib/manticore
tmp: tmp:
image: busybox image: busybox

View File

@ -5,3 +5,5 @@ siphashc
aiohttp[speedups] aiohttp[speedups]
python-dotenv python-dotenv
manticoresearch manticoresearch
numpy
ujson

View File

@ -6,4 +6,5 @@ siphashc
aiohttp[speedups] aiohttp[speedups]
python-dotenv python-dotenv
manticoresearch manticoresearch
numpy
ujson

View File

@ -1,11 +1,9 @@
# Python modules can't start with a number... # Python modules can't start with a number...
import json import ujson
import random import random
import string import string
from datetime import datetime from datetime import datetime
from typing import Any, Dict
import treq
from bs4 import BeautifulSoup from bs4 import BeautifulSoup
from siphashc import siphash from siphashc import siphash
@ -14,6 +12,11 @@ import util
from schemas.ch4_s import ATTRMAP from schemas.ch4_s import ATTRMAP
import aiohttp import aiohttp
import asyncio import asyncio
from numpy import array_split
from math import ceil
from concurrent.futures import ProcessPoolExecutor
p = ProcessPoolExecutor(10)
class Chan4(object): class Chan4(object):
""" """
@ -25,7 +28,8 @@ class Chan4(object):
self.log = util.get_logger(name) self.log = util.get_logger(name)
self.api_endpoint = "https://a.4cdn.org" self.api_endpoint = "https://a.4cdn.org"
self.boards = ["out"] #self.boards = ["out", "g", "a", "3", "pol"] #
self.boards = []
self.thread_list = {} self.thread_list = {}
#self.thread_deferreds = [] #self.thread_deferreds = []
@ -48,13 +52,13 @@ class Chan4(object):
await self.get_board_list() await self.get_board_list()
async def get_board_list(self): async def get_board_list(self):
# responses = await self.api_call({"_": "boards.json"}) responses = await self.api_call({"_": "boards.json"})
# for mapped, response in responses: for mapped, response in responses:
# if not response: if not response:
# continue continue
# for board in response["boards"]: for board in response["boards"]:
# self.boards.append(board["board"]) self.boards.append(board["board"])
# self.log.debug(f"Got boards: {self.boards}") self.log.debug(f"Got boards: {self.boards}")
await self.get_thread_lists(self.boards) await self.get_thread_lists(self.boards)
@ -82,19 +86,43 @@ class Chan4(object):
self.log.debug(f"Getting information for threads: {thread_urls}") self.log.debug(f"Getting information for threads: {thread_urls}")
responses = await self.api_call(thread_urls) responses = await self.api_call(thread_urls)
self.log.debug(f"Got information for threads: {thread_urls}") self.log.debug(f"Got information for threads: {thread_urls}")
all_posts = {}
for mapped, response in responses: for mapped, response in responses:
if not response: if not response:
continue continue
board, thread = mapped board, thread = mapped
self.log.debug(f"Got thread content for thread {thread} on board {board}") self.log.debug(f"Got thread content for thread {thread} on board {board}")
await self.handle_posts(board, thread, response["posts"]) all_posts[mapped] = response["posts"]
async def handle_posts(self, board, thread, posts): # Split into 10,000 chunks
for index, post in enumerate(posts): # split_posts = array_split(all_posts, ceil(len(all_posts) / 10000))
posts[index]["type"] = "msg" # print("SPLIT CHUNK", len(split_posts))
# for posts in split_posts:
# with futures.ThreadPoolExecutor(max_workers=6) as executor:
# print("SUBMITTED THREAD FOR", len(posts))
# executor.submit(self.handle_posts, board, thread, posts)
#await self.handle_posts(board, thread, response["posts"])
#await asyncio.sleep(1)
await self.handle_posts_thread(all_posts)
@asyncio.coroutine
def handle_posts_thread(self, posts):
loop = asyncio.get_event_loop()
print("HANDLE POSTS THREAD", len(posts.keys()))
yield from loop.run_in_executor(p, self.handle_posts, posts)
def handle_posts(self, posts):
print("HANDLE POSTS START")
to_store = []
for key, post_list in posts.items():
board, thread = key
print("PROCESSING BOARD", board, "THREAD", thread)
print("POSTS HERE", len(post_list))
for index, post in enumerate(post_list):
posts[key][index]["type"] = "msg"
# Calculate hash for post # Calculate hash for post
post_normalised = json.dumps(post, sort_keys=True) post_normalised = ujson.dumps(post, sort_keys=True)
hash = siphash(self.hash_key, post_normalised) hash = siphash(self.hash_key, post_normalised)
hash = str(hash) hash = str(hash)
redis_key = f"cache.{board}.{thread}.{post['no']}" redis_key = f"cache.{board}.{thread}.{post['no']}"
@ -102,17 +130,17 @@ class Chan4(object):
if key_content: if key_content:
key_content = key_content.decode("ascii") key_content = key_content.decode("ascii")
if key_content == hash: if key_content == hash:
return continue
else: else:
posts[index]["type"] = "update" posts[key][index]["type"] = "update"
db.r.set(redis_key, hash) #db.r.set(redis_key, hash)
for key, value in list(post.items()): for key2, value in list(post.items()):
if key in ATTRMAP: if key2 in ATTRMAP:
post[ATTRMAP[key]] = posts[index][key] post[ATTRMAP[key2]] = posts[key][index][key2]
del posts[index][key] del posts[key][index][key2]
if "ts" in post: if "ts" in post:
old_time = posts[index]["ts"] old_time = posts[key][index]["ts"]
# '08/30/22(Tue)02:25:37' # '08/30/22(Tue)02:25:37'
time_spl = old_time.split(":") time_spl = old_time.split(":")
if len(time_spl) == 3: if len(time_spl) == 3:
@ -120,21 +148,28 @@ class Chan4(object):
else: else:
old_ts = datetime.strptime(old_time, "%m/%d/%y(%a)%H:%M") old_ts = datetime.strptime(old_time, "%m/%d/%y(%a)%H:%M")
new_ts = old_ts.isoformat() new_ts = old_ts.isoformat()
posts[index]["ts"] = new_ts posts[key][index]["ts"] = new_ts
if "msg" in post: if "msg" in post:
soup = BeautifulSoup(posts[index]["msg"], "html.parser") soup = BeautifulSoup(posts[key][index]["msg"], "html.parser")
msg = soup.get_text(separator="\n") msg = soup.get_text(separator="\n")
posts[index]["msg"] = msg posts[key][index]["msg"] = msg
posts[index]["src"] = "4ch" posts[key][index]["src"] = "4ch"
to_store.append(posts[key][index])
# print({name_map[name]: val for name, val in post.items()}) # print({name_map[name]: val for name, val in post.items()})
#print(f"Got posts: {len(posts)}") #print(f"Got posts: {len(posts)}")
await db.store_message_bulk(posts) print("HANDLE POSTS DONE")
db.store_message_bulk(to_store)
print("STORE DB DONE")
async def fetch(self, url, session, mapped): async def fetch(self, url, session, mapped):
async with session.get(url) as response: async with session.get(url) as response:
try:
return (mapped, await response.json()) return (mapped, await response.json())
except:
print("FETCH ERROR")
return (mapped, None)
async def bound_fetch(self, sem, url, session, mapped): async def bound_fetch(self, sem, url, session, mapped):
@ -143,6 +178,7 @@ class Chan4(object):
try: try:
return await self.fetch(url, session, mapped) return await self.fetch(url, session, mapped)
except: except:
print("BOUND ERROR")
return (mapped, None) return (mapped, None)
async def api_call(self, methods={}): async def api_call(self, methods={}):

View File

@ -3,7 +3,7 @@ import logging
log = logging.getLogger("util") log = logging.getLogger("util")
debug = True debug = False
# Color definitions # Color definitions
BLACK, RED, GREEN, YELLOW, BLUE, MAGENTA, CYAN, WHITE = range(8) BLACK, RED, GREEN, YELLOW, BLUE, MAGENTA, CYAN, WHITE = range(8)