| import multiprocessing |
| import random |
| import traceback |
| from lxml import etree |
| import pathlib |
| import orjson |
| import tqdm |
| import typer |
| import urllib.parse |
| from sqlitedict import SqliteDict |
|
|
| app = typer.Typer() |
|
|
|
|
| def fast_iter(context, func, *args, **kwargs): |
| """ |
| http://lxml.de/parsing.html#modifying-the-tree |
| Based on Liza Daly's fast_iter |
| http://www.ibm.com/developerworks/xml/library/x-hiperfparse/ |
| See also http://effbot.org/zone/element-iterparse.htm |
| """ |
| for event, elem in context: |
| func(elem, *args, **kwargs) |
| |
| |
| elem.clear() |
| |
| for ancestor in elem.xpath("ancestor-or-self::*"): |
| while ancestor.getprevious() is not None: |
| del ancestor.getparent()[0] |
| del context |
|
|
|
|
| @app.command() |
| def decadence(stack_folder: pathlib.Path, output_file: pathlib.Path): |
| post_history = stack_folder / "PostHistory.xml" |
| if not post_history.exists(): |
| raise Exception() |
| |
| print("Working on", stack_folder.name) |
| sql_dict_posts = SqliteDict( |
| f"tmp/{stack_folder.name}.hist.tmp.sqlite", flag="n", tablename="dict_hist" |
| ) |
| try: |
| with tqdm.tqdm() as pbar: |
| ctx = etree.iterparse(post_history, tag="row") |
|
|
| def element_processor2(elem): |
| p_type = int(elem.get("PostHistoryTypeId")) |
| pid: str = elem.get("PostId") |
| if p_type in range(1, 4): |
| if pid in sql_dict_posts: |
| post_data = sql_dict_posts[pid] |
| else: |
| post_data = sql_dict_posts[pid] = {} |
| if p_type == 1: |
| post_data["title"] = urllib.parse.unquote( |
| elem.get("Text", "") |
| ).replace("\r\n", "\n") |
| elif p_type == 2: |
| post_data["body"] = urllib.parse.unquote( |
| elem.get("Text", "") |
| ).replace("\r\n", "\n") |
| elif p_type == 3: |
| post_data["tags"] = urllib.parse.unquote( |
| elem.get("Text", "") |
| ).replace("\r\n", "\n") |
| |
| sql_dict_posts[pid] = post_data |
| pbar.update(1) |
| elif p_type in range(4, 10): |
| if pid in sql_dict_posts: |
| post_data = sql_dict_posts[pid] |
| else: |
| |
| post_data = sql_dict_posts[pid] = {} |
| post_data["partial"] = True |
| if p_type in [4, 7]: |
| post_data["title"] = urllib.parse.unquote( |
| elem.get("Text", "") |
| ).replace("\r\n", "\n") |
| elif p_type in [5, 8]: |
| post_data["body"] = urllib.parse.unquote( |
| elem.get("Text", "") |
| ).replace("\r\n", "\n") |
| elif p_type in [6, 9]: |
| post_data["tags"] = urllib.parse.unquote( |
| elem.get("Text", "") |
| ).replace("\r\n", "\n") |
| sql_dict_posts[pid] = post_data |
| pbar.update(1) |
| elif p_type == 35: |
| if pid in sql_dict_posts: |
| post_data = sql_dict_posts[pid] |
| else: |
| print( |
| f"[W] {pid}, {stack_folder.name} requested to be redirected but redirect doesn't seem to exist?" |
| ) |
| post_data = {} |
| if not post_data.get("partial", False): |
| print(f"[I] {pid}, {stack_folder.name} Not partial?") |
| return |
| print(post_data) |
| elif p_type == 10: |
| if pid not in sql_dict_posts: |
| return |
| post_data = sql_dict_posts[pid] |
| if not post_data: |
| raise Exception |
| post_data["closed"] = True |
| sql_dict_posts[pid] = post_data |
| pbar.update(1) |
| elif p_type == 11: |
| if pid not in sql_dict_posts: |
| return |
| post_data = sql_dict_posts[pid] |
| if not post_data: |
| raise Exception |
| post_data["closed"] = False |
| sql_dict_posts[pid] = post_data |
| pbar.update(1) |
| elif p_type == 12: |
| if pid not in sql_dict_posts: |
| return |
| post_data = sql_dict_posts[pid] |
| if not post_data: |
| raise Exception |
| post_data["deleted"] = True |
| sql_dict_posts[pid] = post_data |
| pbar.update(1) |
| elif p_type == 13: |
| if pid not in sql_dict_posts: |
| return |
| post_data = sql_dict_posts[pid] |
| if not post_data: |
| raise Exception |
| post_data["deleted"] = False |
| sql_dict_posts[pid] = post_data |
| pbar.update(1) |
| elif p_type == 14: |
| if pid not in sql_dict_posts: |
| return |
| post_data = sql_dict_posts[pid] |
| if not post_data: |
| raise Exception |
| post_data["locked"] = True |
| sql_dict_posts[pid] = post_data |
| pbar.update(1) |
| elif p_type == 15: |
| if pid not in sql_dict_posts: |
| return |
| post_data = sql_dict_posts[pid] |
| if not post_data: |
| raise Exception |
| post_data["locked"] = False |
| sql_dict_posts[pid] = post_data |
| pbar.update(1) |
| |
| if pbar.n % 5000 == 0 and pbar.n != 0: |
| |
| sql_dict_posts.commit() |
| |
|
|
| fast_iter(ctx, element_processor2) |
| except Exception as e: |
| print("[!] ERR: ", traceback.format_exception(e)) |
| flagged = set() |
| for k, v in sql_dict_posts.items(): |
| if not v.get("body").strip(): |
| flagged.add(k) |
| if v.get("deleted", False): |
| flagged.add(k) |
| for fg in flagged: |
| del sql_dict_posts[fg] |
| sql_dict_posts.sync() |
| print("Reconstruct done!") |
|
|
| posts = stack_folder / "Posts.xml" |
| if not posts.exists(): |
| raise Exception() |
| |
| main_qn_posts = {} |
| try: |
| |
| with tqdm.tqdm() as bar: |
| ctx2 = etree.iterparse(posts, tag="row") |
|
|
| def element_processor(element): |
| pid = element.get("Id") |
| p_type = int(element.get("PostTypeId")) |
| |
| if p_type == 1: |
| if pid not in sql_dict_posts: |
| main_qn_posts[pid] = {"Stub":True} |
| print( |
| f"[!] Question: {int(element.get('Id'))} {stack_folder.name} does not exist but referanced?" |
| ) |
| return |
| main_qn_posts[pid] = sql_dict_posts[pid] |
| main_qn_posts[pid]["Stub"] = False |
| main_qn_posts[pid]["Id"] = pid |
| main_qn_posts[pid]["CreationDate"] = element.get("CreationDate") |
| main_qn_posts[pid]["Score"] = int(element.get("Score")) |
| main_qn_posts[pid]["Accepted"] = ( |
| int(element.get("AcceptedAnswerId")) |
| if element.get("AcceptedAnswerId") |
| else None |
| ) |
| main_qn_posts[pid]["Counts"] = { |
| "Views": int(element.get("ViewCount")) |
| if element.get("ViewCount") |
| else 0, |
| "Answers": int(element.get("AnswerCount")) |
| if element.get("AnswerCount") |
| else 0, |
| "Comments": int(element.get("CommentCount")) |
| if element.get("CommentCount") |
| else 0, |
| } |
| bar.update(1) |
|
|
| fast_iter(ctx2, element_processor) |
| except Exception as e: |
| print("[!] ERR: ", traceback.format_exception(e)) |
|
|
| print("2nd Pass Posts Done") |
| |
| try: |
| with tqdm.tqdm() as bar: |
| ctx3 = etree.iterparse(posts, tag="row") |
|
|
| def element_processor3(element): |
| pid = element.get("Id") |
| p_type = int(element.get("PostTypeId")) |
| if p_type == 2: |
| parent_id = element.get("ParentId") |
| if parent_id not in main_qn_posts: |
| print( |
| f"[!] Answer: {int(element.get('Id'))} {stack_folder.name} has no parent attached to it!" |
| ) |
| return |
| answers = main_qn_posts[parent_id].setdefault("answers", []) |
| if pid in sql_dict_posts: |
| rec_answer = sql_dict_posts[pid] |
| else: |
| rec_answer = None |
| if rec_answer is None: |
| print( |
| f"[!] Answer: {int(element.get('Id'))} {stack_folder.name} does not have a PostHistory.xml Assoc. with it!" |
| ) |
| main_qn_posts[parent_id]["answers"] = answers |
| return |
| rec_answer["Id"] = int(element.get("Id")) |
| rec_answer["Score"] = ( |
| int(element.get("Score")) if element.get("Score") else 0 |
| ) |
| rec_answer["Counts"] = { |
| "Views": int(element.get("ViewCount")) |
| if element.get("ViewCount") |
| else 0, |
| "Comments": int(element.get("CommentCount")) |
| if element.get("CommentCount") |
| else 0, |
| } |
| answers.append(rec_answer) |
| main_qn_posts[parent_id]["answers"] = answers |
|
|
| fast_iter(ctx3, element_processor3) |
| except Exception as e: |
| print("[!] ERR: ", traceback.format_exception(e)) |
| print("2nd Pass done!") |
| try: |
| with open(output_file, "wb") as f: |
| for k, v in main_qn_posts.items(): |
| f.write(orjson.dumps(v) + b"\n") |
| sql_dict_posts.close() |
| pathlib.Path(f"tmp/{stack_folder.name}.hist.tmp.sqlite").unlink() |
| except Exception as e: |
| print("[!] ERR: ", traceback.format_exception(e)) |
|
|
|
|
| @app.command() |
| def convo_processor(base_file: pathlib.Path, qa_file: pathlib.Path): |
| with open(qa_file, "wb") as fw: |
| with open(base_file, "rb") as f: |
| for line in f: |
| data = orjson.loads(line) |
| if not data.get("body") or not data.get("title"): |
| continue |
| if data["body"].lower().startswith(data["title"].lower()): |
| question = f'{data["title"].strip()}\n\n{data["body"][len(data["title"]) :].strip()}' |
| else: |
| question = f'{data["title"].strip()}\n\n{data["body"].strip()}' |
| if data.get("answers"): |
| for answer in data.get("answers"): |
| if "Score" not in answer: |
| print("no score?",data) |
| continue |
| if answer["Score"] >= 0: |
| zz = { |
| "conversation": [ |
| {"sender": "user", "message": question}, |
| {"sender": "assistant", "message": answer}, |
| ], |
| "meta": { |
| "q_score": data["Score"], |
| "a_score": answer["Score"], |
| "s_score": data["Score"] + answer["Score"] |
| }, |
| } |
| fw.write(orjson.dumps(zz) + b"\n") |
| pass |
|
|
|
|
| @app.command() |
| def qa_processor(base_file: pathlib.Path, qa_file: pathlib.Path, formats: pathlib.Path): |
| question_formats = orjson.loads(formats.read_bytes()) |
| with open(qa_file, "wb") as fw: |
| with open(base_file, "rb") as f: |
| for line in f: |
| data = orjson.loads(line) |
| if not data.get("body") or not data.get("title"): |
| continue |
| if data["body"].lower().startswith(data["title"].lower()): |
| question = f'{data["title"].strip()}\n\n{data["body"][len(data["title"]) :].strip()}' |
| else: |
| question = f'{data["title"].strip()}\n\n{data["body"].strip()}' |
| if data.get("answers"): |
| for answer in data.get("answers"): |
| if answer["Score"] >= 0: |
| fmt = random.choice(question_formats) |
| fmt = fmt.format(question=question, answer=answer["body"]) |
| zz = { |
| "text": fmt, |
| "meta": { |
| "q_score": data["Score"], |
| "a_score": answer["Score"], |
| }, |
| } |
| fw.write(orjson.dumps(zz) + b"\n") |
| pass |
|
|
| def err_handler(e): |
| traceback.print_exception(e) |
|
|
| @app.command() |
| def convo_stackstack(folder: pathlib.Path): |
| with multiprocessing.Pool(processes=64) as pool: |
| fn = [] |
| for item in folder.iterdir(): |
| if item.is_file() and item.suffix.endswith("jsonl"): |
| |
| fn.append( |
| pool.apply_async( |
| convo_processor, |
| ( |
| item, |
| pathlib.Path("convo") |
| / pathlib.Path(f"{item.stem}_convo.jsonl") |
| ), |
| error_callback=err_handler) |
| ) |
| for f in fn: |
| f.wait() |
| pool.close() |
| pool.join |
|
|
| @app.command() |
| def qa_stackstack(folder: pathlib.Path): |
| with multiprocessing.Pool(processes=64) as pool: |
| fn = [] |
| for item in folder.iterdir(): |
| if item.is_file() and item.suffix.endswith("jsonl"): |
| |
| fn.append( |
| pool.apply_async( |
| qa_processor, |
| ( |
| item, |
| pathlib.Path("qa") |
| / pathlib.Path(f"{item.stem}_qa.jsonl"), |
| pathlib.Path("staccato_format.json") |
| ), |
| error_callback=err_handler) |
| ) |
| for f in fn: |
| f.wait() |
| pool.close() |
| pool.join() |
|
|
|
|
| @app.command() |
| def stackstack(folder: pathlib.Path): |
| with multiprocessing.Pool(processes=64) as pool: |
| fn = [] |
| for item in folder.iterdir(): |
| if item.is_dir() and "meta." not in item.name: |
| |
| fn.append( |
| pool.apply_async( |
| decadence, |
| ( |
| item, |
| pathlib.Path("parsed") |
| / pathlib.Path(f"{item.name}_raw.jsonl"), |
| ), |
| ) |
| ) |
| for f in fn: |
| f.wait() |
| pool.close() |
| pool.join() |
|
|
|
|
| app() |
|
|