mirror of
https://github.com/pyt0xic/pablo-bot.git
synced 2024-11-22 02:09:27 +01:00
Using a better script to get drug data and updated requirements.txt
This commit is contained in:
parent
8243a5ba0f
commit
faefd94f2a
@ -1,2 +1,4 @@
|
|||||||
rasa
|
rasa
|
||||||
pandas
|
pandas
|
||||||
|
beautifulsoup4
|
||||||
|
python-graphql-client
|
||||||
|
35201
tripsit/getAllDrugs.json
35201
tripsit/getAllDrugs.json
File diff suppressed because it is too large
Load Diff
@ -1,27 +0,0 @@
|
|||||||
import requests, json
|
|
||||||
import pandas as pd
|
|
||||||
|
|
||||||
# API URL
|
|
||||||
url = "http://tripbot.tripsit.me/api/tripsit/getAllDrugs"
|
|
||||||
r = requests.get(url)
|
|
||||||
data = r.json()
|
|
||||||
|
|
||||||
# Format dict and load into df
|
|
||||||
data = json.dumps(data["data"][0], indent=2, sort_keys=False, ensure_ascii=False)
|
|
||||||
df = pd.DataFrame.from_dict(json.loads(data), orient="index")
|
|
||||||
|
|
||||||
# Add id for each drug for rasa
|
|
||||||
id = []
|
|
||||||
for x in range(0, len(df)):
|
|
||||||
id.append(x)
|
|
||||||
df["id"] = id
|
|
||||||
|
|
||||||
# Write to JSON file
|
|
||||||
with open("tripsit/getAllDrugs.json", "w") as fp:
|
|
||||||
# Clean NaN values
|
|
||||||
clean_data = {
|
|
||||||
k1: {k: v for k, v in v1.items() if v == v and v is not None}
|
|
||||||
for k1, v1 in df.to_dict("index").items()
|
|
||||||
}
|
|
||||||
# Set ensure_ascii to false to ensure we can keep greek letters (like alpha)
|
|
||||||
fp.write(json.dumps(clean_data, indent=2, ensure_ascii=False))
|
|
22162
ts_pn_data/_cached_pw_substances.json
Normal file
22162
ts_pn_data/_cached_pw_substances.json
Normal file
File diff suppressed because it is too large
Load Diff
558
ts_pn_data/getData.py
Normal file
558
ts_pn_data/getData.py
Normal file
@ -0,0 +1,558 @@
|
|||||||
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
|
# downloads and exports data on all substances from psychonautwiki and tripsit factsheets, combining to form master list with standardized format
|
||||||
|
# prioritizes psychonautwiki ROA info (dose/duration) over tripsit factsheets
|
||||||
|
# pip3 install beautifulsoup4 requests python-graphql-client
|
||||||
|
|
||||||
|
import requests
|
||||||
|
from bs4 import BeautifulSoup
|
||||||
|
from time import time, sleep
|
||||||
|
from python_graphql_client import GraphqlClient
|
||||||
|
import json
|
||||||
|
import os
|
||||||
|
import re
|
||||||
|
import traceback
|
||||||
|
|
||||||
|
headers = {
|
||||||
|
"Access-Control-Allow-Origin": "*",
|
||||||
|
"Access-Control-Allow-Methods": "GET",
|
||||||
|
"Access-Control-Allow-Headers": "Content-Type",
|
||||||
|
"Access-Control-Max-Age": "3600",
|
||||||
|
"User-Agent": "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:52.0) Gecko/20100101 Firefox/52.0",
|
||||||
|
}
|
||||||
|
|
||||||
|
ts_api_url = "https://tripbot.tripsit.me/api/tripsit/getAllDrugs"
|
||||||
|
ps_api_url = "https://api.psychonautwiki.org"
|
||||||
|
ps_client = GraphqlClient(endpoint=ps_api_url, headers=headers)
|
||||||
|
|
||||||
|
|
||||||
|
def substance_name_match(name, substance):
|
||||||
|
"""check if name matches any value in keys we care about"""
|
||||||
|
lower_name = name.lower()
|
||||||
|
return any(
|
||||||
|
[
|
||||||
|
lower_name == substance[key].lower()
|
||||||
|
for key in ["name", "pretty_name"]
|
||||||
|
if key in substance
|
||||||
|
]
|
||||||
|
+ [lower_name == alias.lower() for alias in substance.get("aliases", [])]
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def find_substance_in_data(data, name):
|
||||||
|
return next((s for s in data if substance_name_match(name, s)), None)
|
||||||
|
|
||||||
|
|
||||||
|
roa_name_aliases = {
|
||||||
|
"iv": ["intravenous"],
|
||||||
|
"intravenous": ["iv"],
|
||||||
|
"im": ["intramuscular"],
|
||||||
|
"intramuscular": ["im"],
|
||||||
|
"insufflated": ["snorted"],
|
||||||
|
"snorted": ["insufflated"],
|
||||||
|
"vaporized": ["vapourized"],
|
||||||
|
"vapourized": ["vaporized"],
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
def roa_matches_name(roa, name):
|
||||||
|
aliases = roa_name_aliases.get(name.lower(), [])
|
||||||
|
return roa["name"].lower() == name.lower() or roa["name"].lower() in aliases
|
||||||
|
|
||||||
|
|
||||||
|
# get tripsit data
|
||||||
|
|
||||||
|
|
||||||
|
ts_dose_order = ["Threshold", "Light", "Common", "Strong", "Heavy"]
|
||||||
|
ts_combo_ignore = ["benzos"] # duplicate
|
||||||
|
# prettify names in interaction list
|
||||||
|
ts_combo_transformations = {
|
||||||
|
"lsd": "LSD",
|
||||||
|
"mushrooms": "Mushrooms",
|
||||||
|
"dmt": "DMT",
|
||||||
|
"mescaline": "Mescaline",
|
||||||
|
"dox": "DOx",
|
||||||
|
"nbomes": "NBOMes",
|
||||||
|
"2c-x": "2C-x",
|
||||||
|
"2c-t-x": "2C-T-x",
|
||||||
|
"amt": "aMT",
|
||||||
|
"5-meo-xxt": "5-MeO-xxT",
|
||||||
|
"cannabis": "Cannabis",
|
||||||
|
"ketamine": "Ketamine",
|
||||||
|
"mxe": "MXE",
|
||||||
|
"dxm": "DXM",
|
||||||
|
"pcp": "PCP",
|
||||||
|
"nitrous": "Nitrous",
|
||||||
|
"amphetamines": "Amphetamines",
|
||||||
|
"mdma": "MDMA",
|
||||||
|
"cocaine": "Cocaine",
|
||||||
|
"caffeine": "Caffeine",
|
||||||
|
"alcohol": "Alcohol",
|
||||||
|
"ghb/gbl": "GHB/GBL",
|
||||||
|
"opioids": "Opioids",
|
||||||
|
"tramadol": "Tramadol",
|
||||||
|
"benzodiazepines": "Benzodiazepines",
|
||||||
|
"maois": "MAOIs",
|
||||||
|
"ssris": "SSRIs",
|
||||||
|
}
|
||||||
|
|
||||||
|
ts_response = requests.get(ts_api_url)
|
||||||
|
ts_data = ts_response.json()["data"][0]
|
||||||
|
|
||||||
|
ts_substances_data = list(ts_data.values())
|
||||||
|
|
||||||
|
|
||||||
|
# TS has durations split over a few keys, so this finds or creates the duration for the associated ROA
|
||||||
|
# and adds a new line item
|
||||||
|
def ts_add_formatted_duration(ts_roas, formatted_duration, duration_name):
|
||||||
|
units = formatted_duration.get("_unit", "") or ""
|
||||||
|
if "_unit" in formatted_duration:
|
||||||
|
formatted_duration.pop("_unit")
|
||||||
|
|
||||||
|
def add_to_roa(roa, value):
|
||||||
|
if "duration" not in roa:
|
||||||
|
roa["duration"] = []
|
||||||
|
|
||||||
|
roa["duration"].append({"name": duration_name, "value": value})
|
||||||
|
|
||||||
|
for roa_name, value in formatted_duration.items():
|
||||||
|
value_string = f"{value} {units}".strip()
|
||||||
|
|
||||||
|
# if value present (i.e. just one value for all ROA doses provided above), apply to all ROAs
|
||||||
|
if roa_name == "value":
|
||||||
|
# if TS did not add any doses, do nothing with this value
|
||||||
|
# we could theoretically apply this to all PW doses with missing durations, but we can't be sure
|
||||||
|
# if it applies to all ROAs, so just ignore
|
||||||
|
if not len(ts_roas):
|
||||||
|
break
|
||||||
|
|
||||||
|
for ts_roa in ts_roas:
|
||||||
|
add_to_roa(ts_roa, value_string)
|
||||||
|
|
||||||
|
# add to matching ROA or create new ROA if doesn't exist
|
||||||
|
else:
|
||||||
|
ts_roa = next(
|
||||||
|
(ts_roa for ts_roa in ts_roas if roa_matches_name(ts_roa, roa_name)),
|
||||||
|
None,
|
||||||
|
)
|
||||||
|
# if ROA doesn't exist, make new
|
||||||
|
if not ts_roa:
|
||||||
|
ts_roa = {"name": roa_name}
|
||||||
|
ts_roas.append(ts_roa)
|
||||||
|
|
||||||
|
add_to_roa(ts_roa, value_string)
|
||||||
|
|
||||||
|
|
||||||
|
# get psychonautwiki data
|
||||||
|
|
||||||
|
|
||||||
|
def pw_clean_common_name(name):
|
||||||
|
name = re.sub(r'^"', "", name)
|
||||||
|
name = re.sub(r'"$', "", name)
|
||||||
|
name = re.sub(r'"?\[\d*\]$', "", name)
|
||||||
|
name = re.sub(r"\s*More names\.$", "", name)
|
||||||
|
name = re.sub(r"\.$", "", name)
|
||||||
|
return name.strip()
|
||||||
|
|
||||||
|
|
||||||
|
def pw_should_skip(name, soup):
|
||||||
|
return (
|
||||||
|
name.startswith("Experience:") or len(soup.find_all(text="Common names")) == 0
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
pw_substance_data = []
|
||||||
|
|
||||||
|
if os.path.exists("ts_pn_data/_cached_pw_substances.json"):
|
||||||
|
with open("ts_pn_data/_cached_pw_substances.json") as f:
|
||||||
|
pw_substance_data = json.load(f)
|
||||||
|
|
||||||
|
if not len(pw_substance_data):
|
||||||
|
offset = 0
|
||||||
|
pw_substance_urls_query = (
|
||||||
|
f"{{substances(limit: 250 offset: {offset}) {{name url}}}}"
|
||||||
|
)
|
||||||
|
|
||||||
|
pw_substance_urls_data = ps_client.execute(query=pw_substance_urls_query,)["data"][
|
||||||
|
"substances"
|
||||||
|
]
|
||||||
|
|
||||||
|
offset = 252
|
||||||
|
while offset <= 340:
|
||||||
|
pw_substance_urls_query = (
|
||||||
|
f"{{substances(limit: 1 offset: {offset}) {{name url}}}}"
|
||||||
|
)
|
||||||
|
offset += 1
|
||||||
|
temp_data = ps_client.execute(query=pw_substance_urls_query,)["data"][
|
||||||
|
"substances"
|
||||||
|
]
|
||||||
|
print(temp_data)
|
||||||
|
if temp_data is None:
|
||||||
|
continue
|
||||||
|
pw_substance_urls_data.extend(temp_data)
|
||||||
|
|
||||||
|
for idx, substance in enumerate(pw_substance_urls_data):
|
||||||
|
try:
|
||||||
|
url = substance["url"]
|
||||||
|
substance_req = requests.get(url, headers)
|
||||||
|
substance_soup = BeautifulSoup(substance_req.content, "html.parser")
|
||||||
|
|
||||||
|
name = substance_soup.find("h1", id="firstHeading").text
|
||||||
|
if pw_should_skip(name, substance_soup):
|
||||||
|
print(f"Skipping {name} ({idx + 1} / {len(pw_substance_urls_data)})")
|
||||||
|
continue
|
||||||
|
|
||||||
|
# get aliases text
|
||||||
|
common_names_str = substance_soup.find_all(text="Common names")
|
||||||
|
|
||||||
|
cleaned_common_names = (
|
||||||
|
set(
|
||||||
|
map(
|
||||||
|
pw_clean_common_name,
|
||||||
|
common_names_str[0]
|
||||||
|
.parent.find_next_sibling("td")
|
||||||
|
.text.split(", "),
|
||||||
|
)
|
||||||
|
)
|
||||||
|
if len(common_names_str) > 0
|
||||||
|
else set()
|
||||||
|
)
|
||||||
|
cleaned_common_names.add(substance["name"])
|
||||||
|
# don't include name in list of other common names
|
||||||
|
common_names = sorted(filter(lambda n: n != name, cleaned_common_names))
|
||||||
|
|
||||||
|
# scrape ROAs from page
|
||||||
|
|
||||||
|
def get_data_starting_at_row(curr_row):
|
||||||
|
rows = []
|
||||||
|
while curr_row.find("th", {"class": "ROARowHeader"}):
|
||||||
|
row = {}
|
||||||
|
row["name"] = (
|
||||||
|
curr_row.find("th", {"class": "ROARowHeader"}).find("a").text
|
||||||
|
)
|
||||||
|
|
||||||
|
row_values = curr_row.find("td", {"class": "RowValues"})
|
||||||
|
|
||||||
|
row_value_text = row_values.find_all(text=True, recursive=False)
|
||||||
|
if len(row_value_text):
|
||||||
|
row["value"] = "".join(row_value_text).strip()
|
||||||
|
else:
|
||||||
|
row["value"] = None
|
||||||
|
|
||||||
|
row_note = row_values.find("span")
|
||||||
|
if row_note:
|
||||||
|
row["note"] = re.sub(r"\s*\[\d*\]$", "", row_note.text).strip()
|
||||||
|
|
||||||
|
rows.append(row)
|
||||||
|
|
||||||
|
curr_row = curr_row.find_next("tr")
|
||||||
|
return rows, curr_row
|
||||||
|
|
||||||
|
roas = []
|
||||||
|
|
||||||
|
dose_charts = substance_soup.find_all("tr", {"class": "dosechart"})
|
||||||
|
for dose_chart in dose_charts:
|
||||||
|
table = dose_chart.parent.parent
|
||||||
|
roa_name = table.find("tr").find("a").text
|
||||||
|
if not roa_name:
|
||||||
|
continue
|
||||||
|
|
||||||
|
roa = {
|
||||||
|
"name": roa_name,
|
||||||
|
"dosage": [],
|
||||||
|
"duration": [],
|
||||||
|
}
|
||||||
|
|
||||||
|
# dosage
|
||||||
|
|
||||||
|
curr_row = dose_chart.find_next("tr")
|
||||||
|
roa["dosage"], curr_row = get_data_starting_at_row(curr_row)
|
||||||
|
|
||||||
|
# extract bioavailability
|
||||||
|
if len(roa["dosage"]) and roa["dosage"][0]["name"] == "Bioavailability":
|
||||||
|
bioavailability = roa["dosage"].pop(0)
|
||||||
|
roa["bioavailability"] = bioavailability["value"]
|
||||||
|
|
||||||
|
# duration
|
||||||
|
|
||||||
|
if curr_row.find("th", {"class": "ROASubHeader"}):
|
||||||
|
curr_row = curr_row.find_next("tr")
|
||||||
|
roa["duration"], _ = get_data_starting_at_row(curr_row)
|
||||||
|
|
||||||
|
if not len(roa["dosage"]):
|
||||||
|
roa["dosage"] = None
|
||||||
|
if not len(roa["duration"]):
|
||||||
|
roa["duration"] = None
|
||||||
|
|
||||||
|
roas.append(roa)
|
||||||
|
|
||||||
|
# query PS API for more data on substance
|
||||||
|
|
||||||
|
query = (
|
||||||
|
"""
|
||||||
|
{
|
||||||
|
substances(query: "%s") {
|
||||||
|
name
|
||||||
|
class {
|
||||||
|
chemical
|
||||||
|
psychoactive
|
||||||
|
}
|
||||||
|
tolerance {
|
||||||
|
full
|
||||||
|
half
|
||||||
|
zero
|
||||||
|
}
|
||||||
|
toxicity
|
||||||
|
addictionPotential
|
||||||
|
crossTolerances
|
||||||
|
}
|
||||||
|
}
|
||||||
|
"""
|
||||||
|
% substance["name"]
|
||||||
|
)
|
||||||
|
|
||||||
|
data = ps_client.execute(query=query)["data"]["substances"]
|
||||||
|
if len(data) == 0:
|
||||||
|
continue
|
||||||
|
elif len(data) > 1:
|
||||||
|
# should never happen?
|
||||||
|
print(f"{name} has more than one dataset... investigate why")
|
||||||
|
|
||||||
|
data = data[0]
|
||||||
|
if "name" in data:
|
||||||
|
data.pop("name")
|
||||||
|
|
||||||
|
pw_substance_data.append(
|
||||||
|
{
|
||||||
|
"url": url,
|
||||||
|
"name": name,
|
||||||
|
"aliases": common_names,
|
||||||
|
"roas": roas,
|
||||||
|
"data": data,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
print(
|
||||||
|
f"Done with {name} [{len(roas)} ROA(s)] ({idx + 1} / {len(pw_substance_urls_data)})"
|
||||||
|
)
|
||||||
|
|
||||||
|
except KeyboardInterrupt:
|
||||||
|
print("\nScrape canceled")
|
||||||
|
exit(0)
|
||||||
|
except:
|
||||||
|
print(f"{name} failed:")
|
||||||
|
print(traceback.format_exc())
|
||||||
|
exit(1)
|
||||||
|
|
||||||
|
with open(f"ts_pn_data/_cached_pw_substances.json", "w") as f:
|
||||||
|
f.write(json.dumps(pw_substance_data, indent=2, ensure_ascii=False))
|
||||||
|
|
||||||
|
# combine tripsit and psychonautwiki data
|
||||||
|
|
||||||
|
|
||||||
|
all_substance_names = sorted(
|
||||||
|
set(
|
||||||
|
list(map(lambda s: s.get("name", "").lower(), pw_substance_data))
|
||||||
|
+ list(map(lambda s: s.get("name", "").lower(), ts_substances_data))
|
||||||
|
)
|
||||||
|
)
|
||||||
|
substance_data = []
|
||||||
|
x = 0
|
||||||
|
|
||||||
|
for name in all_substance_names:
|
||||||
|
# find PW substance
|
||||||
|
pw_substance = find_substance_in_data(pw_substance_data, name)
|
||||||
|
# remove to get rid of duplicates in final output
|
||||||
|
if pw_substance:
|
||||||
|
pw_substance_data.remove(pw_substance)
|
||||||
|
else:
|
||||||
|
pw_substance = {}
|
||||||
|
|
||||||
|
# find TS substance
|
||||||
|
ts_substance = find_substance_in_data(ts_substances_data, name)
|
||||||
|
# remove to get rid of duplicates in final output
|
||||||
|
if ts_substance:
|
||||||
|
ts_substances_data.remove(ts_substance)
|
||||||
|
else:
|
||||||
|
ts_substance = {}
|
||||||
|
|
||||||
|
# if no substance found in either dataset, skip
|
||||||
|
if not pw_substance and not ts_substance:
|
||||||
|
continue
|
||||||
|
|
||||||
|
ts_properties = ts_substance.get("properties", {})
|
||||||
|
|
||||||
|
# url will always exist for psychonautwiki substance, so tripsit substance must exist if url is None
|
||||||
|
url = pw_substance.get("url") or f"https://drugs.tripsit.me/{ts_substance['name']}"
|
||||||
|
|
||||||
|
ts_links = ts_substance.get("links", {})
|
||||||
|
experiences_url = ts_links.get("experiences")
|
||||||
|
|
||||||
|
# pick display name from available substances found from both datasets
|
||||||
|
names = list(
|
||||||
|
filter(
|
||||||
|
lambda n: n is not None and len(n) > 0,
|
||||||
|
[pw_substance.get("name"), ts_substance.get("pretty_name")],
|
||||||
|
)
|
||||||
|
)
|
||||||
|
# people use shorter names
|
||||||
|
name = min(names, key=len)
|
||||||
|
|
||||||
|
# lowercase list of all names, excluding chosen name above
|
||||||
|
aliases = set(
|
||||||
|
map(
|
||||||
|
lambda n: n.lower(),
|
||||||
|
filter(
|
||||||
|
lambda n: n is not None and len(n) > 0,
|
||||||
|
[pw_substance.get("name"), ts_substance.get("pretty_name")]
|
||||||
|
+ pw_substance.get("aliases", [])
|
||||||
|
+ ts_substance.get("aliases", []),
|
||||||
|
),
|
||||||
|
)
|
||||||
|
)
|
||||||
|
if name.lower() in aliases:
|
||||||
|
aliases.remove(name.lower())
|
||||||
|
aliases = sorted(aliases)
|
||||||
|
|
||||||
|
summary = ts_properties.get("summary", "").strip()
|
||||||
|
if not len(summary):
|
||||||
|
summary = None
|
||||||
|
|
||||||
|
test_kits = ts_properties.get("test-kits", "").strip()
|
||||||
|
if not len(test_kits):
|
||||||
|
test_kits = None
|
||||||
|
|
||||||
|
ts_bioavailability_str = ts_properties.get("bioavailability", "").strip()
|
||||||
|
ts_bioavailability = {}
|
||||||
|
if len(ts_bioavailability_str):
|
||||||
|
matches = re.findall(
|
||||||
|
r"([a-zA-Z\/]+)[.:\s]+([0-9\.%\s\+/\-]+)", ts_bioavailability_str
|
||||||
|
)
|
||||||
|
if len(matches):
|
||||||
|
for roa_name, value in matches:
|
||||||
|
ts_bioavailability[roa_name.lower()] = value.strip(". \t")
|
||||||
|
|
||||||
|
pw_data = pw_substance.get("data", {})
|
||||||
|
|
||||||
|
classes = pw_data.get("class")
|
||||||
|
toxicity = pw_data.get("toxicity")
|
||||||
|
addiction_potential = pw_data.get("addictionPotential")
|
||||||
|
tolerance = pw_data.get("tolerance")
|
||||||
|
cross_tolerances = pw_data.get("crossTolerances")
|
||||||
|
|
||||||
|
roas = []
|
||||||
|
|
||||||
|
# get PW ROAs
|
||||||
|
pw_roas = pw_substance.get("roas", [])
|
||||||
|
|
||||||
|
# process TS ROAs
|
||||||
|
ts_roas = []
|
||||||
|
|
||||||
|
# TS ROA dosage
|
||||||
|
ts_formatted_dose = ts_substance.get("formatted_dose")
|
||||||
|
if ts_formatted_dose:
|
||||||
|
for roa_name, dose_data in ts_formatted_dose.items():
|
||||||
|
dose_levels = []
|
||||||
|
for dose_level in ts_dose_order:
|
||||||
|
value_string = dose_data.get(dose_level)
|
||||||
|
if value_string is None:
|
||||||
|
continue
|
||||||
|
|
||||||
|
dose_levels.append(
|
||||||
|
{"name": dose_level, "value": value_string,}
|
||||||
|
)
|
||||||
|
|
||||||
|
if len(dose_levels):
|
||||||
|
ts_roas.append({"name": roa_name, "dosage": dose_levels})
|
||||||
|
|
||||||
|
# TS ROA durations
|
||||||
|
ts_formatted_onset = ts_substance.get("formatted_onset")
|
||||||
|
if ts_formatted_onset:
|
||||||
|
ts_add_formatted_duration(ts_roas, ts_formatted_onset, "Onset")
|
||||||
|
|
||||||
|
ts_formatted_duration = ts_substance.get("formatted_duration")
|
||||||
|
if ts_formatted_duration:
|
||||||
|
ts_add_formatted_duration(ts_roas, ts_formatted_duration, "Duration")
|
||||||
|
|
||||||
|
ts_formatted_aftereffects = ts_substance.get("formatted_aftereffects")
|
||||||
|
if ts_formatted_aftereffects:
|
||||||
|
ts_add_formatted_duration(ts_roas, ts_formatted_aftereffects, "After effects")
|
||||||
|
|
||||||
|
# merge PW and TS ROAs
|
||||||
|
# prioritize PW for ROAs but use TS to fill in gaps
|
||||||
|
|
||||||
|
roas.extend(pw_roas)
|
||||||
|
for ts_roa in ts_roas:
|
||||||
|
existing_roa = next(
|
||||||
|
(roa for roa in roas if roa_matches_name(roa, ts_roa["name"])), None
|
||||||
|
)
|
||||||
|
# if ROA does not exist, add
|
||||||
|
if not existing_roa:
|
||||||
|
existing_roa = ts_roa
|
||||||
|
roas.append(existing_roa)
|
||||||
|
# we want bioavailability from below, so don't skip
|
||||||
|
|
||||||
|
# if ROA does not already have bioavailability, try to get from TS
|
||||||
|
if not existing_roa.get("bioavailability"):
|
||||||
|
name_lower = ts_roa["name"].lower()
|
||||||
|
name_aliases = roa_name_aliases.get(name_lower, [])
|
||||||
|
|
||||||
|
alias_found = next(
|
||||||
|
(name_alias in ts_bioavailability for name_alias in name_aliases), None
|
||||||
|
)
|
||||||
|
# TS has bioavailability if name or any name alias is found
|
||||||
|
if name_lower in ts_bioavailability or alias_found:
|
||||||
|
existing_roa["bioavailability"] = ts_bioavailability.get(
|
||||||
|
name_lower
|
||||||
|
) or ts_bioavailability.get(alias_found)
|
||||||
|
|
||||||
|
# if existing ROA is missing dosage and TS has dosage, add
|
||||||
|
if (not existing_roa.get("dosage") or not len(existing_roa["dosage"])) and (
|
||||||
|
"dosage" in ts_roa and ts_roa["dosage"] and len(ts_roa["dosage"])
|
||||||
|
):
|
||||||
|
existing_roa["dosage"] = ts_roa["dosage"]
|
||||||
|
|
||||||
|
# if existing ROA is missing duration and TS has duration, add
|
||||||
|
if (not existing_roa.get("duration") or not len(existing_roa["duration"])) and (
|
||||||
|
"duration" in ts_roa and ts_roa["duration"] and len(ts_roa["duration"])
|
||||||
|
):
|
||||||
|
existing_roa["duration"] = ts_roa["duration"]
|
||||||
|
|
||||||
|
interactions = None
|
||||||
|
combos = ts_substance.get("combos")
|
||||||
|
if combos:
|
||||||
|
interactions = []
|
||||||
|
for key, combo_data in combos.items():
|
||||||
|
if key in ts_combo_ignore:
|
||||||
|
continue
|
||||||
|
|
||||||
|
combo_data["name"] = ts_combo_transformations[key]
|
||||||
|
interactions.append(combo_data)
|
||||||
|
interactions = sorted(interactions, key=lambda i: i["name"])
|
||||||
|
|
||||||
|
substance_data.append(
|
||||||
|
{
|
||||||
|
"id": x,
|
||||||
|
"name": name,
|
||||||
|
"aliases": list(aliases),
|
||||||
|
"aliasesStr": ",".join(aliases),
|
||||||
|
"url": url,
|
||||||
|
"experiencesUrl": experiences_url,
|
||||||
|
"summary": summary,
|
||||||
|
"reagents": test_kits,
|
||||||
|
"classes": classes,
|
||||||
|
"toxicity": toxicity,
|
||||||
|
"addictionPotential": addiction_potential,
|
||||||
|
"tolerance": tolerance,
|
||||||
|
"crossTolerances": cross_tolerances,
|
||||||
|
"roas": roas,
|
||||||
|
"interactions": interactions,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
x += 1
|
||||||
|
|
||||||
|
# output
|
||||||
|
|
||||||
|
|
||||||
|
substances_json = json.dumps(substance_data, indent=2, ensure_ascii=False)
|
||||||
|
with open(f"ts_pn_data/substances_{time()}.json", "w") as f:
|
||||||
|
f.write(substances_json)
|
45372
ts_pn_data/substances_1632949161.9573479.json
Normal file
45372
ts_pn_data/substances_1632949161.9573479.json
Normal file
File diff suppressed because it is too large
Load Diff
Loading…
Reference in New Issue
Block a user