mirror of
https://github.com/Brandon-Rozek/website.git
synced 2024-11-23 16:46:34 -05:00
Updated submodules
This commit is contained in:
parent
06c5334df3
commit
890bc42226
3 changed files with 2 additions and 275 deletions
|
@ -1 +1 @@
|
|||
Subproject commit d382a9411c539b36139df5d1d25eb09fe6602fdb
|
||||
Subproject commit c463b7fb1db83679bb0c95e1c7822ccfaa6c0c82
|
|
@ -1 +1 @@
|
|||
Subproject commit b1f6da3c04d68dc8cdbeaaf9392c8f8ef6d80586
|
||||
Subproject commit 04568107105e3663b3f79321d95bfd4f3a6de9af
|
|
@ -1,273 +0,0 @@
|
|||
#!/usr/bin/env python
|
||||
"""
|
||||
Script to create Hugo markdown
|
||||
files from iNaturalist Observations
|
||||
"""
|
||||
|
||||
from urllib import request
|
||||
from glob import glob
|
||||
from http.client import HTTPResponse
|
||||
from pathlib import Path
|
||||
from typing import Any, Dict, Optional
|
||||
import json
|
||||
import sys
|
||||
import time
|
||||
|
||||
CONTENT_LOCATION = "content/observations"
|
||||
USER_ID = "brandonrozek"
|
||||
MIN_OBS_ID = -1
|
||||
|
||||
|
||||
def retrieve_data_from_server():
|
||||
server_data = []
|
||||
server_ids = retrieve_obs_ids_from_server()
|
||||
time.sleep(1) # 60 requests / second cap
|
||||
for id_num in server_ids:
|
||||
# Grab observation from iNaturalist
|
||||
url = f"https://api.inaturalist.org/v1/observations/{id_num}"
|
||||
response: Optional[HTTPResponse] = None
|
||||
|
||||
try:
|
||||
response = request.urlopen(url)
|
||||
except Exception:
|
||||
print(f"Unable to grab observation {id_num} from iNaturalist.")
|
||||
|
||||
if response is None:
|
||||
continue
|
||||
|
||||
# Parse server response
|
||||
server_data_part = None
|
||||
try:
|
||||
server_data_part = json.loads(response.read())['results'][0]
|
||||
# Note: there is only one observation as a result
|
||||
except Exception:
|
||||
print(f"Malformed JSON response for observation {id_num}.")
|
||||
continue
|
||||
|
||||
server_data_part = reformat_obs(id_num, server_data_part)
|
||||
server_data.append(server_data_part)
|
||||
time.sleep(1) # 60 requests / second cap
|
||||
|
||||
print(f"Successfully obtained {len(server_data)} observations from the server.")
|
||||
return server_data
|
||||
|
||||
|
||||
def retrieve_obs_ids_from_server():
|
||||
"""
|
||||
Grabs observation ids from iNaturalist server
|
||||
"""
|
||||
global MIN_OBS_ID
|
||||
server_data = []
|
||||
|
||||
finished_retrieving = False
|
||||
while not finished_retrieving:
|
||||
# Grab observations from iNaturalist
|
||||
id_below = "&id_below=" + str(MIN_OBS_ID) \
|
||||
if MIN_OBS_ID > 0 else ""
|
||||
url = "https://api.inaturalist.org/v1/observations?order=desc&order_by=created_at&only_id=true&user_id=" + USER_ID + id_below
|
||||
response: Optional[HTTPResponse] = None
|
||||
|
||||
try:
|
||||
response = request.urlopen(url)
|
||||
except Exception:
|
||||
print("Unable to grab observations from iNaturalist.")
|
||||
|
||||
if response is None:
|
||||
sys.exit(-1)
|
||||
|
||||
# Parse server response
|
||||
server_data_part: Optional[list] = None
|
||||
try:
|
||||
server_data_part = json.loads(response.read())['results']
|
||||
except Exception:
|
||||
print("Malformed JSON response from server.")
|
||||
|
||||
if server_data is None:
|
||||
sys.exit(-1)
|
||||
|
||||
if not isinstance(server_data_part, list):
|
||||
print("Unexpected JSON response, should be of form list.")
|
||||
sys.exit(-1)
|
||||
|
||||
# No more to retrieve
|
||||
if len(server_data_part) == 0:
|
||||
finished_retrieving = True
|
||||
break
|
||||
|
||||
server_data_part = [d['id'] for d in server_data_part]
|
||||
|
||||
# print(f"Retrieved {len(server_data_part)} observations from server")
|
||||
server_data.extend(server_data_part)
|
||||
MIN_OBS_ID = server_data_part[-1]
|
||||
|
||||
print(f"Parsed a total of {len(server_data)} ids from server")
|
||||
return server_data
|
||||
|
||||
|
||||
def reformat_obs(obsid, obs_json):
|
||||
"""
|
||||
Takes a obs_json and
|
||||
slightly modifies it to match
|
||||
some of the fields Hugo expects.
|
||||
"""
|
||||
obs_data = dict(
|
||||
id=str(obsid),
|
||||
metadata={},
|
||||
content=""
|
||||
)
|
||||
|
||||
# Turn URL -> Syndication
|
||||
obs_data['metadata']['syndication'] = obs_json['uri']
|
||||
|
||||
# Turn Created At -> Date
|
||||
obs_data['metadata']['date'] = obs_json['time_observed_at']
|
||||
|
||||
# Grab some taxonomy information about the organism
|
||||
obs_data['metadata']['taxon'] = dict(
|
||||
name=obs_json['taxon']['name'],
|
||||
common_name=obs_json['taxon']['preferred_common_name']
|
||||
)
|
||||
|
||||
# Grab only a few fields
|
||||
desired_fields = [
|
||||
'quality_grade', 'identifications_most_agree',
|
||||
'species_guess', 'identifications_most_disagree',
|
||||
'captive', 'project_ids',
|
||||
'community_taxon_id', 'geojson',
|
||||
'owners_identification_from_vision',
|
||||
'identifications_count', 'obscured',
|
||||
'num_identification_agreements',
|
||||
'num_identification_disagreements',
|
||||
'place_guess', "photos"
|
||||
]
|
||||
for key in desired_fields:
|
||||
obs_data['metadata'][key] = obs_json[key]
|
||||
|
||||
return obs_data
|
||||
|
||||
############################################################################
|
||||
|
||||
def findall(p, s):
|
||||
"""
|
||||
Yields all the positions of
|
||||
the pattern p in the string s.
|
||||
Source: https://stackoverflow.com/a/34445090
|
||||
"""
|
||||
i = s.find(p)
|
||||
while i != -1:
|
||||
yield i
|
||||
i = s.find(p, i+1)
|
||||
|
||||
def hugo_markdown_to_json(markdown_contents) -> Optional[Dict[Any, Any]]:
|
||||
"""
|
||||
Take the contents from a Hugo markdown
|
||||
file and read the JSON frontmatter if it
|
||||
exists.
|
||||
"""
|
||||
front_matter_indices = list(findall('---', markdown_contents))
|
||||
if len(front_matter_indices) < 2:
|
||||
return None
|
||||
front_matter = markdown_contents[(front_matter_indices[0] + 3):front_matter_indices[1]]
|
||||
json_contents = None
|
||||
try:
|
||||
json_contents = json.loads(front_matter)
|
||||
except Exception:
|
||||
pass
|
||||
if not isinstance(json_contents, dict):
|
||||
json_contents = None
|
||||
html_contents = markdown_contents[front_matter_indices[1] + 19:-17]
|
||||
return json_contents, html_contents
|
||||
|
||||
def create_markdown_str(frontmatter, content):
|
||||
"""
|
||||
Takes a JSON frontmatter
|
||||
and creates a string representing
|
||||
the contents of a Hugo markdown
|
||||
file.
|
||||
"""
|
||||
return "---\n" + \
|
||||
f"{json.dumps(frontmatter)}\n" +\
|
||||
"---\n" +\
|
||||
"{{< unsafe >}}\n" +\
|
||||
f"{content}\n" +\
|
||||
"{{< /unsafe >}}\n"
|
||||
|
||||
def file_from_id(idnum):
|
||||
"""Returns filename from id"""
|
||||
return f"{CONTENT_LOCATION}/{idnum}.md"
|
||||
|
||||
def read_hugo_markdown(idnum) -> Optional[Dict[Any, Any]]:
|
||||
"""
|
||||
Given an id, return the markdown file
|
||||
frontmatter and contents stored in Hugo
|
||||
if it exists.
|
||||
"""
|
||||
try:
|
||||
with open(file_from_id(idnum), "r", encoding="UTF-8") as hugo_file:
|
||||
frontmatter, contents = hugo_markdown_to_json(hugo_file.read())
|
||||
return frontmatter, contents
|
||||
except Exception:
|
||||
return None
|
||||
|
||||
def write_markdown(id_num, frontmatter, contents):
|
||||
"""
|
||||
Takes a frontmatter json
|
||||
and writes it to a hugo
|
||||
markdown content file.
|
||||
"""
|
||||
try:
|
||||
with open(file_from_id(id_num), "w", encoding="UTF-8") as hugo_file:
|
||||
hugo_file.write(create_markdown_str(frontmatter, contents))
|
||||
except Exception as e:
|
||||
print("Failed to write", id_num)
|
||||
|
||||
|
||||
############################################################################
|
||||
|
||||
# Read in saved data
|
||||
saved_filenames = glob(CONTENT_LOCATION + "/*.md")
|
||||
saved_ids = { Path(fname).stem for fname in saved_filenames }
|
||||
|
||||
server_data = retrieve_data_from_server()
|
||||
|
||||
# Data is structured like [{id: '', metadata: '', contents: ''}]
|
||||
# Where metadata is a JSON and contents is HTML
|
||||
|
||||
for data in server_data:
|
||||
id_num = data['id']
|
||||
|
||||
# If the observation already exists
|
||||
if id_num in saved_ids:
|
||||
saved_fm, saved_content = read_hugo_markdown(id_num)
|
||||
if saved_fm is None:
|
||||
print("Unable to read saved data id", id_num)
|
||||
|
||||
# Only update if observation has changed
|
||||
elif saved_fm != data['metadata']:
|
||||
print("Updating id", id_num)
|
||||
write_markdown(id_num, data['metadata'], data['content'])
|
||||
|
||||
# New observation found
|
||||
else:
|
||||
print("Creating id", id_num)
|
||||
write_markdown(id_num, data['metadata'], data['content'])
|
||||
|
||||
print("Completed")
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
Loading…
Reference in a new issue