git update-index --assume-unchanged .taguette_gdocs
```
-I also create a directory called `taguette_backups` like:
+I also create a directory called `taguette_backups` and one called `exported_tags` like:
```
mkdir taguette_backups
+mkdir exported_tags
```
## Step 1: Backing things up
Exporting tags should be as easy as:
```
-python3 taguette-export_tags_to_csv.py > exported_tags.tsv
+python3 taguette-export_tags_to_csv.py
```
This will create a new file called `exported_tags.tsv` which you can manually
#!/usr/bin/env python3
import re
+import json
import sqlite3
from configparser import ConfigParser
import csv
import os
-config_files = [f for f in os.listdir() if f.startswith('.taguette_gdocs_')]
+config_files = [f for f in os.listdir() if f.startswith('.taguette_gdocs')]
for file_path in config_files:
# Run this if you just want tags and no highlights
sql_stmt_get = "SELECT id, path, description FROM tags WHERE project_id = ?"
- # Run this if you want tags AND highlights
- #sql_stmt_get = "SELECT tags.id, tags.path, tags.description, highlights.snippet FROM highlight_tags INNER JOIN tags ON highlight_tags.tag_id = tags.id INNER JOIN highlights ON highlight_tags.highlight_id = highlights.id WHERE project_id = ?"
cur.execute(sql_stmt_get, (project_id,))
-
-
+
with open(output_file_name, 'w', newline='') as output_file:
writer = csv.writer(output_file, delimiter='\t')
+ writer.writerow(['id', 'axial codes', 'tag', 'category', 'description', 'url'])
+
while True:
row = cur.fetchone()
if row == None:
break
-
+
tag_id, path, description = row
- m = re.match(r'^(.+)\_(.*)$', path)
- if m:
- axial = m.group(1)
- tag = m.group(2)
+ tag_match = re.match(r'^(.+)\_(.*)$', path)
+ if tag_match:
+ axial = tag_match.group(1)
+ tag = tag_match.group(2)
else:
axial = ""
tag = path
- writer.writerow([str(tag_id), axial, tag, description])
\ No newline at end of file
+ # look for extra category information stored in the description
+ cat_match = re.match('^(.*)\s*(\{(.*)\})$', description)
+ if cat_match:
+ description = cat_match.group(1)
+ category = json.loads(cat_match.group(2))["category"]
+ else:
+ category = ""
+
+ # create a URL that will link to the list of highlights
+ url = f"https://taguette.communitydata.science/project/{project_id}/highlights/{tag}"
+
+ writer.writerow([str(tag_id), axial, tag, category, description, url])
from csv import DictReader
import sqlite3
from configparser import ConfigParser
+import re
+import json
config = ConfigParser()
config.read('.taguette_gdocs')
for row in DictReader(csv_text.splitlines(), delimiter=","):
#print(row)
- tag_id = row['ID']
+ tag_id = row['id']
new_name = row['tag']
- axial_code = row['Axial Codes']
+ axial_code = row['axial codes']
+ category = row['category']
+ description = row['description']
sql_stmt_get = "SELECT id, path, description from TAGS where id = ? AND project_id = ?"
cur.execute(sql_stmt_get, (tag_id, project_id))
oldname = tag_info[0][1]
old_description = tag_info[0][2]
- if row['Axial Codes']:
- newname = row['Axial Codes'].lower() + "_" + new_name.lower()
+ if axial_code:
+ newname = axial_code.lower() + "_" + new_name.lower()
else:
newname = new_name.lower()
+ new_description = description
+ if description and category:
+ new_description += " "
+ if category:
+ new_description += json.dumps({'category' : category})
+
if not oldname == newname:
- #print(tag_id)
sql_stmt_update = "UPDATE tags SET path = ? WHERE project_id = ? AND id = ?"
cur.execute(sql_stmt_update, (newname, project_id, tag_id))
print(f"UPDATE TAG: {oldname} → {newname}")
- if row["description"].strip() != old_description.strip():
+ if new_description.strip() != old_description.strip():
sql_stmt_update = "UPDATE tags SET description = ? WHERE project_id = ? AND id = ?"
- cur.execute(sql_stmt_update, (row['description'], project_id, tag_id))
- print(f"UPDATE DESC: {old_description} → {row['description']}")
+ cur.execute(sql_stmt_update, (new_description, project_id, tag_id))
+ print(f"UPDATE DESC: {old_description} → {new_description}")
con.commit()
con.close()