I used Alfred (my Clawdbot fork) to organize almost 4000 of my ChatGPT and Claude conversations, emails, GDrive documents into one huge Obsidian Vault that is built based on the Palantir Ontology System. It's weird looking at 80% of the data that describes my entire life.
Then I vectorized the entire thing on a DGX Spark using e5-large, ran sklearn's OPTICS autoclustering to identify relationships based on vector parameters which found 1,052,918 potential relationships.
I selected the top 2000 and running Qwen 2.5 7B on the DGX Spark to label it, giving me a fully enriched knowledge base. This is running right now and takes about 85 minutes to complete.
To make it useful, I set up a few cron jobs via Alfred:
- every conversation, every meeting transcript, every braindump is automatically processed.
- a "janitor" runs through the Obsidian Vault every hour looking for duplicates and ontology violations and automatically fixes them
- a "heartbeat" runs every 30 minutes summarizing the conversation we've been having, flushing context and loading relevant info from the Vault back into context
Now Alfred has real context of my real life that can actually do stuff and not just hallucinate vividly.