Twenty years ago today E and I visited Reboot 7 in Copenhagen. What I wrote a decade ago at the 10th anniversary of that conference still holds true for me.

Over time Reboot 7 became mythical. A myth that can’t return. But one we were part of, participated in and shaped.
Still got the t-shirt.


The yellow t-shirt with red text from the 2005 Reboot 7 conference, on my blue reading chair in my home office 20 years on.

Seventeen years ago today I blogged about a barcamp style event in Amsterdam I co-hosted, called GovCamp_NL. I struck up a conversation there about open government data after having had a similar conversation the week before in Austria. It marked the beginning of my work in this field. We just welcomed the thirteenth team member in the company that over time grew out of that first conversation. Our work at my company is driven by the same thing as the event, something I’ve come to call constructive activism.

These days, the principles and values that drove those events, and have set the tone for the past two decades of everything I’ve done professionally and socially, seem more important than ever. They are elemental in the current geopolitical landscape around everything digital and data. We can look back on our past selves with 20 years hindsight and smile about our one time optimism, because so much exploitation, abuse and surveillance grew out of the platforms and applications that originate in the early 00’s. But not because that optimism was wrong. Naive yes, in thinking that the tech would all take care of itself, by design and by default, and we just needed to nudge it a bit. That optimism in the potential for (networked) agency, for transparency, for inclusion, for diversity, and for global connectedness is still very much warranted, as a celebration of human creativity, of the sense of wonder that wielding complexity for mutual benefit provides, just not singularly attached to the tech involved.
Anything digital is political. The optimism is highly political too.

The time to shape the open web and digital ethics is now, is every day. Time for a reboot.

Y turned nine at the end of May, which we celebrated with a trip to Lego House and Lego Land in Billund, Denmark as it coincided with Y having a few days off from school. But it wasn’t the only Lego related fun we organised around her birthday party. Yesterday with 5 invited friends we visited Roy Scholten‘s workshop at the Hilversum graphics center, where the group tried their hand on Lego printing.

Using flat Lego pieces you create a design, and then ink them up and put them under a press. Next to the 6 kids we were 3 parents, and we all had a lot of fun. When it was time to stop and clean up (before heading home for a small mountain of pancakes for dinner) no one really wanted to quit. I was impressed with how this little group of 8-9yr olds worked with abstract forms, experimented with colors etc, and stayed focused the full time without needing much aid or prompting.


Roy Scholten providing instructions to the group.


Searching for Lego pieces for our designs


A few iterations I made.

I originally met Roy during the pandemic in a conversation about personal knowledge management, and appreciated his bird prints made using Lego. We since acquired a few. Our friend Peter also uses a letterpress, and after making introductions, to my delight came to visit from Canada with his partner L to work with Roy. Yesterday some of their production together still adorned the walls of Roy’s atelier.


Lining up several iterations of my ‘river’ print.

You can book his workshops (and by other members of their collective) for company / team outings, or for training, as well as birthday parties. E has done a training with Roy, and we also gave her mother a workshop with friends for her 80th earlier this year. See Grafisch Atelier Hilversum’s website.


The two prints I like best. At the top one I retro actively dubbed ‘soccer player heading a ball’, that reminds me of De Zaaier by Theo van Doesburg we recently saw in the Drachten DaDa museum. Below the ‘river’, where I flipped the paper 180 degrees before printing again. The result of a much lighter blue second river course reminds me of how old river meanders stay visible as oxbow lakes in the landscape when the river bed has moved on.

How could I not buy these small notebooks? Made by my friend Peter from paper cut-offs from boxes he made and printed in Tuscany, they are made from Magnani 1404 paper. Magnani started making paper in Pescia in 1404 (they ceased operation in recent years, but another Magnani is still making paper, since 1481), right at the moment in time that the literate population of Tuscany started using paper notebooks to make everyday notes, and lots of them. Paper had become affordable and available enough roughly a century earlier, with Tuscany being at the heart of that, and Florentine merchants used their book keeping system and the paper notebooks needed for it to build a continent spanning trade network. After the Black Death personal note taking took off too, and from 1400 onwards it had become commonplace:

At the end of the Middle Ages, urban Tuscans seemed stricken with a writing fever, a desire to note down everything they saw.’ But they remained a peculiarly local phenomenon: there was something uniquely Florentine (or more accurately ‘Tuscan’ as examples also survive from Siena and Lucca) about them,…

Allen, Roland. The Notebook: A History of Thinking on Paper (p. 61).”

Around the turn of the year I gave The Notebook as a present to Peter thinking it would be something to his liking. My own notes have helped me learn and work for decades. E and I when we lived in Lucca for a month, passed through Pescia by train en route to Firenze.

Tuscany, paper from a company that was there from the start of everyday note taking, The Notebook, personal knowledge management, and friendship, all coming together in this piece of craftsmanship. How could I not buy them? So I did.

I have a little over 25 years worth of various notes and writings, and a little over 20 years of blogposts. A corpus that reflects my life, interests, attitude, thoughts, interactions and work over most of my adult life. Wouldn’t it be interesting to run that personal archive as my own chatbot, to specialise a LLM for my own use?

Generally I’ve been interested in using algorithms as personal or group tools for a number of years.

For algorithms to help, like any tool, they need to be ‘smaller’ than us, as I wrote in my networked agency manifesto. We need to be able to control its settings, tinker with it, deploy it and stop it as we see fit.
Me, April 2018, in Algorithms That Work For Me, Not Commoditise Me

Most if not all of our exposure to algorithms online however treats us as a means to manipulate our engagement. I see them as potentially very valuable tools in working with lots of information. But not in their current common incarnations.

Going back to a less algorithmic way of dealing with information isn’t an option, nor something to desire I think. But we do need algorithms that really serve us, perform to our information needs. We need less algorithms that purport to aid us in dealing with the daily river of newsy stuff, but really commodotise us at the back-end.
Me, April 2018, in Algorithms That Work For Me, Not Commoditise Me

Some of the things I’d like my ideal RSS reader to be able to do are along such lines, e.g. to signal new patterns among the people I interact with, or outliers in their writings. Basically to signal social eddies and shifts among my network’s online sharing.

LLMs are highly interesting in that regard too, as in contrast to the engagement optimising social media algorithms, they are focused on large corpora of text and generation thereof, and not on emergent social behaviour around texts. Once trained on a large enough generic corpus, one could potentially tune it with a specific corpus. Specific to a certain niche topic, or to the interests of a single person, small group of people or community of practice. Such as all of my own material. Decades worth of writings, presentations, notes, e-mails etc. The mirror image of me as expressed in all my archived files.

Doing so with a personal corpus, for me has a few prerequisites:

  • It would need to be a separate instance of whatever tech it uses. If possible self-hosted.
  • There should be no feedback to the underlying generic and publicly available model, there should be no bleed-over into other people’s interactions with that model.
  • The separate instance needs an off-switch under my control, where off means none of my inputs are available for use someplace else.

Running your own Stable Diffusion image generator set-up as E currently does complies with this for instance.

Doing so with a LLM text generator would create a way of chatting with my own PKM material, ChatPKM, a way to interact (differently than through search and links, as I do now) with my Avatar (not just my blog though, all my notes). It might adopt my personal style and phrasing in its outputs. When (not if) it hallucinates it would be my own trip so to speak. It would be clear what inputs are in play, w.r.t. the specialisation, so verification and references should be easier to follow up on. It would be a personal prompting tool, to communicate with your own pet stochastic parrot.

Current attempts at chatbots in this style seem to focus on things like customer interaction. Feed it your product manual, have it chat to customers with questions about the product. A fancy version of ‘have you tried switching it off and back on?‘ These services allow you to input one or a handful of docs or sources, and then chat about its contents.
One of those is Chatbase, another is ChatThing by Pixelhop. The last one has the option of continuously adding source material to presumably the same chatbot(s), but more or less on a per file and per URL basis and limited in number of words per month. That’s not like starting out with half a GB in markdown text of notes and writings covering several decades, let alone tens of GBs of e-mail interactions for instance.

Pixelhop is currently working with Dave Winer however to do some of what I mention above: use Dave’s entire blog archives as input. Dave has been blogging since the mid 1990s, so there’s quite a lot of material there.
Checking out ChatThing suggests that they built on OpenAI’s ChatGPT 3.5 through its API. So it wouldn’t qualify per the prerequisites I mentioned. Yet, purposely feeding it a specific online blog archive is less problematic than including my own notes as all the source material involved is public anyway.
The resulting Scripting News bot is a fascinating experiment, the work around which you can follow on GitHub. (As part of that Dave also shared a markdown version of his complete blog archives (33MB), which for fun I loaded into Obsidian to search through. Also for comparison with the generated outputs from the chatbot, such as the question Dave asked the bot when he first wrote about the iPhone on his blog.)

Looking forward to more experiments by Dave and Pixelhop. Meanwhile I’ve joined Pixelhop’s Discord to follow their developments.

Having created a working flow to generate OPML booklists directly from the individual book notes in my PKM system, I did the first actual run in production of those scripts today.

It took a few steps to get to using the scripts in production.

  • I have over 300 book note files in my Obsidian vault.
  • Of course most lacked the templated inline data fields that allow me to create lists. For the 67 fiction books I read in 2021 I already had a manual list with links to the individual files. Where needed I added the templated data fields.
  • Having added those inline fields where they were missing I can easily build lists in Obsidian with the Dataview plugin. Using this code


    results in

  • The same inline data fields are used by my scripts to read the individual files and build the same list in OPML
  • That gets automatically posted to my website where the file is both machine and human readable.

Doing this in production made me discover a small typo in the script that builds the OPML, now fixed (also in the GitHub repository). It also made me realise I want to add a way of ordering the OPML outline entries by month read.

Lists to take into production next are those for currently reading (done), non-fiction 2021, and the anti-library. That last one will be the most work, I have a very long list of books to potentially read. I will approach that not as a task of building the list, but as an ongoing effort of evaluating books I have and why they are potentially of interest to me. A way, in short, to extend my learning, with the list as a useful side effect. The one for currently reading is the least work, and from it the lists for fiction 2022 and non-fiction 2022 will automatically follow. The work is in the backlog, getting history to conform to the convention I came up with, not in moving forward from this point.

In parallel it is great to see that Tom Critchlow is also looking at creating such book lists, in JSON, and at digesting such lists from others. The latter would implement the ‘federated’ part of federated bookshelves. Right now I just point to other people’s list and rss feeds in my ‘list of lists‘. To me getting to federation doesn’t require a ‘standard’. Because JSON, OPML and e.g. schema.org have enough specificity and overlap between them to allow both publishers of lists and parsers or such lists enough freedom to use or discard data fields as they see fit. But there is definitely a discussion to be had on identifying that overlap and how to use it best. Chris Aldrich is planning an IndieWeb event on this and other personal libraries related topics next month. I look forward to participating in that, quite a number of interesting people have expressed interest, and I hope we’ll get to not just talk but also experiment with book lists.

As a form of WAB* I’ve made it easier for myself to update my OPML book lists. I created those lists earlier this year as a proof of concept of publishing federated bookshelves. Updating OPML files residing on my hosted webserver is not a fun manual task. Ultimately I want to automate pushing lists from my personal working environment (notes in Obsidian) to my site. Given my limited coding skills I made a first easier step and created a webform in a php script that allows me to add a book to an opml list. It has a drop-down menu for the various OPML lists I keep (e.g. fiction2021, non-fiction2021, currently reading, anti-library), provides the right fields to add the right OPML data attributes, and then writes them to the correct list (each list is a separate file).

That now works well. Having a way to post to my book lists by submitting a form now, I can take the next step of generating such form submissions to replace manually filling out the form.

* Work Avoiding Behaviour, a continuation of the SAB, Study Avoiding Behaviour that I excelled in at university. WAB seems to fit very well with the current locked down last days until the end of year. The Dutch terms ‘studie/werk ontwijkend gedrag’ SOG/WOG lend themselves to the verb to ‘sog’ and to ‘wog’. Yesterday when Y asked E what she had been doing today, E said ‘I’ve been wogging’, and I realised I had been too.