Zetttelblogging Combines Capture, Journaling, and Reference

As much as we like to complain about distraction and the evils of social media algorithms, I think we need to recognize just how information rich our environment has become. David Allen was inspired to write Getting Things Done at the beginning of the connectedness revolution in the mid 90s when email and platforms like Lotus Notes and GroupWise started creeping into the work environment. (And can you believe that both platforms still exist 30 years later?)

Capture

The first step of GTD is capture. I think the explosion we’re seeing in note taking apps is a reaction to the proliferation of information channels we are tuned into. I love Neuroscience Twitter and Philosophy Twitter. I have set of podcasts that send me off in interesting directions. And RSS feeds and our blogging revival are now growing information flows once again. Did I mention Reddit? And a handful of old-fashioned message boards?

Journal

As I’ve chronicled here over the last few months, when I started using Drafts as my main capture tool, I found myself with a better inbox system that led directly to this casual approach to blogging. I have a stream of inbox notes in Drafts.

Now I’ve learned that if I take some time off my Inbox gets backed up and it’s a project to get it back under control. But the process of working through it is modeled on my process for getting my email inbox empty. I look at each note and evaluate it as a potential entry in the journal here. This entry, for example came from a note that just said this:

  • Capture
  • Journal
  • Explode

I jotted that down when thinking about the workflow. But now I’ve expanded that thought into the idea of capture followed by journaling. And as I’ve mentioned, once I write a journal post like this in Drafts, I publish directly to WordPress. There I review the formatting and publish without trying to edit to perfect. This is a casual writing flow.

Some of the notes don’t get journaled as they are project oriented work. So these get filled out similarly as usable notes if they are telegraphic and they plus the posts get pushed to the DEVONthink database.

Explode and Edit Notes

This final step occurs in Devonthink and is the Zettel part of of the Zettleblogging. I went back and forth about how to write journal posts that were Zettels, but realized that the purpose of this narrative form and the purpose of that note keeping form were just too different to consolidate. The natural order seems to be to narrate what I’m seeing and thinking about then, as a second step, abstract out from the journal what I want to preserve as reference. For example, out of this entry I’ll probably edit down to a simple explanation of the Capture, Journal, Explode concept with some notes as to how I arrived at the conclusion.

Save Reference Notes

One final step.

I’ve learned over 20 years of writing at this site that a blog is not a good way to keep notes. Even if it’s searchable by Google or a tool like Dave Winer’s new Daytona which loads a site into an mySQL database for local search. Since WordPress runs off a mySQL database to begin with, I have a decent search functionality right at the top of site which I use from time to time to find an older post to reference.

But a dedicated notes database in DEVONthink or Tinderbox is a different, curated reference library. It’s the equivalent of my old file cabinet full of the papers I copied at the library for research and my notes torn from the yellow legal pads I used to use for note taking from high school all the way through my years on a University Medical School faculty.

I’ve said it many time before: my current workflows are just digital refinements of those xeroxed papers and sheets from legal pads. Whenever I stray too far from those habits, I tend to spend too much time on the tools and less than I should on doing the work.

Editing progress, emergence, prediction

So it’s been a month away from posting here. Time flies.

After Thanksgiving, I had a break for some important family activities, but on that break I actually got back to editing my manuscript. I finished the first draft back in June and started the first round of editing. I’ve been helped by following the guidance of Tucker Max at Scribe Media in my writing. In his editing method, the first pass is a “Make It Right” edit, where you make sure everything is there and it makes sense.

For me, that’s including some pretty big chapter reorganizations and filling out some key introductory discussions in the first three chapter. Toward the end of the third chapter, discussion where uncertainty comes from, I realized that there wasn’t a really good discussion of emergence and it’s role in making complex systems both unpredictable and at the same time understandable. Depending on you look at it, Sean Carroll had Anil Seth on his podcast which has resulted in a few weeks delving into Seth and others interesting approach to formalizing the idea of emergence in complex systems including ideas around simulation, compressibility and Granger Causality.

Plus, in preparation for editing the next chapter, on the nature of probability, I started to approach a next level appreciation for Bayesian inference and its relation to brain mechanisms. Our perception is an active process where incoming sensory data either matches or doesn’t match the brain’s current model of the world. In other words, we experience a hypothetical world, a set of beliefs that in the language of probablity is a Bayesian prior probability.

Those hypotheses — and not the sensory inputs themselves — give rise to perceptions in our mind’s eye. The more ambiguous the input, the greater the reliance on prior knowledge.

Some important new results comparing machine learning algorithms with neural mechanisms started me reading some of the new literature on cortical analysis and representation— an area that is really making some progress. as summarized in this article in Wired

Computational neuroscientists have built artificial neural networks, with designs inspired by the behavior of biological neurons, that learn to make predictions about incoming information. These models show some uncanny abilities that seem to mimic those of real brains. Some experiments with these models even hint that brains had to evolve as prediction machines to satisfy energy constraints.

So unlike metaphors like your brain is “a switchboard” or “a computer” and speaking of computation, it seems we’re converging on an understanding from two different directions, rather than just using current technology to describe brain function.


Since the idea of writing the manuscript is to collect my own thoughts, I can’t be too hard on myself in trying to make sure it’s all there. I have no deadlines or pressing need to get this out there. It’s a commitment to the process, not the product.

It’s a very long term project for me and as David Peril recently wrote:

Long story short, commitment is undervalued. 

So here’s how I suggest responding to this trend: whatever your tolerance for commitment is, raise it. 

If today you’re comfortable committing to something for two hours, try committing for a weekend. If you’re comfortable committing for two weeks, then raise it to two months; once you’re comfortable with two months, raise it to two years; and once you’re comfortable with two years, raise it to two decades. It’s okay to start small. All big things do. But they have to start somehow and with commitment comes momentum. Commitment happens in stages, and only by embracing it can you stop hugging the X-Axis and climb the compounding curve.

The Explanatory Power of Convergent Models

There’s interesting research emerging comparing our ever-improving machine learning models to data generated from brains. Not surprisingly, we do best at replicating what the brain can do when the computer models begin to model the underlying neural machinery. Now the substrate is entirely different, but the predictive approaches appear to be similar. We used to think that the brain was creating representations of the world, with features being abstracted at each level of cortical processing.

The problem that everyone saw from the beginning with this concept is that there’s no little man in the theater of the mind to look at the representations. Instead, the brain is representing hypotheses and the these predictions are constantly updated by the stream of incoming sensation from the world.

Those hypotheses — and not the sensory inputs themselves — give rise to perceptions in our mind’s eye. The more ambiguous the input, the greater the reliance on prior knowledge.

And so too with language, the brain guesses what word comes next, providing a fixed interpretation when we hear unclear or ambiguous language. Of course, we often make wrong decisions, famously about song lyrics (e.g. “There’s a bathroom on the right”) known as Mondegreens.

I’m beginning to appreciate just how important this is as our ability to look at brain activity improves just as our computational ability to create these models begins to match it. We’re not recreating higher brain function from the bottom up by understanding circuits and connections, but instead from the top down. Perhaps not surprisingly, as this is how physical sciences like chemistry and physics have advanced. They create formulas and equations that are mathematical models of the world that have remarkable predictable powers. Once systems get too complex, these methods seem to fall apart and numerical simulation seems to be needed, but nevertheless, when those models start converging on the behavior of the real thing, they seem to tell us about what’s actually going on in that complex system being modeled. Truly a remarkable time for brain science.

Hard to be a saint in the city

I hadn’t really thought about how our social media environment might affect music and art criticism until I read this Eleanor Halls Interview

Where do you see music journalism headed?

I think we need to have honest conversations about the role of music journalism and whether much of it still has any value. I worry that music journalism—interviews and reviews—is becoming PR to some musicians. Most journalists are freelance and don’t have the support of editors or publishers, and reply on publicists for talent access so they can get work. It’s no wonder they often feel too intimidated by an artist and their team to write what they really think.

There’s always been a bargain between critics and artists regarding access and cooperation. It’s only natural that an artist would share insights with a sympathetic journalist and not one who has little enthusiasm for the style or approach of the artist. Personal relationships have always played a big role in what we read as criticism and commentary.

While some nasty letters from fans may have been the price for a critic to pay for publishing a negative take on something, I can see how the amplification of opinion in social media makes the pressure way more real. But without a publication behind the writer, freelance writers are much more dependent on these relationships for access to artists, creating a competition to curry favor with creators and their fans.

I think its true that the tone of discussion across the internet tends to be more promotional than print publications ever were. Editorial independence is lost. I don’t think its even a real bias, necessarily, but a function of writers choosing to write about what they like. It’s often just another symptom of our fragmentation. Sites team up with companies for synergy.

I like the idea of these personal blogs being islands of authenticity. I try to be positive in general, but that’s a personal bias. We’re all in this together, so my aim has to be to inform and teach a bit so we all do a bit better.

Abstract

For a very long time I had the practice of putting an image at the top of each post here. As decoration for the most part as my images carry very little semantic content. This is another image from the San Francisco trip. I continue to think about the contradictions inherent in abstract photography, but I’ve concluded that the answers lie in making more images, not contemplation.

The Bedrock of Knowledge

I’ve enjoyed Scott Young’s writing since he’s the kind of interested amateur who dips into all kinds of areas without committing to professional work. So it was interesting to read his impression of literature research; What if You Don’t Feel Smart Enough?

The expectation is that as you learn more and more, you’ll eventually hit a bedrock of irrefutable scientific fact. Except usually, the bottom of one’s investigation is muck. Some parts of the original idea get sharpened, others blur as more complications and nuance are introduced.

And it’s true that it’s not well appreciated how tentative scientific explanation is as new areas are explored. It’s been exciting for me to watch COVID-19 science develop in real time, so quickly. Yes, scary and polarized in ways that we generally don’t see in medicine, but a predictable back and forth on the properties of the virus, its propagation, and treatment.

We generally know what we know

Scott misses the important point that there is a bedrock of knowledge, the literature just doesn’t bother to discuss it. In neuroscience, the basic physical architecture and cellular makeup of the brain was established with great clarity over the last 100 years or so. As techniques have been introduced, new areas opened up and took a while to get settled into bedrock, but much of that is done now. In fact, my first published paper in 1983 was part of a major chapter in that story when labs used retrograde tracing techniques to map brain connections. My paper established the identity of all of the areas that sent connections to the motor trigeminal nucleus in the rat. That’s the collection of motor neurons that innervate the jaw closing muscles.

We’re in an in interesting era where cognitive science is successfully exploring its underlying neuronal circuitry. As is typical, the process is messy but the picture is getting filled out, even in some very tricky areas like working memory and perception.

It’s of little importance to my day job in drug development at this point, but these are the kinds of questions that sparked my interest in brain science at the beginning. So while I look on as a spectator, I’m spending time reading papers and developing at least a superficial understanding of the techniques and progress.

Building models to explore the unknown

Neuroscience Twitter is a great resource to keep up with trends across cognitive science. Case in point: I’m reading through Bayesian models of perception and action which is a draft of a book by Wei Ji Ma, Konrad Kording, and Daniel Goldreich, to be published by MIT press. I’ve been dipping into papers published by the three authors to get a feel for the deeper applications of the approach. I learned about it on Twitter

I think this is an important area to watch. I’ve talked about the idea that the brain, in order to control behavior, has to contain a model of the system. One approach is create computer models of circuitry based on observed connectivity and activity in animals when these systems are active. If some models can reproduce the brain activity, then they are candidates for hypothesized mechanisms and be used to make predictions about how the real neural circuits behave. Think about it like a physicist using equations to model physical laws and then testing the predictions from those equations against new observations. Except for the brain we don’t have any such equations, so we can use the immense computer power we have at our disposal to do the same kind of abstraction as the physicist.

Just like the equations of physics describe reality, but aren’t reality, these neural models describe little bits of the brain, they aren’t thinking. But interestingly, some of these brain inspired models can be put to work for real life tasks like image or speech recognition because the escape simple algorithmic approaches to analysis and classification.

Time for Recovery

SF Monochrome

It’s funny how a few days traveling, some dental issues and work can so quickly shift the environment from one of reflection to one of the constant pressure of activity.

As my posting over the last few days shows, I had time to capture images while in San Francisco. My goal was just to get back into visual mode after some months of ignoring the cameras. But the light in the city and the capability of the tools was enough to very quickly get me into that mode of looking that leads to making images. I brought the Leica M10 Monochrom which is a camera with a digital sensor that captures only black and white images since it lacks the color filters needed to reconstruct colors in a digital image. I brought the Monochrom because I wanted to be deliberate in capture, something that the rangfinder focusing M10 brings. And since my final product is monochrome, the B&W camera takes me a step closer from capture to image- The more casual approach compared to the cinematic imagery I’m made in recent years.

So I more or less picked up where I left off, trying to abstract the bits of the city that I can isolate with my lens. I brought my newer 35mm Summicron lens, ending up shooting mostly wide open in that sharp, defining California light. And I definitely enjoyed collecting the images and there’s a pretty high percentage of interesting captures. So I’ve been having fun doing a very quick set of image adjustments and publishing here and Instagram. The 35mm lens opens up the view a bit. While in San Francisco, I stopped in at the SF Leica store to look at the images on display and look over the cameras and lenses. It turned out they had a used electronic viewfinder for the M, the Visoflex, at a good price. So the rest of the trip was variously shot with the EVF, the glass rangefinder or the back screen.

On Sunday, before heading back to the airport for the red-eye flight back to Baltimore, I spent a few hours at SFMOMA, the wonderful San Francisco Museum of Modern Art. Not much going on in the way of photography because of new exhibit being hung. So I spent more time looking at paintings this time there. I’ve noted here a number of times that my formative experience with art was with the Abstract Expressionists, particularly Motherwell, Johns, Rothko, Diebenkorn. The headline show at the museum were the paintings of Joan Mitchell. I’d seen them before, but there’s nothing like a retrospective like this to get to know a artist well. Of interest to me was how she wasn’t afraid as an abstract painter to let here images drift back to the landscape enough that the underlying structure of nature starts to emerge from the abstraction. And color. Color is emotional.

Joan Mitchell, La Grande Vallée XVI, Pour Iva, 1983

While viewing the show and since I’ve been thinking about how abstract expressionism informs the images I make. I like the tension between my formal abstraction and the concreteness of the photographic image. I can’t hide the fact that what I’ve photographed a fire plug. But you have to ask why did I capture the image? What did I see in that moment that compelled me to capture an image and publish it here. One of the ideas that I took away from Joan Mitchell’s paintings is that the view wants some challenge. Enough to make viewing a way of participating in the creative act. You see, art presents ambiguous, noisy sensations allowing the viewer to participate in finishing the creation through inference. If there’s no sense of participation with the artist, looking is boring and no fun.