Reading across books with Claude Code
117 points
19 hours ago
| 12 comments
| pieterma.es
| HN
zkmon
4 minutes ago
[-]
I used AI for accelerating my reading a book recently. This is a interesting usecase. But it same as racing for the destination instead enjoying the journey.

It kills the tone, pace and the expressions of the author. It is pretty much same as an assistant summarizing the whole book for you, if that's what you want. It misses the entire experience delivered by the author.

reply
Ronsenshi
10 hours ago
[-]
For me this looks like a great way to build connections between books in order to create a recommendation engine - something better than what Goodreads & Co provides. Something actually useful.

The cost of indexing using third party API is extremely high, however. This might work out well with an open source model and a cluster of raspberry pi for large library indexing?

reply
padolsey
40 minutes ago
[-]
The incumbants Goodreads and their owner Amazon have indeed done such a poor job at this. Seven years ago I tried creating a basic graph using collaborative-filtering (effectively using our actual reading patterns as the embeddings space instead of semantics [human X likes book Y so likers of Y might like other things that human X has enjoyed]). It works well to this day (ablf.io) but the codebase is so ugly I've not had the bravery to update its data in a couple of years.
reply
nubskr
4 hours ago
[-]
I've been using Claude Code for my research notes and had the same realization, it's less about perfecting prompts and more about building tools so it can surprise you. The moment I stopped treating it like a function and started treating it like a coworker who reads at 1000 wpm, everything clicked
reply
duck
17 hours ago
[-]
reply
jszymborski
17 hours ago
[-]
This is all interesting, however I find myself most interested in how the topic tree is created. It seems super useful for lots of things. Anyone can point me to something similar with details?

EDIT: Whoops, I found more details at the very end of the article.

reply
ebiester
14 hours ago
[-]
I did a similar thing with productivity books early last year, but never released it because it wasn't high enough quality. I keep meaning to get back to that project but it had a much more rigid hypothesis in mind - trying to get the kind of classification from this is pretty difficult and even more so to get high value from it.
reply
lloydatkinson
33 minutes ago
[-]
How can anyone even trust crap like this? It was only a few days ago Claude and ChatGPT hallucinated a bunch of stuff from actual docs I sent them links to. When asked about it, they just apologised.
reply
mpalmer
27 minutes ago
[-]
Synthesizing 500 words at a time into digestible topics is significantly less prone to error. You're giving it a lot of info and asking for an organized subset. It's good at following such direction.

In your example, you're doing the inverse (give me a lot of text based on a little), and that's where LLMs have no problem hallucinating the new information.

reply
doytch
12 hours ago
[-]
The mental model I had of this was actually on the paragraph or page level, rather than words like the post demos. I think it'd be really interesting if you're reading a take on a concept in one book and you can immediately fan-out and either read different ways of presenting the same information/argument, or counters to it.
reply
voidhorse
15 hours ago
[-]
This was posted before and there were many good criticisms raised in the comments thread.

I'd just reiterate two general points of critique:

1. The point of establishing connections between texts is semantic and terms can have vastly different semantic meanings dependent on the sphere of discourse in which they occur. Because of the way LLMs work, the really novel connections probably won't be found by an LLM since the way they function is quite literally to uncover what isn't novel.

2. Part of the point in making these connections is the process that acts on the human being making the connections. Handing it all off to an LLM is no better than blindly trusting authority figures. If you want to use LLMs as generators of possible starting points or things to look at and verify and research yourself, that seems totally fine.

reply
skeptrune
15 hours ago
[-]
I really like the idea of the topic tree. That intuitively resonates.
reply
kylehotchkiss
16 hours ago
[-]
In several years, IMO the most interesting people are going to be the ones still actually reading paper books and not trying to shove everything into a LLM
reply
hungryhobbit
15 hours ago
[-]
I don't think the Venn diagram of those people and everyone else is as separate as you imagine.

I'm a Literature major and avid reader, but projects like this are still incredibly exciting to me. I salivate at the thought of new kinds of literary analysis that AI is going to open up.

reply
imdsm
15 hours ago
[-]
the people most likely to analyse books like this are those of us who are more likely to read them as well
reply
pradmatic
15 hours ago
[-]
Sure but those people don't have to be mutually exclusive. At the very least, a tool like this can help me decide what to read next.
reply
gulugawa
16 hours ago
[-]
[flagged]
reply
dang
11 hours ago
[-]
"Please don't post shallow dismissals, especially of other people's work. A good critical comment teaches us something."

"Don't be curmudgeonly. Thoughtful criticism is fine, but please don't be rigidly or generically negative."

https://news.ycombinator.com/newsguidelines.html

reply
gjm11
16 hours ago
[-]
I agree that we should be reading books with our eyes and that feeding a book into an LLM doesn't constitute reading it and confers few of the same benefits.

But this thing isn't (so far as I can tell) even slightly proposing that we feed books into an LLM instead of reading them. It looks to me more like a discovery mechanism: you run this thing, it shows you some possible links between books, and maybe you think "hmm, that little snippet seems well written" or "well, I enjoyed book X, let's give book Y a try" or whatever.

I don't think it would work particularly well for me; I'd want longer excerpts to get a sense of whether a book is interesting, and "contains a fragment that has some semantic connection with a fragment of a book I liked" doesn't feel like enough recommendation. Maybe it is indeed a huge waste of time. But if it is, it isn't because it's encouraging people to substitute LLM use for reading.

reply
imdsm
16 hours ago
[-]
commenter above probably didn't read the post, ironically
reply
ryan_n
14 hours ago
[-]
Guess we need “reading across hacker news articles with Claude code.”
reply
stavros
15 hours ago
[-]
I need a name for people who dismiss an entirely new and revolutionary class of technology without even trying it, so much so that they'll not even read about any new ideas that involve it.
reply
dang
10 hours ago
[-]
The HN guidelines include the term "curmudgeonly", which IMO is fair.
reply
imdsm
15 hours ago
[-]
we call them luddites
reply
lsaferite
14 hours ago
[-]
I'm not entirely sure that's a fair association. The Luddites weren't against technology in general, they were fighting for their livelihoods. There very well could be a fresh luddite movement centered around the use of AI tools, but I don't think "luddite" is the right term in this specific case.
reply
ironbound
11 hours ago
[-]
No that was a labor issue, abusive factory owners got targeted.
reply
mikkupikku
16 hours ago
[-]
I zgrep my epubs, is that a problem too?
reply