Saturday, March 14, 2026

Where is the LLLine?

My personal uses and struggles with large language models (LLMs) cannot be fully expounded upon in one post and I will make no such attempt here. However, I have been mulling over the genuine utility of LLMs versus the undeniable impact they have on my personal ability to process and understand information. There are generally two diametrically opposed camps: 

  1.  The AI-haters: these people believe that there is absolutely no place for LLMs regardless of utility. They point to not only the impact they have on people, but also the disastrous moral, ethical, environmental, and economical impact they could be argued to have. I won't evaluate that impact here; but this is their rationale. These people use generative LLMs for absolutely nothing and absolutely despise it in all forms. 
  2. The AI die-hards: these people believe that true AGI is right around the corner, and that these tools will completely reshape the way we use technology, and perhaps our role as human beings. They point to the impressive capabilities of LLMs and argue that a massive upheaval is right around the corner, as companies lay off workers left and right to be seemingly replaced by LLMs. I won't evaluate these claims here, but this is their rationale. These people use LLMs for everything they can think of, from genuinely useful work to helping choose their clothes to writing text messages to their moms. 

I do not find either of these camps truly appealing. While I think I lean more hater than die-hard--as an artist I am extremely alarmed by the rise of slop and the willingness with which LLM companies disregard notions of intellectual property--I also do use LLMs myself, on a near-daily basis. They are objectively immensely useful, even after just a few years of development, and have made certain things about my work more enjoyable and efficient. 

However, as with all things that seem too good to be true, there is a tradeoff. With each task I use LLMs for, I feel a token of practice escape, a chance to become better evaporate. The time I saved on the task is replaced by the weight of the knowledge that I have just made myself easier to replace, or robbed myself of an opportunity to learn. This feeling is not so strong on tasks I am already very good at--for example, making plots in Matplotlib--but it dramatically waxes when I am using LLMs to save time on the learning process itself; i.e., using an LLM to automate a task that I am not fully capable of myself. My wonder at the technology contorts into an intense, epistemic discomfort. Like a student cheating on an exam or a person skipping a workout, I know deep down that in the long run, I may be saving time and energy, but I am only hurting myself.

I have been struggling with these two extremes, knowing these tools can help me, but not knowing when I am taking it too far. Today, I had a discussion with my very close friend and colleague, the brilliant and handsome Joaquin. He presented an analogy for LLM usage that clicked so well I genuinely stopped in my tracks. 

Joaquin compares using an LLM to using a chess engine (e.g., Stockfish, or Komodo). This comparison has an immediate appeal, as while they are both breathtaking examples of modern technology and innovation, they are just about equally close to being true artificial intelligence (that is to say, not at all). However, you might be fooled into thinking otherwise, whether you are seemingly "chatting" with Claude, or watching Magnus Carlsen get flattened by AlphaZero. Upon reflection and discussion with Joaquin, I realized this analogy was unbelievably 1:1, and indeed questions about LLMs that seem impossible to answer become almost trivial when applied to a chess engine--but the answers are plainly fungible. 

 

A picture of me learning chess with my father in the early 2000s. Chess has always been a huge part of my life from a very young age, and thinking about LLMs terms of something that comes to me very intuitively has been a game-changer. 

The particular question I am interested in answering is: in my own work as an academic, where does the line between "using LLMs to help me" and "abusing LLMs in a way that hurts me" exist? I consider the following litmus tests, based on how I have used chess engines as a player myself to get to a 1700 ELO. 

  1. You must use these tools to learn, and not to cheat. The analogy here is extremely obvious. Every serious chess player in the world uses engines to analyze their games; understand their weaknesses; hone their strengths. But there is a clear and intuitive point where this becomes cheating--when they take the engine onto the battlefield. Why is this so hard to transfer to LLMs? Because it is not clear where, as an academic, our battlefield is. Is it at our desk? In the conference room? In a presentation? Within the journal submission? The line is vague and unclear. I believe answering this question--where is the battlefield of academia?--is the key to understanding where the line is. My best answer is that the battlefield is in our ability as academics to communicate with our peers--in any format. Every successful academic has an intuitive understanding of understanding--a sense, a click, when you really understand something, when the concept goes from something you are regurgitating to something you have internalized. When you are reading a textbook, struggling to understand a concept, an LLM can be an extremely useful personal tutor, just as a chess engine can help you analyze a game (this also plays to an LLM's strength, since textbook knowledge tends to be over-represented in training data compared to more niche bleeding-edge concepts). But when it's time to take what you've learned to the battlefield, if you cannot leave the engine behind, then you are likely relying on it too much. When you stand up and talk to your colleague, can you defend the ideas you've learned with LLM assistance? Can you teach it to someone yourself? When your colleague asks you questions about your idea, can you answer with an innate sense of the boundary of your own knowledge? Or are you left stalling as you make a mental note to forward the question to your LLM of choice later? Framing the question in this way has helped me draw the line--your goal of using the LLM to learn should be to eventually not need it. These machines are the most powerful learning tools ever made. But when you ask it to teach you a concept, the goal must be true understanding, not automation or substitution. Every time, no exceptions.
  2.  LLMs can do some things better and faster, and that's okay. I struggled the most with this concept, as anything I offloaded onto LLMs felt like a skill I was giving up, or an opportunity to improve myself I was missing. But the reality is, there are just some things at which LLMs are fantastic. The chess engine analogy has helped me realize this isn't the end of the world. There are some things in chess we have offloaded to engines, likely permanently, such as the study of openings. It is simply not feasible to expect players to calculate the outcomes of common openings from scratch, every single time. So we use powerful tools like Chess.com's Opening Explorer to walk through every opening sequence imaginable, learning the right moves to make, the wrong moves to avoid. The engine helps us experiment--what happens if we try this? Make that move? Take that piece? Every question, answered immediately, with the cold confidence of a calculating machine vastly beyond our own capabilities. And yes, sometimes we just memorize some of these openings, like a Cuber memorizing a solution algorithm for a particular sequence of patterns. In the end, it doesn't change the fun of the game; it just helps you get a better start more often than not. In the same way, LLMs can genuinely improve your efficiency and productivity, when used carefully and in the right situations. And when you find those situations, it's okay to use them freely. However, I will finish this point on an interesting word of caution. While engine-explored openings are useful for 99% of situations, you will be vulnerable against someone who is a true master. Magnus Carlsen is well known for deliberately playing bizarre openings to force engine-reliant opponents into unfamiliar game scripts, where he has the upper hand due to his unmatched calculational ability. For whatever area you allow LLMs to supplement you more fully, you will have this same vulnerability. A real expert will be able to tell. Proceed with caution. 
  3. Sometimes, human intuition IS better. The reality is, LLMs have their limitations. This plays both ways--in one scenario, you may not want to use an LLM because they are simply incapable of performing a task (such as making a genuinely new discovery outside its training data in the way humans can), and in another scenario, you may not want to allow an LLM to teach you something a particular way because its enormous dataset-based approach is simply not ideal compared to the way humans think about things. There is a really direct parallel with chess engines here as well. Despite chess engines being many, many times stronger than any human opponent could ever hope to be, they can still fail or be useless in several key ways. First, their computations boil down to brute force (this is a drastic oversimplification, but it's largely what they do better than us). We can't do what they do--and that means, we can't always rely on their solutions! If I'm trying to understand the best move in a position, sometimes it isn't the best move mathematically, which the computer had to check 12 moves into the future to notice. Sometimes the best move is the one that is maybe 0.05 points less optimal, but I understand deeply and intuitively why it is a good move, and I know next time I am in this position I will be able to find the move right away. This reality is an accommodation of our own failings, but the limitations of engines go beyond this. There are real scenarios where humans are just better at assessing the situation than machines. Sometimes the human and machine disagree, and the human ends up being right, even though the machine should be able to calculate far more extensively than our soft wrinkly brains. I've drawn out this analogy a bit too far, but I do find it fascinating--and the conclusions are directly transferable to LLMs. Turning to an LLM will not always be the best decision, especially at the elite level of anything. Knowing where those gray areas and blind spots are requires developing your own expertise and understanding--but it is effort that will pay dividends in the short- and long-term. 

I have more thoughts, but I will stop here for now. Reflecting on the question posed above in this way has given me significantly more confidence in how I choose to use LLMs and siginficantly more peace in the emotions and uncertainty this struggle has created within me. Thanks again to Joaquin for suggesting this amazing analogy! 

Sunday, December 21, 2025

The Forest Unseen: A Year's Watch in Nature by David George Haskell

Goodreads:

A biologist reveals the secret world hidden in a single square meter of forest.

In this wholly original book, biologist David Haskell uses a one-square-meter patch of old-growth Tennessee forest as a window onto the entire natural world. Visiting it almost daily for one year to trace nature’s path through the seasons, he brings the forest and its inhabitants to vivid life.

Each of this book’s short chapters begins with a simple observation: a salamander scuttling across the leaf litter; the first blossom of spring wildflowers. From these, Haskell spins a brilliant web of biology and ecology, explaining the science that binds together the tiniest microbes and the largest mammals and describing the ecosystems that have cycled for thousands—sometimes millions—of years. Each visit to the forest presents a nature story in miniature as Haskell elegantly teases out the intricate relationships that order the creatures and plants that call it home.

Written with remarkable grace and empathy, The Forest Unseen is a grand tour of nature in all its profundity. Haskell is a perfect guide into the world that exists beneath our feet and beyond our backyards.
 

The Forest Unseen by Haskell scratched just the right itch for me, and pushed all the right buttons.

As documented in some of my previous posts, I have been searching for and enjoying media that can help me relax a bit more effectively. I was entranced by the meditative prose and descriptions in The Sound of a Wild Snail Eating by Bailey, and enthusiastically searched for the next book that would give me a similar vibe. I quickly found Winter World by Heinrich, which I discarded around halfway through. While his descriptions are certainly beautiful and relaxing, he reveals a little too much about some of the queasy and sometimes downright gory methods biologists of his era and before used when learning about the animals they study. I frequently found myself more disgusted than relaxed (for those that have read the book, the turtles of Chapter 11 are when I seriously began to consider putting the book down, and did so soon after).

So, my search continued. Through some online recommendations, I discovered The Forest Unseen and immediately found it to be the perfect combination of natural (as in, of the natural world), reflective, and tranquil. Haskell's prose is absolutely outstanding, with vivid descriptions of the mandala as well as its inhabitants. You can feel through his words the tension he experiences attempting to observe as much as possible while simultaneously minimizing his own footprint in the little forest ring, and the struggle as he confronts the fact that his mere presence there is a disturbance. He communicates this so effectively to the reader. You come away believing that he and the mandala have made a sacrifice, an offering: the sanctity of this natural forest temple disturbed, to offer us human beings a chance to see and appreciate the delicate tapestry of life, death, and balance that we so often take for granted.

While I frankly would have been satisfied with Haskell's tranquil and meticulous descriptions of the mandala in every season, I was deeply moved by his engagement with the philosophical ramifications of his expedition. Even the choice to name the forest space a mandala--invoking the geometric and cosmic shrines found in Eastern philosophies such as Buddhism--is a thoughtful means to remind you with every mention that you are joining him in a sacred space. You are not here to touch, you are here to look; to observe silently; to reflect. And reflect he does. Every chapter masterfully connects a biological or evolutionary principle unveiled by the mandala to a spiritual revelation of the self and the collective nature of humanity. 

The book is enjoyable completely on its own merits--Haskell has genuinely created a masterpiece here. As for my quest to find relaxing media, it wildly succeeds. Reading a chapter before falling asleep became a favorite part of my nightly routine, and my biggest complaint about the book was that it ended.

 

Previously on this blog:

 

The Forest Unseen: A Year's Watch in Nature: Haskell, David George:  9780143122944: Amazon.com: Books
Book cover

 

Thursday, November 20, 2025

Alba: A Wildlife Adventure

 

Gameplay showing capture of Imperial Spanish Eagle

UsTwo Games:

Join Alba as she visits her grandparents on a Mediterranean island. She is ready for a peaceful summer of wildlife exploration with her friend Ines, but when she sees an animal in danger, she realises she needs to do something about it!

My quest to relax more effectively continues. I made a list of relaxing games which I plan to work through in order to decompress; the first on this list was a little game called Alba: A Wildlife Adventure. I played on iOS with the Backbone controller, with no issues. 

The premise of this game is fairly simple: you play as Alba, a child exploring an island with her friend as part of their "organization" (AIWRL), dedicated to restoring the nature preserve on the island and stopping the construction of a hotel that will flatten and commoditize the beautiful local biome. The undertones are clear and the message isn't particularly subtle, but well-received nonetheless.  As you play, you listen for rare and interesting animals, which you photograph and identify. The game takes place in stages, with each stage corresponding to one day on the island. 

Overall, I really enjoyed the game. Out of the box, it immediately helped me achieve my primary aim, which was to relax. The game is very cheerful, with a simple but beautiful design. The gameplay was very simple to pick up, with a non-existent learning curve and not too many things to remember, which is good for someone trying to discharge after a long day of remembering things and thinking too hard. Underlying the entire experience is an upbeat and light-hearted soundtrack, with the tranquil sounds of the island's twinkling in the background. The leaves rustle in the trees, the wind is gentle, and in the distance, waves crash on the shore. As each stage winds down, the Sun gradually gets lower and you get lulled into a sleepy sunset, which helped slow me down enough to fall asleep a little faster whenever I played it.

View from above the nature preserve, which you repair as part of the game. I greatly enjoyed some of the "view" spots, where you could sit and watch the sun set or keep an eye out for the next animal on your list.

The mild list of chores and animals to track down never felt overwhelming or discouraging, and you can progress the story without finding everything. I might try to go back and find everything at some point as I think I only found ~80% of the animals in Alba's field guide. I so got used to listening for birds in the game that I started doing it in real life too, which I admit surprised me! The stages are nicely scoped to take 20-30 minutes each, so it never felt like a huge commitment to complete one as I was winding down for the evening. I was pleased to find that despite it technically counting as more screen time, I was still able to relax, which I attribute to the peaceful soundscape.

All in all, I had a good time and certainly felt more relaxed after playing than I did before. I will keep an eye out for similar games in the future, and if you have any recommendations please let me know.  

I appreciate the mission of the game, and want to plug some of the organizations promoted in the game here as well: AIWRLWWF

PCGamer

"Everybody understands trees," says Maria Sayans, CEO of Ustwo games. It's maybe not the kind of pitch we're used to hearing from developers, but that sentiment does get to the roots of Ustwo's mission. In just two months this indie developer has helped plant more than half a million trees with its 2020 release Alba: A Wildlife Adventure.

Alba has become the mascot for the tree-planting project and Ustwo's goal for a more environmentally kind world. Her face is the logo on Ustwo's Ecologi page, which also happens to be named Alba's Forest. Her in-game conservation group, AIWRL, is used to promote initiatives that Ustwo games work with.


Previously on this blog:

Sunday, July 20, 2025

The Sound of a Wild Snail Eating by Elisabeth Tova Bailey

Amazon:

While an illness keeps her bedridden, Bailey watches a wild snail that has taken up residence on her nightstand. As a result, she discovers the solace and sense of wonder that this mysterious creature brings and comes to a greater understanding of her own place in the world.

Intrigued by the snail’s molluscan anatomy, cryptic defenses, clear decision making, hydraulic locomotion, and courtship activities, Bailey becomes an astute and amused observer, offering a candid and engaging look into the curious life of this underappreciated small animal.

The Sound of a Wild Snail Eating is a remarkable journey of survival and resilience, showing us how a small part of the natural world can illuminate our own human existence, while providing an appreciation of what it means to be fully alive. 

I picked up and recently finished this book, recommended by a friend to help me relax a little before falling asleep. Without spoilers, the book tells the story of a woman who has a severe illness which forces her to stay bedridden for many years. During that time, she adopts a snail who accidentally found its way inside her home, and the snail becomes a both a vehicle for companionship and an allegory for her illness-induced isolation.  

The storytelling and writing is really outstanding; when describing her own medical situation, Bailey has a bleak and gripping tone that makes you almost feel like you are sick yourself and suffering alongside her. This makes the extended dives into the comings and goings of the titular snail that much more refreshing; you feel the relief and escape she felt gazing into her terrarium and watching the snail live its own life as she could not live hers, and yet, just like her, isolated and alone.

I originally picked up the book to help unwind and decompress before falling asleep; I heard the vivid imagery and plodding descriptions of the snail would be well suited to this. And I definitely found this to be true; Bailey's descriptions have a relaxing, almost meditative quality that fully immerses you in her small two-creature world.

I was astonished to learn in the notes at the end that the experience Bailey describes in the book was completely real; I honestly didn't look too much into the book before picking it up and assumed it was fictional or only based in part on a real story. 

The book was really excellent and I can definitely recommend if you want something relaxing to help decompress from your day-to-day stress. Despite having finished it, I occasionally revisit some bookmarked sections with particularly wonderful descriptions--like the building of the terrarium or the snail's ever-longer journeys from its original clay-pot oasis--to help me relax and fall asleep.   

Book cover

Thursday, May 15, 2025

Chess improvement: 5/14/2025

Disclaimer: I'm not very good at chess, leaving some personal notes here to try and get better.  

Played a few bullet games and won with estimated ELO between 1600-1700, which I'm pleased to see. I'm trying to get back to my old strength at least (was between 1500-1600) and I'm performing a little better under time pressure. I want to analyze the one longer time control game I played, against a human:

Game 1

Against human (level 1). Opened with the main line Vienna game. Game outcome: win. Accuracy: 91%, ELO for game: 2150. My opponent played pretty well, estimated ELO 1750. I played a standard opening, Black countered with Owen's defense and a double fianchetto:

End of opening.

My weakest opening move was this exchange:

Move 7.

Analysis favors dxe5. The following line doesn't make it super clear why so I want to consider from a fundamentals perspective. Having the pawn on e5 is good, it is strongly backed up by the knight and bishop. It also probes farther into Black's territory, forcing them to either deal with the more limited space or exchange, breaking their pawn structure (since they will be left with a lone pawn on the e file). By contrast, having the knight there is nice for a second but Bxe5 8. dxe5 leaves with me with a weaker pawn and Black has a strong three pawn chain ready to go and disrupt. I think the lesson to learn here is, when faced with multiple pieces to perform a capture, strengthen as many balances simultaneously as possible. From here on, I played pretty well. Ironically, having the knight on e5 led to this really strong position:

Move 17.

It wouldn't have been possible without the knight where it was, but that's why Black should have kicked the knight as soon as it could, to avoid exactly this scenario. I realize that this tactic wasn't set up because Nxe5 was a good move from me but because Black didn't punish me for it properly, so lesson still learned. From there on I didn't make any suboptimal moves and we arrived at

Move 26.

At this point I was able to force exchanges of the knight and bishop, and advanced a pawn to promotion. Black resigned.

Lessons learned:

  • When faced with multiple pieces to perform a capture, strengthen as many balances simultaneously as possible.


Previously on this blog:

Where is the LLLine?

My personal uses and struggles with large language models (LLMs) cannot be fully expounded upon in one post and I will make no such attempt ...