This post is a living document. Last updated on 2026-04-12.

Computational research and AI

At the Wolfram Institute, I explore how ideas from mathematical physics can be carried into the framework of Wolfram models. My work focuses on designing the least complex discrete model capable of reproducing a macro phenomenon, with the potential to branch out and converge to known continuous structures in the limit of many computational steps. After implementing such a model and setting up a computational experiment, the empirical phase begins: making statistical observations, searching for emergent patterns, and identifying laws between well-designed observables as the complexity of the computational substrate generated by the experiment increases. It is exciting to think about how the seemingly continuous, unlimited, and unique world might emerge from a discrete one characterized by indivisibility, boundedness, and ambiguity.

Before committing to a concrete project, I brainstorm ideas, gather papers and other resources, distill suitable constructions, implement them in Wolfram Language, and create notebooks for experimentation and preliminary presentation to colleagues. If a direction looks promising, I zoom in and develop it further, eventually contributing to Wolfram Institute projects, paclets, or independent writing. In practice, three parts of this workflow are especially suitable for AI assistance: the exploratory phase, polishing phase, and repo maintenance and organization.

My utilization of AI

For two years I used AI in a basic way—for chat, reformats, and corrections. When working in a corporate environment, I was anxious about sharing unwanted context and responsibility for correctness. The peak of my AI usage was taking advantage of most of CodeCompanion’s functionality for Neovim, except agents.

A turning point came with the release of Claude Opus 4.5. I began experimenting with agentic workflows and got used to offloading tasks to Claude using only guidance and specifications, without touching the code directly. The results were surprisingly good. My first “wow” moment was vibe-coding a Swift app to manage and time-track my weekly tasks—generated entirely from a screenshot of my wife’s Excel table. I have never programmed in Swift and have never looked at the generated code, yet the app works perfectly. It feels like we are approaching a world where anyone can define the app they need and let the AI build it, like selecting a song on a jukebox.

Rapid prototyping—getting from A to B without deep dives—also seems essentially solved. When improvements are needed, the simplest approach is to direct the AI to reuse fragments of earlier code or templates. When I have time or need specific architecture, I still refine things myself, but AI gives me the option to strategically postpone that work.

Given this shift, I think it is time to consider agentic workflows and researcher best practices for computational research in the AI era. Since general standards and guidelines still seem to be missing, and I don’t have time to dive too deep, I decided to stay within the Claude ecosystem and began writing a plugin that captures my workflow.

Computational research plugin

The plugin (GitHub) is installed from the WolframInstitute marketplace. Core principles:

See the LLM-updated version history below for a changelog after each major release.

Some philosophy and future direction

AI will surely make certain hard skills obsolete. Paradigm shifts are exciting, and letting things go and starting anew deepens life experience. But several human roles remain essential:

A few ideas for missing pieces:

Version history (LLM updated)

Version 3 (2026-04-12)

Version 3.0.0 expands the plugin from a wiki-centric research tool into a full project lifecycle manager: scaffolding for research projects, paclet repos, and standalone paclets; skills for building and publishing paclets to Wolfram Cloud; LaTeX paper scaffolding with amsart and biblatex; and search across the Wolfram ecosystem (documentation, Function Repository, Community, writings, Physics Project).

Version 2 (2026-04-05)

Version 2.0.0 introduces a plain-markdown wiki as the central knowledge base — readable by both LLMs and humans, version-controlled, with cross-references the LLM navigates instead of scanning every file. Resources and notebooks are stored as Markdown with recovery instructions, built on demand (idea by sw1sh). A revision protocol prevents the LLM from overwriting human-edited content. A tour skill walks through the project from simple to advanced.

The core MCP servers remain Wolfram MCP (or the unofficial wolfram-mcp with LSP support) and arXiv-mcp (plus arxiv-latex-mcp for reading LaTeX source).

Version 1 (2026-03-04)

The first version of the plugin bundled a few Claude skills: wolfram-notebook for creating Wolfram notebooks from prompts via Markdown import (an idea by sw1sh), and computational-exploration for scaffolding a structured research project.

The skill searched arXiv and Wolfram Community for papers, downloaded them, and produced organized notes with citations. Planned skills included notes-to-article, list-topics, setup-experiment, and polish-research.

After using this on several projects, I found the design too broad and not goal-oriented enough. Exploration, resource management, and knowledge organization were tangled together. Knowledge was spread across CLAUDE.md, notebooks, LaTeX notes, and resources. Generated notebooks were redundant to store, as they can be imported from Markdown anyway. Resources had no recovery mechanism, so a fresh clone lost all downloaded papers. And there was no revision protocol, so the LLM could overwrite user-edited content.