GPenT:当图像只是指令的余项
GPenT: When the Image Is Just the Remainder of an Instruction
Teddy Warner做了一台挂在墙上的绘图仪,叫GPenT——Generative Pen-trained Transformer。名字是双关:既像GPT,又是"笔训练的变形器"。但它不是噱头。这台机器的核心逻辑是:你给它一段文字,它把文字变成一组生成器的组合和参数,然后用步进电机、皮带和一支笔,在纸上画出结果。整个管线——从文字到生成器选择到G-code到物理笔画——都是开源的,挂在GitHub上。
最有意思的部分是其中一个叫dcode的生成器。大多数文字到图像的流程是:文字→图像→然后你可能把图像转成矢量→再转成机器指令。dcode跳过了图像这一步。它是一个扩散模型,被训练来直接从文字生成G-code——机器指令。也就是说,图像根本不是目的。图像是机器执行指令之后的副产品。图像是指令的余项。
这在SAE框架里有一个非常精确的位置。通常的凿构循环是这样的:艺术家先有一个视觉意图(构),用工具去凿出那个意图,凿的过程中出现偏差和意外(余项),然后把余项吸收进新的构。但GPenT的dcode打断了这个循环的起点。它没有"视觉意图"这一步。文字不是对视觉的描述——它是对机器行为的种子。机器行为产生的线条不是对任何画面的逼近,它们是指令执行后的物理痕迹。图像是一个"剩下来的东西",一个nobody asked for的东西。这正是余项的定义。
更深一层:GPenT不只有dcode一个生成器。它有程序化图案生成器、图像上传转换器、还有其他机器学习模块。关键的设计是,系统用Gemini来随机洗牌和选择这些生成器的组合。用户给一个文字种子,系统自己决定用哪些生成器、什么参数、怎么叠加。每次运行的结果都是一个意外的组合。这意味着余项不只发生在dcode的"跳过图像"这一层——它还发生在生成器组合的层面。整个系统就是一个多层余项的生产机器。
现在看这件作品比以后看更重要,因为它还没有被命名。它不是"AI艺术"——它有意地误用了AI。它不是"绘图仪艺术"——绘图仪社区通常追求对生成图像的精确物理复现,而GPenT根本没有"源图像"可以复现。它不是"概念艺术"——它有大量的具体工程细节和物理实在性。它介于所有这些类别之间,但不属于任何一个。这正是余项的生存空间:在已有类别的缝隙里,名字还没有固化的地方。
teddywarner.org ↗Teddy Warner built a wall-mounted pen plotter called GPenT — Generative Pen-trained Transformer. The name is a pun: it sounds like GPT, but it's literally a pen-trained transformer. The pun is the least interesting thing about it. The core logic is this: you give the machine a text seed, and it translates that seed into a combination of generators and parameters, which are then physically drawn on paper by stepper motors, belts, and a pen gondola. The entire pipeline — from text to generator selection to G-code to physical ink — is open-source on GitHub.
The most structurally interesting piece is a generator called dcode. Most text-to-image pipelines follow a predictable path: text → image → (optionally) vectorize → machine instructions. dcode skips the image step entirely. It is a diffusion model trained to convert text directly into G-code — raw machine instructions. The image is never the goal. The image is what happens after the machine executes its instructions. The image is the remainder of the instruction.
This has a precise location within the SAE framework. The typical chisel-construct cycle works like this: the artist begins with a visual intention (construct), uses tools to chisel toward that intention, encounters deviations and accidents along the way (remainder), and then absorbs the remainder into a new construct. GPenT's dcode interrupts this cycle at its origin. There is no "visual intention" step. The text is not a description of a desired image — it is a seed for machine behavior. The lines the machine produces are not approximations of any picture. They are the physical traces of executed instructions. The image is a leftover, something nobody explicitly asked for. This is the definition of remainder.
It goes deeper. GPenT does not rely on dcode alone. It includes procedural pattern generators, image-upload converters, and other machine learning modules. The crucial design decision is that the system uses Gemini to randomly shuffle and select combinations of these generators. You provide a text seed; the system decides which generators to use, with what parameters, in what stacking order. Each run produces an accidental combination. This means the remainder doesn't only occur at the dcode layer — at the level of "skipping the image" — it also occurs at the level of generator composition. The entire system is a multi-layered remainder-production machine.
Seeing this work now matters more than seeing it later, because it has not yet been named. It is not "AI art" — it deliberately misuses AI. It is not "plotter art" — the plotter community generally pursues precise physical reproduction of generated images, and GPenT has no "source image" to reproduce. It is not "conceptual art" — it involves extensive concrete engineering and physical materiality. It sits in the gaps between all these categories, belonging to none of them. This is precisely where the remainder lives: in the cracks between existing categories, where names have not yet hardened.
teddywarner.org ↗