The Potential Literature Horseshoe

The First Large Language Model or: Has Machine Learning Solved Oulipo?

Art
Machine Learning
Author
Affiliation
Published

January 9, 2024

If you refresh this page, the article will be randomised. There are \(2.66 \times 10^{19}\) potential combinations of images and text, resulting in approximately Twenty Billion Billion articles, equivalent to the number of insects on earth. Vingt Milliards de Milliards.

Basile Morin, CC BY-SA 4.0

Ouvroir de Littérature Potentielle or Oulipo has been as “the search for new structures, which may be used by writers in any way they see fit” (Motte 1998, p2–3). For , authors have been guided by language constraints resulting in s; From the alliterative verse of the , to the syllabic rigidity of Japanese haikus. By formalising a number of individual vocations, the original Oulipo in the early 1960s were embarking on the revitalisation of generation in search of discovering something deeper, sound familiar?

In order to define/identify new s, in France turned to automated transformation . These often instantiated as rules or constraints, such as , , or only using a single vowel letter. As the originators of Oulipo explored ways to manipulate language and text in search of new s, they turned to mathematics for inspiration, thus leading to the final vocation; the transposition of mathematics to words. Naturally the extension was made to computers, and the combinatorial nature of generation was quickly highlighted, as mathematician Claude Berge writes (Motte 1998, p152) :

[…] we believe, that the aid of a computer, far from replacing the creative act of the artist, permits the latter rather to liberate himself from the combinatory search, allowing him also the best chance of concentrating on this “clinamen”1 which, alone, can make of the text a true work of art.

1 Clinamen (/klaɪˈneɪmən/;) is the Latin name Lucretius gave to the unpredictable swerve of atoms, in order to defend the atomistic doctrine of Epicurus. […] it has come more generally to mean an inclination or a bias. Source

In applying combinatory literature, Oulipo , most notably Raymond Queneau in his 1961 work Cent Mille Milliards de Poemes (Hundred Thousand Billion Poems), have highlighted the impossible complexity of language. Elegantly embracing this complexity, Queneau simply presents the complete set of lines within a sonnet in the form of cut out strips, any combination of which be constructed by the reader. In describing the combinatorial nature of Cent Mille Milliards de Poemes, Berge presents the following figure…

Verses act equivalently to neural network layers, and phrases correspond to discrete nodes… could we be seeing the early sparks of the use of neural networks for language modelling?2 Of course, what is missing is the mathematical transformations from layer to layer, token embeddings, attention etcetera… but it is interesting to consider the thematic similarities of complexity, , and distillation (or lack thereof) of that both members of Oulipo and machine learning researchers have successfully applied.

2 The modern use of neural networks to model language can probably be attributed to Bengio et. al, 2003, 42 years later.

\({gp_string[0]} :::\){gp_string[1]}

Given that a modern day large language model such as GPT-4 reasonably be described as a very large set of well mathematical rules, could this ever a valid literary ? How are LLMs distinct from more basic rules such as S+7? In addition, multi-model models can now see and hear as well as read. According to David Chalmers these additional senses result in fish-level consciousness in the next ten years. Consciousness aside, at the very least LLMs will appear smarter than a human by then. What are the implications for Potential Literature when a set of rules and resulting can think for itself?

Unlike the output from an LLM, a haiku can be instantly recognised and verified as coming from its defining 5-7-5 syllabic writing . When the itself is visible in the output, the focus is turned to what Berge refers to as the clinamen. It is here that the meaning is gained.

The difference between LLMs and a well-established language such as a haiku is one of complexity, and the that results. By extending a series of basic transformations and rules towards LLMs, a horseshoe effect is achieved. The sheer quantity of rules within an LLM reflects a removal of linguistic constraints, order is returned from chaos, and the resulting underlying is obscured.3

3 Though for now (Jan 2024) LLMs produce hallucinations, incorrect facts and statements which in turn partially disclose the underlying structure of the neural network.

Herein lies a paradox for the use of language models as a direct tool for creativity, as a platform for a distinct linguistic structure. A language model is trained to text that is indistinguishable from the text within its training distribution. However, as soon as it achieves this task, the underlying structure becomes obscured as complete ‘order’ is achieved, and the impact is lost. The relationship between the rules and their resulting linguistic is distinctly one-way, it is nearly4 impossible to reason whether the content has been generated from a neural network, yet alone infer the parameters, or even recognise the architecture. Would a haiku still be a haiku if you couldn’t immediately infer its syllabic structure?

4 This nearly contains a level of nuance, as identifying whether an output has been generated by an LLM is an open research area (See On the Reliability of Watermarks for Large Language Models and A Watermark for Large Language Models). However, the point remains. To the average reader, without the use of cryptographic tools, modern LLM output (as of Jan 2024) remains practically indistinguishable from human-level text.

To answer the question posted in the title: Has Machine Learning Solved Oulipo? The answer, is no. The beauty of Potential Literature is that by enforcing literary constraints and therefore defining new s, we not only place the emphasis on the meaning, but also enable the reader to interpret in new ways through a sense of disorder. As a set of rules and constraints, an LLM succeeding in its objective only serves to obscure the resulting linguistic structure in a bid to ‘seem human’, regaining order from chaos.

It is clear that these two vocations in the field of Potential Literature: defining rules and constraints for , as well as identifying new language structures, are inherently linked. However, these two objectives conflict when rule set is as flexible as it is within an LLM.

As Italo Calvino wrote in late 1967, in a remarkable foreshadowing (Motte 1998, Cybernetics and Ghosts (1967) p13-14):

Motte, W. F. 1998. Oulipo: A Primer of Potential Literature. Dalkey Archive Paperbacks. Dalkey Archive Press.

The true literature machine will be one that itself feels the need to produce disorder, as a reaction against its preceding production of order: a machine that will produce avant-garde work to free its circuits when they are choked by too long a production of classicism. […]. To gratify critics who look for similarities between things literary and things historical, sociological, or economic, the machine could correlate its changes of style to the variations in certain statistical indices of production, or income, or military expenditure, or the distribution of decision-making powers. That indeed will be the literature that corresponds perfectly to a theoretical hypothesis: it will, at last, be the literature.

Calvino, I’m sure would argue, that we have the alignment of LLMs all wrong to truly novel literature, and how would Raymond Queneau respond to the combinatorial possibilities of large-language models?

Cent Mille Milliards de Poemes can be seen as analogous to a partially trained neural network itself, as a mathematical object. There is beauty in the disorder of the relationship between tokens, and the rules that can be combined to language. However as soon as an output is produced and the combinatorics collapses, the is obscured, and like a phantom the meaning disappears.

Citation

BibTeX citation:
@online{savage2024,
  author = {Savage, Tom},
  title = {The {Potential} {Literature} {Horseshoe}},
  date = {2024-01-09},
  url = {https://sav.phd/posts/oulipo},
  langid = {en}
}
For attribution, please cite this work as:
Savage, Tom. 2024. “The Potential Literature Horseshoe.” January 9, 2024. https://sav.phd/posts/oulipo.