Adema
The Ethics of Emergent Creativity: Can We Move Beyond Writing as Human Enterprise, Commodity and Innovation?
2019


# 3\. The Ethics of Emergent Creativity: Can We Move Beyond Writing as Human
Enterprise, Commodity and Innovation?

Janneke Adema

© 2019 Janneke Adema, CC BY 4.0
[https://doi.org/10.11647/OBP.0159.03](https://doi.org/10.11647/OBP.0159.03)

In 2013, the Authors’ Licensing & Collecting Society
(ALCS)[1](ch3.xhtml#footnote-152) commissioned a survey of its members to
explore writers’ earnings and contractual issues in the UK. The survey, the
results of which were published in the summary booklet ‘What Are Words Worth
Now?’, was carried out by Queen Mary, University of London. Almost 2,500
writers — from literary authors to academics and screenwriters — responded.
‘What Are Words Worth Now?’ summarises the findings of a larger study titled
‘The Business Of Being An Author: A Survey Of Authors’ Earnings And
Contracts’, carried out by Johanna Gibson, Phillip Johnson and Gaetano Dimita
and published in April 2015 by Queen Mary University of
London.[2](ch3.xhtml#footnote-151) The ALCS press release that accompanies the
study states that this ‘shocking’ new research into authors’ earnings finds a
‘dramatic fall, both in incomes, and the number of those working full-time as
writers’.[3](ch3.xhtml#footnote-150) Indeed, two of the main findings of the
study are that, first of all, the income of a professional author (which the
research defines as those who dedicate the majority of their time to writing)
has dropped 29% between 2005 and 2013, from £12,330 (£15,450 in real terms) to
just £11,000. Furthermore, the research found that in 2005 40% of professional
authors earned their incomes solely from writing, where in 2013 this figure
had dropped to just 11.5%.[4](ch3.xhtml#footnote-149)

It seems that one of the primary reasons for the ALCS to conduct this survey
was to collect ‘accurate, independent data’ on writers’ earnings and
contractual issues, in order for the ALCS to ‘make the case for authors’
rights’ — at least, that is what the ALCS Chief Executive Owen Atkinson writes
in the introduction accompanying the survey, which was sent out to all ALCS
members.[5](ch3.xhtml#footnote-148) Yet although this research was conducted
independently and the researchers did not draw conclusions based on the data
collected — in the form of policy recommendations for example — the ALCS did
frame the data and findings in a very specific way, as I will outline in what
follows; this framing includes both the introduction to the survey and the
press release that accompanies the survey’s findings. Yet to some extent this
framing, as I will argue, is already apparent in the methodology used to
produce the data underlying the research report.

First of all, let me provide an example of how the research findings have been
framed in a specific way. Chief Executive Atkinson mentions in his
introduction to the survey that the ALCS ‘exists to ensure that writers are
treated fairly and remunerated appropriately’. He continues that the ALCS
commissioned the survey to collect ‘accurate, independent data,’ in order to
‘make the case for writers’ rights’.[6](ch3.xhtml#footnote-147) Now this focus
on rights in combination with remuneration is all the more noteworthy if we
look at an earlier ALCS funded report from 2007, ‘Authors’ Earnings from
Copyright and Non-Copyright Sources: a Survey of 25,000 British and German
Writers’. This report is based on the findings of a 2006 writers’ survey,
which the 2013 survey updates. The 2007 report argues conclusively that
current copyright law has empirically failed to ensure that authors receive
appropriate reward or remuneration for the use of their
work.[7](ch3.xhtml#footnote-146) The data from the subsequent 2013 survey show
an even bleaker picture as regards the earnings of writers. Yet Atkinson
argues in the press release accompanying the findings of the 2013 survey that
‘if writers are to continue making their irreplaceable contribution to the UK
economy, they need to be paid fairly for their work. This means ensuring
clear, fair contracts with equitable terms and a copyright regime that support
creators and their ability to earn a living from their
creations’.[8](ch3.xhtml#footnote-145) Atkinson does not outline what this
copyright regime should be, nor does he draw attention to how this model could
be improved. More importantly, the fact that a copyright model is needed to
ensure fair pay stands uncontested for Atkinson and the ALCS — not surprising
perhaps, as protecting and promoting the rights of authors is the primary
mission of this member society. If there is any culprit to be held responsible
for the study’s ‘shocking’ findings, it is the elusive and further undefined
notion of ‘the digital’. According to Atkinson, digital technology is
increasingly challenging the mission of the ALCS to ensure fair remuneration
for writers, since it is ‘driving new markets and leading the copyright
debate’.[9](ch3.xhtml#footnote-144) The 2013 study is therefore, as Atkinson
states ‘the first to capture the impact of the digital revolution on writers’
working lives’.[10](ch3.xhtml#footnote-143) This statement is all the more
striking if we take into consideration that none of the questions in the 2013
survey focus specifically on digital publishing.[11](ch3.xhtml#footnote-142)
It therefore seems that — despite earlier findings — the ALCS has already
decided in advance what ‘the digital’ is and that a copyright regime is the
only way to ensure fair remuneration for writers in a digital context.

## Creative Industries

This strong uncontested link between copyright and remuneration can be traced
back to various other aspects of the 2015 report and its release. For example,
the press release draws a strong connection between the findings of the report
and the development of the creative industries in the UK. Again, Atkinson
states in the press release:

These are concerning times for writers. This rapid decline in both author
incomes and in the numbers of those writing full-time could have serious
implications for the economic success of the creative industries in the
UK.[12](ch3.xhtml#footnote-141)

This connection to the creative industries — ‘which are now worth £71.4
billion per year to the UK economy’,[13](ch3.xhtml#footnote-140) Atkinson
points out — is not surprising where the discourse around creative industries
maintains a clear bond between intellectual property rights and creative
labour. As Geert Lovink and Ned Rossiter state in their MyCreativity Reader,
the creative industries consist of ‘the generation and exploitation of
intellectual property’.[14](ch3.xhtml#footnote-139) Here they refer to a
definition created as part of the UK Government’s Creative Industries Mapping
Document,[15](ch3.xhtml#footnote-138) which states that the creative
industries are ‘those industries which have their origin in individual
creativity, skill and talent and which have a potential for wealth and job
creation through the generation and exploitation of intellectual property’.
Lovink and Rossiter point out that the relationship between IP and creative
labour lies at the basis of the definition of the creative industries where,
as they argue, this model of creativity assumes people only create to produce
economic value. This is part of a larger trend Wendy Brown has described as
being quintessentially neoliberal, where ‘neoliberal rationality disseminates
the model of the market to all domains and activities’ — and this includes the
realm of politics and rights.[16](ch3.xhtml#footnote-137) In this sense the
economization of culture and the concept of creativity is something that has
become increasingly embedded and naturalised. The exploitation of intellectual
property stands at the basis of the creative industries model, in which
cultural value — which can be seen as intricate, complex and manifold —
becomes subordinated to the model of the market; it becomes economic
value.[17](ch3.xhtml#footnote-136)

This direct association of cultural value and creativity with economic value
is apparent in various other facets of the ALCS commissioned research and
report. Obviously, the title of the initial summary booklet, as a form of
wordplay, asks ‘What are words worth?’. It becomes clear from the context of
the survey that the ‘worth’ of words will only be measured in a monetary
sense, i.e. as economic value. Perhaps even more important to understand in
this context, however, is how this economic worth of words is measured and
determined by focusing on two fixed and predetermined entities in advance.
First of all, the study focuses on individual human agents of creativity (i.e.
creators contributing economic value): the value of writing is established by
collecting data and making measurements at the level of individual authorship,
addressing authors/writers as singular individuals throughout the survey.
Secondly, economic worth is further determined by focusing on the fixed and
stable creative objects authors produce, in other words the study establishes
from the outset a clear link between the worth and value of writing and
economic remuneration based on individual works of
writing.[18](ch3.xhtml#footnote-135) Therefore in this process of determining
the economic worth of words, ‘writers’ and/or ‘authors’ are described and
positioned in a certain way in this study (i.e. as the central agents and
originators of creative objects), as is the form their creativity takes in the
shape of quantifiable outputs or commodities. The value of both these units of
measurement (the creator and the creative objects) are then set off against
the growth of the creative industries in the press release.

The ALCS commissioned survey provides some important insights into how
authorship, cultural works and remuneration — and ultimately, creativity — is
currently valued, specifically in the context of the creative industries
discourse in the UK. What I have tried to point out — without wanting to
downplay the importance either of writers receiving fair remuneration for
their work or of issues related to the sustainability of creative processes —
is that the findings from this survey have both been extracted and
subsequently framed based on a very specific economic model of creativity (and
authorship). According to this model, writing and creativity are sustained
most clearly by an individual original creator (an author) who extracts value
from the work s/he creates and distributes, aided by an intellectual property
rights regime. As I will outline more in depth in what follows, the enduring
liberal and humanist presumptions that underlie this survey continuously
reinforce the links between the value of writing and established IP and
remuneration regimes, and support a vision in which authorship and creativity
are dependent on economic incentives and ownership of works. By working within
this framework and with these predetermined concepts of authorship and
creativity (and ‘the digital’) the ALCS is strongly committed to the upkeep of
a specific model and discourse of creativity connected to the creative
industries. The ALCS does not attempt to complicate this model, nor does it
search for alternatives even when, as the 2007 report already implies, the
existing IP model has empirically failed to support the remuneration of
writers appropriately.

I want to use this ALCS survey as a reference point to start problematising
existing constructions of creativity, authorship, ownership, and
sustainability in relation to the ethics of publishing. To explore what ‘words
are worth’ and to challenge the hegemonic liberal humanist model of creativity
— to which the ALCS adheres — I will examine a selection of theoretical and
practical publishing and writing alternatives, from relational and posthuman
authorship to radical open access and uncreative writing. These alternatives
do not deny the importance of fair remuneration and sustainability for the
creative process; however, they want to foreground and explore creative
relationalities that move beyond the individual author and her ownership of
creative objects as the only model to support creativity and cultural
exchange. By looking at alternatives while at the same time complicating the
values and assumptions underlying the dominant narrative for IP expansion, I
want to start imagining what more ethical, fair and emergent forms of
creativity might entail. Forms that take into consideration the various
distributed and entangled agencies involved in the creation of cultural
content — which are presently not being included in the ALCS survey on fair
remuneration, for example. As I will argue, a reconsideration of the liberal
and humanist model of creativity might actually create new possibilities to
consider the value of words, and with that perhaps new solutions to the
problems pointed out in the ALCS study.

## Relational and Distributed Authorship

One of the main critiques of the liberal humanist model of authorship concerns
how it privileges the author as the sole source and origin of creativity. Yet
the argument has been made, both from a historical perspective and in relation
to today’s networked digital environment, that authorship and creativity, and
with that the value and worth of that creativity, are heavily
distributed.[19](ch3.xhtml#footnote-134) Should we therefore think about how
we can distribute notions of authorship and creativity more ethically when
defining the worth and value of words too? Would this perhaps mean a more
thorough investigation of what and who the specific agencies involved in
creative production are? This seems all the more important given that, today,
‘the value of words’ is arguably connected not to (distributed) authors or
creative agencies, but to rights holders (or their intermediaries such as
agents).[20](ch3.xhtml#footnote-133) From this perspective, the problem with
the copyright model as it currently functions is that the creators of
copyright don’t necessarily end up benefiting from it — a point that was also
implied by the authors of the 2007 ALCS commissioned report. Copyright
benefits rights holders, and rights holders are not necessarily, and often not
at all, involved in the production of creative work.

Yet copyright and the work as object are knit tightly to the authorship
construct. In this respect, the above criticism notwithstanding, in a liberal
vision of creativity and ownership the typical unit remains either the author
or the work. This ‘solid and fundamental unit of the author and the work’ as
Foucault has qualified it, albeit challenged, still retains a privileged
position.[21](ch3.xhtml#footnote-132) As Mark Rose argues, authorship — as a
relatively recent cultural formation — can be directly connected to the
commodification of writing and to proprietorship. Even more it developed in
tandem with the societal principle of possessive individualism, in which
individual property rights are protected by the social
order.[22](ch3.xhtml#footnote-131)

Some of the more interesting recent critiques of these constructs of
authorship and proprietorship have come from critical and feminist legal
studies, where scholars such as Carys Craig have started to question these
connections further. As Craig, Turcotte and Coombe argue, IP and copyright are
premised on liberal and neoliberal assumptions and constructs, such as
ownership, private rights, self-interest and
individualism.[23](ch3.xhtml#footnote-130) In this sense copyright,
authorship, the work as object, and related discourses around creativity
continuously re-establish and strengthen each other as part of a self-
sustaining system. We have seen this with the discourse around creative
industries, as part of which economic value comes to stand in for the creative
process itself, which, according to this narrative, can only be sustained
through an IP regime. Furthermore, from a feminist new materialist position,
the current discourse on creativity is very much a material expression of
creativity rather than merely its representation, where this discourse has
been classifying, constructing, and situating creativity (and with that,
authorship) within a neoliberal framework of creative industries.

Moving away from an individual construct of creativity therefore immediately
affects the question of the value of words. In our current copyright model
emphasis lies on the individual original author, but in a more distributed
vision the value of words and of creative production can be connected to a
broader context of creative agencies. Historically there has been a great
discursive shift from a valuing of imitation or derivation to a valuing of
originality in determining what counts as creativity or creative output.
Similar to Rose, Craig, Turcotte and Coombe argue that the individuality and
originality of authorship in its modern form established a simple route
towards individual ownership and the propertisation of creative achievement:
the original work is the author’s ownership whereas the imitator or pirate is
a trespasser of thief. In this sense original authorship is
‘disproportionately valued against other forms of cultural expression and
creative play’, where copyright upholds, maintains and strengthens the binary
between imitator and creator — defined by Craig, Turcotte and Coombe as a
‘moral divide’.[24](ch3.xhtml#footnote-129) This also presupposes a notion of
creativity that sees individuals as autonomous, living in isolation from each
other, ignoring their relationality. Yet as Craig, Turcotte and Coombe argue,
‘the act of writing involves not origination, but rather the adaptation,
derivation, translation and recombination of “raw material” taken from
previously existing texts’.[25](ch3.xhtml#footnote-128) This position has also
been explored extensively from within remix studies and fan culture, where the
adaptation and remixing of cultural content stands at the basis of creativity
(what Lawrence Lessig has called Read/Write culture, opposed to Read/Only
culture).[26](ch3.xhtml#footnote-127) From the perspective of access to
culture — instead of ownership of cultural goods or objects — one could also
argue that its value would increase when we are able to freely distribute it
and with that to adapt and remix it to create new cultural content and with
that cultural and social value — this within a context in which, as Craig,
Turcotte and Coombe point out, ‘the continuous expansion of intellectual
property rights has produced legal regimes that restrict access and downstream
use of information resources far beyond what is required to encourage their
creation’[27](ch3.xhtml#footnote-126)

To move beyond Enlightenment ideals of individuation, detachment and unity of
author and work, which determine the author-owner in the copyright model,
Craig puts forward a post-structuralist vision of relational authorship. This
sees the individual as socially situated and constituted — based also on
feminist scholarship into the socially situated self — where authorship in
this vision is situated within the communities in which it exists, but also in
relation to the texts and discourses that constitute it. Here creativity takes
place from within a network of social relations and the social dimensions of
authorship are recognised, as connectivity goes hand in hand with individual
autonomy. Craig argues that copyright should not be defined out of clashing
rights and interests but should instead focus on the kinds of relationships
this right would structure; it should be understood in relational terms: ‘it
structures relationships between authors and users, allocating powers and
responsibilities amongst members of cultural communities, and establishing the
rules of communication and exchange’.[28](ch3.xhtml#footnote-125) Cultural
value is then defined within these relationships.

## Open Access and the Ethics of Care

Craig, Turcotte and Coombe draw a clear connection between relational
authorship, feminism and (the ideals of) the open access movement, where as
they state, ‘rather than adhering to the individuated form of authorship that
intellectual property laws presuppose, open access initiatives take into
account varying forms of collaboration, creativity and
development’.[29](ch3.xhtml#footnote-124) Yet as I and others have argued
elsewhere,[30](ch3.xhtml#footnote-123) open access or open access publishing
is not a solid ideological block or model; it is made up of disparate groups,
visions and ethics. In this sense there is nothing intrinsically political or
democratic about open access, practitioners of open access can just as well be
seen to support and encourage open access in connection with the neoliberal
knowledge economy, with possessive individualism — even with CC licenses,
which can be seen as strengthening individualism —[31](ch3.xhtml#footnote-122)
and with the unity of author and work.[32](ch3.xhtml#footnote-121)

Nevertheless, there are those within the loosely defined and connected
‘radical open access community’, that do envision their publishing outlook and
relationship towards copyright, openness and authorship within and as part of
a relational ethics of care.[33](ch3.xhtml#footnote-120) For example Mattering
Press, a scholar-led open access book publishing initiative founded in 2012
and launched in 2016, publishes in the field of Science and Technology Studies
(STS) and works with a production model based on cooperation and shared
scholarship. As part of its publishing politics, ethos and ideology, Mattering
Press is therefore keen to include various agencies involved in the production
of scholarship, including ‘authors, reviewers, editors, copy editors, proof
readers, typesetters, distributers, designers, web developers and
readers’.[34](ch3.xhtml#footnote-119) They work with two interrelated feminist
(new materialist) and STS concepts to structure and perform this ethos:
mattering[35](ch3.xhtml#footnote-118) and care.[36](ch3.xhtml#footnote-117)
Where it concerns mattering, Mattering Press is conscious of how their
experiment in knowledge production, being inherently situated, puts new
relationships and configurations into the world. What therefore matters for
them are not so much the ‘author’ or the ‘outcome’ (the object), but the
process and the relationships that make up publishing:

[…] the way academic texts are produced matters — both analytically and
politically. Dominant publishing practices work with assumptions about the
conditions of academic knowledge production that rarely reflect what goes on
in laboratories, field sites, university offices, libraries, and various
workshops and conferences. They tend to deal with almost complete manuscripts
and a small number of authors, who are greatly dependent on the politics of
the publishing industry.[37](ch3.xhtml#footnote-116)

For Mattering Press care is something that extends not only to authors but to
the many other actants involved in knowledge production, who often provide
free volunteer labour within a gift economy context. As Mattering Press
emphasises, the ethics of care ‘mark vital relations and practices whose value
cannot be calculated and thus often goes unacknowledged where logics of
calculation are dominant’.[38](ch3.xhtml#footnote-115) For Mattering Press,
care can help offset and engage with the calculative logic that permeates
academic publishing:

[…] the concept of care can help to engage with calculative logics, such as
those of costs, without granting them dominance. How do we calculate so that
calculations do not dominate our considerations? What would it be to care for
rather than to calculate the cost of a book? This is but one and arguably a
relatively conservative strategy for allowing other logics than those of
calculation to take centre stage in publishing.[39](ch3.xhtml#footnote-114)

This logic of care refers, in part, to making visible the ‘unseen others’ as
Joe Deville (one of Mattering Press’s editors) calls them, who exemplify the
plethora of hidden labour that goes unnoticed within this object and author-
focused (academic) publishing model. As Endre Danyi, another Mattering Press
editor, remarks, quoting Susan Leigh Star: ‘This is, in the end, a profoundly
political process, since so many forms of social control rely on the erasure
or silencing of various workers, on deleting their work from representations
of the work’.[40](ch3.xhtml#footnote-113)

## Posthuman Authorship

Authorship is also being reconsidered as a polyvocal and collaborative
endeavour by reflecting on the agentic role of technology in authoring
content. Within digital literature, hypertext and computer-generated poetry,
media studies scholars have explored the role played by technology and the
materiality of text in the creation process, where in many ways writing can be
seen as a shared act between reader, writer and computer. Lori Emerson
emphasises that machines, media or technology are not neutral in this respect,
which complicates the idea of human subjectivity. Emerson explores this
through the notion of ‘cyborg authorship’, which examines the relation between
machine and human with a focus on the potentiality of in-
betweenness.[41](ch3.xhtml#footnote-112) Dani Spinosa talks about
‘collaboration with an external force (the computer, MacProse, technology in
general)’.[42](ch3.xhtml#footnote-111) Extending from the author, the text
itself, and the reader as meaning-writer (and hence playing a part in the
author function), technology, she states, is a fourth term in this
collaborative meaning-making. As Spinosa argues, in computer-generated texts
the computer is more than a technological tool and becomes a co-producer,
where it can occur that ‘the poet herself merges with the machine in order to
place her own subjectivity in flux’.[43](ch3.xhtml#footnote-110) Emerson calls
this a ‘break from the model of the poet/writer as divinely inspired human
exemplar’, which is exemplified for her in hypertext, computer-generated
poetry, and digital poetry.[44](ch3.xhtml#footnote-109)

Yet in many ways, as Emerson and Spinosa also note, these forms of posthuman
authorship should be seen as part of a larger trend, what Rolf Hughes calls an
‘anti-authorship’ tradition focused on auto-poesis (self-making), generative
systems and automatic writing. As Hughes argues, we see this tradition in
print forms such as Oulipo and in Dada experiments and surrealist games
too.[45](ch3.xhtml#footnote-108) But there are connections here with broader
theories that focus on distributed agency too, especially where it concerns
the influence of the materiality of the text. Media theorists such as N.
Katherine Hayles and Johanna Drucker have extensively argued that the
materiality of the page is entangled with the intentionality of the author as
a further agency; Drucker conceptualises this through a focus on ‘conditional
texts’ and ‘performative materiality’ with respect to the agency of the
material medium (be it the printed page or the digital
screen).[46](ch3.xhtml#footnote-107)

Where, however, does the redistribution of value creation end in these
narratives? As Nick Montfort states with respect to the agency of technology,
‘should other important and inspirational mechanisms — my CD player, for
instance, and my bookshelves — get cut in on the action as
well?’[47](ch3.xhtml#footnote-106) These distributed forms of authorship do
not solve issues related to authorship or remuneration but further complicate
them. Nevertheless Montfort is interested in describing the processes involved
in these types of (posthuman) co-authorship, to explore the (previously
unexplored) relationships and processes involved in the authoring of texts
more clearly. As he states, this ‘can help us understand the role of the
different participants more fully’.[48](ch3.xhtml#footnote-105) In this
respect a focus on posthuman authorship and on the various distributed
agencies that play a part in creative processes is not only a means to disrupt
the hegemonic focus on a romantic single and original authorship model, but it
is also about a sensibility to (machinic) co-authorship, to the different
agencies involved in the creation of art, and playing a role in creativity
itself. As Emerson remarks in this respect: ‘we must be wary of granting a
(romantic) specialness to human intentionality — after all, the point of
dividing the responsibility for the creation of the poems between human and
machine is to disrupt the singularity of human identity, to force human
identity to intermingle with machine identity’.[49](ch3.xhtml#footnote-104)

## Emergent Creativity

This more relational notion of rights and the wider appreciation of the
various (posthuman) agencies involved in creative processes based on an ethics
of care, challenges the vision of the single individualised and original
author/owner who stands at the basis of our copyright and IP regime — a vision
that, it is worth emphasising, can be seen as a historical (and Western)
anomaly, where collaborative, anonymous, and more polyvocal models of
authorship have historically prevailed.[50](ch3.xhtml#footnote-103) The other
side of the Foucauldian double bind, i.e. the fixed cultural object that
functions as a commodity, has however been similarly critiqued from several
angles. As stated before, and as also apparent from the way the ALCS report
has been framed, currently our copyright and remuneration regime is based on
ownership of cultural objects. Yet as many have already made clear, this
regime and discourse is very much based on physical objects and on a print-
based context.[51](ch3.xhtml#footnote-102) As such the idea of ‘text’ (be it
print or digital) has not been sufficiently problematised as versioned,
processual and materially changing within an IP context. In other words, text
and works are mostly perceived as fixed and stable objects and commodities
instead of material and creative processes and entangled relationalities. As
Craig et al. state, ‘the copyright system is unfortunately employed to
reinforce the norms of the analog world’.[52](ch3.xhtml#footnote-101) In
contrast to a more relational perspective, the current copyright regime views
culture through a proprietary lens. And it is very much this discursive
positioning, or as Craig et al. argue ‘the language of “ownership,”
“property,” and “commodity”’, which ‘obfuscates the nature of copyright’s
subject matter, and cloaks the social and cultural conditions of its
production and the implications of its
protection’.[53](ch3.xhtml#footnote-100) How can we approach creativity in
context, as socially and culturally situated, and not as the free-standing,
stable product of a transcendent author, which is very much how it is being
positioned within an economic and copyright framework? This hegemonic
conception of creativity as property fails to acknowledge or take into
consideration the manifold, distributed, derivative and messy realities of
culture and creativity.

It is therefore important to put forward and promote another more emergent
vision of creativity, where creativity is seen as both processual and only
ever temporarily fixed, and where the work itself is seen as being the product
of a variety of (posthuman) agencies. Interestingly, someone who has written
very elaborately about a different form of creativity relevant to this context
is one of the authors of the ALCS commissioned report, Johanna Gibson. Similar
to Craig, who focuses on the relationality of copyright, Gibson wants to pay
more attention to the networking of creativity, moving it beyond a focus on
traditional models of producers and consumers in exchange for a ‘many-to-many’
model of creativity. For Gibson, IP as a system aligns with a corporate model
of creativity, one which oversimplifies what it means to be creative and
measures it against economic parameters alone.[54](ch3.xhtml#footnote-099) In
many ways in policy driven visions, IP has come to stand in for the creative
process itself, Gibson argues, and is assimilated within corporate models of
innovation. It has thus become a synonym for creativity, as we have seen in
the creative industries discourse. As Gibson explains, this simplified model
of creativity is very much a ‘discursive strategy’ in which the creator is
mythologised and output comes in the form of commodified
objects.[55](ch3.xhtml#footnote-098) In this sense we need to re-appropriate
creativity as an inherently fluid and uncertain concept and practice.

Yet this mimicry of creativity by IP and innovation at the same time means
that any re-appropriation of creativity from the stance of access and reuse is
targeted as anti-IP and thus as standing outside of formal creativity. Other,
more emergent forms of creativity have trouble existing within this self-
defining and sustaining hegemonic system. This is similar to what Craig
remarked with respect to remixed, counterfeit and pirated, and un-original
works, which are seen as standing outside the system. Gibson uses actor
network theory (ANT) as a framework to construct her network-based model of
creativity, where for her ANT allows for a vision that does not fix creativity
within a product, but focuses more on the material relationships and
interactions between users and producers. In this sense, she argues, a network
model allows for plural agencies to be attributed to creativity, including
those of users.[56](ch3.xhtml#footnote-097)

An interesting example of how the hegemonic object-based discourse of
creativity can be re-appropriated comes from the conceptual poet Kenneth
Goldsmith, who, in what could be seen as a direct response to this dominant
narrative, tries to emphasise that exactly what this discourse classifies as
‘uncreative’, should be seen as valuable in itself. Goldsmith points out that
appropriating is creative and that he uses it as a pedagogical method in his
classes on ‘Uncreative Writing’ (which he defines as ‘the art of managing
information and representing it as writing’[57](ch3.xhtml#footnote-096)). Here
‘uncreative writing’ is something to strive for and stealing, copying, and
patchwriting are elevated as important and valuable tools for writing. For
Goldsmith the digital environment has fostered new skills and notions of
writing beyond the print-based concepts of originality and authorship: next to
copying, editing, reusing and remixing texts, the management and manipulation
of information becomes an essential aspect of
creativity.[58](ch3.xhtml#footnote-095) Uncreative writing involves a
repurposing and appropriation of existing texts and works, which then become
materials or building blocks for further works. In this sense Goldsmith
critiques the idea of texts or works as being fixed when asking, ‘if artefacts
are always in flux, when is a historical work determined to be
“finished”?’[59](ch3.xhtml#footnote-094) At the same time, he argues, our
identities are also in flux and ever shifting, turning creative writing into a
post-identity literature.[60](ch3.xhtml#footnote-093) Machines play important
roles in uncreative writing, as active agents in the ‘managing of
information’, which is then again represented as writing, and is seen by
Goldsmith as a bridge between human-centred writing and full-blown
‘robopoetics’ (literature written by machines, for machines). Yet Goldsmith is
keen to emphasise that these forms of uncreative writing are not beholden to
the digital medium, and that pre-digital examples are plentiful in conceptual
literature and poetry. He points out — again by a discursive re-appropriation
of what creativity is or can be — that sampling, remixing and appropriation
have been the norm in other artistic and creative media for decades. The
literary world is lagging behind in this respect, where, despite the
experiments by modernist writers, it continues neatly to delineate avant-garde
from more general forms of writing. Yet as Goldsmith argues the digital has
started to disrupt this distinction again, moving beyond ‘analogue’ notions of
writing, and has fuelled with it the idea that there might be alternative
notions of writing: those currently perceived as
uncreative.[61](ch3.xhtml#footnote-092)

## Conclusion

There are two addendums to the argument I have outlined above that I would
like to include here. First of all, I would like to complicate and further
critique some of the preconceptions still inherent in the relational and
networked copyright models as put forward by Craig et al. and Gibson. Both are
in many ways reformist and ‘responsive’ models. Gibson, for example, does not
want to do away with IP rights, she wants them to develop and adapt to mirror
society more accurately according to a networked model of creativity. For her,
the law is out of tune with its public, and she wants to promote a more
inclusive networked (copy) rights model.[62](ch3.xhtml#footnote-091) For Craig
too, relationalities are established and structured by rights first and
foremost. Yet from a posthuman perspective we need to be conscious of how the
other actants involved in creativity would fall outside such a humanist and
subjective rights model.[63](ch3.xhtml#footnote-090) From texts and
technologies themselves to the wider environmental context and to other
nonhuman entities and objects: in what sense will a copyright model be able to
extend such a network beyond an individualised liberal humanist human subject?
What do these models exclude in this respect and in what sense are they still
limited by their adherence to a rights model that continues to rely on
humanist nodes in a networked or relational model? As Anna Munster has argued
in a talk about the case of the monkey selfie, copyright is based on a logic
of exclusion that does not line up with the assemblages of agentic processes
that make up creativity and creative expression.[64](ch3.xhtml#footnote-089)
How can we appreciate the relational and processual aspects of identity, which
both Craig and Gibson seem to want to promote, if we hold on to an inherently
humanist concept of subjectification, rights and creativity?

Secondly, I want to highlight that we need to remain cautious of a movement
away from copyright and the copyright industries, to a context of free culture
in which free content — and the often free labour it is based upon — ends up
servicing the content industries (i.e. Facebook, Google, Amazon). We must be
wary when access or the narrative around (open) access becomes dominated by
access to or for big business, benefitting the creative industries and the
knowledge economy. The danger of updating and adapting IP law to fit a
changing digital context and to new technologies, of making it more inclusive
in this sense — which is something both Craig and Gibson want to do as part of
their reformative models — is that this tends to be based on a very simplified
and deterministic vision of technology, as something requiring access and an
open market to foster innovation. As Sarah Kember argues, this technocratic
rationale, which is what unites pro-and anti-copyright activists in this
sense, essentially de-politicises the debate around IP; it is still a question
of determining the value of creativity through an economic perspective, based
on a calculative lobby.[65](ch3.xhtml#footnote-088) The challenge here is to
redefine the discourse in such a way that our focus moves away from a dominant
market vision, and — as Gibson and Craig have also tried to do — to emphasise
a non-calculative ethics of relations, processes and care instead.

I would like to return at this point to the ALCS report and the way its
results have been framed within a creative industries discourse.
Notwithstanding the fact that fair remuneration and incentives for literary
production and creativity in general are of the utmost importance, what I have
tried to argue here is that the ‘solution’ proposed by the ALCS does not do
justice to the complexities of creativity. When discussing remuneration of
authors, the ALCS seems to prefer a simple solution in which copyright is seen
as a given, the digital is pointed out as a generalised scapegoat, and
binaries between print and digital are maintained and strengthened.
Furthermore, fair remuneration is encapsulated by the ALCS within an economic
calculative logic and rhetoric, sustained by and connected to a creative
industries discourse, which continuously recreates the idea that creativity
and innovation are one. Instead I have tried to put forward various
alternative visions and practices, from radical open access to posthuman
authorship and uncreative writing, based on vital relationships and on an
ethics of care and responsibility. These alternatives highlight distributed
and relational authorship and/or showcase a sensibility that embraces
posthuman agencies and processual publishing as part of a more complex,
emergent vision of creativity, open to different ideas of what creativity is
and can become. In this vision creativity is thus seen as relational, fluid
and processual and only ever temporarily fixed as part of our ethical decision
making: a decision-making process that is contingent on the contexts and
relationships with which we find ourselves entangled. This involves asking
questions about what writing is and does, and how creativity expands beyond
our established, static, or given concepts, which include copyright and a
focus on the author as a ‘homo economicus’, writing as inherently an
enterprise, and culture as commodified. As I have argued, the value of words,
indeed the economic worth and sustainability of words and of the ‘creative
industries’, can and should be defined within a different narrative. Opening
up from the hegemonic creative industries discourse and the way we perform it
through our writing practices might therefore enable us to explore extended
relationalities of emergent creativity, open-ended publishing processes, and a
feminist ethics of care and responsibility.

This contribution has showcased examples of experimental, hybrid and posthuman
writing and publishing practices that are intervening in this established
discourse on creativity. How, through them, can we start to performatively
explore a new discourse and reconfigure the relationships that underlie our
writing processes? How can the worth of writing be reflected in different
ways?

## Works Cited

(2014) ‘New Research into Authors’ Earnings Released’, Authors’ Licensing and
Collecting Society,
Us/News/News/What-are-words-worth-now-not-much.aspx>

Abrahamsson, Sebastian, Uli Beisel, Endre Danyi, Joe Deville, Julien McHardy,
and Michaela Spencer (2013) ‘Mattering Press: New Forms of Care for STS
Books’, The EASST Review 32.4, volume-32-4-december-2013/mattering-press-new-forms-of-care-for-sts-books/>

Adema, Janneke (2017) ‘Cut-Up’, in Eduardo Navas (ed.), Keywords in Remix
Studies (New York and London: Routledge), pp. 104–14,


— (2014) ‘Embracing Messiness’, LSE Impact of Social Sciences,
adema-pdsc14/>

— (2015) ‘Knowledge Production Beyond The Book? Performing the Scholarly
Monograph in Contemporary Digital Culture’ (PhD dissertation, Coventry
University), f4c62c77ac86/1/ademacomb.pdf>

— (2014) ‘Open Access’, in Critical Keywords for the Digital Humanities
(Lueneburg: Centre for Digital Cultures (CDC)),


— and Gary Hall (2013) ‘The Political Nature of the Book: On Artists’ Books
and Radical Open Access’, New Formations 78.1, 138–56,


— and Samuel Moore (2018) ‘Collectivity and Collaboration: Imagining New Forms
of Communality to Create Resilience in Scholar-Led Publishing’, Insights 31.3,


ALCS, Press Release (8 July 2014) ‘What Are Words Worth Now? Not Enough’,


Barad, Karen (2007) Meeting the Universe Halfway: Quantum Physics and the
Entanglement of Matter and Meaning (Durham, N.C., and London: Duke University
Press).

Boon, Marcus (2010) In Praise of Copying (Cambridge, MA: Harvard University
Press).

Brown, Wendy (2015) Undoing the Demos: Neoliberalism’s Stealth Revolution
(Cambridge, MA: MIT Press).

Chartier, Roger (1994) The Order of Books: Readers, Authors, and Libraries in
Europe Between the 14th and 18th Centuries, 1st ed. (Stanford, CA: Stanford
University Press).

Craig, Carys J. (2011) Copyright, Communication and Culture: Towards a
Relational Theory of Copyright Law (Cheltenham, UK, and Northampton, MA:
Edward Elgar Publishing).

— Joseph F. Turcotte, and Rosemary J. Coombe (2011) ‘What’s Feminist About
Open Access? A Relational Approach to Copyright in the Academy’, Feminists@law
1.1,

Cramer, Florian (2013) Anti-Media: Ephemera on Speculative Arts (Rotterdam and
New York, NY: nai010 publishers).

Drucker, Johanna (2015) ‘Humanist Computing at the End of the Individual Voice
and the Authoritative Text’, in Patrik Svensson and David Theo Goldberg
(eds.), Between Humanities and the Digital (Cambridge, MA: MIT Press), pp.
83–94.

— (2014) ‘Distributed and Conditional Documents: Conceptualizing
Bibliographical Alterities’, MATLIT: Revista do Programa de Doutoramento em
Materialidades da Literatura 2.1, 11–29.

— (2013) ‘Performative Materiality and Theoretical Approaches to Interface’,
Digital Humanities Quarterly 7.1 [n.p.],


Ede, Lisa, and Andrea A. Lunsford (2001) ‘Collaboration and Concepts of
Authorship’, PMLA 116.2, 354–69.

Emerson, Lori (2008) ‘Materiality, Intentionality, and the Computer-Generated
Poem: Reading Walter Benn Michaels with Erin Moureacute’s Pillage Land’, ESC:
English Studies in Canada 34, 45–69.

— (2003) ‘Digital Poetry as Reflexive Embodiment’, in Markku Eskelinen, Raine
Koskimaa, Loss Pequeño Glazier and John Cayley (eds.), CyberText Yearbook
2002–2003, 88–106,

Foucault, Michel, ‘What Is an Author?’ (1998) in James D. Faubion (ed.),
Essential Works of Foucault, 1954–1984, Volume Two: Aesthetics, Method, and
Epistemology (New York: The New Press).

Gibson, Johanna (2007) Creating Selves: Intellectual Property and the
Narration of Culture (Aldershot, England and Burlington, VT: Routledge).

— Phillip Johnson and Gaetano Dimita (2015) The Business of Being an Author: A
Survey of Author’s Earnings and Contracts (London: Queen Mary University of
London), [https://orca.cf.ac.uk/72431/1/Final Report - For Web
Publication.pdf](https://orca.cf.ac.uk/72431/1/Final%20Report%20-%20For%20Web%20Publication.pdf)

Goldsmith, Kenneth (2011) Uncreative Writing: Managing Language in the Digital
Age (New York: Columbia University Press).

Hall, Gary (2010) ‘Radical Open Access in the Humanities’ (presented at the
Research Without Borders, Columbia University),
humanities/>

— (2008) Digitize This Book!: The Politics of New Media, or Why We Need Open
Access Now (Minneapolis, MN: University of Minnesota Press).

Hayles, N. Katherine (2004) ‘Print Is Flat, Code Is Deep: The Importance of
Media-Specific Analysis’, Poetics Today 25.1, 67–90,


Hughes, Rolf (2005) ‘Orderly Disorder: Post-Human Creativity’, in Proceedings
of the Linköping Electronic Conference (Linköpings universitet: University
Electronic Press).

Jenkins, Henry, and Owen Gallagher (2008) ‘“What Is Remix Culture?”: An
Interview with Total Recut’s Owen Gallagher’, Confessions of an Aca-Fan,


Johns, Adrian (1998) The Nature of the Book: Print and Knowledge in the Making
(Chicago, IL: University of Chicago Press).

Kember, Sarah (2016) ‘Why Publish?’, Learned Publishing 29, 348–53,


— (2014) ‘Why Write?: Feminism, Publishing and the Politics of Communication’,
New Formations: A Journal of Culture/Theory/Politics 83.1, 99–116.

Kretschmer, M., and P. Hardwick (2007) Authors’ Earnings from Copyright and
Non-Copyright Sources : A Survey of 25,000 British and German Writers (Poole,
UK: CIPPM/ALCS Bournemouth University),
[https://microsites.bournemouth.ac.uk/cippm/files/2007/07/ALCS-Full-
report.pdf](https://microsites.bournemouth.ac.uk/cippm/files/2007/07/ACLS-
Full-report.pdf)

Lessig, Lawrence (2008) Remix: Making Art and Commerce Thrive in the Hybrid
Economy (New York: Penguin Press).

Lovink, Geert, and Ned Rossiter (eds.) (2007) MyCreativity Reader: A Critique
of Creative Industries (Amsterdam: Institute of Network Cultures),


McGann, Jerome J. (1992) A Critique of Modern Textual Criticism
(Charlottesville, VA: University of Virginia Press).

McHardy, Julien (2014) ‘Why Books Matter: There Is Value in What Cannot Be
Evaluated.’, Impact of Social Sciences [n.p.],


Mol, Annemarie (2008) The Logic of Care: Health and the Problem of Patient
Choice, 1st ed. (London and New York: Routledge).

Montfort, Nick (2003) ‘The Coding and Execution of the Author’, in Markku
Eskelinen, Raine Kosimaa, Loss Pequeño Glazier and John Cayley (eds.),
CyberText Yearbook 2002–2003, 2003, 201–17,
, pp. 201–17.

Moore, Samuel A. (2017) ‘A Genealogy of Open Access: Negotiations between
Openness and Access to Research’, Revue Française des Sciences de
l’information et de la Communication 11,

Munster, Anna (2016) ‘Techno-Animalities — the Case of the Monkey Selfie’
(presented at the Goldsmiths University, London),


Navas, Eduardo (2012) Remix Theory: The Aesthetics of Sampling (Vienna and New
York: Springer).

Parikka, Jussi, and Mercedes Bunz (11 July 2014) ‘A Mini-Interview: Mercedes
Bunz Explains Meson Press’, Machinology,
meson-press/>

Richards, Victoria (7 January 2016) ‘Monkey Selfie: Judge Rules Macaque Who
Took Grinning Photograph of Himself “Cannot Own Copyright”’, The Independent,
macaque-who-took-grinning-photograph-of-himself-cannot-own-
copyright-a6800471.html>

Robbins, Sarah (2003) ‘Distributed Authorship: A Feminist Case-Study Framework
for Studying Intellectual Property’, College English 66.2, 155–71,


Rose, Mark (1993) Authors and Owners: The Invention of Copyright (Cambridge,
MA: Harvard University Press).

Spinosa, Dani (14 May 2014) ‘“My Line (Article) Has Sighed”: Authorial
Subjectivity and Technology’, Generic Pronoun,


Star, Susan Leigh (1991) ‘The Sociology of the Invisible: The Primacy of Work
in the Writings of Anselm Strauss’, in Anselm Leonard Strauss and David R.
Maines (eds.), Social Organization and Social Process: Essays in Honor of
Anselm Strauss (New York: A. de Grutyer).

* * *

[1](ch3.xhtml#footnote-152-backlink) The Authors’ Licensing and Collecting
Society is a [British](https://en.wikipedia.org/wiki/United_Kingdom)
membership organisation for writers, established in 1977 with over 87,000
members, focused on protecting and promoting authors’ rights. ALCS collects
and pays out money due to members for secondary uses of their work (copying,
broadcasting, recording etc.).

[2](ch3.xhtml#footnote-151-backlink) This survey was an update of an earlier
survey conducted in 2006 by the Centre of Intellectual Property Policy and
Management (CIPPM) at Bournemouth University.

[3](ch3.xhtml#footnote-150-backlink) ‘New Research into Authors’ Earnings
Released’, Authors’ Licensing and Collecting Society, 2014,
Us/News/News/What-are-words-worth-now-not-much.aspx>

[4](ch3.xhtml#footnote-149-backlink) Johanna Gibson, Phillip Johnson, and
Gaetano Dimita, The Business of Being an Author: A Survey of Author’s Earnings
and Contracts (London: Queen Mary University of London, 2015), p. 9,
[https://orca.cf.ac.uk/72431/1/Final Report - For Web Publication.pdf
](https://orca.cf.ac.uk/72431/1/Final%20Report%20-%20For%20Web%20Publication.pdf)

[5](ch3.xhtml#footnote-148-backlink) ALCS, Press Release. What Are Words Worth
Now? Not Enough, 8 July 2014, worth-now-not-enough>

[6](ch3.xhtml#footnote-147-backlink) Gibson, Johnson, and Dimita, The Business
of Being an Author, p. 35.

[7](ch3.xhtml#footnote-146-backlink) M. Kretschmer and P. Hardwick, Authors’
Earnings from Copyright and Non-Copyright Sources: A Survey of 25,000 British
and German Writers (Poole: CIPPM/ALCS Bournemouth University, 2007), p. 3,
[https://microsites.bournemouth.ac.uk/cippm/files/2007/07/ALCS-Full-
report.pdf](https://microsites.bournemouth.ac.uk/cippm/files/2007/07/ACLS-
Full-report.pdf)

[8](ch3.xhtml#footnote-145-backlink) ALCS, Press Release, 8 July 2014,
[https://www.alcs.co.uk/news/what-are-words-](https://www.alcs.co.uk/news
/what-are-words-worth-now-not-enough)
worth-now-not-enough

[9](ch3.xhtml#footnote-144-backlink) Gibson, Johnson, and Dimita, The Business
of Being an Author, p. 35.

[10](ch3.xhtml#footnote-143-backlink) Ibid.

[11](ch3.xhtml#footnote-142-backlink) In the survey, three questions that
focus on various sources of remuneration do list digital publishing and/or
online uses as an option (questions 8, 11, and 15). Yet the data tables
provided in the appendix to the report do not provide the findings for
questions 11 and 15 nor do they differentiate according to type of media for
other tables related to remuneration. The only data table we find in the
report related to digital publishing is table 3.3, which lists ‘Earnings
ranked (1 to 7) in relation to categories of work’, where digital publishing
ranks third after books and magazines/periodicals, but before newspapers,
audio/audio-visual productions and theatre. This lack of focus on the effect
of digital publishing on writers’ incomes, for a survey that is ‘the first to
capture the impact of the digital revolution on writers’ working lives’, is
quite remarkable. Gibson, Johnson, and Dimita, The Business of Being an
Author, Appendix 2.

[12](ch3.xhtml#footnote-141-backlink) Ibid., p. 35.

[13](ch3.xhtml#footnote-140-backlink) Ibid.

[14](ch3.xhtml#footnote-139-backlink) Geert Lovink and Ned Rossiter (eds.),
MyCreativity Reader: A Critique of Creative Industries (Amsterdam: Institute
of Network Cultures, 2007), p. 14,


[15](ch3.xhtml#footnote-138-backlink) See:
estimates-january-2015/creative-industries-economic-estimates-january-2015
-key-findings>

[16](ch3.xhtml#footnote-137-backlink) Wendy Brown, Undoing the Demos:
Neoliberalism’s Stealth Revolution (Cambridge, MA: MIT Press, 2015), p. 31.

[17](ch3.xhtml#footnote-136-backlink) Therefore Lovink and Rossiter make a
plea to, ‘redefine creative industries outside of IP generation’. Lovink and
Rossiter, MyCreativity Reader, p. 14.

[18](ch3.xhtml#footnote-135-backlink) Next to earnings made from writing more
in general, the survey on various occasions asks questions about earnings
arising from specific categories of works and related to the amount of works
exploited (published/broadcast) during certain periods. Gibson, Johnson, and
Dimita, The Business of Being an Author, Appendix 2.

[19](ch3.xhtml#footnote-134-backlink) Roger Chartier, The Order of Books:
Readers, Authors, and Libraries in Europe Between the 14th and 18th Centuries,
1st ed. (Stanford: Stanford University Press, 1994); Lisa Ede and Andrea A.
Lunsford, ‘Collaboration and Concepts of Authorship’, PMLA 116.2 (2001),
354–69; Adrian Johns, The Nature of the Book: Print and Knowledge in the
Making (Chicago, IL: University of Chicago Press, 1998); Jerome J. McGann, A
Critique of Modern Textual Criticism (Charlottesville, VA, University of
Virginia Press, 1992); Sarah Robbins, ‘Distributed Authorship: A Feminist
Case-Study Framework for Studying Intellectual Property’, College English 66.2
(2003), 155–71,

[20](ch3.xhtml#footnote-133-backlink) The ALCS survey addresses this problem,
of course, and tries to lobby on behalf of its authors for fair contracts with
publishers and intermediaries. That said, the survey findings show that only
42% of writers always retain their copyright. Gibson, Johnson, and Dimita, The
Business of Being an Author, p. 12.

[21](ch3.xhtml#footnote-132-backlink) Michel Foucault, ‘What Is an Author?’,
in James D. Faubion (ed.), Essential Works of Foucault, 1954–1984, Volume Two:
Aesthetics, Method, and Epistemology (New York: The New Press, 1998), p. 205.

[22](ch3.xhtml#footnote-131-backlink) Mark Rose, Authors and Owners: The
Invention of Copyright (Cambridge, MA: Harvard University Press, 1993).

[23](ch3.xhtml#footnote-130-backlink) Carys J. Craig, Joseph F. Turcotte, and
Rosemary J. Coombe, ‘What’s Feminist About Open Access? A Relational Approach
to Copyright in the Academy’, Feminists@law 1.1 (2011),


[24](ch3.xhtml#footnote-129-backlink) Ibid., p. 8.

[25](ch3.xhtml#footnote-128-backlink) Ibid., p. 9.

[26](ch3.xhtml#footnote-127-backlink) Lawrence Lessig, Remix: Making Art and
Commerce Thrive in the Hybrid Economy (New York: Penguin Press, 2008); Eduardo
Navas, Remix Theory: The Aesthetics of Sampling (Vienna and New York:
Springer, 2012); Henry Jenkins and Owen Gallagher, ‘“What Is Remix Culture?”:
An Interview with Total Recut’s Owen Gallagher’, Confessions of an Aca-Fan,
2008,

[27](ch3.xhtml#footnote-126-backlink) Craig, Turcotte, and Coombe, ‘What’s
Feminist About Open Access?, p. 27.

[28](ch3.xhtml#footnote-125-backlink) Ibid., p. 14.

[29](ch3.xhtml#footnote-124-backlink) Ibid., p. 26.

[30](ch3.xhtml#footnote-123-backlink) Janneke Adema, ‘Open Access’, in
Critical Keywords for the Digital Humanities (Lueneburg: Centre for Digital
Cultures (CDC), 2014), ; Janneke Adema,
‘Embracing Messiness’, LSE Impact of Social Sciences, 2014,
adema-pdsc14/>; Gary Hall, Digitize This Book!: The Politics of New Media, or
Why We Need Open Access Now (Minneapolis, MN: University of Minnesota Press,
2008), p. 197; Sarah Kember, ‘Why Write?: Feminism, Publishing and the
Politics of Communication’, New Formations: A Journal of
Culture/Theory/Politics 83.1 (2014), 99–116; Samuel A. Moore, ‘A Genealogy of
Open Access: Negotiations between Openness and Access to Research’, Revue
Française des Sciences de l’information et de la Communication, 2017,


[31](ch3.xhtml#footnote-122-backlink) Florian Cramer, Anti-Media: Ephemera on
Speculative Arts (Rotterdam and New York: nai010 publishers, 2013).

[32](ch3.xhtml#footnote-121-backlink) Especially within humanities publishing
there is a reluctance to allow derivative uses of one’s work in an open access
setting.

[33](ch3.xhtml#footnote-120-backlink) In 2015 the Radical Open Access
Conference took place at Coventry University, which brought together a large
array of presses and publishing initiatives (often academic-led) in support of
an ‘alternative’ vision of open access and scholarly communication.
Participants in this conference subsequently formed the loosely allied Radical
Open Access Collective: [radicaloa.co.uk](https://radicaloa.co.uk/). As the
conference concept outlines, radical open access entails ‘a vision of open
access that is characterised by a spirit of on-going creative experimentation,
and a willingness to subject some of our most established scholarly
communication and publishing practices, together with the institutions that
sustain them (the library, publishing house etc.), to rigorous critique.
Included in the latter will be the asking of important questions about our
notions of authorship, authority, originality, quality, credibility,
sustainability, intellectual property, fixity and the book — questions that
lie at the heart of what scholarship is and what the university can be in the
21st century’. Janneke Adema and Gary Hall, ‘The Political Nature of the Book:
On Artists’ Books and Radical Open Access’, New Formations 78.1 (2013),
138–56, ; Janneke Adema and Samuel
Moore, ‘Collectivity and Collaboration: Imagining New Forms of Communality to
Create Resilience In Scholar-Led Publishing’, Insights 31.3 (2018),
; Gary Hall, ‘Radical Open Access in the
Humanities’ (presented at the Research Without Borders, Columbia University,
2010), humanities/>; Janneke Adema, ‘Knowledge Production Beyond The Book? Performing
the Scholarly Monograph in Contemporary Digital Culture’ (PhD dissertation,
Coventry University, 2015),
f4c62c77ac86/1/ademacomb.pdf>

[34](ch3.xhtml#footnote-119-backlink) Julien McHardy, ‘Why Books Matter: There
Is Value in What Cannot Be Evaluated’, Impact of Social Sciences, 2014, n.p.,
[http://blogs.lse.ac.uk/impactofsocial sciences/2014/09/30/why-books-
matter/](http://blogs.lse.ac.uk/impactofsocialsciences/2014/09/30/why-books-
matter/)

[35](ch3.xhtml#footnote-118-backlink) Karen Barad, Meeting the Universe
Halfway: Quantum Physics and the Entanglement of Matter and Meaning (Durham,
N.C. and London: Duke University Press, 2007).

[36](ch3.xhtml#footnote-117-backlink) Annemarie Mol, The Logic of Care: Health
and the Problem of Patient Choice, 1st ed. (London and New York: Routledge,
2008).

[37](ch3.xhtml#footnote-116-backlink) Sebastian Abrahamsson and others,
‘Mattering Press: New Forms of Care for STS Books’, The EASST Review 32.4
(2013), press-new-forms-of-care-for-sts-books/>

[38](ch3.xhtml#footnote-115-backlink) McHardy, ‘Why Books Matter’.

[39](ch3.xhtml#footnote-114-backlink) Ibid.

[40](ch3.xhtml#footnote-113-backlink) Susan Leigh Star, ‘The Sociology of the
Invisible: The Primacy of Work in the Writings of Anselm Strauss’, in Anselm
Leonard Strauss and David R. Maines (eds.), Social Organization and Social
Process: Essays in Honor of Anselm Strauss (New York: A. de Gruyter, 1991).
Mattering Press is not alone in exploring an ethics of care in relation to
(academic) publishing. Sarah Kember, director of Goldsmiths Press is also
adamant in her desire to make the underlying processes of publishing (i.e.
peer review, citation practices) more transparent and accountable Sarah
Kember, ‘Why Publish?’, Learned Publishing 29 (2016), 348–53,
. Mercedes Bunz, one of the editors running
Meson Press, argues that a sociology of the invisible would incorporate
‘infrastructure work’, the work of accounting for, and literally crediting
everybody involved in producing a book: ‘A book isn’t just a product that
starts a dialogue between author and reader. It is accompanied by lots of
other academic conversations — peer review, co-authors, copy editors — and
these conversations deserve to be taken more serious’. Jussi Parikka and
Mercedes Bunz, ‘A Mini-Interview: Mercedes Bunz Explains Meson Press’,
Machinology, 2014, mercedes-bunz-explains-meson-press/>. For Open Humanities Press authorship is
collaborative and even often anonymous: for example, they are experimenting
with research published in wikis to further complicate the focus on single
authorship and a static marketable book object within academia (see their
living and liquid books series).

[41](ch3.xhtml#footnote-112-backlink) Lori Emerson, ‘Digital Poetry as
Reflexive Embodiment’, in Markku Eskelinen, Raine Koskimaa, Loss Pequeño
Glazier and John Cayley (eds.), CyberText Yearbook 2002–2003, 2003, 88–106,


[42](ch3.xhtml#footnote-111-backlink) Dani Spinosa, ‘“My Line (Article) Has
Sighed”: Authorial Subjectivity and Technology’, Generic Pronoun, 2014,


[43](ch3.xhtml#footnote-110-backlink) Spinosa, ‘My Line (Article) Has Sighed’.

[44](ch3.xhtml#footnote-109-backlink) Emerson, ‘Digital Poetry as Reflexive
Embodiment’, p. 89.

[45](ch3.xhtml#footnote-108-backlink) Rolf Hughes, ‘Orderly Disorder: Post-
Human Creativity’, in Proceedings of the Linköping Electronic Conference
(Linköpings universitet: University Electronic Press, 2005).

[46](ch3.xhtml#footnote-107-backlink) N. Katherine Hayles, ‘Print Is Flat,
Code Is Deep: The Importance of Media-Specific Analysis’, Poetics Today 25.1
(2004), 67–90, ; Johanna Drucker,
‘Performative Materiality and Theoretical Approaches to Interface’, Digital
Humanities Quarterly 7.1 (2013),
; Johanna
Drucker, ‘Distributed and Conditional Documents: Conceptualizing
Bibliographical Alterities’, MATLIT: Revista do Programa de Doutoramento em
Materialidades da Literatura 2.1 (2014), 11–29.

[47](ch3.xhtml#footnote-106-backlink) Nick Montfort, ‘The Coding and Execution
of the Author’, in Markku Eskelinen, Raine Kosimaa, Loss Pequeño Glazier and
John Cayley (eds.), CyberText Yearbook 2002–2003, 2003, 201–17 (p. 201),


[48](ch3.xhtml#footnote-105-backlink) Montfort, ‘The Coding and Execution of
the Author’, p. 202.

[49](ch3.xhtml#footnote-104-backlink) Lori Emerson, ‘Materiality,
Intentionality, and the Computer-Generated Poem: Reading Walter Benn Michaels
with Erin Moureacute’s Pillage Land’, ESC: English Studies in Canada 34
(2008), 66.

[50](ch3.xhtml#footnote-103-backlink) Marcus Boon, In Praise of Copying
(Cambridge, MA: Harvard University Press, 2010); Johanna Drucker, ‘Humanist
Computing at the End of the Individual Voice and the Authoritative Text’, in
Patrik Svensson and David Theo Goldberg (eds.), Between Humanities and the
Digital (Cambridge, MA: MIT Press, 2015), pp. 83–94.

[51](ch3.xhtml#footnote-102-backlink) We have to take into consideration here
that print-based cultural products were never fixed or static; the dominant
discourses constructed around them just perceive them to be so.

[52](ch3.xhtml#footnote-101-backlink) Craig, Turcotte, and Coombe, ‘What’s
Feminist About Open Access?’, p. 2.

[53](ch3.xhtml#footnote-100-backlink) Ibid.

[54](ch3.xhtml#footnote-099-backlink) Johanna Gibson, Creating Selves:
Intellectual Property and the Narration of Culture (Aldershot, UK, and
Burlington: Routledge, 2007), p. 7.

[55](ch3.xhtml#footnote-098-backlink) Gibson, Creating Selves, p. 7.

[56](ch3.xhtml#footnote-097-backlink) Ibid.

[57](ch3.xhtml#footnote-096-backlink) Kenneth Goldsmith, Uncreative Writing:
Managing Language in the Digital Age (New York: Columbia University Press,
2011), p. 227.

[58](ch3.xhtml#footnote-095-backlink) Ibid., p. 15.

[59](ch3.xhtml#footnote-094-backlink) Goldsmith, Uncreative Writing, p. 81.

[60](ch3.xhtml#footnote-093-backlink) Ibid.

[61](ch3.xhtml#footnote-092-backlink) It is worth emphasising that what
Goldsmith perceives as ‘uncreative’ notions of writing (including
appropriation, pastiche, and copying), have a prehistory that can be traced
back to antiquity (thanks go out to this chapter’s reviewer for pointing this
out). One example of this, which uses the method of cutting and pasting —
something I have outlined more in depth elsewhere — concerns the early modern
commonplace book. Commonplacing as ‘a method or approach to reading and
writing involved the gathering and repurposing of meaningful quotes, passages
or other clippings from published books by copying and/or pasting them into a
blank book.’ Janneke Adema, ‘Cut-Up’, in Eduardo Navas (ed.), Keywords in
Remix Studies (New York and London: Routledge, 2017), pp. 104–14,


[62](ch3.xhtml#footnote-091-backlink) Gibson, Creating Selves, p. 27.

[63](ch3.xhtml#footnote-090-backlink) For example, animals cannot own
copyright. See the case of Naruto, the macaque monkey that took a ‘selfie’
photograph of itself. Victoria Richards, ‘Monkey Selfie: Judge Rules Macaque
Who Took Grinning Photograph of Himself “Cannot Own Copyright”’, The
Independent, 7 January 2016, /monkey-selfie-judge-rules-macaque-who-took-grinning-photograph-of-himself-
cannot-own-copyright-a6800471.html>

[64](ch3.xhtml#footnote-089-backlink) Anna Munster, ‘Techno-Animalities — the
Case of the Monkey Selfie’ (presented at the Goldsmiths University, London,
2016),

[65](ch3.xhtml#footnote-088-backlink) Sarah Kember, ‘Why Write?: Feminism,
Publishing and the Politics of Communication’, New Formations: A Journal of
Culture/Theory/Politics 83.1 (2014), 99–116.

Sollfrank
The Surplus of Copying
2018


## essay #11

The Surplus of Copying
How Shadow Libraries and Pirate Archives Contribute to the
Creation of Cultural Memory and the Commons
By Cornelia Sollfrank

Digital artworks tend to have a problematic relationship with the white
cube—in particular, when they are intended and optimized for online
distribution. While curators and exhibition-makers usually try to avoid
showing such works altogether, or at least aim at enhancing their sculptural
qualities to make them more presentable, the exhibition _Top Tens_ featured an
abundance of web quality digital artworks, thus placing emphasis on the very
media condition of such digital artifacts. The exhibition took place at the
Onassis Cultural Center in Athens in March 2018 and was part of the larger
festival _Shadow Libraries: UbuWeb in Athens_ ,1 an event to introduce the
online archive UbuWeb2 to the Greek audience and discuss related cultural,
ethical, technical, and legal issues. This text takes the event—and the
exhibition in particular—as a starting point for a closer look at UbuWeb and
the role an artistic approach can play in building cultural memory within the
neoliberal knowledge economy.

_UbuWeb—The Cultural Memory of the Avant-Garde_

Since Kenneth Goldsmith started Ubu in 1997 the site has become a major point
of reference for anyone interested in exploring twentieth-century avant-garde
art. The online archive provides free and unrestricted access to a remarkable
collection of thousands of artworks—among them almost 700 films and videos,
over 1000 sound art pieces, dozens of filmed dance productions, an
overwhelming amount of visual poetry and conceptual writing, critical
documents, but also musical scores, patents, electronic music resources, plus
an edition of vital new literature, the /ubu editions. Ubu contextualizes the
archived objects within curated sections and also provides framing academic
essays. Although it is a project run by Goldsmith without a budget, it has
built a reputation for making all the things available one would not find
elsewhere. The focus on “avant-garde” may seem a bit pretentious at first, but
when you look closer at the project, its operator and the philosophy behind
it, it becomes obvious how much sense this designation makes. Understanding
the history of the twentieth-century avant-garde as “a history of subversive
takes on creativity, originality, and authorship,”3 such spirit is not only
reflected in terms of the archive’s contents but also in terms of the project
as a whole. Theoretical statements by Goldsmith in which he questions concepts
such as authorship, originality, and creativity support this thesis4—and with
that a conflictual relationship with the notion of intellectual property is
preprogrammed. Therefore it comes as no surprise that the increasing
popularity of the project goes hand-in-hand with a growing discussion about
its ethical justification.

At the heart of Ubu, there is the copy! Every item in the archive is a digital
copy, either of another digital item or, in fact, it is the digitized version
of an analog object.5 That is to say, the creation of a digital collection is
inevitably based on copying the desired archive records and storing them on
dedicated media. However, making a copy is in itself a copyright-relevant act,
if the respective item is an original creation and as such protected under
copyright law.6 Hence, “any reproduction of a copyrighted work infringes the
copyright of the author or the corresponding rights of use of the copyright
holder”.7 Whether the existence of an artwork within the Ubu collection is a
case of copyright infringement varies with each individual case and depends on
the legal status of the respective work, but also on the way the rights
holders decide to act. As with all civil law, there is no judge without a
plaintiff, which means even if there is no express consent by the rights
holders, the work can remain in the archive as long as there is no request for
removal.8 Its status, however, is precarious. We find ourselves in the
notorious gray zone of copyright law where nothing is clear and many things
are possible—until somebody decides to challenge this status. Exploring the
borders of this experimental playground involves risk-taking, but, at the same
time, it is the only way to preserve existing freedoms and make a case for
changing cultural needs, which have not been considered in current legal
settings. And as the 20 years of Ubu’s existence demonstrate, the practice may
be experimental and precarious, but with growing cultural relevance and
reputation it is also gaining in stability.

_Fair Use and Public Interest_

At all public appearances and public presentations Goldsmith and his
supporters emphasize the educational character of the project and its non-
commercial orientation.9 Such a characterization is clearly intended to take
the wind out of the sails of its critics from the start and to shift the
attention away from the notion of piracy and toward questions of public
interest and the common good.

From a cultural point of view, the project unquestionably is of inestimable
value; a legal defense, however, would be a difficult undertaking. Copyright
law, in fact, has a built-in opening, the so-called copyright exceptions or
fair use regulations. They vary according to national law and cultural
traditions and allow for the use of copyrighted works under certain, defined
provisions without permission of the owner. The exceptions basically apply to
the areas of research and private study (both non-commercial), education,
review, and criticism and are described through general guidelines. “These
defences exist in order to restore the balance between the rights of the owner
of copyright and the rights of society at large.”10

A very powerful provision in most legislations is the permission to make
“private copies”, digital and analog ones, in small numbers, but they are
limited to non-commercial and non-public use, and passing on to a third party
is also excluded.11 As Ubu is an online archive that makes all of its records
publicly accessible and, not least, also provides templates for further
copying, it exceeds the notion of a “private copy” by far. Regarding further
fair use provisions, the four factors that are considered in a decision-making
process in US copyright provisions, for instance, refer to: 1) the purpose and
character of the use, including whether such use is of a commercial nature or
is for non-profit educational purposes; 2) the nature of the copyrighted work;
3) the amount and substantiality of the portion used in relation to the
copyrighted work as a whole; and 4) the effect of the use upon the potential
market for the value of the copyrighted work (US Copyright Act, 1976, 17 USC.
§107, online, n.pag.). Applying these fair use provisions to Ubu, one might
consider that the main purposes of the archive relate to education and
research, that it is by its very nature non-commercial, and it largely does
not collide with any third party business interests as most of the material is
not commercially available. However, proving this in detail would be quite an
endeavor. And what complicates matters even more is that the archival material
largely consists of original works of art, which are subject to strict
copyright law protection, that all the works have been copied without any
transformative or commenting intention, and last but not least, that the
aspect of the appropriateness of the amount of used material becomes absurd
with reference to an archive whose quality largely depends on
comprehensiveness: the more the merrier. As Simon Stokes points out, legally
binding decisions can only be made on a case-by-case basis, which is why it is
difficult to make a general evaluation of Ubu’s legal situation.12 The ethical
defense tends to induce the cultural value of the archive as a whole and its
invaluable contribution to cultural memory, while the legal situation does not
consider the value of the project as a whole and necessitates breaking it down
into all the individual items within the collection.

This very brief, when not abridged discussion of the possibilities of fair use
already demonstrates how complex it would be to apply them to Ubu. How
pointless it would be to attempt a serious legal discussion for such a
privately run archive becomes even clearer when looking at the problems public
libraries and archives have to face. While in theory such official
institutions may even have a public mission to collect, preserve, and archive
digital material, in practice, copyright law largely prevents the execution of
this task, as Steinhauer explains.13 The legal expert introduces the example
of the German National Library, which was assigned the task since 2006 to make
back-up copies of all websites published within the .de sublevel domain, but
it turned out to be illegal.14 Identifying a deficiently legal situation when
it comes to collecting, archiving, and providing access to digital cultural
goods, Steinhauer even speaks of a “legal obligation to amnesia”.15 And it is
particularly striking that, from a legal perspective, the collecting of
digitalia is more strictly regulated than the collecting of books, for
example, where the property status of the material object comes into play.
Given the imbalance between cultural requirements, copyright law, and the
technical possibilities, it is not surprising that private initiatives are
being founded with the aim to collect and preserve cultural memory. These
initiatives make use of the affordability and availability of digital
technology and its infrastructures, and they take responsibility for the
preservation of cultural goods by simply ignoring copyright induced
restrictions, i.e. opposing the insatiable hunger of the IP regime for
control.

_Shadow Libraries_

Ubu was presented and discussed in Athens at an event titled _Shadow
Libraries: UbuWeb in Athens_ , thereby making clear reference to the ecosystem
of shadow libraries. A library, in general, is an institution that collects,
orders, and makes published information available while taking into account
archival, economic, and synoptic aspects. A shadow library does exactly the
same thing, but its mission is not an official one. Usually, the
infrastructure of shadow libraries is conceived, built, and run by a private
initiative, an individual, or a small group of people, who often prefer to
remain anonymous for obvious reasons. In terms of the media content provided,
most shadow libraries are peer-produced in the sense that they are based on
the contributions of a community of supporters, sometimes referred to as
“amateur librarians”. The two key attributes of any proper library, according
to Amsterdam-based media scholar Bodo Balazs, are the catalog and the
community: “The catalogue does not just organize the knowledge stored in the
collection; it is not just a tool of searching and browsing. It is a critical
component in the organisation of the community of librarians who preserve and
nourish the collection.”16 What is specific about shadow libraries, however,
is the fact that they make available anything their contributors consider to
be relevant—regardless of its legal status. That is to say, shadow libraries
also provide unauthorized access to copyrighted publications, and they make
the material available for download without charge and without any other
restrictions. And because there is a whole network of shadow libraries whose
mission is “to remove all barriers in the way of science,”17 experts speak of
an ecosystem fostering free and universal access to knowledge.

The notion of the shadow library enjoyed popularity in the early 2000s when
the wide availability of digital networked media contributed to the emergence
of large-scale repositories of scientific materials, the most famous one
having been Gigapedia, which later transformed into library.nu. This project
was famous for hosting approximately 400,000 (scientific) books and journal
articles but had to be shut down in 2012 as a consequence of a series of
injunctions from powerful publishing houses. The now leading shadow library in
the field, Library Genesis (LibGen), can be considered as its even more
influential successor. As of November 2016 the database contained 25 million
documents (42 terabytes), of which 2.1 million were books, with digital copies
of scientific articles published in 27,134 journals by 1342 publishers.18 The
large majority of the digital material is of scientific and educational nature
(95%), while only 5% serves recreational purposes.19 The repository is based
on various ways of crowd-sourcing, i.e. social and technical forms of
accessing and sharing academic publications. Despite a number of legal cases
and court orders, the site is still available under various and changing
domain names.20

The related project Sci-Hub is an online service that processes requests for
pay-walled articles by providing systematic, automized, but unauthorized
backdoor access to proprietary scholarly journal databases. Users requesting
papers not present in LibGen are advised to download them through Sci-Hub; the
respective PDF files are served to users and automatically added to LibGen (if
not already present). According to _Nature_ magazine, Sci-Hub hosts around 60
million academic papers and was able to serve 75 million downloads in 2016. On
a daily basis 70,000 users access approximately 200,000 articles.

The founder of the meta library Sci-Hub is Kazakh programmer Alexandra
Elbakyan, who has been sued by large publishing houses and was convicted twice
to pay almost 20 million US$ in compensation for the losses her activities
allegedly have caused, which is why she had to go underground in Russia. For
illegally leaking millions of documents the _New York Times_ compared her to
Edward Snowden in 2016: “While she didn’t reveal state secrets, she took a
stand for the public’s right to know by providing free online access to just
about every scientific paper ever published, ranging from acoustics to
zymology.” 21 In the same year the prestigious _Nature_ magazine elected her
as one of the ten most influential people in science. 22 Unlike other
persecuted people, she went on the offensive and started to explain her
actions and motives in court documents and blog posts. Sci-Hub encourages new
ways of distributing knowledge, beyond any commercial interests. It provides a
radically open infrastructure thus creating an inviting atmosphere. “It is a
knowledge infrastructure that can be freely accessed, used and built upon by
anyone.”23

As both projects LibGen and Sci-Hub are based in post-Soviet countries, Balazs
reconstructed the history and spirit of Russian reading culture and brings
them into connection.24 Interestingly, the author also establishes a
connection to the Kolhoz (Russian: колхо́з), an early Soviet collective farm
model that was self-governing, community-owned, and a collaborative
enterprise, which he considers to be a major inspiration for the digital
librarians. He also identifies parallels between this Kolhoz model and the
notion of the “commons”—a concept that will be discussed in more detail with
regards to shadow libraries further below.

According to Balazs, these sorts of libraries and collections are part of the
Guerilla Open Access movement (GOA) and thus practical manifestations of Aaron
Swartz’s “Guerilla Open Access Manifesto”.25 In this manifesto the American
hacker and activist pointed out the flaws of open access politics and aimed at
recruiting supporters for the idea of “radical” open access. Radical in this
context means to completely ignore copyright and simply make as much
information available as possible. “Information is power” is how the manifesto
begins. Basically, it addresses the—what he calls—“privileged”, in the sense
that they do have access to information as academic staff or librarians, and
he calls on their support for building a system of freely available
information by using their privilege, downloading and making information
available. Swartz and Elbakyan both have become the “iconic leaders”26 of a
global movement that fights for scientific knowledge to be(come) freely
accessible and whose protagonists usually prefer to operate unrecognized.
While their particular projects may be of a more or less temporary nature, the
discursive value of the work of the “amateur librarians” and their projects
will have a lasting impact on the development of access politics.

_Cultural and Knowledge Commons_

The above discussion illustrates that the phenomenon of shadow libraries
cannot be reduced to its copyright infringing aspects. It needs to be
contextualized within a larger sociopolitical debate that situates the demand
for free and unrestricted access to knowledge within the struggle against the
all-co-opting logic of capital, which currently aims to economize all aspects
of life.

In his analysis of the Russian shadow libraries Balazs has drawn a parallel to
the commons as an alternative mode of ownership and a collective way of
dealing with resources. The growing interest in the discourses around the
commons demonstrates the urgency and timeliness of this concept. The
structural definition of the commons conceived by political economist Massimo
de Angelis allows for its application in diverse fields: “Commons are social
systems in which resources are pooled by a community of people who also govern
these resources to guarantee the latter’s sustainability (if they are natural
resources) and the reproduction of the community. These people engage in
‘commoning,’ that is a form of social labour that bears a direct relation to
the needs of the people, or the commoners”.27 While the model originates in
historical ways of sharing natural resources, it has gained new momentum in
relation to very different resources, thus constituting a third paradigm of
production—beyond state and private—however, with all commoning activities
today still being embedded in the surrounding economic system.

As a reason for the newly aroused interest in the commons, de Angelis provides
the crisis of global capital, which has maneuvered itself into a systemic
impasse. While constantly expanding through its inherent logic of growth and
accumulation, it is the very same logic that destroys the two systems capital
relies on: non-market-shaped social reproduction and the ecological system.
Within this scenario de Angelis describes capital as being in need of the
commons as a “fix” for the most urgent systemic failures: “It needs a ‘commons
fix,’ especially in order to deal with the devastation of the social fabric as
a result of the current crisis of reproduction. Since neoliberalism is not
about to give up its management of the world, it will most likely have to ask
the commons to help manage the devastation it creates. And this means: if the
commons are not there, capital will have to promote them somehow.”28

This rather surprising entanglement of capital and the commons, however, is
not the only perspective. Commons, at the same time, have the potential to
create “a social basis for alternative ways of articulating social production,
independent from capital and its prerogatives. Indeed, today it is difficult
to conceive emancipation from capital—and achieving new solutions to the
demands of _buen vivir_ , social and ecological justice—without at the same
time organizing on the terrain of commons, the non-commodified systems of
social production. Commons are not just a ‘third way’ beyond state and market
failures; they are a vehicle for emerging communities of struggle to claim
ownership to their own conditions of life and reproduction.”29 It is their
purpose to satisfy people’s basic needs and empower them by providing access
to alternative means of subsistence. In that sense, commons can be understood
as an _experimental zone_ in which participants can learn to negotiate
responsibilities, social relations, and peer-based means of production.

_Art and Commons_

Projects such as UbuWeb, Monoskop,30 aaaaarg,31 Memory of the World,32 and
0xdb33 vary in size, they have different forms of organization and foci, but
they all care for specific cultural goods and make sure these goods remain
widely accessible—be it digital copies of artworks and original documents,
books and other text formats, videos, film, or sound and music. Unlike the
large shadow libraries introduced above, which aim to provide access to
hundreds of thousands, if not millions of mainly academic papers and books,
thus trying to fully cover the world of scholarly and academic works, the
smaller artist-run projects are of different nature. While UbuWeb’s founder,
for instance, also promotes a generally unrestricted access to cultural goods,
his approach with UbuWeb is to build a curated archive with copies of artworks
that he considers to be relevant for his very context.34 The selection is
based on personal assessment and preference and cared for affectionately.
Despite its comprehensiveness, it still can be considered a “personal website”
on which the artist shares things relevant to him. As such, he is in good
company with similar “artist-run shadow libraries”, which all provide a
technical infrastructure with which they share resources, while the resources
are of specific relevance to their providers.

Just like the large pirate libraries, these artistic archiving and library
practices challenge the notion of culture as private property and remind us
that it is not an unquestionable absolute. As Jonathan Lethem contends,
“[culture] rather is a social negotiation, tenuously forged, endlessly
revised, and imperfect in its every incarnation.”35 Shadow libraries, in
general, are symptomatic of the cultural battles and absurdities around access
and copyright within an economic logic that artificially tries to limit the
abundance of digital culture, in which sharing does not mean dividing but
rather multiplying. They have become a cultural force, one that can be
represented in Foucauldian terms, as symptomatic of broader power struggles as
well as systemic failures inherent in the cultural formation. As Marczewska
puts it, “Goldsmith moves away from thinking about models of cultural
production in proprietary terms and toward paradigms of creativity based on a
culture of collecting, organizing, curating, and sharing content.”36 And by
doing so, he produces major contradictions, or rather he allows the already
existing contradictions to come to light. The artistic archives and libraries
are precarious in terms of their legal status, while it is exactly due to
their disregard of copyright that cultural resources could be built that
exceed the relevance of most official archives that are bound to abide the
law. In fact, there are no comparable official resources, which is why the
function of these projects is at least twofold: education and preservation.37

Maybe UbuWeb and the other, smaller or larger, shadow libraries do not qualify
as commons in the strict sense of involving not only a non-market exchange of
goods but also a community of commoners who negotiate the terms of use among
themselves. This would require collective, formalized, and transparent types
of organization. Furthermore, most of the digital items they circulate are
privately owned and therefore cannot simply be transferred to become commons
resources. These projects, in many respects, are in a preliminary stage by
pointing to the _ideal of culture as a commons_. By providing access to
cultural goods and knowledge that would otherwise not be available at all or
inaccessible for large parts of the general public, they might even fulfill
the function of a “commons fix”, to a certain degree, but at the same time
they are the experimental zone needed to unlearn copyright and relearn new
ways of cultural production and dissemination beyond the property regime. In
any case, they can function as perfect entry points for the discussion and
investigation of the transformative force art can have within the current
global neoliberal knowledge society.

_Top Tens—Showcasing the Copy as an Aesthetic and Political Statement_

The exhibition _Top Tens_ provided an experimental setting to explore the
possibilities of translating the abundance of a digital archive into a “real
space”, by presenting one hundred artworks from the Ubu archive. 38 Although
all works were properly attributed in the exhibition, the artists whose works
were shown neither had a say about their participation in the exhibition nor
about the display formats. Tolerating the presence of a work in the archive is
one thing; tolerating its display in such circumstances is something else,
which might even touch upon moral rights and the integrity of the work.
However, the exhibition was not so much about the individual works on display
but the archiving condition they are subject to. So the discussion here has
nothing to do the abiding art theory question of original and copy.
Marginally, it is about the question of high-quality versus low-quality
copies. In reproducible media the value of an artwork cannot be based on its
originality any longer—the core criterion for sales and market value. This is
why many artists use the trick of high-resolution and limited edition, a kind
of distributed originality status for several authorized objects, which all
are not 100 percent original but still a bit more original than an arbitrary
unlimited edition. Leaving this whole discussion aside was a clear indication
that something else was at stake. The conceptual statement made by the
exhibition and its makers foregrounded the nature of the shadow library, which
visitors were able to experience when entering the gallery space. Instead of
viewing the artworks in the usual way—online—they had the opportunity to
physically immerse themselves in the cultural condition of proliferated acts
of copying, something that “affords their reconceptualization as a hybrid
creative-critical tool and an influential aesthetic category.”39

Appropriation and copying as longstanding methods of subversive artistic
production, where the reuse of existing material serves as a tool for
commentary, social critique, and a means of making a political statement, has
expanded here to the art of exhibition-making. The individual works serve to
illustrate a curatorial concept, thus radically shifting the avant-garde
gesture which copying used to be in the twentieth century, to breathe new life
in the “culture of collecting, organizing, curating, and sharing content.”
Organizing this conceptually concise exhibition was a brave and bold statement
by the art institution: The Onassis Cultural Centre, one of Athens’ most
prestigious cultural institutions, dared to adopt a resolutely political
stance for a—at least in juridical terms—questionable project, as Ubu lives
from the persistent denial of copyright. Neglecting the concerns of the
individual authors and artists for a moment was a necessary precondition in
order to make space for rethinking the future of cultural production.

________________
Special thanks to Eric Steinhauer and all the artists and amateur librarians
who are taking care of our cultural memory.

1 Festival program online: Onassis Cultural Centre, “Shadow Libraries: UbuWeb
in Athens,” (accessed on Sept. 30, 2018).
2 _UbuWeb_ is a massive online archive of avant-garde art created over the
last two decades by New York-based artist and writer Kenneth Goldsmith.
Website of the archive: (accessed on Sept. 30, 2018).
3 Kaja Marczewska, _This Is Not a Copy. Writing at the Iterative Turn_ (New
York: Bloomsbury Academic, 2018), 22.
4 For further reading: Kenneth Goldsmith, _Uncreative Writing: Managing
Language in the Digital Age_ (New York: Columbia University Press, 2011).
5 Many works in the archive stem from the pre-digital era, and there is no
precise knowledge of the sources where Ubu obtains its material, but it is
known that Goldsmith also digitizes a lot of material himself.
6 In German copyright law, for example, §17 and §19a grant the exclusive right
to reproduce, distribute, and make available online to the author. See also:
(accessed on Sept. 30,
2018).
7 Eric Steinhauer, “Rechtspflicht zur Amnesie: Digitale Inhalte, Archive und
Urheberrecht,” _iRightsInfo_ (2013), /rechtspflicht-zur-amnesie-digitale-inhalte-archive-und-urheberrecht/18101>
(accessed on Sept. 30, 2018).
8 In particularly severe cases of copyright infringement also state
prosecutors can become active, which in practice, however, remains the
exception. The circumstances in which criminal law must be applied are
described in §109 of German copyright law.
9 See, for example, “Shadow Libraries” for a video interview with Kenneth
Goldsmith.
10 Paul Torremans, _Intellectual Property Law_ (Oxford: Oxford University
Press, 2010), 265.
11 See also §53 para. 1–3 of the German Act on Copyright and Related Rights
(UrhG), §42 para. 4 in the Austrian UrhG, and Article 19 of Swiss Copyright
Law.
12 Simon Stokes, _Art & Copyright_ (Oxford: Hart Publishing, 2003).
13 Steinhauer, “Rechtspflicht zur Amnesie”.
14 This discrepancy between a state mandate for cultural preservation and
copyright law has only been fixed in 2018 with the introduction of a special
law, §16a DNBG.
15 Steinhauer, “Rechtspflicht zur Amnesie”.
16 Bodo Balazs, “The Genesis of Library Genesis: The Birth of a Global
Scholarly Shadow Library,” Nov. 4, 2014, _SSRN_ ,
, (accessed on
Sept. 30, 2018).
17 Motto of Sci-Hub: “Sci-Hub,” _Wikipedia_ , /Sci-Hub> (accessed on Sept. 30, 2018).
18 Guillaume Cabanac, “Bibliogifts in LibGen? A study of a text-sharing
platform driven by biblioleaks and crowdsourcing,” _Journal of the Association
for Information Science and Technology_ , 67, 4 (2016): 874–884.
19 Ibid.
20 The current address is (accessed on Sept. 30, 2018).
21 Kate Murphy, “Should All Research Papers Be Free?” _New York Times Sunday
Review_ , Mar. 12, 2016, /should-all-research-papers-be-free.html> (accessed on Sept. 30, 2018).
22 Richard Van Noorden, “Nature’s 10,” _Nature_ , Dec. 19, 2016,
(accessed on Sept. 30,
2018).
23 Bodo Balazs, “Pirates in the library – an inquiry into the guerilla open
access movement,” paper for the 8th Annual Workshop of the International
Society for the History and Theory of Intellectual Property, CREATe,
University of Glasgow, UK, July 6–8, 2016. Online available at: https
://adrien-chopin.weebly.com/uploads/2/1/7/6/21765614/2016_bodo_-_pirates.pdf
(accessed on Sept. 30, 2018).
24 Balazs, “The Genesis of Library Genesis”.
25 Aaron Swartz, “Guerilla Open Access Manifesto,” _Internet Archive_ , July
2008,

(accessed on Sept. 30, 2018).
26 Balazs, “Pirates in the library”.
27 Massimo De Angelis, “Economy, Capital and the Commons,” in: _Art,
Production and the Subject in the 21st Century_ , eds. Angela Dimitrakaki and
Kirsten Lloyd (Liverpool: Liverpool University Press, 2015), 201.
28 Ibid., 211.
29 Ibid.
30 See: (accessed on Sept. 30, 2018).
31 Accessible with invitation. See:
[https://aaaaarg.fail/](https://aaaaarg.fail) (accessed on Sept. 30, 2018).
32 See: (accessed on Sept. 30, 2018).
33 See: (accessed on Sept. 30, 2018).
34 Kenneth Goldsmith in conversation with Cornelia Sollfrank, _The Poetry of
Archiving_ , 2013, (accessed on Sept. 30, 2018).
35 Jonathan Lethem, _The Ecstasy of Influence: Nonfictions, etc._ (London:
Vintage, 2012), 101.
36 Marczewska, _This Is Not a Copy_ , 2.
37 The research project _Creating Commons_ , based at Zurich University of the
Arts, is dedicated to the potential of art projects for the creation of
commons: “creating commons,” (accessed on
Sept. 30, 2018).
38 One of Ubu’s features online has been the “top ten”, the idea to invite
guests to pick their ten favorite works from the archive and thus introduce a
mix between chance operation and subjectivity in order to reveal hidden
treasures. The curators of the festival in Athens, Ilan Manouach and Kenneth
Goldsmith, decided to elevate this principle to the curatorial concept of the
exhibition and invited ten guests to select their ten favorite works. The
Athens-based curator Elpida Karaba was commissioned to work on an adequate
concept for the realization, which turned out to be a huge black box divided
into ten small cubicles with monitors and seating areas, supplemented by a
large wall projection illuminating the whole space.
39 Marczewska, _This Is Not a Copy_ , 7.

This text is under a _Creative Commons_ license: CC BY NC SA 3.0 Austria

Medak, Sekulic & Mertens
Book Scanning and Post-Processing Manual Based on Public Library Overhead Scanner v1.2
2014


PUBLIC LIBRARY
&
MULTIMEDIA INSTITUTE

BOOK SCANNING & POST-PROCESSING MANUAL
BASED ON PUBLIC LIBRARY OVERHEAD SCANNER

Written by:
Tomislav Medak
Dubravka Sekulić
With help of:
An Mertens

Creative Commons Attribution - Share-Alike 3.0 Germany

TABLE OF CONTENTS

Introduction
3
I. Photographing a printed book
7
I. Getting the image files ready for post-processing
11
III. Transformation of source images into .tiffs
13
IV. Optical character recognition
16
V. Creating a finalized e-book file
16
VI. Cataloging and sharing the e-book
16
Quick workflow reference for scanning and post-processing
18
References
22

INTRODUCTION:
BOOK SCANNING - FROM PAPER BOOK TO E-BOOK
Initial considerations when deciding on a scanning setup
Book scanning tends to be a fragile and demanding process. Many factors can go wrong or produce
results of varying quality from book to book or page to page, requiring experience or technical skill
to resolve issues that occur. Cameras can fail to trigger, components to communicate, files can get
corrupted in the transfer, storage card doesn't get purged, focus fails to lock, lighting conditions
change. There are trade-offs between the automation that is prone to instability and the robustness
that is prone to become time consuming.
Your initial choice of book scanning setup will have to take these trade-offs into consideration. If
your scanning community is confined to your hacklab, you won't be risking much if technological
sophistication and integration fails to function smoothly. But if you're aiming at a broad community
of users, with varying levels of technological skill and patience, you want to create as much timesaving automation as possible on the condition of keeping maximum stability. Furthermore, if the
time of individual members of your scanning community can contribute is limited, you might also
want to divide some of the tasks between users and their different skill levels.
This manual breaks down the process of digitization into a general description of steps in the
workflow leading from the printed book to a digital e-book, each of which can be in a concrete
situation addressed in various manners depending on the scanning equipment, software, hacking
skills and user skill level that are available to your book scanning project. Several of those steps can
be handled by a single piece of equipment or software, or you might need to use a number of them your mileage will vary. Therefore, the manual will try to indicate the design choices you have in the
process of planning your workflow and should help you make decisions on what design is best for
you situation.
Introducing book scanner designs
The book scanning starts with the capturing of digital image files on the scanning equipment. There
are three principle types of book scanner designs:
 flatbed scanner
 single camera overhead scanner
 dual camera overhead scanner
Conventional flatbed scanners are widely available. However, given that they require the book to be
spread wide open and pressed down with the platen in order to break the resistance of the book
binding and expose sufficiently the inner margin of the text, it is the most destructive approach for
the book, imprecise and slow.
Therefore, book scanning projects across the globe have taken to custom designing improvised
setups or scanner rigs that are less destructive and better suited for fast turning and capturing of
pages. Designs abound. Most include:




one or two digital photo cameras of lesser or higher quality to capture the pages,
transparent V-shaped glass or Plexiglas platen to press the open book against a V-shape
cradle, and
a light source.

The go-to web resource to help you make an informed decision is the DIY book scanning
community at http://diybookscanner.org. A good place to start is their intro
(http://wiki.diybookscanner.org/ ) and scanner build list (http://wiki.diybookscanner.org/scannerbuild-list ).
The book scanners with a single camera are substantially cheaper, but come with an added difficulty
of de-warping the distorted page images due to the angle that pages are photographed at, which can
sometimes be difficult to correct in the post-processing. Hence, in this introductory chapter we'll
focus on two camera designs where the camera lens stands relatively parallel to the page. However,
with a bit of adaptation these instructions can be used to work with any other setup.
The Public Library scanner
In the focus of this manual is the scanner built for the Public Library project, designed by Voja
Antonić (see Illustration 1). The Public Library scanner was built with the immediate use by a wide
community of users in mind. Hence, the principle consideration in designing the Public Library
scanner was less sophistication and more robustness, facility of use and distributed process of
editing.
The board designs can be found here: http://www.memoryoftheworld.org/blog/2012/10/28/ourbeloved-bookscanner. The current iterations are using two Canon 1100 D cameras with the kit lens
Canon EF-S 18-55mm 1:3.5-5.6 IS. Cameras are auto-charging.

Illustration 1: Public Library Scanner
The scanner operates by automatically lowering the Plexiglas platen, illuminating the page and then
triggering camera shutters. The turning of pages and the adjustments of the V-shaped cradle holding

the book are manual.
The scanner is operated by a two-button controller (see Illustration 2). The upper, smaller button
breaks the capture process in two steps: the first click lowers the platen, increases the light level and
allows you to adjust the book or the cradle, the second click triggers the cameras and lifts the platen.
The lower button has
two modes. A quick
click will execute the
whole capture process in
one go. But if you hold
it pressed longer, it will
lower the platen,
allowing you to adjust
the book and the cradle,
and lift it without
triggering cameras when
you press again.

Illustration 2: A two-button controller

More on this manual: steps in the book scanning process
The book scanning process in general can be broken down in six steps, each of which will be dealt
in a separate chapter in this manual:
I. Photographing a printed book
I. Getting the image files ready for post-processing
III. Transformation of source images into .tiffs
IV. Optical character recognition
V. Creating a finalized e-book file
VI. Cataloging and sharing the e-book
A step by step manual for Public Library scanner
This manual is primarily meant to provide a detailed description and step-by-step instructions for an
actual book scanning setup -- based on the Voja Antonić's scanner design described above. This is a
two-camera overhead scanner, currently equipped with two Canon 1100 D cameras with EF-S 1855mm 1:3.5-5.6 IS kit lens. It can scan books of up to A4 page size.
The post-processing in this setup is based on a semi-automated transfer of files to a GNU/Linux
personal computer and on the use of free software for image editing, optical character recognition
and finalization of an e-book file. It was initially developed for the HAIP festival in Ljubljana in
2011 and perfected later at MaMa in Zagreb and Leuphana University in Lüneburg.
Public Library scanner is characterized by a somewhat less automated yet distributed scanning
process than highly automated and sophisticated scanner hacks developed at various hacklabs. A
brief overview of one such scanner, developed at the Hacker Space Bruxelles, is also included in
this manual.
The Public Library scanning process proceeds thus in following discrete steps:

1. creating digital images of pages of a book,
2. manual transfer of image files to the computer for post-processing,
3. automated renaming of files, ordering of even and odd pages, rotation of images and upload to a
cloud storage,
4. manual transformation of source images into .tiff files in ScanTailor
5. manual optical character recognition and creation of PDF files in gscan2pdf
The detailed description of the Public Library scanning process follows below.
The Bruxelles hacklab scanning process
For purposes of comparison, here we'll briefly reference the scanner built by the Bruxelles hacklab
(http://hackerspace.be/ScanBot). It is a dual camera design too. With some differences in hardware functionality
(Bruxelles scanner has automatic turning of pages, whereas Public Library scanner has manual turning of pages), the
fundamental difference between the two is in the post-processing - the level of automation in the transfer of images
from the cameras and their transformation into PDF or DjVu e-book format.
The Bruxelles scanning process is different in so far as the cameras are operated by a computer and the images are
automatically transferred, ordered and made ready for further post-processing. The scanner is home-brew, but the
process is for advanced DIY'ers. If you want to know more on the design of the scanner, contact Michael Korntheuer at
contact@hackerspace.be.
The scanning and post-processing is automated by a single Python script that does all the work
http://git.constantvzw.org/?
p=algolit.git;a=tree;f=scanbot_brussel;h=81facf5cb106a8e4c2a76c048694a3043b158d62;hb=HEAD
The scanner uses two Canon point and shoot cameras. Both cameras are connected to the PC with USB. They both run
PTP/CHDK (Canon Hack Development Kit). The scanning sequence is the following:
1. Script sends CHDK command line instructions to the cameras
2. Script sorts out the incoming files. This part is tricky. There is no reliable way to make a distinction between the left
and right camera, only between which camera was recognized by USB first. So the protocol is to always power up the
left camera first. See the instructions with the source code.
3. Collect images in a PDF file
4. Run script to OCR a .PDF file to plain .TXT file: http://git.constantvzw.org/?
p=algolit.git;a=blob;f=scanbot_brussel/ocr_pdf.sh;h=2c1f24f9afcce03520304215951c65f58c0b880c;hb=HEAD

I. PHOTOGRAPHING A PRINTED BOOK
Technologically the most demanding part of the scanning process is creating digital images of the
pages of a printed book. It's a process that is very different form scanner design to scanner design,
from camera to camera. Therefore, here we will focus strictly on the process with the Public Library
scanner.
Operating the Public Library scanner
0. Before you start:
Better and more consistent photographs lead to a more optimized and faster post-processing and a
higher quality of the resulting digital e-book. In order to guarantee the quality of images, before you
start it is necessary to set up the cameras properly and prepare the printed book for scanning.
a) Loosening the book
Depending on the type and quality of binding, some books tend to be too resistant to opening fully
to reveal the inner margin under the pressure of the scanner platen. It is thus necessary to “break in”
the book before starting in order to loosen the binding. The best way is to open it as wide as
possible in multiple places in the book. This can be done against the table edge if the book is more
rigid than usual. (Warning – “breaking in” might create irreversible creasing of the spine or lead to
some pages breaking loose.)
b) Switch on the scanner
You start the scanner by pressing the main switch or plugging the power cable into the the scanner.
This will also turn on the overhead LED lights.

c) Setting up the cameras
Place the cameras onto tripods. You need to move the lever on the tripod's head to allow the tripod
plate screwed to the bottom of the camera to slide into its place. Secure the lock by turning the lever
all the way back.
If the automatic chargers for the camera are provided, open the battery lid on the bottom of the
camera and plug the automatic charger. Close the lid.
Switch on the cameras using the lever on the top right side of the camera's body and place it into the
aperture priority (Av) mode on the mode dial above the lever (see Illustration 3). Use the main dial
just above the shutter button on the front side of the camera to set the aperture value to F8.0.

Illustration 3: Mode and main dial, focus mode switch, zoom
and focus ring
On the lens, turn the focus mode switch to manual (MF), turn the large zoom ring to set the value
exactly midway between 24 and 35 mm (see Illustration 3). Try to set both cameras the same.
To focus each camera, open a book on the cradle, lower the platen by holding the big button on the
controller, and turn on the live view on camera LCD by pressing the live view switch (see
Illustration 4). Now press the magnification button twice and use the focus ring on the front of the
lens to get a clear image view.

Illustration 4: Live view switch and magnification button

d) Connecting the cameras
Now connect the cameras to the remote shutter trigger cables that can be found lying on each side
of the scanner. They need to be plugged into a small round port hidden behind a protective rubber
cover on the left side of the cameras.
e) Placing the book into the cradle and double-checking the cameras
Open the book in the middle and place it on the cradle. Hold pressed the large button on the
controller to lower the Plexiglas platen without triggering the cameras. Move the cradle so that the
the platen fits into with the middle of the book.
Turn on the live view on the cameras' LED to see if the the pages fit into the image and if the
cameras are positioned parallel to the page.
f) Double-check storage cards and batteries
It is important that both storage cards on cameras are empty before starting the scanning in order
not to mess up the page sequence when merging photos from the left and the right camera in the
post-processing. To double-check, press play button on cameras and erase if there are some photos
left from the previous scan -- this you do by pressing the menu button, selecting the fifth menu from
the left and then select 'Erase Images' -> 'All images on card' -> 'OK'.
If no automatic chargers are provided, double-check on the information screen that batteries are
charged. They should be fully charged before starting with the scanning of a new book.

g) Turn off the light in the room
Lighting conditions during scanning should be as constant as possible, to reduce glare and achieve
maximum quality remove any source of light that might reflect off the Plexiglas platen. Preferably
turn off the light in the room or isolate the scanner with the black cloth provided.

1. Photographing a book
Now you are ready to start scanning. Place the book closed in the cradle and lower the platen by
holding the large button on the controller pressed (see Illustration 2). Adjust the position of the
cradle and lift the platen by pressing the large button again.
To scan you can now either use the small button on the controller to lower the platen, adjust and
then press it again to trigger the cameras and lift the platen. Or, you can just make a short press on
the large button to do it in one go.
ATTENTION: When the cameras are triggered, the shutter sound has to be heard coming
from both cameras. If one camera is not working, it's best to reconnect both cameras (see
Section 0), make sure the batteries are charged or adapters are connected, erase all images
and restart.
A mistake made in the photographing requires a lot of work in the post-processing, so it's
much quicker to repeat the photographing process.
If you make a mistake while flipping pages, or any other mistake, go back and scan from the page
you missed or incorrectly scanned. Note down the page where the error occurred and in the postprocessing the redundant images will be removed.
ADVICE: The scanner has a digital counter. By turning the dial forward and backward, you
can set it to tell you what page you should be scanning next. This should help you avoid
missing a page due to a distraction.
While scanning, move the cradle a bit to the left from time to time, making sure that the tip of Vshaped platen is aligned with the center of the book and the inner margin is exposed enough.

II. GETTING THE IMAGE FILES READY FOR POST-PROCESSING
Once the book pages have been photographed, they have to be transfered to the computer and
prepared for post-processing. With two-camera scanners, the capturing process will result in two
separate sets of images -- odd and even pages -- coming from the left and right cameras respectively
-- and you will need to rename and reorder them accordingly, rotate them into a vertical position
and collate them into a single sequence of files.
a) Transferring image files
For the transfer of files your principle process design choices are either to copy the files by
removing the memory cards from the cameras and copying them to the computer via a card reader
or to transfer them via a USB cable. The latter process can be automated by remote operating your
cameras from a computer, however this can be done only with a certain number of Canon cameras
(http://bit.ly/16xhJ6b) that can be hacked to run the open Canon Hack Development Kit firmware
(http://chdk.wikia.com).
After transferring the files, you want to erase all the image files on the camera memory card, so that
they would not end up messing up the scan of the next book.
b) Renaming image files
As the left and right camera are typically operated in sync, the photographing process results in two
separate sets of images, with even and odd pages respectively, that have completely different file
names and potentially same time stamps. So before you collate the page images in the order how
they appear in the book, you want to rename the files so that the first image comes from the right
camera, the second from the left camera, the third comes again from the right camera and so on.
You probably want to do a batch renaming, where your right camera files start with n and are offset
by an increment of 2 (e.g. page_0000.jpg, page_0002.jpg,...) and your left camera files start with
n+1 and are also offset by an increment of 2 (e.g. page_0001.jpg, page_0003.jpg,...).
Batch renaming can be completed either from your file manager, in command line or with a number
of GUI applications (e.g. GPrename, rename, cuteRenamer on GNU/Linux).
c) Rotating image files
Before you collate the renamed files, you might want to rotate them. This is a step that can be done
also later in the post-processing (see below), but if you are automating or scripting your steps this is
a practical place to do it. The images leaving your cameras will be positioned horizontally. In order
to position them vertically, the images from the camera on the right will have to be rotated by 90
degrees counter-clockwise, the images from the camera on the left will have to be rotated by 90
degrees clockwise.
Batch rotating can be completed in a number of photo-processing tools, in command line or
dedicated applications (e.g. Fstop, ImageMagick, Nautilust Image Converter on GNU/Linux).
d) Collating images into a single batch
Once you're done with the renaming and rotating of the files, you want to collate them into the same
folder for easier manipulation later.

Getting the image files ready for post-processing on the Public Library scanner
In the case of Public Library scanner, a custom C++ script was written by Mislav Stublić to
facilitate the transfer, renaming, rotating and collating of the images from the two cameras.
The script prompts the user to place into the card reader the memory card from the right camera
first, gives a preview of the first and last four images and provides an entry field to create a subfolder in a local cloud storage folder (path: /home/user/Copy).
It transfers, renames, rotates the files, deletes them from the card and prompts the user to replace the
card with the one from the left camera in order to the transfer the files from there and place them in
the same folder. The script was created for GNU/Linux system and it can be downloaded, together
with its source code, from: https://copy.com/nLSzflBnjoEB
If you have other cameras than Canon, you can edit the line 387 of the source file to change to the
naming convention of your cameras, and recompile by running the following command in your
terminal: "gcc scanflow.c -o scanflow -ludev `pkg-config --cflags --libs gtk+-2.0`"
In the case of Hacker Space Bruxelles scanner, this is handled by the same script that operates the cameras that can be
downloaded from: http://git.constantvzw.org/?
p=algolit.git;a=tree;f=scanbot_brussel;h=81facf5cb106a8e4c2a76c048694a3043b158d62;hb=HEAD

III. TRANSFORMATION OF SOURCE IMAGES INTO .TIFFS
Images transferred from the cameras are high definition full color images. You want your cameras
to shoot at the largest possible .jpg resolution in order for resulting files to have at least 300 dpi (A4
at 300 dpi requires a 9.5 megapixel image). In the post-processing the size of the image files needs
to be reduced down radically, so that several hundred images can be merged into an e-book file of a
tolerable size.
Hence, the first step in the post-processing is to crop the images from cameras only to the content of
the pages. The surroundings around the book that were captured in the photograph and the white
margins of the page will be cropped away, while the printed text will be transformed into black
letters on white background. The illustrations, however, will need to be preserved in their color or
grayscale form, and mixed with the black and white text. What were initially large .jpg files will
now become relatively small .tiff files that are ready for optical character recognition process
(OCR).
These tasks can be completed by a number of software applications. Our manual will focus on one
that can be used across all major operating systems -- ScanTailor. ScanTailor can be downloaded
from: http://scantailor.sourceforge.net/. A more detailed video tutorial of ScanTailor can be found
here: http://vimeo.com/12524529.
ScanTailor: from a photograph of a page to a graphic file ready for OCR
Once you have transferred all the photos from cameras to the computer, renamed and rotated them,
they are ready to be processed in the ScanTailor.
1) Importing photographs to ScanTailor
- start ScanTailor and open ‘new project’
- for ‘input directory’ chose the folder where you stored the transferred and renamed photo images
- you can leave ‘output directory’ as it is, it will place your resulting .tiffs in an 'out' folder inside
the folder where your .jpg images are
- select all files (if you followed the naming convention above, they will be named
‘page_xxxx.jpg’) in the folder where you stored the transferred photo images, and click 'OK'
- in the dialog box ‘Fix DPI’ click on All Pages, and for DPI choose preferably '600x600', click
'Apply', and then 'OK'
2) Editing pages
2.1 Rotating photos/pages
If you've rotated the photo images in the previous step using the scanflow script, skip this step.
- Rotate the first photo counter-clockwise, click Apply and for scope select ‘Every other page’
followed by 'OK'
- Rotate the following photo clockwise, applying the same procedure like in the previous step
2.2 Deleting redundant photographs/pages
- Remove redundant pages (photographs of the empty cradle at the beginning and the end of the
book scanning sequence; book cover pages if you don’t want them in the final scan; duplicate pages
etc.) by right-clicking on a thumbnail of that page in the preview column on the right side, selecting
‘Remove from project’ and confirming by clicking on ‘Remove’.

# If you by accident remove a wrong page, you can re-insert it by right-clicking on a page
before/after the missing page in the sequence, selecting 'insert after/before' (depending on which
page you selected) and choosing the file from the list. Before you finish adding, it is necessary to
again go through the procedure of fixing DPI and Rotating.
2.3 Adding missing pages
- If you notice that some pages are missing, you can recapture them with the camera and insert them
manually at this point using the procedure described above under 2.2.
3) Split pages and deskew
Steps ‘Split pages’ and ‘Deskew’ should work automatically. Run them by clicking the ‘Play’ button
under the 'Select content' function. This will do the three steps automatically: splitting of pages,
deskewing and selection of content. After this you can manually re-adjust splitting of pages and deskewing.
4) Selecting content
Step ‘Select content’ works automatically as well, but it is important to revise the resulting selection
manually page by page to make sure the entire content is selected on each page (including the
header and page number). Where necessary, use your pointer device to adjust the content selection.
If the inner margin is cut, go back to 'Split pages' view and manually adjust the selected split area. If
the page is skewed, go back to 'Deskew' and adjust the skew of the page. After this go back to
'Select content' and readjust the selection if necessary.
This is the step where you do visual control of each page. Make sure all pages are there and
selections are as equal in size as possible.
At the bottom of thumbnail column there is a sort option that can automatically arrange pages by
the height and width of the selected content, making the process of manual selection easier. The
extreme differences in height should be avoided, try to make selected areas as much as possible
equal, particularly in height, across all pages. The exception should be cover and back pages where
we advise to select the full page.
5) Adjusting margins
For best results select in the previous step content of the full cover and back page. Now go to the
'Margins' step and set under Margins section both Top, Bottom, Left and Right to 0.0 and do 'Apply
to...' → 'All pages'.
In Alignment section leave 'Match size with other pages' ticked, choose the central positioning of
the page and do 'Apply to...' → 'All pages'.
6) Outputting the .tiffs
Now go to the 'Output' step. Ignore the 'Output Resolution' section.
Next review two consecutive pages from the middle of the book to see if the scanned text is too
faint or too dark. If the text seems too faint or too dark, use slider Thinner – Thicker to adjust. Do
'Apply to' → 'All pages'.
Next go to the cover page and select under Mode 'Color / Grayscale' and tick on 'White Margins'.
Do the same for the back page.
If there are any pages with illustrations, you can choose the 'Mixed' mode for those pages and then

under the thumb 'Picture Zones' adjust the zones of the illustrations.
Now you are ready to output the files. Just press 'Play' button under 'Output'. Once the computer is
finished processing the images, just do 'File' → 'Save as' and save the project.

IV. OPTICAL CHARACTER RECOGNITION
Before the edited-down graphic files are finalized as an e-book, we want to transform the image of
the text into an actual text that can be searched, highlighted, copied and transformed. That
functionality is provided by Optical Character Recognition. This a technically difficult task dependent on language, script, typeface and quality of print - and there aren't that many OCR tools
that are good at it. There is, however, a relatively good free software solution - Tesseract
(http://code.google.com/p/tesseract-ocr/) - that has solid performance, good language data and can
be trained for an even better performance, although it has its problems. Proprietary solutions (e.g.
Abby FineReader) sometimes provide superior results.
Tesseract supports as input format primarily .tiff files. It produces a plain text file that can be, with
the help of other tools, embedded as a separate layer under the original graphic image of the text in
a PDF file.
With the help of other tools, OCR can be performed also against other input files, such as graphiconly PDF files. This produces inferior results, depending again on the quality of graphic files and
the reproduction of text in them. One such tool is a bashscript to OCR a ODF file that can be found
here: https://github.com/andrecastro0o/ocr/blob/master/ocr.sh
As mentioned in the 'before scanning' section, the quality of the original book will influence the
quality of the scan and thus the quality of the OCR. For a comparison, have a look here:
http://www.paramoulipist.be/?p=1303
Once you have your .txt file, there is still some work to be done. Because OCR has difficulties to
interpret particular elements in the lay-out and fonts, the TXT file comes with a lot of errors.
Recurrent problems are:
- combinations of specific letters in some fonts (it can mistake 'm' for 'n' or 'I' for 'i' etc.);
- headers become part of body text;
- footnotes are placed inside the body text;
- page numbers are not recognized as such.

V. CREATING A FINALIZED E-BOOK FILE
After the optical character recognition has been completed, the resulting text can be merged with
the images of pages and output into an e-book format. While increasingly the proper e-book file
formats such as ePub have been gaining ground, PDFs still remain popular because many people
tend to read on their computers, and they retain the original layout of the book on paper including
the absolute pagination needed for referencing in citations. DjVu is also an option, as an alternative
to PDF, used because of its purported superiority, but it is far less popular.
The export to PDF can be done again with a number of tools. In our case we'll complete the optical
character recognition and PDF export in gscan2pdf. Again, the proprietary Abbyy FineReader will
produce a bit smaller PDFs.
If you prefer to use an e-book format that works better with e-book readers, obviously you will have
to remove some of the elements that appear in the book - headers, footers, footnotes and pagination.

This can be done earlier in the process of cropping down the original .jpg image files (see under III)
or later by transforming the PDF files. This can be done in Calibre (http://calibre-ebook.com) by
converting the PDF into an ePub, where it can be further tweaked to better accommodate or remove
the headers, footers, footnotes and pagination.
Optical character recognition and PDF export in Public Library workflow
Optical character recognition with the Tesseract engine can be performed on GNU/Linux by a
number of command line and GUI tools. Much of those tools exist also for other operating systems.
For the users of the Public Library workflow, we recommend using gscan2pdf application both for
the optical character recognition and the PDF or DjVu export.
To do so, start gscan2pdf and open your .tiff files. To OCR them, go to 'Tools' and select 'OCR'. In
the dialog box select the Tesseract engine and your language. 'Start OCR'. Once the OCR is
finished, export the graphic files and the OCR text to PDF by selecting 'Save as'.
However, given that sometimes the proprietary solutions produce better results, these tasks can also
be done, for instance, on the Abbyy FineReader running on a Windows operating system running
inside the Virtual Box. The prerequisites are that you have both Windows and Abbyy FineReader
you can install in the Virtual Box. If using Virtual Box, once you've got both installed, you need to
designate a shared folder in your Virtual Box and place the .tiff files there. You can now open them
from the Abbyy FineReader running in the Virtual Box, OCR them and export them into a PDF.
To use Abbyy FineReader transfer the output files in your 'out' out folder to the shared folder of the
VirtualBox. Then start the VirtualBox, start Windows image and in Windows start Abbyy
FineReader. Open the files and let the Abbyy FineReader read the files. Once it's done, output the
result into PDF.

VI. CATALOGING AND SHARING THE E-BOOK
Your road from a book on paper to an e-book is complete. If you want to maintain your library you
can use Calibre, a free software tool for e-book library management. You can add the metadata to
your book using the existing catalogues or you can enter metadata manually.
Now you may want to distribute your book. If the work you've digitized is in the public domain
(https://en.wikipedia.org/wiki/Public_domain), you might consider contributing it to the Gutenberg
project
(http://www.gutenberg.org/wiki/Gutenberg:Volunteers'_FAQ#V.1._How_do_I_get_started_as_a_Pr
oject_Gutenberg_volunteer.3F ), Wikibooks (https://en.wikibooks.org/wiki/Help:Contributing ) or
Arhive.org.
If the work is still under copyright, you might explore a number of different options for sharing.

QUICK WORKFLOW REFERENCE FOR SCANNING AND
POST-PROCESSING ON PUBLIC LIBRARY SCANNER
I. PHOTOGRAPHING A PRINTED BOOK
0. Before you start:
- loosen the book binding by opening it wide on several places
- switch on the scanner
- set up the cameras:
- place cameras on tripods and fit them tigthly
- plug in the automatic chargers into the battery slot and close the battery lid
- switch on the cameras
- switch the lens to Manual Focus mode
- switch the cameras to Av mode and set the aperture to 8.0
- turn the zoom ring to set the focal length exactly midway between 24mm and 35mm
- focus by turning on the live view, pressing magnification button twice and adjusting the
focus to get a clear view of the text
- connect the cameras to the scanner by plugging the remote trigger cable to a port behind a
protective rubber cover on the left side of the cameras
- place the book into the crade
- double-check storage cards and batteries
- press the play button on the back of the camera to double-check if there are images on the
camera - if there are, delete all the images from the camera menu
- if using batteries, double-check that batteries are fully charged
- switch off the light in the room that could reflect off the platen and cover the scanner with the
black cloth
1. Photographing
- now you can start scanning either by pressing the smaller button on the controller once to
lower the platen and adjust the book, and then press again to increase the light intensity, trigger the
cameras and lift the platen; or by pressing the large button completing the entire sequence in one
go;
- ATTENTION: Shutter sound should be coming from both cameras - if one camera is not
working, it's best to reconnect both cameras, make sure the batteries are charged or adapters
are connected, erase all images and restart.
- ADVICE: The scanner has a digital counter. By turning the dial forward and backward,
you can set it to tell you what page you should be scanning next. This should help you to
avoid missing a page due to a distraction.

II. Getting the image files ready for post-processing
- after finishing with scanning a book, transfer the files to the post-processing computer
and purge the memory cards
- if transferring the files manually:
- create two separate folders,
- transfer the files from the folders with image files on cards, using a batch
renaming software rename the files from the right camera following the convention
page_0001.jpg, page_0003.jpg, page_0005.jpg... -- and the files from the left camera
following the convention page_0002.jpg, page_0004.jpg, page_0006.jpg...
- collate image files into a single folder
- before ejecting each card, delete all the photo files on the card
- if using the scanflow script:
- start the script on the computer
- place the card from the right camera into the card reader
- enter the name of the destination folder following the convention
"Name_Surname_Title_of_the_Book" and transfer the files
- repeat with the other card
- script will automatically transfer the files, rename, rotate, collate them in proper
order and delete them from the card
III. Transformation of source images into .tiffs
ScanTailor: from a photograph of page to a graphic file ready for OCR
1) Importing photographs to ScanTailor
- start ScanTailor and open ‘new project’
- for ‘input directory’ chose the folder where you stored the transferred photo images
- you can leave ‘output directory’ as it is, it will place your resulting .tiffs in an 'out' folder
inside the folder where your .jpg images are
- select all files (if you followed the naming convention above, they will be named
‘page_xxxx.jpg’) in the folder where you stored the transferred photo images, and click
'OK'
- in the dialog box ‘Fix DPI’ click on All Pages, and for DPI choose preferably '600x600',
click 'Apply', and then 'OK'
2) Editing pages
2.1 Rotating photos/pages
If you've rotated the photo images in the previous step using the scanflow script, skip this step.
- rotate the first photo counter-clockwise, click Apply and for scope select ‘Every other
page’ followed by 'OK'
- rotate the following photo clockwise, applying the same procedure like in the previous
step

2.2 Deleting redundant photographs/pages
- remove redundant pages (photographs of the empty cradle at the beginning and the end;
book cover pages if you don’t want them in the final scan; duplicate pages etc.) by rightclicking on a thumbnail of that page in the preview column on the right, selecting ‘Remove
from project’ and confirming by clicking on ‘Remove’.
# If you by accident remove a wrong page, you can re-insert it by right-clicking on a page
before/after the missing page in the sequence, selecting 'insert after/before' and choosing the file
from the list. Before you finish adding, it is necessary to again go the procedure of fixing DPI and
rotating.
2.3 Adding missing pages
- If you notice that some pages are missing, you can recapture them with the camera and
insert them manually at this point using the procedure described above under 2.2.
3)

Split pages and deskew
- Functions ‘Split Pages’ and ‘Deskew’ should work automatically. Run them by
clicking the ‘Play’ button under the 'Select content' step. This will do the three steps
automatically: splitting of pages, deskewing and selection of content. After this you can
manually re-adjust splitting of pages and de-skewing.

4)

Selecting content and adjusting margins
- Step ‘Select content’ works automatically as well, but it is important to revise the
resulting selection manually page by page to make sure the entire content is selected on
each page (including the header and page number). Where necessary use your pointer device
to adjust the content selection.
- If the inner margin is cut, go back to 'Split pages' view and manually adjust the selected
split area. If the page is skewed, go back to 'Deskew' and adjust the skew of the page. After
this go back to 'Select content' and readjust the selection if necessary.
- This is the step where you do visual control of each page. Make sure all pages are there
and selections are as equal in size as possible.
- At the bottom of thumbnail column there is a sort option that can automatically arrange
pages by the height and width of the selected content, making the process of manual
selection easier. The extreme differences in height should be avoided, try to make
selected areas as much as possible equal, particularly in height, across all pages. The
exception should be cover and back pages where we advise to select the full page.

5) Adjusting margins
- Now go to the 'Margins' step and set under Margins section both Top, Bottom, Left and
Right to 0.0 and do 'Apply to...' → 'All pages'.
- In Alignment section leave 'Match size with other pages' ticked, choose the central

positioning of the page and do 'Apply to...' → 'All pages'.
6) Outputting the .tiffs
- Now go to the 'Output' step.
- Review two consecutive pages from the middle of the book to see if the scanned text is
too faint or too dark. If the text seems too faint or too dark, use slider Thinner – Thicker to
adjust. Do 'Apply to' → 'All pages'.
- Next go to the cover page and select under Mode 'Color / Grayscale' and tick on 'White
Margins'. Do the same for the back page.
- If there are any pages with illustrations, you can choose the 'Mixed' mode for those
pages and then under the thumb 'Picture Zones' adjust the zones of the illustrations.
- To output the files press 'Play' button under 'Output'. Save the project.
IV. Optical character recognition & V. Creating a finalized e-book file
If using all free software:
1) open gscan2pdf (if not already installed on your machine, install gscan2pdf from the
repositories, Tesseract and data for your language from https://code.google.com/p/tesseract-ocr/)
- point gscan2pdf to open your .tiff files
- for Optical Character Recognition, select 'OCR' under the drop down menu 'Tools',
select the Tesseract engine and your language, start the process
- once OCR is finished and to output to a PDF, go under 'File' and select 'Save', edit the
metadata and select the format, save
If using non-free software:
2) open Abbyy FineReader in VirtualBox (note: only Abby FineReader 10 installs and works with some limitations - under GNU/Linux)
- transfer files in the 'out' folder to the folder shared with the VirtualBox
- point it to the readied .tiff files and it will complete the OCR
- save the file

REFERENCES
For more information on the book scanning process in general and making your own book scanner
please visit:
DIY Book Scanner: http://diybookscannnner.org
Hacker Space Bruxelles scanner: http://hackerspace.be/ScanBot
Public Library scanner: http://www.memoryoftheworld.org/blog/2012/10/28/our-belovedbookscanner/
Other scanner builds: http://wiki.diybookscanner.org/scanner-build-list
For more information on automation:
Konrad Voeckel's post-processing script (From Scan to PDF/A):
http://blog.konradvoelkel.de/2013/03/scan-to-pdfa/
Johannes Baiter's automation of scanning to PDF process: http://spreads.readthedocs.org
For more information on applications and tools:
Calibre e-book library management application: http://calibre-ebook.com/
ScanTailor: http://scantailor.sourceforge.net/
gscan2pdf: http://sourceforge.net/projects/gscan2pdf/
Canon Hack Development Kit firmware: http://chdk.wikia.com
Tesseract: http://code.google.com/p/tesseract-ocr/
Python script of Hacker Space Bruxelles scanner: http://git.constantvzw.org/?
p=algolit.git;a=tree;f=scanbot_brussel;h=81facf5cb106a8e4c2a76c048694a3043b158d62;hb=HEA
D


 

Display 200 300 400 500 600 700 800 900 1000 ALL characters around the word.