<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Empirical Evidence of the Limits of Automatic Assessment of Fictional Ideation</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">A</forename><surname>Tapscott</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Facultad de Informática</orgName>
								<orgName type="institution">Universidad Complutense de Madrid</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">J</forename><surname>Gómez</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Facultad de Informática</orgName>
								<orgName type="institution">Universidad Complutense de Madrid</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">C</forename><surname>León</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Facultad de Informática</orgName>
								<orgName type="institution">Universidad Complutense de Madrid</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">J</forename><surname>Smailović</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Department of Knowledge Technologies</orgName>
								<orgName type="institution">Jožef Stefan Institute</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">M</forename><surname>Žnidaršič</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Department of Knowledge Technologies</orgName>
								<orgName type="institution">Jožef Stefan Institute</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">P</forename><surname>Gervás</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Facultad de Informática</orgName>
								<orgName type="institution">Universidad Complutense de Madrid</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Empirical Evidence of the Limits of Automatic Assessment of Fictional Ideation</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">4557B448362D758C732ABC0D1A608FF6</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T03:11+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Automatic evaluation</term>
					<term>ideation</term>
					<term>empirical study</term>
					<term>narrative</term>
					<term>computational creativity</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Automatic evaluation of fictional ideation systems and their output is a topic relevant to Computational Creativity. Models and techniques have been proposed for this task, but their applicability is limited to the field of fictional ideation. In this paper we describe an evaluation procedure for fictional ideation, which compares human validation of the ideas with a number of automatically generated metrics obtained from them. We report on the observed limits of this procedure. The results suggest that, besides technical limitations, providing a stable evaluation method is fundamentally incomplete unless the full creative phenomenon is modelled, including aspects that are beyond current technical capabilities.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Evaluation of creative processes and artefacts is key to computational creativity. Explicitly reflecting on the relative value and novelty is crucial if machines are to produce content that would be deemed creative <ref type="bibr" target="#b5">[6]</ref>. As such, addressing evaluation is fundamental for computational creativity that can successfully fulfill human needs.</p><p>This crucial aspect contrasts with the relative scarcity of systems explicitly generating rich evaluation of their own generated material or inner processes. Some systems arguably control the quality of their artifacts by carrying out a process that ensures a minimum relative quality, but an explicit evaluation arguably represents a qualitative advantage, both theoretical (as studied by computational creativity frameworks <ref type="bibr" target="#b28">[29]</ref>) and practical ( <ref type="bibr" target="#b3">[4]</ref>).</p><p>Although the semantics of creativity are elusive and usually problematic, the vision that quality and novelty influence the perception of the creativity of an artifact (at least from the point of view of observation) is commonly accepted. Still, quality and novelty vary depending on the domain and context. Theoretical</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Previous Work</head><p>While all scientific exploration requires thorough evaluation of the steps taken, doing so in creativity represents a challenge. How to assess creativity itself is a commonly discussed aspect of the whole phenomena of creative generation. While most authors agree on the correlation between a number of features and the perception of creativity, there is no consensus either on what these features are or how they really correlate. Moreover, adding computers to the problem makes it even more difficult to know whether a system has been successful or not. There is still a debate on what parts should be evaluated, the influence of the programmer on the output, the very definition of creative behavior, the decision of whether to focus on the process or the artifacts (or both), and many others.</p><p>The few examples present in the literature describing actual evaluation of automatic creative systems usually focus on less ambitious, more measurable aspects. This makes these systems less useful from a general perspective, but they nonetheless provide insight on the current capabilities of computer systems to assess their own production.</p><p>There is, however, a number of proposals that try to provide guidelines to evaluate creative systems. For instance, Ritchie <ref type="bibr" target="#b23">[24,</ref><ref type="bibr" target="#b24">25]</ref> addresses the issue of evaluating when a program can be considered creative by outlining a set of empirical criteria to measure the creativity of the program in terms of its output. He makes it very clear that he is restricting his analysis to the questions of what factors are to be observed, and how these might relate to creativity, specifically stating that he does not intend to build a model of creativity. Ritchie's criteria are defined in terms of two observable properties of the results produced by the program: novelty (to what extent is the produced item dissimilar to existing examples of that genre) and quality (to what extent is the produced item a highquality example of that genre). To measure these aspects, two rating schemes are introduced, which rate the typicality of a given item (item is typical) and its quality (item is good). Another important issue that affects the assessment of creativity in creative programs is the concept of inspiring set, the set of (usually highly valued) artifacts that the programmer is guided by when designing a creative program. Ritchie's criteria are phrased in terms of: what proportion of the results rates well according to each rating scheme, ratios between various subsets of the result (defined in terms of their ratings), and whether the elements in these sets were already present or not in the inspiring set. Ritchie's criteria have been used in subsequent evaluations of creative systems output <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b20">21,</ref><ref type="bibr" target="#b7">8]</ref>.</p><p>Pease et al. <ref type="bibr" target="#b18">[19]</ref> discuss relevant factors to evaluating systems in terms of creativity. The proposed framework mainly takes into account input provided, output produced and process employed. Each of these categories are detailed in depth, detailing their required measures. Before detailing the measurement methods, Pease et al. provide assumptions regarding creativity, also admitting their 'somewhat arbitrary' nature. The evaluation tests proposed deal with two main aspects: how close does the test predict human evaluation of creativity and how possible and practical it is to apply the test to a system. Overall, this work suggests that the very definition of creativity is subjective and that evaluating systems in a general way is problematic.</p><p>Colton et al. <ref type="bibr" target="#b4">[5]</ref> propose an extension of Ritchie's criteria <ref type="bibr" target="#b23">[24]</ref> that attempts to determine the impact of the input data on the creative artifact produced by a system. This more agnostic approach attempts to obtain an objective measure by comparing the output of the system to the inspirational material used as input. This investigation attempts to discriminate systems that overfit or shuffle input data (fine-tuning) instead of producing genuine novel artifacts. Among other conclusions, the authors state that comparing creative systems might not be viable, suggesting their criteria to be used as guidelines for program construction rather than post-hoc evaluation.</p><p>The creative tripod framework, proposed by Colton <ref type="bibr" target="#b2">[3]</ref>, is built around the premise that a creative system must demonstrate skill, imagination and appreciation. These qualities are not required to be possessed by the system, but rather to be perceived as possessed by the system. This is an important remark by Colton to avoid debates around the definition of creativity. The framework also includes the programmer, the system and the consumer, however Colton is only interested in the program's behavior.</p><p>Pease and Colton <ref type="bibr" target="#b17">[18]</ref> propose an alternative to the Turing Test to assess computational systems' creativity, the FACE (Frame, Aesthetic, Concept, Expression of concept) and IDEA (Iterative Development Execution Appreciation) model. The model includes creative acts and audiences, with relevant measures such as popularity, appeal, provocation, opinion, subversion and shock. Putting the focus on the reaction produced by the creative artifact, this model attempts to avoid the shortcomings of the Turing Test by going further than merely assessing the capacity of a creative system to imitate human behavior. By including the audience into the model, this approach acknowledges the highly subjective nature of creativity evaluation.</p><p>SPECS <ref type="bibr" target="#b8">[9]</ref>, introduced by Jordanous as "a standardised and systematic methodology for evaluating computational creativity", represents a substantial effort to provide a standard for evaluating the creativity of a system in the field of computational creativity and address the multi-faceted and subjective nature of creativity. Its flexible nature allows SPECS to adapt to the demands of the researchers' field, applying the required demands and standards. The methodology informs researchers of their system's strength and weaknesses, providing useful feedback for achieving creative results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Evaluation of Automatically Generated Narrative</head><p>Automatic generation of narratives has been a long-standing goal of Artificial Intelligence since its very beginning. There are a number of systems described in the literature, but the evaluation of these systems -be it its output, its creative process or whatever other aspect -is seldom found. This is most likely due to the fact that the average quality or variety of the generated stories is not really comparable to those written by most humans, not necessarily professional writers.</p><p>The Mexica system <ref type="bibr" target="#b22">[23]</ref> includes procedures for the dynamic assessment of the novelty of a story in progress with respect to previously known stories. Novelty is considered in terms of how the stories differ in terms of the actions they include and their frequency of appearance.</p><p>In Pérez et al <ref type="bibr" target="#b21">[22]</ref> three different characteristics are considered as relevant for measuring story novelty: sequence of actions, structure of the story, and use of characters and actions.</p><p>Peinado &amp; Gervás <ref type="bibr" target="#b19">[20]</ref> carried out an empirical study of how generated stories were perceived by a set of human volunteer evaluators. Human judges blindly compared one of the generated basic stories to two alternatives: one rendered directly from a stored fabula of the knowledge base and another randomly generated. Values were collected for: linguistic quality (how well is the text written), coherence (how well is the sequence of events linked), interest (how interesting is the topic of the story for the reader) and originality (how different is the story from others).</p><p>León &amp; Gervás <ref type="bibr" target="#b10">[11]</ref> propose a model, intended as a tool to drive automatic story generation, of how quality is evaluated in stories. This paper proposes a computational model for story evaluation in which an evaluation function receives stories and outputs a value as the rating for that story. The value for this function is computed from values assigned to: accumulation of contributions from individual events depending on the meaning of the event -aspects such as whether the reader wants to continue reading the story, or how much danger or love the reader perceives in the story -, appearance of patterns or relationships between the events of a story -aspects such as causality, humour or relative chronology -and inference -which captures the ability to interpret stories by adding material to explain what they are told even if it is not explicitly present in the story. The evaluation function has been implemented as a rule based system.</p><p>Ware, Young et. al. <ref type="bibr" target="#b26">[27]</ref> propose a formal model for narrative conflict with seven dimensions from various narratological sources meant to aid in distinguishing one conflict from another: participant, subject, duration, balance, directness, intensity and resolution. Their experimental results <ref type="bibr" target="#b27">[28]</ref> suggest the model predicts these seven dimensions of narrative conflict similarly to human criteria. Their good results predicting human-perceived narrative conflict suggest a similar approach may be viable for measures related to creativity.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Evaluating Automatic Ideation</head><p>Original ideation is central to any creative process. Coming up with innovative ideas that potentially trigger the creation of new material is fundamental to human creativity. It is not uncommon to focus creative processes on the identification of a single, valuable idea that unlocks new paths leading to finished artifacts. Although human creative teams usually rely on pure ideation to foster creativity, there have only been a few small, ad-hoc studies of how to automate ideation until recent times. Section 3.1 describes an effort to provide a system able to produce novel ideas.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">The What-If Machine</head><p>Llano et al. have recently proposed an automatic ideation system <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b13">14,</ref><ref type="bibr" target="#b11">12]</ref>. This computational system is designed to produce relatively valuable and novel ideas autonomously. This system, the What-If Machine<ref type="foot" target="#foot_0">1</ref> , includes a module for analysing the ideas and generating narrative metrics, and a module for computing a predictive machine learning model. This model is trained against collected human evaluations of what-ifs, and is intended to learn a robust function from narrative metrics to perceived overall quality. Two main hypotheses guide the design of the What-if Machine and the presented research:</p><p>1. There is a strong correlation between the perceived overall quality and the perceived narrative potential, in the sense that if the audience perceives high narrative potential, it will also perceive a high overall quality. A subset of the What-If Machine (modules 1, 2 and 3) was used to generate the material for the study, which is described in detail in Section 4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Study</head><p>A pilot study was performed to determine the feasibility of predicting the perceived quality and narrative potential in the artifacts created by a computable creative system. Both magnitudes have been introduced in the previous section, and in order to avoid influencing our subjects, no definition for them is provided in the questionnaires (as seen in Fig. <ref type="figure" target="#fig_0">1</ref>). This naive approach is a result of our focus on the model and its capability to predict human assessment instead of introducing our own views or definitions. The study was conducted to obtain the human rating of perceived quality and narrative potential.</p><p>Using both measures, a machine learning process will search for correlations between some metrics (detailed in the next section) and the perceived quality and perceived narrative potential. This should allow us to determine what measures are relevant to predict human-perceived quality and narrative potential to produce what-ifs that present both qualities to human observers.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Metrics</head><p>Since we have no certainty about what metrics extracted from each what-if's mini-narrative may impact over the perceived quality and narrative potential, we focused on generating the maximum amount of computable features. The impact of these features on the perceived quality and narrative potential may be obtained with machine learning techniques (we refer to these features as metrics). This approach is similar to the one used by Nowak for image classification <ref type="bibr" target="#b16">[17]</ref> that generates a high number of arbitrary features from each image.</p><p>A mini-narrative is a structure that contains a set of narrative points linked to schemas like setting or resolution. Each narrative point is a set of narrative statements that provide information about characters or events through predicates (e.g. dog is old or dog learns to play a piano). Narrative statements may be related to one another (caused by or inferred by another statement).</p><p>The next list includes the set of implemented features along with their description:</p><p>-Length: mini-narrative narrative points amount.</p><p>-SettingQuality: Amount of schemas divided by 3.</p><p>-ExplicitFact: the amount of narrative statements in the mini-narrative.</p><p>-RatioCharacters: the character/statement ratio.</p><p>-Originality: hits returned by the full text of the mini-narrative in the Bing search engine. -OriginalityAccurate: hits returned by the exact full text of the mininarrative in the Bing search engine. -Divergence: average hits returned by the mini-narrative statements in the Bing search engine. -DivergenceMinimum: minimum hits returned by the mini-narrative statements in the Bing search engine. -Evolution: amount of learnTo predicates found in the mini-narrative.</p><p>-Handicap: amount of negated capableOf predicates found in the mininarrative. -InterestingLife: amount of negated doesFor predicates found in the mininarrative. -TotalStoriesGenerated: amount of stories generated by the story generator from the current mini-narrative. -StoryCharacters: average number of characters in the generated stories.</p><p>-Names: StanfordNLP <ref type="bibr" target="#b15">[16]</ref> queries for the what-if's names.</p><p>-NamesRatio: Names/ExplicitFact ratio.</p><p>-Valence: Sum per statement, each statement codified as +1 if a fact is positive, -1 if negative and 0 otherwise). -ValenceAverage: Valence/ExplicitFact ratio.</p><p>-JointWordsProbability: joint probability average for each set of words using ngrams. For this metric we use the Project Oxford<ref type="foot" target="#foot_1">2</ref> services. -JointWordsProbabilityMinimum: the minimum joint probability for the set of words using ngrams from Project Oxford. -RealityDistortionRatio: events in the mini-narrative that negate a fact from the knowledge base are considered a reality distortion. This metric provides the reality distortion amount/ExplicitFact ratio.</p><p>-FictionalAdditionsRatio: any event in the mini-narrative that is missing from the knowledge base is considered a fictional addition. This metric provides the fictional addition amount/ExplicitFact ratio. -FictionalRatio: reality distortion amount plus fictional addition amount/ExplicitFact. -ResolutionTriggerRatio: resolution events solve conflicts from the mininarrative. Provides the resolution event amount/ExplicitFact ratio. -MainCharacterEventsRatio: protagonist statements are statements in which this actor plays any role. This metric provides the protagonist statement amount/ExplicitFact ratio.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Methodology</head><p>A set of 890 what-ifs were generated by the What-If Machine. All of their source mini-narratives were processed by the metric generation system. A total of 15 different questionnaires were created, each including 10 what-ifs rendered as text from the original set of 890. 150 what-ifs were included in the evaluation set. 101 volunteers received a link that randomly redirects to one of the 15 possible questionnaires through email. Given the simplicity of the questions, Google Forms was our platform of choice. The platform was robust and stable and all of the answers were successfully stored in a Google Sheet document automatically. There was no active supervision for each subject given the remote nature and limitations of the Google Forms platform.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Questionnaire</head><p>The questionnaire informed subjects about their participation in a study related to computer-generated content (Figure <ref type="figure" target="#fig_0">1</ref>). Some demographic information was queried (age, gender and English level) and then they were asked to evaluate the overall quality (on a 0-5 Likert scale) of each what-ifs plus its narrative potential (yes/no binary answer). A text box accepting any comment was also provided in order to gather additional qualitative information.</p><p>You are about to evaluate some of the preliminary results of the "WHIM: The What-If Machine" research project from the European Union. The overall objective of the What-If Machine is to automatically generate fictional ideas with cultural value. You will be presented a number of what-if style ideas and we kindly ask you to rate them according to the following features:</p><p>-Overall quality: from 0 (no quality) to 5 (superb quality). -Narrative potential (yes/no). -Any observation you can provide.</p><p>Completing the questionnaire should not take more than 10 minutes. We really appreciate your contribution to the project. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">Results</head><p>101 subjects participated in the study. Statistical analysis of the results revealed no significant differences between evaluators in terms of English level, age or gender. For instance, the quality (Q) for gender yielded µ(Q) male = 2.66, σ(Q) male = 0.75; µ(Q) f emale = 2.69, σ(Q) f emale = 0.89. The corresponding results for English and age are comparable.</p><p>Questionnaires provided 1,007 Quality and 1,004 Narrative Potential rankings for the 150 What-Ifs used. What-Ifs were ranked between 1 and 27 times. For the Narrative Potential (P ) measurements, we mapped "Yes" to +1, "Not sure" to 0, and "No" to -1. Overall measures resulted in µ(Q) = 2, 4 and σ(Q) = 1, 3 for Quality and µ(P ) = −0, 05 and σ(P ) = 0, 89 for Narrative Potential. Individual What-Ifs aggregated ranking values were used for calculating:</p><p>-Pairwise correlations between perceived Quality and perceived Narrative Potential, perceived Quality or perceived Narrative Potential and the metrics, and between individual metrics. -Global measure of attribute importance for these metrics in predictive modeling of the average perceived Quality or perceived Narrative Potential.</p><p>Pairwise correlations Metrics that provided the same values for all What-Ifs in the dataset were discarded. Correlation coefficients were calculated with the Pearson Product-Moment. There is a strong positive correlation between Quality and Narrative Potential averages (0.83) and medians (0.758). As seen in table <ref type="table" target="#tab_1">1</ref>, both measures correlate positively with some metrics, such as MainCharac-terEventsRatio and RatioCharacters and correlate negatively with others, such as ExplicitFact and Length. Importance for Predictive Modeling In order to determine the importance of each metric in predicting perceived Quality and Narrative Potential we used the Relief measure <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b25">26]</ref>, which is a method commonly used for feature selection in machine learning. This measure does not assume independence among the metrics, but takes their possible interdependence into account. The more the Relief scores are positive, the more a metric contributes to prediction of a target value (in our case, the value of average Quality or the average Potential). The ones that scored close to zero or negative are irrelevant and those with negative values have even a negative impact.</p><p>According to the results in Table <ref type="table" target="#tab_2">2</ref> it seems that most of the metrics have no use in predictive models of average Quality. For the average Narrative Potential, however, most of the metrics seem to be slightly informative . According to Relief ranks for the metrics results, usefulness of the metrics for average Quality is to some extent inversely proportional to their usefulness for the average Narrative Potential. The absolute values of the Relief scores depend on the characteristics of data and the parameters of the assessment, which makes it difficult to use absolute thresholds for judgements on the relevance of features. However, a strong correlation among the Quality and Narrative Potential values and a mismatch of the Relief scores of metrics for these two targets provide an indication that also the contributions of the positively scored metrics are likely to be too low to be considered relevant. The results previously presented evidence that there is a strong correlation between narrative potential and perceived overall quality of a what-if, which indicates that focusing on narrative plausibility as one of the main factors of quality can lead to better results. Moreover, some of the metrics are weakly correlated to narrative potential. However, these results are still inconclusive, and there is a number of aspects worth mentioning for their influence on the results. Automatically generating stories and computing useful values for metrics is heavily dependent on the available knowledge. The outcome of the system is constrained by the use of ConceptNet. The amount of relations that can be safely used in ConceptNet is small and the richness and depth of the chains of properties is limited regarding to its use as a source for narrative processing. This makes it necessary to address knowledge management from a different perspective. The WHIM project currently includes a whole module for providing robust knowledge to the rest of the modules, and the impact of the application of this subsystem on the creation and evaluation of what-if ideas will be reported once the results are ready.</p><p>The generation process (for the what-ifs, the stories and the metrics) strongly influences the overall outcome. Many design decisions have been taken in order to provide a working, implemented prototype able to generate actual what-ifs, and these decisions set the kind of what-ifs generated, the complexity of the stories and many other aspects. The provided results are then the outcome of a specific implementation which does not claim any generality. However, the approach itself (namely the generation-metric computation-evaluation process) is presented as a generally applicable method for producing novel what-if ideas.</p><p>The used metrics for labeling narrative properties do not cover all computable features. There is a large number of aspects that can be extracted from a whatif, and the narrative-based feature extraction module of the What-If Machine does not currently provide coverage for all of them. This is considered to be not strictly relevant with regard to the methodology and scope of the study. To test the second hypothesis (the existence of a correlation between a certain set of metrics and the overall quality and plausibility), the metrics must be improved. For that purpose, the presented study gives valuable insight on which direction to go next.</p><p>The weak correlation between our metrics and the quality perceived by humans suggested that considering more sophisticated metrics was necessary. Some of them were considered:</p><p>1. Humanization: An approximation of how much human-like the main character is, assuming that fictional scenarios use characters that, while behaving like humans, can be non-human. 2. Empathy: How much empathy will a reader feel about the characters. 3. Tragedy: The amount of tragedy in the story. 4. Reality: How real and current the context is. An approximation of fictionally in terms of context.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">TimeSpan:</head><p>The time span the story covers. It could be minutes, days or years.</p><p>Modelling and implementing these metrics proved to be beyond technical capabilities because it required complex, rich knowledge bases <ref type="bibr" target="#b0">(1,</ref><ref type="bibr" target="#b3">4)</ref>, reliable text understanding systems <ref type="bibr" target="#b4">(5)</ref>, sophisticated emotional models <ref type="bibr" target="#b1">(2)</ref> or formal versions of narratological models (3). All of these resources are currently not available.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusions</head><p>The current paper has presented a pilot study trying to gain insight on two hypotheses, namely that (1) human evaluation on overall quality of what-if ideas correlates to the perception of narrative potential and that (2) there is a set of computable metrics that also correlate to this perception. The study has evidenced that there is a strong correlation between quality and narrative potential for humans <ref type="bibr" target="#b0">(1)</ref>, but failed to prove such a strong correlation between the current metrics and the human ratings. These results have been analysed and discussed in terms of the limited potential of the current implementation of both the fictional ideation procedure and the method employed to evaluate it. Actual implementations lack the required complexity to approximate evaluations with a relatively acceptable level of accuracy, mainly due to the limited technical capabilities of current computational solutions.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Information presented to the user in the evaluation questionnaire.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>The overall quality is defined in terms of the analyzed response from humans (i.e. no specific model beyond what humans say about quality is assumed), and the narrative potential is assumed to be directly proportional to the amount and quality of the stories a certain what-if can trigger or inspire. 2. There is a set of computable metrics whose values correlate (directly or indirectly) with the overall quality and the narrative potential.The What-If Machine is, to the best of our knowledge, the only attempt to implement a computer system able to produce novel what-if ideas. The What-If Machine is a distributed computer system in which several modules collaborate in order to output rendered what-ifs. Five modules compose the system:1. The ideation module produces, using a knowledge base, what-if ideas formalized as mini-narratives. 2. The mini-narratives are fed into the narrative-based metric generation, which generates values for a set of metrics which hypothetically have a correlation with human perception of quality. These metrics are based on narrative properties of the what-ifs. 3. The mini-narratives, now enriched with its corresponding metrics, are sent to a crowd-sourcing evaluation module, which applies machine learning to create and refine models for predicting overall quality against human ratings. 4. The world view creation, providing knowledge for what-if generation, story creation and metric computation. 5. The finished, filtered what-ifs are finally passed to a rendering module, which creates artifacts from the final what-ifs (stories, texts or images, for instance).</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1 .</head><label>1</label><figDesc>The correlation coefficient between average/median Quality (Q) or Narrative Potential (P) labels and the metrics. The values are sorted by correlation coefficient values of the average Quality.</figDesc><table><row><cell></cell><cell cols="4">Avg Q Mdn Q Avg P Mdn P</cell></row><row><cell cols="5">MainCharEventsRatio 0.371 0.346 0.379 0.329</cell></row><row><cell>RatioCharacters</cell><cell cols="4">0.354 0.296 0.368 0.307</cell></row><row><cell cols="5">ResolutionTriggerRatio 0.342 0.303 0.305 0.261</cell></row><row><cell cols="5">TotalStoriesGenerated 0.312 0.250 0.321 0.264</cell></row><row><cell>JointWordsProbMin</cell><cell cols="4">0.308 0.289 0.367 0.314</cell></row><row><cell>. . .</cell><cell>. . .</cell><cell>. . .</cell><cell>. . .</cell><cell>. . .</cell></row><row><cell>ValenceAverage</cell><cell cols="4">-0.219 -0.188 -0.296 -0.249</cell></row><row><cell>ValenceSum</cell><cell cols="4">-0.258 -0.234 -0.323 -0.276</cell></row><row><cell>StoryCharacters</cell><cell cols="4">-0.283 -0.269 -0.327 -0.285</cell></row><row><cell>ExplicitFact</cell><cell cols="4">-0.379 -0.336 -0.406 -0.345</cell></row><row><cell>Length</cell><cell cols="4">-0.379 -0.336 -0.406 -0.345</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2 .</head><label>2</label><figDesc>Relief measure results for average Quality (Relief Avg Q) and average Narrative Potential (Relief Avg P ). Rows sorted by Relief Avg Q. The best three results are in bold and the worst three are in italics.</figDesc><table><row><cell>Metric</cell><cell cols="2">Relief Avg Q Relief Avg P</cell></row><row><cell>Handicap</cell><cell>0.027</cell><cell>-0.009</cell></row><row><cell>MainCharacterEventsRatio</cell><cell>0.007</cell><cell>0.004</cell></row><row><cell>NamesRatio</cell><cell>0.001</cell><cell>0.006</cell></row><row><cell>DivergenceMinimum</cell><cell>0.000</cell><cell>0.000</cell></row><row><cell>JointWordsProbabilityMinimum</cell><cell>0.000</cell><cell>0.000</cell></row><row><cell>Divergence</cell><cell>0.000</cell><cell>0.000</cell></row><row><cell>Originality</cell><cell>-0.006</cell><cell>0.013</cell></row><row><cell>. . .</cell><cell>. . .</cell><cell>. . .</cell></row><row><cell>FictionalAdditionsRatio</cell><cell>-0.075</cell><cell>0.028</cell></row><row><cell>InterestingLife</cell><cell>-0.116</cell><cell>0.045</cell></row><row><cell>TotalStoriesGenerated</cell><cell>-0.116</cell><cell>0.045</cell></row><row><cell>OriginalityAccurate</cell><cell>-0.126</cell><cell>0.024</cell></row><row><cell>FictionalRatio</cell><cell>-0.142</cell><cell>0.039</cell></row><row><cell>RatioCharacters</cell><cell>-0.142</cell><cell>0.039</cell></row><row><cell>SettingQuality</cell><cell>-0.147</cell><cell>0.024</cell></row><row><cell>Names</cell><cell>-0.147</cell><cell>0.024</cell></row><row><cell>ValenceSum</cell><cell>-0.174</cell><cell>0.033</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">The What-if Machine: http://www.whim-project.eu/.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://www.projectoxford.ai/</note>
		</body>
		<back>

			<div type="funding">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Supported by the project WHIM (611560) and PROSECCO (600653), funded by the European Commission, Framework Program 7, the ICT theme, and the Future Emerging Technologies FET program.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Computational Models of Creativity</title>
		<author>
			<persName><forename type="first">M</forename><surname>Boden</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Handbook of Creativity</title>
				<imprint>
			<date type="published" when="1999">1999</date>
			<biblScope unit="page" from="351" to="373" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Creative Mind: Myths and Mechanisms</title>
		<author>
			<persName><forename type="first">M</forename><surname>Boden</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2003">2003</date>
			<publisher>Routledge</publisher>
			<biblScope unit="page">10001</biblScope>
			<pubPlace>New York, NY</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Creativity Versus the Perception of Creativity in Computational Systems</title>
		<author>
			<persName><forename type="first">S</forename><surname>Colton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AAAI Spring Symposium on Creative Systems</title>
				<meeting>the AAAI Spring Symposium on Creative Systems<address><addrLine>Colton</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2002">2002. 2008</date>
			<biblScope unit="page" from="14" to="20" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">The painting fool: Stories from building an automated painter</title>
		<author>
			<persName><forename type="first">S</forename><surname>Colton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computers and Creativity</title>
		<imprint>
			<biblScope unit="volume">9783642317</biblScope>
			<biblScope unit="page" from="3" to="38" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">The effect of input knowledge on creativity</title>
		<author>
			<persName><forename type="first">S</forename><surname>Colton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pease</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Ritchie</surname></persName>
		</author>
		<ptr target="http://www.inf.ed.ac.uk/publications/online/0055.pdf" />
		<imprint>
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
	<note type="report_type">Technical Reports of the Navy Center for</note>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Computational creativity: The final frontier? ECAI</title>
		<author>
			<persName><forename type="first">S</forename><surname>Colton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Wiggins</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Linguistic creativity at different levels of decision in sentence production</title>
		<author>
			<persName><forename type="first">P</forename><surname>Gervás</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AISB 02 Symposium on AI and Creativity in Arts and Science</title>
		<title level="s">Imperial College</title>
		<meeting>the AISB 02 Symposium on AI and Creativity in Arts and Science</meeting>
		<imprint>
			<date type="published" when="2002-04-05">3rd-5th April 2002. 2002</date>
			<biblScope unit="page" from="79" to="88" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Investigating artificial creativity by generating melodies, using connectionist knowledge representation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Haenen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Rauchas</surname></persName>
		</author>
		<ptr target="http://ccg.doc.gold.ac.uk/events/ecai06/proceedings/Haenen.pdf" />
	</analytic>
	<monogr>
		<title level="m">The Third Joint Workshop on Computational Creativity</title>
				<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">A Standardised Procedure for Evaluating Creative Systems: Computational Creativity Evaluation Based on What it is to be Creative</title>
		<author>
			<persName><forename type="first">A</forename><surname>Jordanous</surname></persName>
		</author>
		<ptr target="http://dblp.uni-trier.de/db/journals/cogcom/cogcom4.html#Jordanous12" />
	</analytic>
	<monogr>
		<title level="j">Cognitive Computation</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="246" to="279" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">A practical approach to feature selection</title>
		<author>
			<persName><forename type="first">K</forename><surname>Kira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Rendell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ninth international workshop on Machine learning</title>
				<meeting>the ninth international workshop on Machine learning</meeting>
		<imprint>
			<date type="published" when="1992">1992</date>
			<biblScope unit="page" from="249" to="256" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">The Role of Evaluation-Driven rejection in the Successful Exploration of a Conceptual Space of Stories</title>
		<author>
			<persName><forename type="first">C</forename><surname>León</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Gervás</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Minds and Machines</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="615" to="634" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Automated Fictional Ideation via Knowledge Base Manipulation</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Llano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Colton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Hepworth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gow</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Cognitive Computation</title>
		<imprint>
			<biblScope unit="page" from="1" to="22" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Towards the automatic generation of fictional ideas for games</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Llano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cook</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Guckelsberger</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Experimental AI in</title>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Automating fictional ideation using ConceptNet</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Llano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Hepworth</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proceedings</title>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>Machado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Martins</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Amaro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Abreu</surname></persName>
		</author>
		<title level="m">Beyond interactive evolution: Expressing intentions through fitness functions</title>
				<imprint>
			<publisher>Leonardo</publisher>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">The stanford corenlp natural language processing toolkit</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Surdeanu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bauer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Finkel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bethard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Mcclosky</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACL (System Demonstrations)</title>
				<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="55" to="60" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Sampling Strategies for Bag-of-Features Image Classification</title>
		<author>
			<persName><forename type="first">E</forename><surname>Nowak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Jurie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Triggs</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2006">2006</date>
			<publisher>Springer</publisher>
			<biblScope unit="page" from="490" to="503" />
			<pubPlace>Berlin Heidelberg</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">On impact and evaluation in computational creativity: A discussion of the Turing test and an alternative proposal</title>
		<author>
			<persName><forename type="first">A</forename><surname>Pease</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Colton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computing and Philosophy</title>
		<imprint>
			<biblScope unit="page" from="15" to="22" />
			<date type="published" when="2011">2011. 2011</date>
		</imprint>
	</monogr>
	<note>AISB</note>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Evaluating machine creativity</title>
		<author>
			<persName><forename type="first">A</forename><surname>Pease</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Winterstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Colton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Workshop on Creative Systems</title>
				<imprint>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page">4</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Evaluation of Automatic Generation of Basic Stories. New Generation Computing, Computational Paradigms and Computational Intelligence</title>
		<author>
			<persName><forename type="first">F</forename><surname>Peinado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Gervás</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Special issue: Computational Creativity</title>
		<imprint>
			<biblScope unit="volume">24</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="289" to="302" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">A Multiagent Text Generator with Simple Rhetorical Habilities</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">C</forename><surname>Pereira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Hervás</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Gervás</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Cardoso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the AAAI-06 Workshop on Computational Aesthetics: AI Approaches to Beauty and Happiness</title>
				<meeting>of the AAAI-06 Workshop on Computational Aesthetics: AI Approaches to Beauty and Happiness</meeting>
		<imprint>
			<publisher>AAAI Press</publisher>
			<date type="published" when="2006-07">July 2006. 2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">A system for evaluating novelty in computer generated narratives</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">Y</forename><surname>Pérez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Ortiz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Luna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Negrete</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Creativity</title>
		<imprint>
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<title level="m" type="main">MEXICA: A Computer Model of Creativity in Writing</title>
		<author>
			<persName><forename type="first">R</forename><surname>Pérez Y Pérez</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1999">1999</date>
		</imprint>
		<respStmt>
			<orgName>The University of Sussex</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Ph.D. thesis</note>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Assessing creativity</title>
		<author>
			<persName><forename type="first">G</forename><surname>Ritchie</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AISB Symposium on AI and Creativity in Arts and Science</title>
				<meeting>the AISB Symposium on AI and Creativity in Arts and Science<address><addrLine>York, UK</addrLine></address></meeting>
		<imprint>
			<biblScope unit="page" from="3" to="11" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Some Empirical Criteria for Attributing Creativity to a Computer Program</title>
		<author>
			<persName><forename type="first">G</forename><surname>Ritchie</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Minds &amp; Machines</title>
		<imprint>
			<biblScope unit="volume">17</biblScope>
			<biblScope unit="page" from="67" to="99" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">An adaptation of relief for attribute estimation in regression</title>
		<author>
			<persName><forename type="first">M</forename><surname>Robnik-Šikonja</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Kononenko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Machine Learning: Proceedings of the Fourteenth International Conference (ICML</title>
				<imprint>
			<date type="published" when="1997">1997. 1997</date>
			<biblScope unit="page" from="296" to="304" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Validating a Plan-Based Model of Narrative Conflict</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">G</forename><surname>Ware</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">M</forename><surname>Young</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Conference on the Foundations of Digital Games</title>
				<meeting>the International Conference on the Foundations of Digital Games<address><addrLine>New York, New York, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM Press</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="220" to="227" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<title level="m" type="main">Four Quantitative Metrics Describing Narrative Conflict</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">G</forename><surname>Ware</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">M</forename><surname>Young</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Harrison</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">L</forename><surname>Roberts</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2012">2012</date>
			<publisher>Springer</publisher>
			<biblScope unit="page" from="18" to="29" />
			<pubPlace>Berlin Heidelberg</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">A preliminary framework for description, analysis and comparison of creative systems</title>
		<author>
			<persName><forename type="first">G</forename><surname>Wiggins</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Knowledge-Based Systems</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="issue">7</biblScope>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Searching for Computational Creativity. New Generation Computing</title>
		<author>
			<persName><forename type="first">G</forename><surname>Wiggins</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Computational Paradigms and Computational Intelligence</title>
				<imprint>
			<date type="published" when="2006">2006</date>
			<biblScope unit="volume">24</biblScope>
			<biblScope unit="page" from="209" to="222" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
