=Paper=
{{Paper
|id=Vol-2276/paper5
|storemode=property
|title=Bounding Ambiguity: Experiences with an Image Annotation System
|pdfUrl=https://ceur-ws.org/Vol-2276/paper5.pdf
|volume=Vol-2276
|authors=Margaret Warren,Pat Hayes
|dblpUrl=https://dblp.org/rec/conf/hcomp/WarrenH18
}}
==Bounding Ambiguity: Experiences with an Image Annotation System==
Bounding Ambiguity: Experiences with an Image Annotation System Margaret Warren1, Patrick Hayes2 1 Metadata Authoring Systems 2 Florida IHMC, Abstract. This paper reports on a web interface for creating and editing rich metadata descriptions for images using RDF triples. We discuss the roles of ambiguity, disagreement and subjectivity in knowledge formation arising from an ongoing experiment in which users create semantic annotation of images, and we discuss the ways these elements have influenced the design of the system. 1 Background ImageSnippets ( http://www.imagesnippets.com ) is a web interface designed for the construction of machine-computable image descriptions by lay users. Prompted by image cues, image annotators and/or subject matter experts can disambiguate entities from public corpora or define new terms through the interface. The subsequent captured knowledge is then stored in triple based graphs for inference, findability and reuse by semantically aware processes. The output formats conform to W3C linked data standards: (RDFa, JSON-LD) and utilize terms from the large and growing web- based concept datasets including DBpedia, YAGO and the Art & Architecture Thesaurus, without requiring users to be aware of this machinery. The system has a number of features for capturing the highly subjective things people might say about images, but it also allows them to disambiguate their meanings in a suitably precise way, thereby providing a process for formalizing intuitive and expert knowledge. In our research with the system, we have found disagreements about meanings to be a centrally important methodological tool to create more useful and intuitive conceptual distinctions and that ambiguity is unavoidable and often useful, and should be considered something to be controlled and bounded appropriately, rather than eliminated; and that shared subjectivity is more important than objective truth or correctness when using semantic markup for inference and retrieval. Our work originally began as an exploration of how to bring together precise machine-readable descriptions in Semantic Web notations such as RDF and OWL with the informal language of how people actually describe images, especially (but not exclusively) the language of artists and deliberately considered the artists themselves as the subject matter experts on their own work (Eskridge et al, 2006, Warren & Hayes 2007, 2010). But this early knowledge representation work also had a secondary goal: that we could use our work to design a system with a (hopefully) intuitive interface that would capture this highly subjective, personalized and informal knowledge in such a way that the resulting precise and disambiguated metadata could improve findability and then be distributed on the web with as much readability, interoperability and persistence as possible. So the ImageSnippets system was designed first and foremost as a research tool to observe how users or image annotators would interact with the system, but we also considered that the metadata created through this research could also be accessed by any semantically aware processes and search engines and further, that the data could be properly attributed with provenance and publishing authority. At the core of the ImageSnippets system is an intentionally small, lightweight ontology which provides a basic vocabulary of properties used to relate an image - the subject of the descriptions (or a region in the image) - to object values as entities which are then selected from public data sets and which, in other image annotation systems, would normally be referred to as keywords or tags. The annotators are given these properties as intuitive guidelines and encouraged to use terms from LIO first, but with some training, they can create new properties as well. The ability for users to create new properties is a highly useful way for the system to capture and analyze subtle distinctions and use these distinctions to build both new, domain specific ontologies and to extend LIO to allow annotators to talk, not just about images, but about the things that are seen in the images. LIO has evolved over many thousands of hours of use and testing, and the history of this evolution provides some suggestive lessons in the use of contradiction and ambiguity to discover intuitive concepts. The design of the main ImageSnippets interface by necessity had to be concerned with capturing highly subjective and ambiguous sentiments from the naturally expressed ways people describe images, but incorporated in the design was also the creation of the core terms which make up LIO and this process was, in itself an iterative exercise in using something we call bounded ambiguity. Outside of mathematics and some related pursuits, the meanings of words or symbols are always ultimately determined by the way that those words are actually used to pragmatically convey content, rather than by exact formalizable definitions. Even though Web protocols give IRIs an unprecedented degree of global exactness when used to retrieve content, this does not apply to their use as names in the Semantic Web, where their meanings are just as socially defined as words in natural languages. Widely cited calls to remove all ambiguity from IRIs (such as https://www.w3.org/wiki/GoodURIs ) are doomed to failure. Nevertheless, while ambiguity cannot be eliminated, it can be bounded. Published OWL ontologies and thoroughly documented public catalogs of IRI meanings such as DBPedia can both give IRIs a tighter. more exactly restricted meaning than is commonly found in natural-language words, and sometimes can adequately capture useful meaning distinctions which have not (yet) been adopted in normal English. Much of the art in designing useful Web data seems to involve locating the most useful bounds on ambiguity and we believe that bounded ambiguity has yielded some highly useful and encouraging search results. Naturalness of use, and plausibility of the resulting descriptions, have been central to the goal of our project, which aimed to take image description to a level impossible to reach by the use of simple ‘keywords’ or ‘tags’. Image tags or keywords cannot overcome the sometimes extreme lexical ambiguity of English words in isolation, and cannot record how the indicated idea is related to the image. It should be noted here that there are many existing annotation systems and crowdsourcing efforts that employ both complex schemas and keyword disambiguation as well as the ability to group keywords into more useful machine processable conventions; however the subject of our research has always been to push the edges of keyword meaning in relationship to the image. Therefore, ours is a very different kind of activity from conventional ontology engineering. It is mostly performed by people experienced in the use of the software but with no philosophical or technical background, and in some cases with a high- school level of education. It rarely strays into ‘upper-level’ decisions about how things are to be classified, beyond a very basic and shallow taxonomy of classes of things. It is driven by the immediate needs of describing things seen in images, and failure to say something is not an acceptable option, so the subject matter is open- ended. Naturalness is essential, but descriptions cannot go beyond the restricted grammatical patterns of RDF linked triples, so concepts must sometimes be invented to express complex ideas. When this is done, the results are subjected to a Darwinian process of selection: if a new idea – usually a new relation name – is found to be useful in other, subsequent, annotations, it is retained, and this re-use in itself provides a growing body of evidence illustrating its actual meaning. We believe that, in this way, a slowly growing corpus of useful concepts - mostly RDF properties - are being accumulated which seem to allow quite rich expressiveness within the confines of the simple RDF triple graph syntax. One important idea in our methodology has been that of the information recording point, where a user has a clear thought about an image and should be able to express this naturally, in an annotation which is detailed and structured enough to transmit that intended meaning to other users downstream. The information recording point is the moment that the annotator has something to say and is ready to say it. Our aim is to provide a tool that can be used at that point, without the annotator being required to see or think about anything beyond ideas expressed in English words and a readable rendering of their content. (The interface provides this as a grammatically primitive but readable ‘pidgin’ pseudo-English.) Coupled with this is an idea borrowed from software training usually in corporate cultures, that of ‘just-in-time learning’. In a way, it could be said that we are giving annotators ‘just enough’ structure and just enough guidance in the interface to choose from one of the core concepts at the very moment the annotator is deciding what they want to say. Most users of the system can be trained in the basic construction of triples using the core LIO concepts in just a few minutes and after they have gained sufficient experience in the basics, they can easily then be trained to extend the triples to allow even greater expressivity. The inherent ambiguity of the core LIO vocabulary seems to be part of the reason for its success. If we had imposed very ‘strict’ meanings on these relations, making finer-grained distinctions based on very exact meanings, it would of necessity have been several orders of magnitude larger, greatly increasing the cost of development but also the cost of use, since users would be required to make very precise distinctions at the point of use, with low tolerance for error. It is difficult to talk about the design of LIO without also discussing some of the interface design decisions, because the vocabulary was deliberately allowed to grow somewhat organically around a combination of user workflow and some key philosophical decisions as starting points. So we first created the editor to enable the triples to be written and almost immediately designed a complementary search and sort function to test the effectiveness of the properties. Additionally, as part of building the triples, users were provided with a look-up for finding and disambiguating the object values of the triples from public datasets. Choosing entities from sources such DBpedia, Yago, The Art & Architecture Thesaurus is also a highly subjective process which itself influences the choices made for properties. To date, the basic LIO ontology consists of eleven properties (including a small number used to classify images into categories, such as being a photograph or a pencil drawing, and placing them in collections) have been found to be sufficient, together with the concepts from DBPedia and other online corpora, to create rich and useful descriptions of a wide variety of images. 2 Building LIO Almost every aspect of the LIO ontology has its origin in a disagreement, either between ourselves or between us and some of our users. Sometimes these were resolved by careful disambiguation; but more often, by realizing that we were all using words differently, leading to the creation of a new concept name, whose true meaning was often decided by the patterns of its actual use. Some examples follow. 2.1 Images vs Works The root subject IRI of all our descriptions identifies the digital image shown to the user as a thumbnail, but exactly what this IRI is supposed to denote in our descriptions is not entirely obvious, and was heavily debated. For a digital photograph the root subject is simply the image itself, but for a photograph of a painting in a museum catalog it is natural to say that ‘the image’ refers to the painting, and in some cases –for example, a cropped image of a sculpture in a museum catalog – it may even be a physical object. In general, we use the terminology of a work, as in ‘work of art’ and the subject of our descriptions is always the work. But this introduces an ambiguity: is our subject IRI referring to the image itself or to a work seen in the image? At an early stage of the design of our interface we resolved this ambiguity by providing a carefully described set of alternatives, but this proved to be unworkable, as it required users to carefully make artificial-seeming distinctions, and to create awkward descriptions. If one delves into these distinctions deeply, one quickly descends into an intellectual quicksand. Who exactly is the creator of what you see in a photo of graffiti? Most people would say the photographer, unless it is something recognizable by an artist like Bansky. In our experience, the way an image is cropped seems to have some bearing on who might be said to be the creator. In some cases, it is the Contact Volume Editor that checks all the pdfs. In such cases, the authors are not involved in the checking phase. We simply assume that the subject of our descriptions is a work, without recording explicitly whether this means the image itself or something it illustrates. (Intuitive guidelines are that a cropped, sharp image of a single object shown in an idealized studio-lighting format is a picture of the work, while a simple photographic image of a landscape is the work These distinctions are not recorded explicitly, so the image IRI might refer to itself or to a work it illustrates. This is of course the classical philosophical confusion of use versus mention. This is an ambiguity that we have decided to embrace rather than disambiguate, which has has not turned out to be a problem in practice. Users seem to follow the required discipline intuitively without needing to be trained or corrected. We think that there are probably good reasons why some famous philosophical confusions are widespread in human affairs, even though they have been noted by scholars for thousands of years. The time or effort they save, compared to more accurate but more pedantic descriptions, is enormous, while the ambiguity they embody is easily resolved in practice. In our case, the use of multiple sources of information in descriptions, including the use of a ‘creator’ field, or the Dublin Core dc:creator property, or both, seems adequate. 2.2 Depiction vs Showing The most basic LIO property is depicts. We debated reusing the depiction properties used in FOAF (Brickley & Miller 2010). But in spite of the widely recognized best- practice utility of re-use, we decided to introduce and use lio:depicts as a new property on the grounds that almost all FOAF uses will be images of people, whereas LIO is intended to apply to a much wider range of images. While nothing in the published FOAF vocabulary description requires this restriction to people, it seems clear that in actual use, the domain of foaf:depiction is people rather than owl:Thing; and we believe that use determines actual meaning. Real images often show many entities that are incidental to the main subject (if there is one). Are they all depicted? A disagreement about pictures which ‘accidentally’ reveal a celebrity, while being aimed at other subjects, led us to introduce a distinction between depicting and merely showing. We noted that other systems exist which introduce a concept of weighting how central each thing depicted is in a scene, so we introduced a property lio:shows to be used for noting things in images which are in a sense incidental or not the ‘main subject’. The distinction between lio:depicts and lio:shows is vague and underdetermined, but it seems natural, and there is a high degree of agreement in the use of these relations by different users in all the cases we have tested. Also, people use these terms with very little training and they use them rapidly, in much the same amount of time as it would take to create bare keyword tags. Figure 1 illustrates some examples of how the distinction works. Subsequently, all of the other LIO properties seem to be split away from ‘depiction’ where a visual ambiguity needs to be resolved. Figure 1: Depicts, Shows, UsesPictorially 2.3 Location vs. Setting Early uses of a draft of the ontology revealed a form of use that we had not anticipated. It can be illustrated by the example of a photograph of some newly prepared food on a table. The photographer had added the description lio:shows dbpedia:Pensacola,_Florida on the grounds, when asked, that the photo had been taken in Pensacola. Clearly he was not using our property in the way we had intended. When we explained his error, his response was to say that he felt the ‘setting’ of the photograph was important. It is worth noting that the issue of location in image annotation is in and of itself highly ambiguous and probably hotly contested in many circles. Early on, we had added a static location field (re-using an Adobe XMP location that can often be found embedded in image data) to attempt to capture at least one version of location initially; but there are at least 4 notions of location: the location of the camera or input device that captured the digital image, the location where a work might have been created, the location of a scene in the work and the location of where the work may currently be located. In all of these cases, attempts have been made to classify the distinction between scene location versus image location — and even the names of the location fields themselves become ambiguous and tend to change over time. So we created a new property, lio:hasSetting, to relate an image to the broader place where it was created while still providing in other places in the interface, a way for annotators to deal with camera or scene locations. This judicious use of bounded ambiguity was rapidly overtaken by our users, who quickly extended its use further than we had imagined. They included events (the 1939 NY World’s Fair), periods of time (the 17th century, when a portrait was painted) and even such things as ’kitchen’, ‘wedding’, and even a person – for the setting of a tattoo. While these uses were not what we had in mind, they seem to work, and indeed we now find them natural. We therefore decided to simply allow all such uses, and rather than regard this as a situation where the range of the property is ambiguous, think rather that the users of LIO are, by their use of this property on thousands of images, creating a category, which might be called the class of Settings. We have no way to exactly characterize this class in a formal ontology, and it does not occur in any published ePedia, but we can say that it has a large overlap with what one might call spatiotemporal envelopes, and includes a wide range of things, such as twilight and dusk, which are hard to classify. It need not be shown or depicted in the image itself (although it some cases it can be, as in a room or a park or someplace one might be ‘in’ when they make the image) And this is enough to make it a useful and intuitive element of the LIO vocabulary. The key point is that users find it natural to use, and they use it consistently; which we take as evidence that it corresponds to a widely shared concept. 2.4 Foreground and Two Backgrounds The notions of foreground and background are of course familiar, and we provided relations hasInForeground and hasInBackground to allow users to talk of objects in the foreground or background. But we quickly discovered that there are two additional notions of ‘background’ that could be utilized with a fair amount of ease once an annotator had learned they could use the terms. First, there is the flat 2-D background field of a rendered image versus the 3-D background of a depicted scene, the stuff ‘behind’ the main object depicted in an image. We needed the former in order to state precise search criteria for works of art (a painting by Keith Haring with a white background) when most ‘backgrounds’ were distant areas in photographic images. In this case, a real (but unforeseen) ambiguity had to be bounded by providing distinct IRIs for two subtly different but visually important senses of an English word; even though the meanings of our current relations lio:hasPictoriaBackground, lio:hasDepictedBackground are (of course) themselves still somewhat ambiguous, we find that when used, they are used consistently, perhaps because the very presence of the distinction in the small vocabulary draws user’s attention to fact that they can make this distinction explicit. HasPictorialBackground can refer to any kind of pattern or texture (such as ‘checkerboard’ or ‘swirls’) while hasDepictedBackground might be something like ‘sand’ behind a crab. These distinctions are valuable if perhaps one is searching for a pelican completely surrounded by water as opposed to sky (See Figure 2.) Figure 2. Pelicans with Water as Depicted Background 2.5 Looking Like Early in the LIO project we had a photograph of clouds and sunlight that some people said, depicted an ‘angel’. This gave rise to a debate. Did it depict an angel? Some said, obviously yes, since this was the whole point of the photograph. Some said, obviously no, since there are is no actual angel to be depicted, only clouds, sky and sunlight. Some said angels did not exist and others emphatically said that medieval religious art was full of depictions of angels. We resolved this contradiction by introducing the notion of one thing looking like another. The image depicts clouds, but lio:looksLike an angel. Fortunately, the argument about the existence of angels was resolved by http://dbpedia.org/resource/Angel. What seemed at first to be a simple conceptual hack for a limited class of ‘illusions’ turned out to be immediately useful in a large variety of descriptive situations, and lio:looksLike is now easily used by annotators. When reading plain- text descriptions of images, we often see phrases “looks like a bird” or “looks like a face”. One might call this the visual equivalent of metaphoric language. 2.6 Conveys Early in the project, marking up a blue period Picasso, someone wanted to say that it shows sadness. This produced immediate objections, such as where exactly was this shown in the image? Again, the resulting disagreement/discussion led to a new conceptual distinction, and the addition of lio:conveys to the vocabulary, intended to refer to emotions and moods that images do not in any sense show or depict. But just as with lio:hasSetting, lio:conveys gets used in ways we had not initially anticipated and yet make perfect sense. So, for example an image can lio:conveys emptiness, friendship or even a ‘party’ (as in a ‘party atmosphere’) and so we realized that annotators might also use it to refer to an idea. So far there seems to be no reason to further delineate this concept. By using the search interface in the system, we have seen that the use of the ‘conveys’ property is basically never used to ‘convey’ something ‘depicted’. 2.7 Artistic considerations The final two properties added to LIO were: hasArtisticElement and lio:usesPictorially. Throughout the design of the interface and the design of the ontology, we were using teams of annotators who would annotate small groups of images and then run multiple types of searches over the freshly created data so that we could examine the usage of both the vocabulary and the similarity in the object terms chosen from (mostly) DBpedia and Yago. During this testing phase, we began to observe patterns in the usage of the terms and realized that the teams were naturally aligning themselves to their own conventions on edge cases. Take, for example a photo of a guitar in which the guitar takes up a large area of the image, but is still clearly not the only thing in the image. Does the image ‘depict’ a guitar or ‘show’ a guitar? In the beginning we might have seen annotators choose ‘depicts’ until they ran a search and observed that there was clearly a visible difference in the search results and thus, they aligned themselves using the images themselves as guides. While these testing sessions were occurring, we additionally discovered two other properties that were eagerly embraced, one was for the notion of an Artistic Element. We began searching for terms like: spirals, movement, shadow and blur; words the annotators had been using, and realized that a category of general concepts related to classical art theory design elements: line, form, shape and color attributes could be useful. Once the property existed, annotators quickly began using it to describe things like ‘vanishing points’ and ‘symmetry’. The last property added was one called: lio:UsesPictorially. In practical use, while annotators were “triple-tagging” images and then testing their use of the properties, another distinction between objects was observed in the search results that could not be handled by any of the other properties. The distinction was a subtle difference in the way one would say they ‘normally’ see something real in the world depicted and then those items that seem to be visually represented in an unusual way – perhaps an artistic sense, or maybe just photographed in an unconventional way. This distinction most often includes artistic renderings of objects in the form of paintings or illustrations, but it often goes beyond that into abstract photos of everyday objects or unusual angles. Almost every new annotator learns this property very quickly and though there can be ambiguous edge cases here as well, this property has seemed to bear a great deal of utility in search results. Figure 1 illustrates the distinctions between depicts, shows and usesPictorially in images of guitars 3. Describing what is depicted While working on the design of LIO, we were also working on the design of the system itself. The core of the ImageSnippets interface is an editor window called the ‘triple editor’ in which most of the user activity occurs. Over the years we have experimented with several design options and consider this part of the system a work in progress. Our goal has been to make the process of finding, and sometimes of coining, concept IRIs as quick and as painless as possible and at the moment, we have multiple ways of accomplishing the creation of the triples including ways to select multiple images at one time to annotate all of them with the same triple(s) simultaneously as well as entity extraction from free text input provided by subject-matter experts (SMEs) and direct image-to-word suggestions made by Clarifai and CloudSight, as shown in figure 3. Figure 3: Clarifai (AI generated keyword suggestions) and the free text description by the SME The basic creation of a triple involves the user typing in a word or phrase. At this point, we have a large number of concepts that annotators have previously found in all of the public corpora we have used to date. These concepts are auto-completed and presented to the user who can hover over the concepts to make sure they are disambiguating the term correctly. This auto-complete function in itself aids in the alignment of ambiguous results, since it is natural for a user to select an entity that has previously been chosen if it matches their intended sense of the concept they are describing. However, if it is a new concept – not previously used in the system, or if the user chooses – they can go to a full lookup and scan through a look-up window that shows all of the possible entities that might match the term the user has typed to find a correct definition that matches the sense of the word they had in mind. This can often be the most time consuming stage of creating the triples and one of the most challenging aspects of the design of the system. It was our original intent to use just one public dataset to look up terms; however, we quickly found that not anyone dataset by itself contained enough range of concepts to allow for the expressiveness of real annotation. So, while the system allows both custom datasets to be used and the creation of new datasets; the default look-up displays results from DBpedia, Yago, Art and Architecture Thesaurus and Wikidata. This, in and of itself can lead to duplications in triples across different datasets, but also occasionally forces annotators to make arbitrary choices among needlessly fine distinctions of meaning. The word ‘curve’ for example yields two senses in Yago: a curved section of a road or railway (the object itself) or the property of something (such as a roadway) being curved. This excess of conceptual distinctions based on fine philosophical or grammatical considerations or simple due to duplication is a source of difficulty for our project. Still, we cannot ‘not’ show all results, because in some cases – that fine grained distinction is necessary. While the various corpora have many owl:sameAs links identifying similar concepts, these are not useful in practice. They are unreliable; but in any case, reasoning with equality is expensive and slow. What is needed is to have a single source of canonical names for concepts at just the right level of ambiguity and for now, we have made some headway in resolving this sometimes immense ‘embarrassment of concepts’ by hand over the course of hundreds of person-hours to build the cached concepts found by auto-completion. We also use some programmatic tricks that organize the results in the look-up such that the most likely matches will be at the top of each dataset. In any case, over time, power annotators quickly learn the fastest paths to disambiguation and this leads to increasingly more fluid look-up overall. 4. Visual Ontology Building The development of the system has also given us multiple opportunities for experimentation in ontology development. First, we added the ability for the annotators to build more complex descriptions by chaining triples together by use of a ‘which’ and ‘and’ statements. In this way, more skilled annotators can not only describe the objects seen in the image, but they are also creating a slowly growing corpus of useful concepts - mostly RDF properties which seem to allow quite rich expressiveness within the confines of the simple RDF triple graph syntax. As images can depict just about anything, even mythological or illusory things, this process often pushes the edge of the expressive abilities of the published vocabularies, requiring power users to create new properties as well as new nominal concepts. Figure 4 illustrates an example of triples chained together with new ‘properties’ which are being accumulated in the system for further analysis. Often these are concepts arising in specialized domains, such as cookingMethod and FoodStage used to describe images illustrating recipes, isWearing used to talk about people in an image, and hasVIN to identify particular cars. But some are of more general utility, such as isPartOf, isMadeOf, and hasColor. Some of these have become extended in use, just as in the case of lio:hasSetting. One in particular, hasCondition, was introduced for describing stages in a process, such as a building under construction or a car engine being assembled, but its use has extended to things like a curved road, made up from the available concepts of dbpedia:Road and yago:curve. It can be used now to indicate a complex concept in which one descriptor modifies another. In many ways this seems to capture the adjectival construction in English, while staying within the simple triple-graph syntax of RDF. This process of Darwinian selection (by frequency of re- use) and conceptual generalization is evolving a general-purpose conceptual language for describing things, which builds on and extends the basic semantic corpora; and is of course itself, uniquely, documented and illustrated by the images themselves. Another style of usage for ontology development has also been found especially productive in which subject-matter experts (SMEs) provide image descriptions in free text, which is then processed by trained ImageSnippets power annotators who create the actual triple, using multiple techniques for more rapid triple building including the use of an auto-entity extractor from DBpedia over the free text which users can then automatically add to a list of pre-disambiguated concepts for attaching to LIO or their own user-created properties. In cases like this, the concepts used by the SME are often more specialized, reflecting their expertise, than those found in the conceptual corpus. For example, a vehicle may be identified as a rare 1953 Fletcher prototype (one of only two still in existence) rather than simply a military vehicle, which is the best suggestion made by Figure 4. Use of chained triples in the Triple Editor Clarifai. In these cases, the annotator can generate new IRI’s for such terms, while noting and recording the subclass or instance relationships between these and the existing terms in the public corpus. In this way, a specialized ontological vocabulary can be created as a byproduct of the image annotation process. 5. Conclusion In conclusion, our long-term image annotation project, predominantly a labor of love by the first author, has had a number of goals, the first and foremost of which has been the construction of a linked-data, image annotation ‘sandbox’ for extended research and development in the areas of knowledge capture and representation using images as stimulus. Secondary goals involve the more practical use of sharing, managing and publishing the images and their metadata as well as an interest in using the system for the curation and preservation of historically significant digital image assets, particularly those from niche or at-risk domains with extremely deep and precise metadata. Much of our work has involved highly subjective and ambiguous distinctions, trial and error, testing and re-testing while simultaneously making interface changes to improve usability and the training of annotators to work in unfamiliar mental territory. Our methodology, in favor of intuitive usability has embraced a process of creating boundaries around ambiguity eschewing rigid definitions. Through a process of debating and resolving disagreements by splitting concepts, we attempt to identify useful conceptual generalizations as they emerge and adjust the boundaries of natural meanings. This is work in progress, but we are optimistic. 6. References Eskridge, T., Hoffman,R., Hayes, P. and Warren, M. 2006 Formalizing the informal: a Confluence of Concept Mapping and the Semantic Web. Proc. Second International Conference on Concept Mapping, A J Cañas & J. Novak, eds. Costa Rica, 2006 Warren, M., and Hayes, P. 2007 Artspeak: The Contemporary Artist Meets the Semantic Web. Creating Formal Semantic Web Ontologies from the Language of Artists Electronic Techtonics - Thinking at the Interface; HASTAC (Humanities, Arts, Science, and Technology Advanced Collaboratory) - Duke University, 2007 Brickley, D. and Miller, L. 2010 FOAF Vocabulary Specification 0.97 http://xmlns.com/foaf/spec/20100101.html Warren, M, 2016 A Visual Guide to the ImageSnippets Properties. http://www.imagesnippets.com/ArtSpeak/help/properties.html