<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">The road to highlights is paved with good intentions: envisioning a paradigm shift in OLAP modeling</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Panos</forename><surname>Vassiliadis</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Univ. Ioannina</orgName>
								<address>
									<settlement>Ioannina</settlement>
									<country key="GR">Greece</country>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><forename type="first">Patrick</forename><surname>Marcel</surname></persName>
							<email>patrick.marcel@univ-tours.fr</email>
							<affiliation key="aff1">
								<orgName type="institution">University of Tours Tours</orgName>
								<address>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">The road to highlights is paved with good intentions: envisioning a paradigm shift in OLAP modeling</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073)</idno>
					</monogr>
					<idno type="MD5">26ED987066268EC1C244A939C57C1CAB</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T03:34+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this vision paper we structure a vision for the Business Intelligence of the near future in terms of a model with novel concepts and operators. We envision systems where the end-user requests information at a very high level, expressed as his intention to discover information, and the system transforms this request to the concrete execution of algorithms in order to compute, visualize and comment data and important highlights among them as an answer to the information request made by the end-user.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">INTRODUCTION</head><p>How will BI look like 10 years from now? What foundations should academia build in order to rigorously support the building of tools, the optimization of OLAP sessions, and the training of new scientists around a logical Paradigm? In this vision paper, we make a first attempt to revisit the foundations of OLAP and BI in an attempt to address the aforementioned questions.</p><p>To start with, it is worth to shortly revisit the evolution of OLAP and BI modeling so far.</p><p>• At the beginning of time, people would be working with relational queries and recordsets, returned by these queries. This treatment of BI was very DBMS-oriented, as the focus of attention was on what the DBMS can do for the users <ref type="bibr" target="#b7">[8]</ref>. • Then, both the scientific community and the industry understood that it is possible to provide a layer of abstraction on top of the database. So, users would deal (on-line) with cubes, rather than with traditional database data, which gives a very elegant simplification of the data to the user. The operations would be cube-oriented, one level of abstraction above the database operations and would involve roll-ups, drill-downs, etc. (see for example <ref type="bibr" target="#b14">[15]</ref>). • Rapidly, these abstractions became even closer to the way the multidimensional data space was navigated. Research proposed advanced operators permitting discovery driven analysis were basically combinations of OLAP primitives <ref type="bibr" target="#b15">[16]</ref><ref type="bibr" target="#b16">[17]</ref><ref type="bibr" target="#b17">[18]</ref>. What is the vision for the new generation of BI tools? Broadly speaking, we envision the redefinition of what an OLAP query is, both with respect to what users ask the system and with respect to what the answer entails.</p><p>• Concerning the aspect of what users can ask, we wish to further simplify the operations and choices that the users face and add an extra layer of abstraction. Specifically, we propose to replace cube operators with intentional operators over the data -in other words, with an "algebra" of operators closely to the user's intentions. Instead of the users operating in terms of roll-up and drill-down with the cube, we wish to empower them to state their intentions with respect to an operation. For example, instead of saying "drill down this cell one level", the user might ask "explain the drop in the sales of this product", or "is the main drop in sales observed at the best selling city also observed in similar cities?"; as another example, instead of issuing a roll-up operator, the user can state "verify whether a trend that I observe in a certain context still holds in general". • Concerning the aspect of what the answer to a query is, we believe we can no longer remain with "sets of tuples" as the answers to queries. We can envision BI tools where the answer to a query includes (i) a set of tuples accompanied with appropriate visualizations, (ii) the automatic mining of models and patterns, (iii) the extraction of important "jewels" hidden in the result (which we call highlights), and naturally, (iv) advanced, intuitive reporting.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Contributions and Benefits.</head><p>The main contribution of this paper is that it structures a vision for the Business Intelligence of the near future in terms of a model with novel concepts and operators. We aim our definitions to be broad enough, yet as precise as possible; at the same time, we want to link them as much as possible to the intentional nature of the next generation of BI tools, where the end-user requests information at a very high level and the system transforms these requests to concrete execution of algorithms in order to compute, visualize and comment data and important highlights among them as an answer to the information request made by the end-user.</p><p>Why bother? We are convinced that after 50 years of naive query answering, it is now time to replace it with effortless, automatic insight gaining from the user. Instead of making the end user dig into sets of records, we can increase productivity and the understanding of the essence of the data by using two pillars, one devoted to querying and another devoted to answering, specifically, firstly, by allowing the user to focus on high-level goals of information acquisition, rather than details of what data to bring in, and secondly, by providing focus-points in the answers that will move the user effort from manual "jewel mining" to addressing the insights gained.</p><p>Several other practical benefits can be envisioned. Our motivation was partially triggered from the problem of generating meaningful, artificial session logs for experimental reasons (e.g., like <ref type="bibr" target="#b13">[14]</ref>). Moreover, such efforts also make sense in other domains, e.g., to search Web data sources <ref type="bibr" target="#b2">[3]</ref>. Finally, note that this algebra can be seen as a first step towards addressing some of the open challenges proposed by the research community, like e.g., in <ref type="bibr" target="#b8">[9]</ref>, namely the lack of declarative exploration languages to present and reason about popular navigational idioms, or the various challenges raised by <ref type="bibr" target="#b4">[5]</ref> around the benchmarking of interactive data exploration by measuring user's gain in terms of insights.</p><p>An OLAP session is a sequence of dashboards that the analyst sees, each with its own information, including data, charts and informative summaries of KPI performance. The sequence is produced by the actions of the analyst that changes the contents of the dashboard by requesting more information on the basis of a set of operations made available to him by the tool.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Preliminaries on multidimensional modeling</head><p>For lack of space, we do not go to the details of data modeling.</p><p>We closely follow the model of <ref type="bibr" target="#b6">[7]</ref> and constrain ourselves to mention a textual summary. As typically happens with multidimensional models, we assume that dimensions provide a context for facts <ref type="bibr" target="#b10">[11]</ref>. This is especially important if combined with the fact that dimension values come in hierarchies; therefore, every single fact can be simultaneously placed in multiple hierarchically structured contexts, providing thus the ability to analyze sets of facts from multiple perspectives. The underlying data sets include measures that are characterized with respect to these dimensions. Cube queries involve measure aggregations at specific levels of granularity per dimension, along with filtering of data for specific values of interest.</p><p>We model an OLAP session as a directed graph G(V ,E). The set of vertices of the graph represent states of the user's session -practically the dashboard screens the user is receiving by the OLAP tool-and the arcs represent transitions among states.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Intentional Queries: The Transitions of a Session</head><p>Before describing in detail our vision on dashboards and their internals, we start with the deeper essence of the intentional nature of our proposal: user operations, or, equivalently transitions between the states of an OLAP session. The main idea behind transitions is that we move from a concrete model of logical operators like roll-up's and drill down's, to an intentional model where the user expresses in terms of operators for high-level requirements like "explain a certain phenomenon", "predict the future values" and this high-level requirement has to be automatically translated to specific OLAP and Data Mining operators/algorithms that will carry the answer. This can also facilitate greatly the extraction of highlights, as the user's goal is explicitly stated to the system. We organize the transition operators that can express the users' requirements for extra information, in families, i.e., a taxonomy of types for our transitions, which we call transition types. In contrast to other models where a transition would practically be a query (relational or multidimensional), in our model, a transition characterizes the intention of the user with respect to her information need. Each family is a collection of transition operators with the same high-level intention, but of course, different semantics at the details. This set of families intends to cover (a) traditional OLAP operators, (b) various contributions of the literature, in particular discovery driven analysis operators <ref type="bibr" target="#b15">[16]</ref><ref type="bibr" target="#b16">[17]</ref><ref type="bibr" target="#b17">[18]</ref>, Cinecubes acts <ref type="bibr" target="#b6">[7]</ref>, and operators for exploratory search in the web <ref type="bibr" target="#b2">[3]</ref>, as well as (c) novel operators that can be developed by the research community in the future. Before detailing these families, we mention Figure <ref type="figure" target="#fig_0">1</ref> that intuitively describes an analysis phrased with some transitions, using the Star Schema Benchmark data cube <ref type="bibr" target="#b12">[13]</ref>. The annotation and highlight generation of the data is explained in Section 2.3.</p><p>Compare. A transition of type compare is used to retrieve more data to be compared with what the current dashboard displays. Intuitively, transitions in this family ask for bringing in more data to the ongoing analysis. The family includes OLAP operators pivot, drill-across, the augment operator of <ref type="bibr" target="#b2">[3]</ref>, the put-in-context act of Cinecubes <ref type="bibr" target="#b6">[7]</ref>. For instance, on the example of Figure <ref type="figure" target="#fig_0">1</ref>, the put-in-context <ref type="bibr" target="#b6">[7]</ref> is used twice (top left vertical path in Figure <ref type="figure" target="#fig_0">1</ref>), first, to compare the sales for a category of products (MFGR#5) in a given supplier country (Argentina) to that of past years (indicated as ❷), and second to compare these past results to the results for sibling countries in the dimension Supplier (indicated as ❺).</p><p>FocusOn. A transition of type FocusOn is used to exclude data from the current analysis. Transitions in this family focus on a particular cell, or group of cells. The family includes the OLAP slice/dice operation, the take operator of <ref type="bibr" target="#b2">[3]</ref>, and also personalization operators like skyline, winnow, etc. <ref type="bibr" target="#b9">[10]</ref>. For instance, in Figure <ref type="figure" target="#fig_0">1</ref>, a slice/dice operation is done to focus on sales results for part MFRG#51 in years 2015 and 2016 (indicated as ❼).</p><p>Abstract. The transitions of this family are used to reduce the information load of the dashboard, by abstracting them in a more compact form. The family includes the traditional roll-up operator, as well as data mining operations like clustering, frequent itemset mining, etc. For instance, in Figure <ref type="figure" target="#fig_0">1</ref>, sales results for years 2011 (not all shown in previous dashboards) to 2016 in Argentina are abstracted in 2 classes based on a result threshold (indicated as ❹).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Analyze and Explain. A transition of type Analyze and Explain</head><p>is used to understand the cause of a phenomenon displayed in the current dashboard and to explore what is displayed at finer levels of detail. With respect to the analysis part, the family includes the drill-down OLAP operation, and the details act of Cinecubes <ref type="bibr" target="#b6">[7]</ref>. With respect to the explain part, the family includes Diff <ref type="bibr" target="#b15">[16]</ref> (interesting drill downs at various granularities explaining the content of two cells) and the operators of Cariou &amp; al. <ref type="bibr" target="#b3">[4]</ref>. For instance, at the top of Figure <ref type="figure" target="#fig_0">1</ref>, a drill-down from countries to cities is used to analyze sale results by cities (indicated as ❶), and, at the left bottom of the figure, a Diff is used to explain an important drop of sales in Argentina in year 2016 compared to year 2015, for the specific category of product MFGR#5 (indicated as ❻).</p><p>Verify. A transition of type Verify is used to check that a potential trend or pattern currently observed can be generalized to broader contexts, e.g., we can check whether it occurs at coarser levels of detail. The family includes Relax operator <ref type="bibr" target="#b17">[18]</ref> (interesting rollups at various granularities agreeing or contradicting a trend observed in the content of two cells), the verification of outlier data points in broader context than the current, etc. For instance, in Figure <ref type="figure" target="#fig_0">1</ref>, a Relax operation is used to verify whether the drop of sale results for category MFGR#5 in Argentina, in 2016, also holds for all products sold (indicated as ❸).</p><p>Predict. A transition of type Predict is used for predictive analysis of the displayed data. Typically this involves the entire set of timeseries forecasting methods, as well as classification methods used for assessing future events <ref type="bibr" target="#b0">[1]</ref>. For instance, in Figure <ref type="figure" target="#fig_0">1</ref>, sales results of product MFGR#51 for year 2015 and 2016 are used to issue predictions for year 2017 in Argentina cities (indicated as ❽). Suggest. A transition of type Suggest is used to ask for guidance with respect to where to navigate next in the multidimensional data space. The family includes Sarawagi's Inform <ref type="bibr" target="#b16">[17]</ref> (Interesting drill downs at various granularities, deviating from trends that what was already observed), or query recommendation techniques <ref type="bibr" target="#b11">[12]</ref>. We note that this kind of transition should incorporate risk control mechanisms to prevent false discoveries <ref type="bibr" target="#b1">[2]</ref>. For instance, in Figure <ref type="figure" target="#fig_0">1</ref>, a suggest operation is used to automatically find neighboring sale results deviating from the results of the previous operation (indicated as ❾).</p><p>Concluding with transitions, we emphasize that an open research challenge is to define an algebraically closed language of intentional operators, each of which can be translated automatically to the execution of concrete query and data mining operators. The synthesis of these operators can then create entire stories. Assume the user, being at state ❷, would like to: "verify whether the distribution of sales for mfgr#5 in Argentina from 2011 to 2016 still holds in general for all parts, and build a 2 classes model for it, then backtrack to parts and compare with sibling countries, and finally explain the highest country-wise difference. " This request would correspond to the following statement in the intentional language: explain hiдhl iдht :Max Dif f er ence ( compare sibl inдCount r ies ( backtrack showT r end ( abstract 2Cl asses ( veri f y all P ar t s,showT r end (❷))))). An optimizer would then automatically translate this request into the sequence of operations: roll up, cluster, backtrack, Cinecube's put in context, Diff and optimize it to produce a data story using dashboards ❷ to ❻.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Dashboards: The states of a session</head><p>A state is a dashboard that the user sees. In principle, each dashboard is ultimately based on the generating data provided by as a finite collection of queries, posed to the underlying database. A sharp distinction of our approach compared to previous models is that we do not restrict ourselves to data for each state but accompany them with a set of interesting findings, which come in two flavors, specifically, (a) in terms of models, i.e., results of data mining or machine learning algorithms applied over the data of a dashboard's state, and, (b) important findings that accompany the dashboard, to which we refer as the highlights of the state.</p><p>Take for example, the small scenario of Fig. <ref type="figure" target="#fig_1">2</ref>. The dashboard is based on a simple cube, as its generating data, involving Product, Reдion and Year as dimension levels and Sales and Cost as (aggregate) measures. Then, the dashboard automatically computes the Bene f it as the application of a simple arithmetic function, specifically, the difference of the two measures. A second step Steps. In order to construct and visualize a dashboard, we envision several computations taking place. Here is the sequence of the performed actions:</p><p>(1) First, the queries of the state's dashboard are issued and their results, the generating data of the dashboard, are computed. Any straightforward computations for extra, derived columns of the dashboard (e.g., дain=price-cost) are performed too. (2) Then, the available data are fed to model extraction algorithms for the computation of models abstract, summarize and provide patterns and insights for the data. (3) The potentially large amount of data and models computed has to be ranked and assessed on their interestingness for the analyst; the most important findings are classified as the dashboards highlights to be used for providing the main insights and the main directions for future transitions by the analyst. (4) The above are accompanied by visualization, text construction and reporting tasks that aim the process of understanding and communicating the main findings.</p><p>The generating data of the dashboard. The results of the underlying queries are the basis for the subsequent computations that take place for the construction of the dashboard: this is why we refer to them as the generating data of the dashboard.</p><p>Models. In difference from the state of the art in OLAP modeling, in our approach, we believe that the static results of aggregate queries, and their visualizations with charts and speedometers will simply be not enough in the near future. The automatic assessment and critical characterization of the presented data will be part of the BI of the near-future. See some simple cases based on the example of observing sales data of an international company:</p><p>• Sales data will be automatically characterized with respect to a decision tree that classifies them (e.g., as "successful", "risky", "potentially hazardous" etc) • Sales per country will be automatically clustered to reveal similarities and differences, as a first step towards understanding outliers and non-expected behavior • Aggregate sales over significant periods will be fed into time series analysis and forecasting methods to automatically detect trends, seasonalities and to deduce future values</p><p>We consider the plugging of data analysis algorithms in the backstage of a dashboard as an indispensable part of BI. These algorithms can range from very simple ones (e.g., finding the top values of a cuboid, or detecting whether a dimension value is systematically related to top or bottom sales) to very complicated ones (like, for example, outlier detection, dimensionality reduction, etc). Most importantly, as the operation of the algorithms will likely be as transparent as possible to the end user, their execution will require an almost automatic tuning of their parameters. The findings of these algorithms will be models of the data that are typically (not always) used to annotate the existing data with characterizations and offer focus points to the visualization of the dashboard (forecasts, outliers, dimension values that dominate top or bottom measures, . . .). The models themselves give a multitude of results. However, some of these results indicate that a part of a dashboard's data are of important interestingness value to the end user. Due to that, we collectively refer to the important results of the execution of these algorithms as highlights, in an attempt to show that the aim is to enrich the current data-intensive dashboards with knowledge that is worth exploring or using for decision making.</p><p>We are going to treat model extraction algorithms as "blackbox" algorithm without probing into their internals, and, most importantly, without assuming any particular properties for their output. What does a model extraction algorithm do? Basically, the algorithm receives as input (a) a set of input data, and, (b) a set of execution parameters that have to be fixed for the algorithm's execution. Without loss of generality, we can assume that a subset of these parameters will be bound to string or numerical values and the rest will be mapped to attributes of the input data. The output of a model extraction algorithm is a model of the input data. Depending on the algorithm, the result differs. For instance, a descriptive model built using unsupervised clustering is basically just a labeling of each cube's cells, while a predictive one allows enriching the cube with predictions and comes with an accuracy score. In summary, the main properties of a model extraction algorithm is outlined as follows:</p><p>(1) Input: a set of input data, which is the result of an extended cube query set of the dashboard, along with a fixing of the algorithm's input parameters (2) Output: a (possibly complex) result composed of (a) a model of the input data, and, (b) several characterizations of it (precision, strength, p-value, . . . )</p><p>Each model type can be arbitrarily complex and consists of a set of model components of versatile nature. Examples:</p><p>• A time series splits each of its points to 3 measurements, specifically error, trend and seasonality (practically creating 3 times series in the place of one, whose sum reconstructs the original one). • A clustering scheme includes a set of clusters, each with a set of tuples that constitute it, along with a centroid. • A classification decision tree includes a tree structure, best expressed as the composition of a set of paths, leading to characterization classes; again, each class comprises a set of generating tuples that pertain to it.</p><p>Typically, each such component, as well as the model on its entirety is accompanied with a set of metrics of its statistical power. Given a model type T , we denote its components with T C = {tmc 1 . . .tmc m }. We avoid resolving the internals of the composition of the tmc parts and treat T C as a set of components. This is realistic, as we can always mask a part-of relation as a set of constituting components, even a recursive one, by allowing an extra composition relation to provide the semantics of the part of-relationship. Apart from this simple modeling, attempts to relationally code mining results already exist <ref type="bibr" target="#b5">[6]</ref>. In other words: assuming a model M of type T (e.g., a particular cluster set M of type ClusterSet), we can use its components (e.g., the clusters of M, M = {cluster 1 , . . . , cluster m }) as fundamental parts of subsequent analyses, and, at the same time, link each of these components to their underlying data.</p><p>But, then, how do we handle the heterogeneity of different model types? Clearly, a cluster is inherently different from a decision tree or the formula for a trend. Is there a unifying model to cover them all? The unifying essence of all the plethora of diverse model types is that, at the end of the day, all of them are annotations of the original data. At the end of the day, every component of a complex model type (be it a cluster id, a path in a decision tree and a resulting class, a characterization of the top-k tuples, or a trend formula): (a) refers to a subset of the input data and vice-versa, and, (b) refers to the overall model via a part-of relationship. So, once a model of the underlying data is available, our solution to the problem is to provide a distinct identity (an id) to the components of a complex type T , retain the membership/part-of relationships of T separately and annotate or characterize the data with respect to the part of T that pertains to them. In fact, this step can be blended within the model extraction itself. Examples of such annotation follow:</p><p>• Assuming a time series model that splits a time series to trend, seasonality and noise, these attributes can be appended to the generating data set. • Assuming a cluster model, the generating data can be annotated with the id of the cluster to which they belong. • Assuming a classification model, the input data can be labeled via an extra attribute with respect to the class(es) of the model to which they belong. • Assuming a model of top-k values of a measure, the input data can be annotated with their rank, if they belong in the top-k set and they have been ranked.</p><p>A notable property of our modeling is that we require model components to be directly mapped and linked to their generating data in a bidirectional mapping, so that the end-user can navigate back and forth between cube cells and their models.</p><p>Highlight Production. As already mentioned, the set of highlights of the dashboard is a set of important findings that accompany the dashboard. These can be findings of any nature, e.g., important outliers in the contents of the dashboard's data, all the tuples belonging to a certain class of a classification scheme, the top or bottom values of a measure, etc. It would be straightforward (and doable with our modeling) to treat all the contents of an extended query as highlights. However, we would like to stress that we want to restrict our attention to the ones that are really important for the end-user.</p><p>Other "local' operations. Once all computations are done, there are several tasks to be taken care of before automatically constructing the dashboard. We envision that the development of principled methods for (a) the visualization of the dashboard data in various ways and (b) the creation of reports, via data storytelling methods will present major challenges for the future.</p><p>Visualize. Operations in this family compute the contents of visual representations (bar charts, speedometers, scatterplots, ...) on the basis of the current contents of the dashboard. We consider them to be local operations as they are performed without any further interaction with data external to the ones already retrieved from the dashboard. Data storytelling. Data storytelling is a novel trend that seems to carry significant weight in the way the future BI will look like. The main idea involves the automatic generation of a textual report from the data of a particular dashboard or of an entire OLAP sessions. There are already some academic efforts, like Cinecubes <ref type="bibr" target="#b6">[7]</ref> as well as some tools (see <ref type="bibr" target="#b18">[19]</ref> for a nice discussion).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">PATHS FOR FUTURE RESEARCH</head><p>This paper is a vision paper describing, in broad terms, a potential future for OLAP, to strengthen its place as the corner stone of BI. Our call to arms to the research community can be structured along several lines. Population of the families of the transition operators with concrete operators. Each new operator (or each new formalization of existing operators) should hopefully carry (a) clear semantics, (b) execution algorithms as well as their optimizations, (c) fine tuning algorithms for the parameter fixing of the introduced algorithms and, equally importantly, (d) a graceful linkage to the overall model proposed here (in an attempt to be able to gracefully plug it in the respective BI tools). Automation of transitions and tuning. We believe this to be the most important piece of the puzzle. How do we fully automate what model and highlight extraction methods should be employed? This includes the possibility of predicting interesting results for the end-user, which in turn, requires appropriate models of interestingness. Along with the necessary run-time optimizations, all these tasks provide challenging and open problems of significance practical importance. Key Performance Indicators (at least in the way they are used today) are examples of simple model extraction, and by definition, KPI's are highlights (or else they wouldn't be "Key" indicators). Linking them explicitly to a model for OLAP is a necessary add-on for a comprehensive view of what OLAP does. Benchmarking and tools. A reference free, open-source tool and a reference benchmark for the future BI (involving data, model and highlight extraction requests and sessions) can be a really handy tool for the research community (otherwise, each new paper will need to improvise on its experimental assessment). A tool will also trigger other research directions, like e.g., the incorporation of research results on natural language processing to accept the user requests, visualizations to depict the models and highlights, etc.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: A BI session phrased with some of our transitions. Numbers show the sequence of dashboard generation. Colored cells are results of data analytics operations annotating the dashboard data.</figDesc><graphic coords="3,53.80,83.68,487.68,365.76" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Model extraction, data annotation and highlight production: an example</figDesc><graphic coords="4,139.14,83.69,316.99,259.92" type="bitmap" /></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">C</forename><surname>Charu</surname></persName>
		</author>
		<author>
			<persName><surname>Aggarwal</surname></persName>
		</author>
		<title level="m">Data Mining -The Textbook</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Toward Sustainable Insights, or Why Polygamy is Bad for You</title>
		<author>
			<persName><forename type="first">Carsten</forename><surname>Binnig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lorenzo</forename><forename type="middle">De</forename><surname>Stefani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tim</forename><surname>Kraska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Eli</forename><surname>Upfal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Emanuel</forename><surname>Zgraggen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zheguang</forename><surname>Zhao</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CIDR 2017, 8th Biennial Conference on Innovative Data Systems Research</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Exploratory search framework for Web data sources</title>
		<author>
			<persName><forename type="first">Alessandro</forename><surname>Bozzon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Marco</forename><surname>Brambilla</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Stefano</forename><surname>Ceri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Davide</forename><surname>Mazza</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">VLDB J</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="641" to="663" />
			<date type="published" when="2013">2013. 2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Embedded indicators to facilitate the exploration of a data cube</title>
		<author>
			<persName><forename type="first">Véronique</forename><surname>Cariou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jérôme</forename><surname>Cubillé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christian</forename><surname>Derquenne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sabine</forename><surname>Goutier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Françoise</forename><surname>Guisnel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Henri</forename><surname>Klajnmic</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IJBIDM</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="329" to="349" />
			<date type="published" when="2009">2009. 2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Towards a Benchmark for Interactive Data Exploration</title>
		<author>
			<persName><forename type="first">Philipp</forename><surname>Eichmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Emanuel</forename><surname>Zgraggen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zheguang</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Carsten</forename><surname>Binnig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tim</forename><surname>Kraska</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Data Eng. Bull</title>
		<imprint>
			<biblScope unit="volume">39</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="50" to="61" />
			<date type="published" when="2016">2016. 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A Relational View of Pattern Discovery</title>
		<author>
			<persName><forename type="first">Arnaud</forename><surname>Giacometti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Patrick</forename><surname>Marcel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Arnaud</forename><surname>Soulet</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Database Systems for Advanced Applications -16th International Conference, DASFAA</title>
				<imprint>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="153" to="167" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">CineCubes: Aiding data workers gain insights from OLAP queries</title>
		<author>
			<persName><forename type="first">Dimitrios</forename><surname>Gkesoulis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Panos</forename><surname>Vassiliadis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Petros</forename><surname>Manousis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Inf. Syst</title>
		<imprint>
			<biblScope unit="volume">53</biblScope>
			<biblScope unit="page" from="60" to="86" />
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Data Cube: A Relational Aggregation Operator Generalizing Group-By, Cross-Tab, and Sub-Total</title>
		<author>
			<persName><forename type="first">Jim</forename><surname>Gray</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Adam</forename><surname>Bosworth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andrew</forename><surname>Layman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hamid</forename><surname>Pirahesh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Twelfth International Conference on Data Engineering</title>
				<meeting>the Twelfth International Conference on Data Engineering</meeting>
		<imprint>
			<date type="published" when="1996">1996</date>
			<biblScope unit="page" from="152" to="159" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Overview of Data Exploration Techniques</title>
		<author>
			<persName><forename type="first">Stratos</forename><surname>Idreos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Olga</forename><surname>Papaemmanouil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Surajit</forename><surname>Chaudhuri</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2015 ACM SIGMOD International Conference on Management of Data</title>
				<meeting>the 2015 ACM SIGMOD International Conference on Management of Data</meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="277" to="281" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">A survey of top-k query processing techniques in relational database systems</title>
		<author>
			<persName><forename type="first">F</forename><surname>Ihab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">George</forename><surname>Ilyas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mohamed</forename><forename type="middle">A</forename><surname>Beskales</surname></persName>
		</author>
		<author>
			<persName><surname>Soliman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Comput. Surv</title>
		<imprint>
			<biblScope unit="volume">40</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page">58</biblScope>
			<date type="published" when="2008">2008. 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">Multidimensional Databases and Data Warehousing</title>
		<author>
			<persName><forename type="first">Christian</forename><forename type="middle">S</forename><surname>Jensen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Torben</forename><surname>Bach Pedersen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christian</forename><surname>Thomsen</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2010">2010</date>
			<publisher>Morgan &amp; Claypool Publishers</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">A survey of query recommendation techniques for data warehouse exploration</title>
		<author>
			<persName><forename type="first">Patrick</forename><surname>Marcel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Elsa</forename><surname>Negre</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Actes des 7èmes journées francophones sur les Entrepôts de Données et l&apos;Analyse en ligne</title>
				<meeting>s des 7èmes journées francophones sur les Entrepôts de Données et l&apos;Analyse en ligne<address><addrLine>Clermont-Ferrand, France; EDA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2011-06">2011. 2011. Juin 2011</date>
			<biblScope unit="page" from="119" to="134" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">The Star Schema Benchmark and Augmented Fact Table Indexing</title>
		<author>
			<persName><forename type="first">Patrick</forename><forename type="middle">E</forename><surname>O'neil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Elizabeth</forename><forename type="middle">J</forename><surname>O'neil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Xuedong</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Stephen</forename><surname>Revilak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">TPCTC</title>
		<imprint>
			<biblScope unit="page" from="237" to="252" />
			<date type="published" when="2009">2009. 2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">CubeLoad: A Parametric Generator of Realistic OLAP Workloads</title>
		<author>
			<persName><forename type="first">Stefano</forename><surname>Rizzi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Enrico</forename><surname>Gallinucci</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advanced Information Systems Engineering -26th International Conference, CAiSE 2014</title>
				<meeting><address><addrLine>Thessaloniki, Greece</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2014-06-16">2014. June 16-20, 2014</date>
			<biblScope unit="page" from="610" to="624" />
		</imprint>
	</monogr>
	<note>Proceedings.</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">On the Need of a Reference Algebra for OLAP</title>
		<author>
			<persName><forename type="first">Oscar</forename><surname>Romero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alberto</forename><surname>Abelló</surname></persName>
		</author>
		<idno>Proceed- ings. 99-110</idno>
	</analytic>
	<monogr>
		<title level="m">Data Warehousing and Knowledge Discovery, 9th International Conference, DaWaK 2007</title>
				<meeting><address><addrLine>Regensburg, Germany</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2007-09-03">2007. September 3-7, 2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Explaining Differences in Multidimensional Aggregates</title>
		<author>
			<persName><forename type="first">Sunita</forename><surname>Sarawagi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of 25th International Conference on Very Large Data Bases (VLDB&apos;99)</title>
				<meeting>25th International Conference on Very Large Data Bases (VLDB&apos;99)</meeting>
		<imprint>
			<date type="published" when="1999">1999</date>
			<biblScope unit="page" from="42" to="53" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">User-Adaptive Exploration of Multidimensional Data</title>
		<author>
			<persName><forename type="first">Sunita</forename><surname>Sarawagi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of 26th International Conference on Very Large Data Bases (VLDB</title>
				<meeting>26th International Conference on Very Large Data Bases (VLDB</meeting>
		<imprint>
			<date type="published" when="2000">2000. 2000</date>
			<biblScope unit="page" from="307" to="316" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Intelligent Rollups in Multidimensional OLAP Data</title>
		<author>
			<persName><forename type="first">Gayatri</forename><surname>Sathe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sunita</forename><surname>Sarawagi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of 27th International Conference on Very Large Data Bases (VLDB</title>
				<meeting>27th International Conference on Very Large Data Bases (VLDB</meeting>
		<imprint>
			<date type="published" when="2001">2001. 2001</date>
			<biblScope unit="page" from="531" to="540" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Algorithmic authors</title>
		<author>
			<persName><forename type="first">Alex</forename><surname>Wright</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Commun. ACM</title>
		<imprint>
			<biblScope unit="volume">58</biblScope>
			<biblScope unit="page" from="12" to="14" />
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
