Abstract

Expert Systems were proposed in the mid 1970s (Arnott & Pervan, 2014) with the goal of building computerized systems that mimic human behavior to solve real-world tasks. Such systems were based on artificial intelligence (AI) techniques, typically by adopting explicit (human understandable) knowledge, extracted from domain experts (e.g. by using interviews) and that was stored in a knowledge base (Buchanan, 1986). In the last decades, the world as changed due to advances in information and communication technology (e.g. massive usage of computers and personal mobile devices, Internet and social media, usage of digital cameras and other sensors). In effect, we are now in the age of data, where a large portion of organizational, societal or personal activities is captured digitally (Mojsilovic, 2014). Following this change, ES have evolved to include data-driven models, either solely or complemented by expert-driven knowledge. Such change is reflected in the ES journal, which currently publishes several articles related with data analysis fields (e.g. analytics and business intelligence, data mining and knowledge discovery, big data and data science). In this special issue, we highlight two of data related terms: knowledge discovery (KD) and Business Intelligence (BI). KD is often used as a synonym of data mining, and it consists in an AI subfield that uses machine learning algorithms to extract high-level interesting knowledge from raw data (Fayyad et al. 1996). BI is a popular management term (Arnott & Pervan, 2014), and it represents several technologies (e.g. data warehouses, KD and dashboards) that store and process organizational data in order to support managerial decision making (Delen et al. 2014). The ‘Knowledge Discovery and Business Intelligence’ (KDBI) thematic track was proposed for the EPIA conference on AI in 2009 with the goal of promoting the interaction between the KD and BI areas. Since then, the track has been included in all EPIA biennial conferences. After 2011, the KDBI track has been associated with special ES journal issues, which included extended versions of the best KDBI papers. The first special issue was published in 2013, and it included the best KDBI 2011 track papers (Cortez & Santos, 2013), while the second special issue appeared in 2015, and it encompassed the best KDBI 2013 track papers (Cortez & Santos, 2015). This issue, entitled ‘Third special issue on Knowledge Discovery and Business Intelligence’, contains recent KD and BI contributions that can be used in ES to produce a valuable impact in real-world applications. It includes extended versions of papers from the 4th KDBI thematic track, of the 17th EPIA conference on AI (EPIA 2015), held in Coimbra, Portugal. The track received 18 paper submissions and the authors of the best papers were invited to extend their works for this special issue. After two rounds of reviews, which included reviewers from the KDBI track and also ES journal, the best six papers were accepted, corresponding to an overall acceptance rate of 33%. Due to the interest in data-driven models, in the last years, there has been several interesting developments in the KDBI area. Despite this progress, there are still many challenges and opportunities. For instance, most KD algorithms were targeted for single label classification tasks and often these algorithms cannot deal adequately with label ranking, which is useful in several real-world applications (e.g. modeling user preferences). Also, the financial domain still rises many challenges: it is not clear what is the best approach (regression or classification) to forecast trading actions; and easy to interpret tools are needed to better disclose the relationships among price variations from distinct financial products. Moreover, most of the real-world data has a temporal dimension and there is still room for proposing specialized algorithms that use this dimension in the KD or BI process (e.g. time series retrieval, visual representation of temporal changes). Furthermore, the ‘Extract, Transform and Load’ (ETL) is a vital component of BI systems, but it often requires a substantial manual effort in terms of its design and implementation, even when there are several common ETL processes that are repeated through distinct BI projects. All these challenges and opportunities are addressed in the six papers accepted in this special issue, which we will briefly detail in the next section. In the first paper ‘Label Ranking Forests’, de Sá et al (2016) propose a novel KD algorithm for label ranking (LR), which is a classification variant task. The goal is to learn the implicit function that performs a mapping between a set of inputs, which characterize an item, and a ranking of labels (instead of a single label, as in standard classification). LR is used in several real-world applications, such as microarray analysis, image categorization or modeling user preferences. The proposed label ranking forests (LRF) algorithm uses an ensemble of decision tree methods for LR and, it is considered a natural adaptation of the popular random forest algorithm for LR. Several experiments were held using 16 datasets from the KEBI Data Repository. Overall, competitive LR results were achieved by the LRF algorithm when compared with LR decision tree methods. In ‘A comparative study of approaches to forecast the correct trading actions’, Baía and Torgo (2016) perform an extensive comparison of two main approaches to forecast trading actions of financial markets. The first approach uses standard regression models to predict the daily variation on prices and then uses pre-defined decision rules in order to transform the numeric predictions into trading actions. The second approach uses classification models that directly forecast the trading decision (hold, buy and sell). The data analyzed included assets prices of 12 companies, ranging from 7 to 30 years of closing prices daily data. Several machine learning models were tested (e.g. neural networks, support vector machines, random forests). The main conclusion of the paper is that there is no significant difference between the two main approaches: numeric price variation prediction and and direct classification of trading actions. The study contains two additional recommendations: re-sampling strategies (for regression or classification) are not recommended in this financial domain, even if the data is imbalanced; also, the usage of cost-benefit matrices is promising to enhance the classification models. Also approaching the financial context, in ‘Ramex-Forum: a tool for displaying and analyzing complex sequential patterns of financial products’, Tiple et al. (2016) present an improved version of the Ramex-Forum algorithm, leading to a KD method that is capable of extracting knowledge from multivariate time series in a visual way and that is easy to interpret. The algorithm was applied in two real-world applications: petroleum production prices and risk analysis of European financial institutions. The obtained results attest the algorithm capability to show relevant price variations in financial markets. In ‘Aggressive pruning strategy for time series retrieval using a multi-resolution representation based on vector quantization coupled with discrete wavelet transform’, Muhammad Fuad (2016) proposes a novel KD algorithm for time series indexing and retrieval. The algorithm adopts a multi-resolution representation of the series based on Haar wavelets and vector quantization. The experiments were conducted using 10-time series with distinct sizes and from different repositories. The proposed algorithm obtained competitive results when compared with a two other representation methods (single-resolution and other multi-resolution algorithm). Temporal data was also studied in the paper of Gérk (2016), entitled ‘Visual analytics of educational time-dependent data using interactive dynamic visualization’. The paper presents a new web-based visualization framework for BI and that is targeted for an interactive exploration by the decision maker. In particular, the framework is based on motion charts and clustering techniques, allowing to reveal changes over time in terms of two-dimensional space animations. The framework was demonstrated using real-world educational data. Sixteen human participants evaluated the quality of the framework, confirming the proposed animations as an interesting contribution to the analytic process. In the last paper, Oliveira and Belo (2016) approach ETL processes, which play a key role in BI by extracting data from several sources into a data warehousing system. The paper, entitled ‘On the specification of ETL patterns behavior, a domain-specific language approach’, proposes the use of patterns to represent common ETL tasks (e.g. surrogate key pipelining or intensive data loading). Using the business process modeling notation (BPMN), it is shown how a typical and crucial ETL process (data quality enhancement) can be improved in terms of its design and implementation. We would like to thank the other KDBI 2015 track (of EPIA) co-organizers, Luís Cavique, João Gama and Nuno Marques. Also, we thank the authors, who contributed with their papers, and the reviewers (from the KDBI 2015 program committee and the ES journal). This work has been supported by COMPETE: POCI-01-0145-FEDER-007043 and FCT – Fundação para a Ciência e Tecnologia within the Project Scope: UID/CEC/00319/2013.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call