scispace - formally typeset
Search or ask a question
Proceedings ArticleDOI

Automated faceted reporting for web analytics

TL;DR: This work proposes a business-metric driven scheme for automatic selection of facets, a dimension along which a document could be represented in the context of web analytics, where aggregated business metrics are reported grouped by a facet.
Abstract: Traditionally, web analytics has focused on analysis and reporting of business metrics of interest to marketers, such as page views and revenue, by various dimensions of session characteristics, that can be obtained from user request. We introduce the notion of faceted reporting in the context of web analytics, where aggregated business metrics are reported grouped by a facet, a dimension along which a document could be represented. For example, in the case of e-Commerce sites, facets are typically various product attributes such as price, color, manufacturer, etc. For a typical website one could think of thousands of facets, but not all of them are equally important for the marketer in all reporting scenarios. In this work, we propose a business-metric driven scheme for automatic selection of facets for various reporting scenarios. The facet selection is done based on optimizing an objective function involving business metrics and we present our evaluation results based on multiple objective functions. We observe that, marketers' intuitive selection of useful facets is inaccurate. On the other hand automated methods proposed in this paper can highlight insights from the data.
Citations
More filters
Journal Article
TL;DR: A natural way of extending the current paradigm employed by traditional search systems, the exploratory search is suggested, which is to provide a framework or a platform which is extensible with plugins and able to provide instances tunable to a particular document collection of choice.
Abstract: Modern populations rely heavily on the worldwide web in searching information because it is the largest human repository of knowledge. However, finding relevant information on the web is often challenging. In the current work, we review analyses and optimize the performance of exploratory and faceted search techniques. Search behavior that is characterized by a large amount of uncertainty about the goals of the search is common in exploratory search. On the other hand, faceted search technique refines search results by a faceted taxonomy in an iterative manner. In addition, facets provide an efficient way to analyze and navigate the search result space. However, we believe that facet selection has been guided by the properties of suboptimal facet and facet term. As a consequence, users may need technical support while searching information. Thus, this paper suggests a natural way of extending the current paradigm employed by traditional search systems, the exploratory search. Our main objective is to provide a framework or a platform which is extensible with plugins and able to provide instances tunable to a particular document collection of choice. In addition, this paper presents a research model based on the prototype that will be developed.

10 citations


Cites background from "Automated faceted reporting for web..."

  • ...appearances of the top-level facet terms in data objects are rare [41]....

    [...]

  • ...happens whenever an observer fails to notice visual changes [41]....

    [...]

Journal ArticleDOI
TL;DR: REQAnalytics is presented, a recommender system that collects information on the usage of a web service, relates that information back to the requirements, and generates reports with recommendations and change suggestions that can increase the quality of that service.
Abstract: In the context of SaaS, where the change requests can be frequent, there is the need for a systematic requirements management process so as to maintain requirements updated and ease the management of changes required to improve the service to provide. Changes to perform need to be prioritized and their impact on the system should be assessed. The extraction and analysis of the use of the services provided through the web and their relationship to the requirements can help identify improvements and help keep the service useful for longer period of time. This paper presents REQAnalytics, a recommender system that collects information on the usage of a web service, relates that information back to the requirements, and generates reports with recommendations and change suggestions that can increase the quality of that service. The proposed approach aims to provide reports of the analysis made in a language closer to the business where, for example, it indicates new workflows and navigation paths, identifies the features that can be removed and presents the relationship between requirements and the proposed changes helping to maintain the software requirements specification updated and useful.

8 citations


Cites background from "Automated faceted reporting for web..."

  • ...Either, web analytics has focused on analysis and reporting of business metrics of interest mainly to marketers [6]....

    [...]

Book ChapterDOI
01 Jan 2016
TL;DR: An approach through a recommender system that collects the information about the usage of a website using a web analytics tools and generate recommendations reports that may help the requirements maintenance and increase the quality of the software requirements specification of the website is presented.
Abstract: The requirements of a website are constantly changing and evolving. In this context, managing the changes related to the baseline and reviewing the requirements’ prioritization is critical. Nowadays, web analytics tools are used to analyze the behaviour of users on a website in order to improve its usability and user experience. The analysis of the usage of websites may also help software requirements maintenance which can be a contribution to the overall quality of the service provided. This paper presents an approach through a recommender system that collects the information about the usage of a website using a web analytics tools and generate recommendations reports that may help the requirements maintenance and increase the quality of the software requirements specification of the website. This paper presents also an experimental evaluation of a case study based on an online newspaper website. The results of the proposed approach provide more readable reports in a language more closer to the business and suggest requirements priority change, identify new requirements and functionalities that may be removed.

7 citations

Journal ArticleDOI
TL;DR: The results showed that REQAnalytics can produce reports in a language closer to the business, identify most and less used functionalities, requirements that can be split in two or more requirements and give support to the maintenance of requirements of the website being analyzed.

6 citations

Jorge Garcia1
07 Apr 2016
TL;DR: Along the research work described in this dissertation, an approach and supporting tool is developed to collect the information about the usage of the website, to analyse it and to build reports with recommendations that can contribute to increase the quality of the requirements specification and the website itself.
Abstract: In recent years, the use of the World Wide Web (WWW) has had a huge growth and there is a greater variety of web applications with an increasing importance in society and in supporting the development to all kinds of business. Often, most of websites are providing support services that must be maintained and improved over time. This maintenance and upgrade can be difficult because frequently the requirements are no longer actual and/or often not even exist documented. Furthermore, it can also be difficult to assess what are the most critical features in order to define the changes to implement first (in the case of several requests). Websites are increasingly monitoring usage data, and this type of information is increasingly abundant as the meta-data that is possible to extract from the web navigation. However, commonly, the data about the usage of web sites is used only for reporting and for analysing the usage traffic not adding any value to the strengthening and improvement of web sites. Extracting and analysing the information about the usage of the websites can help identify improvements on the website and help to maintain requirements updated which can be a contribution to the overall quality of the service provided. Along the research work described in this dissertation, we develop an approach and supporting tool (called REQAnalytics) to collect the information about the usage of the website, to analyse it and to build reports with recommendations that can, hopefully, contribute to increase the quality of the requirements specification and the website itself. The contributions of this research work are: a set of rules to analyse usage data; an high-level mapping tool that through a web-based application, maps

3 citations

References
More filters
Journal ArticleDOI
Jacob Cohen1
TL;DR: In this article, the authors present a procedure for having two or more judges independently categorize a sample of units and determine the degree, significance, and significance of the units. But they do not discuss the extent to which these judgments are reproducible, i.e., reliable.
Abstract: CONSIDER Table 1. It represents in its formal characteristics a situation which arises in the clinical-social-personality areas of psychology, where it frequently occurs that the only useful level of measurement obtainable is nominal scaling (Stevens, 1951, pp. 2526), i.e. placement in a set of k unordered categories. Because the categorizing of the units is a consequence of some complex judgment process performed by a &dquo;two-legged meter&dquo; (Stevens, 1958), it becomes important to determine the extent to which these judgments are reproducible, i.e., reliable. The procedure which suggests itself is that of having two (or more) judges independently categorize a sample of units and determine the degree, significance, and

34,965 citations

Proceedings Article
01 Jan 1997
TL;DR: This work introduces wrapper induction, a method for automatically constructing wrappers, and identifies hlrt, a wrapper class that is e ciently learnable, yet expressive enough to handle 48% of a recently surveyed sample of Internet resources.
Abstract: Many Internet information resources present relational data|telephone directories, product catalogs, etc. Because these sites are formatted for people, mechanically extracting their content is di cult. Systems using such resources typically use hand-coded wrappers, procedures to extract data from information resources. We introduce wrapper induction, a method for automatically constructing wrappers, and identify hlrt, a wrapper class that is e ciently learnable, yet expressive enough to handle 48% of a recently surveyed sample of Internet resources. We use PAC analysis to bound the problem's sample complexity, and show that the system degrades gracefully with imperfect labeling knowledge.

1,182 citations

Proceedings ArticleDOI
05 Apr 2003
TL;DR: An alternative based on enabling users to navigate along conceptual dimensions that describe the images is presented, which makes use of hierarchical faceted metadata and dynamically generated query previews.
Abstract: There are currently two dominant interface types for searching and browsing large image collections: keyword-based search, and searching by overall similarity to sample images. We present an alternative based on enabling users to navigate along conceptual dimensions that describe the images. The interface makes use of hierarchical faceted metadata and dynamically generated query previews. A usability study, in which 32 art history students explored a collection of 35,000 fine arts images, compares this approach to a standard image search interface. Despite the unfamiliarity and power of the interface (attributes that often lead to rejection of new search interfaces), the study results show that 90% of the participants preferred the metadata approach overall, 97% said that it helped them learn more about the collection, 75% found it more flexible, and 72% found it easier to use than a standard baseline system. These results indicate that a category-based approach is a successful way to provide access to image collections.

1,074 citations

Book
01 Sep 2009
TL;DR: This book summarizes developments of the state of the art of search interface design, both in academic research and in deployment in commercial systems, presenting the most broadly acceptable make their way into major web search engines.
Abstract: This book focuses on the human users of search engines and the tool they use to interact with them: the search user interface. The truly worldwide reach of the Web has brought with it a new realization among computer scientists and laypeople of the enormous importance of usability and user interface design. In the last ten years, much has become understood about what works in search interfaces from a usability perspective, and what does not. Researchers and practitioners have developed a wide range of innovative interface ideas, but only the most broadly acceptable make their way into major web search engines. This book summarizes these developments, presenting the state of the art of search interface design, both in academic research and in deployment in commercial systems. Many books describe the algorithms behind search engines and information retrieval systems, but the unique focus of this book is specifically on the user interface. It will be welcomed by industry professionals who design systems that use search interfaces as well as graduate students and academic researchers who investigate information systems.

702 citations

Book ChapterDOI
23 Mar 1998
TL;DR: A method for locating data anomalies in a k dimensional data cube that includes the steps of associating a surprise value with each cell of a data cube, and indicating a data anomaly when the surprise value associated with a cell exceeds a predetermined exception threshold.
Abstract: A method for locating data anomalies in a k dimensional data cube that includes the steps of associating a surprise value with each cell of a data cube, and indicating a data anomaly when the surprise value associated with a cell exceeds a predetermined exception threshold. According to one aspect of the invention, the surprise value associated with each cell is a composite value that is based on at least one of a Self-Exp value for the cell, an In-Exp value for the cell and a Path-Exp value for the cell. Preferably, the step of associating the surprise value with each cell includes the steps of determining a Self-Exp value for the cell, determining an In-Exp value for the cell, determining a Path-Exp value for the cell, and then generating the surprise value for the cell based on the Self-Exp value, the In-Exp value and the Path-value.

455 citations