scispace - formally typeset
Search or ask a question

Showing papers on "Web page published in 2001"


Journal ArticleDOI
TL;DR: The authors propose the markup of Web services in the DAML family of Semantic Web markup languages, which enables a wide variety of agent technologies for automated Web service discovery, execution, composition and interoperation.
Abstract: The authors propose the markup of Web services in the DAML family of Semantic Web markup languages. This markup enables a wide variety of agent technologies for automated Web service discovery, execution, composition and interoperation. The authors present one such technology for automated Web service composition.

1,978 citations


Journal ArticleDOI
TL;DR: By allowing image reconstruction to continue even after a packet is lost, this type of representation can prevent a Web browser from becoming dormant, and the source can be approximated from any subset of the chunks.
Abstract: This article focuses on the compressed representations of pictures. The representation does not affect how many bits get from the Web server to the laptop, but it determines the usefulness of the bits that arrive. Many different representations are possible, and there is more involved in their choice than merely selecting a compression ratio. The techniques presented represent a single information source with several chunks of data ("descriptions") so that the source can be approximated from any subset of the chunks. By allowing image reconstruction to continue even after a packet is lost, this type of representation can prevent a Web browser from becoming dormant.

1,533 citations


Journal ArticleDOI
TL;DR: It is found that most people use few search terms, few modified queries, view few Web pages, and rarely use advanced search features, and the language of Web queries is distinctive.
Abstract: In studying actual Web searching by the public at large, we analyzed over one million Web queries by users of the Excite search engine. We found that most people use few search terms, few modified queries, view few Web pages, and rarely use advanced search features. A small number of search terms are used with high frequency, and a great many terms are unique; the language of Web queries is distinctive. Queries about recreation and entertainment rank highest. Findings are compared to data from two other large studies of Web queries. This study provides an insight into the public practices and choices in Web searching.

1,153 citations


Patent
06 Aug 2001
TL;DR: In this article, a server-side rasterization of visual content is used to display visual content on a client device using a client-side interface. But the server does not provide the interface.
Abstract: The present invention relates to display of visual content on a client device using server-side rasterization of visual content. Visual content is rendered on a server system, transformed into bitmaps compatible with the display attributes of a client device, and transmitted for display on the client device. The invention allows the server to perform, in effect, as a remote browser for displaying Web pages, e-mail, e-mail attachments, electronic document and forms, database queries and results, drawings, presentations, and images at the client device. The approach is “remote” because the server does the rendering and the client provides the interface; “multi-level” because rendered visual content is represented as a multi-level set of raster representations; and constitutes a “browsing system” because the client and server share data about the source visual content element being browsed, and the client performs a specific browsing function assisted by the server.

799 citations


Patent
15 Nov 2001
TL;DR: In this article, a method of allowing navigation of a plurality of data elements residing within a category-tree database includes providing, on a single primary “web” page, a distinct data element symbol for each data element and a distinct parent category symbols for each parent category in the category paths of the data elements.
Abstract: In some embodiments, a method of allowing navigation of a plurality of data elements residing within a category-tree database includes providing, on a single primary “web” page, a distinct data element symbol for each data element and a distinct parent category symbol for each parent category in the category paths of the data elements. For each illustrated parent category, a distinct first level web page is provided that includes (i) a first level subcategory symbol for each first level subcategory in the category paths of the data element(s) associated with that parent category and (ii) the data element symbol for each such data element shown associated with its respective first level subcategory symbol.

778 citations


Proceedings ArticleDOI
01 Jan 2001
TL;DR: It was found that the time spent on a pages, the amount of scrolling on a page and the combination of time and scrolling had a strong correlation with explicit interest, while individual scrolling methods and mouse-clicks were ineffective in predicting explicit interest.
Abstract: Recommender systems provide personalized suggestions about items that users will find interesting. Typically, recommender systems require a user interface that can ``intelligently'' determine the interest of a user and use this information to make suggestions. The common solution, ``explicit ratings'', where users tell the system what they think about a piece of information, is well-understood and fairly precise. However, having to stop to enter explicit ratings can alter normal patterns of browsing and reading. A more ``intelligent'' method is to useimplicit ratings, where a rating is obtained by a method other than obtaining it directly from the user. These implicit interest indicators have obvious advantages, including removing the cost of the user rating, and that every user interaction with the system can contribute to an implicit rating.Current recommender systems mostly do not use implicit ratings, nor is the ability of implicit ratings to predict actual user interest well-understood. This research studies the correlation between various implicit ratings and the explicit rating for a single Web page. A Web browser was developed to record the user's actions (implicit ratings) and the explicit rating of a page. Actions included mouse clicks, mouse movement, scrolling and elapsed time. This browser was used by over 80 people that browsed more than 2500 Web pages.Using the data collected by the browser, the individual implicit ratings and some combinations of implicit ratings were analyzed and compared with the explicit rating. We found that the time spent on a page, the amount of scrolling on a page and the combination of time and scrolling had a strong correlation with explicit interest, while individual scrolling methods and mouse-clicks were ineffective in predicting explicit interest.

768 citations


Proceedings ArticleDOI
01 Apr 2001
TL;DR: The paper presents the overall design of Annotea and describes some of the issues the project faced and how it has solved them, including combining RDF with XPointer, XLink, and HTTP.
Abstract: Annotea is a Web-based shared annotation system based on a general-purpose open RDF infrastructure, where annotations are modeled as a class of metadata. Annotations are viewed as statements made by an author about a Web document. Annotations are external to the documents and can be stored in one or more annotation servers. One of the goals of this project has been to re-use as much existing W3C technology as possible. We have reached it mostly by combining RDF with XPointer, XLink, and HTTP. We have also implemented an instance of our system using the Amaya editor/browser and a generic RDF database, accessible through an Apache HTTP server. In this implementation, the merging of annotations with documents takes place within the client. The paper presents the overall design of Annotea and describes some of the issues we have faced and how we have solved them.

765 citations


01 Jan 2001
TL;DR: A generic operational model of a hidden Web crawler is introduced and how this model is realized in HiWE (Hidden Web Exposer), a prototype crawler built at Stanford is described.
Abstract: Current-day crawlers retrieve content only from the publicly indexable Web, i.e., the set of Web pages reachable purely by following hypertext links, ignoring search forms and pages that require authorization or prior registration. In particular, they ignore the tremendous amount of high quality content “hidden” behind search forms, in large searchable electronic databases. In this paper, we address the problem of designing a crawler capable of extracting content from this hidden Web. We introduce a generic operational model of a hidden Web crawler and describe how this model is realized in HiWE (Hidden Web Exposer), a prototype crawler built at Stanford. We introduce a new Layout-based Information Extraction Technique (LITE) and demonstrate its use in automatically extracting semantic information from search forms and response pages. We also present results from experiments conducted to test and validate our techniques.

720 citations


Proceedings Article
11 Sep 2001
TL;DR: In this paper, the authors address the problem of designing a crawler capable of extracting content from the hidden web, i.e., the set of web pages reachable purely by following hypertext links, ignoring search forms and pages that require authorization or prior registration.
Abstract: Current-day crawlers retrieve content only from the publicly indexable Web, i.e., the set of Web pages reachable purely by following hypertext links, ignoring search forms and pages that require authorization or prior registration. In particular, they ignore the tremendous amount of high quality content “hidden” behind search forms, in large searchable electronic databases. In this paper, we address the problem of designing a crawler capable of extracting content from this hidden Web. We introduce a generic operational model of a hidden Web crawler and describe how this model is realized in HiWE (Hidden Web Exposer), a prototype crawler built at Stanford. We introduce a new Layout-based Information Extraction Technique (LITE) and demonstrate its use in automatically extracting semantic information from search forms and response pages. We also present results from experiments conducted to test and validate our techniques.

698 citations


Patent
06 Mar 2001
TL;DR: In this paper, a method and system for retrieving information through the use of a multi-stage interaction with a client to identify particular knowledge content associated with a knowledge map is described.
Abstract: A method and system are disclosed for retrieving information through the use of a multi-stage interaction with a client to identify particular knowledge content associated with a knowledge map. The present invention is an application program running on a server accessed via the world-wide web or other data network using standard Internet protocols, a web browser and web server software. In addition to an automated portion, the present invention allows a human dialog designer to model the way the system elicits information, giving a human feel to the dialog and a better customer experience. In operation, users start a dialog by directing their web browser to a designated web page. This web page asks the user some initial questions that are then passed to a dialog engine. The dialog engine then applies its methods and algorithms to a knowledge map, using dialog control information\ and the user's responses to provide feedback to the user. The feedback may include follow-up questions, relevant documents, and instructions to the user (e.g., instructions to contact a human customer service representative). This dialog engine response is rendered as a web page and returned to the user's web browser. The user can then respond further to the follow-up questions he or she is presented, and the cycle repeats. The invention can be implemented so that it can interact with customers through a wide variety of communication channels including the Internet, wireless devices (e.g., telephone, pager, etc.), handheld devices such as a Personal Data Assistant (PDA), email, and via a telephone where the automated system is delivered using an interactive voice response (IVR) and/or speech-recognition system.

590 citations


Proceedings Article
11 Sep 2001
TL;DR: Lixto, of which a working prototype has been implemented, assists the user to semi-automatically create wrapper programs by providing a fully visual and interactive user interface that helps to create very expressive extraction programs.
Abstract: We present new techniques for supervised wrapper generation and automated web information extraction, and a system called Lixto implementing these techniques. Our system can generate wrappers which translate relevant pieces of HTML pages into XML. Lixto, of which a working prototype has been implemented, assists the user to semi-automatically create wrapper programs by providing a fully visual and interactive user interface. In this convenient user-interface very expressive extraction programs can be created. Internally, this functionality is reected by the new logicbased declarative language Elog. Users never have to deal with Elog and even familiarity with HTML is not required. Lixto can be used to create an \XML-Companion" for an HTML web page with changing content, containing the continually updated XML translation of the relevant information.

Proceedings ArticleDOI
01 Apr 2001
TL;DR: IEPAD is proposed, a system that automatically discovers extraction rules from Web pages that can automatically identify record boundary by repeated pattern mining and multiple sequence alignment and can achieve 97 percent extraction over fourteen popular search engines.
Abstract: research in information extraction (IE) regards the generation of wrappers that can extract particular information from semi- structured Web documents. Similar to compiler generation, the extractor is actually a driver program, which is accompanied with the generated extraction rule. Previous work in this field aims to learn extraction rules from users' training example. In this paper, we propose IEPAD, a system that automatically discovers extraction rules from Web pages. The system can automatically identify record boundary by repeated pattern mining and multiple sequence alignment. The discovery of repeated patterns are realized through a data structure call PAT trees. Additionally, repeated patterns are further extended by pattern alignment to comprehend all record instances. This new track to IE involves no human effort and content-dependent heuristics. Experimental results show that the constructed extraction rules can achieve 97 percent extraction over fourteen popular search engines.

Journal ArticleDOI
TL;DR: Perceived usefulness is associated with the four activities, while perceived ease of use and perceived enjoyment are associated with messaging, browsing and downloading activities.
Abstract: Examines demographic variables (gender, age, educational level) and motivation variables (perceived ease of use, perceived enjoyment, perceived usefulness) associated with Internet usage activities (defined in terms of messaging, browsing, downloading and purchasing). A total of 1,370 usable responses were obtained using a Web page survey. Results showed that males are more likely to engage in downloading and purchasing activities while females are more likely to engage in messaging activities. Younger users engage in messaging and downloading activities to a greater extent than older users. Perceived usefulness is associated with the four activities, while perceived ease of use and perceived enjoyment are associated with messaging, browsing and downloading activities.

Patent
20 Jul 2001
TL;DR: In this article, a method and system for interactively responding to queries from a remotely located user includes a computer server system configured to receive an instant message query or request from the user over the Internet, and appropriate action is taken such as accessing a local or remote data resource and formulating an answer to the user's query.
Abstract: A method and system for interactively responding to queries from a remotely located user (18) includes a computer server system (22) configured to receiving an instant message query or request (20) from the user over the Internet. The query or request is interpreted and appropriate action is taken, such as accessing a local or remote data resource and formulating an answer to the user's query. The answer is formatted as appropriate and returned to the user as an instant message or via another route specified by the user. A method and system of providing authenticated access to a given web page via instant messaging is also disclosed.

Patent
Michael Radwin1
10 Sep 2001
TL;DR: In this paper, an advertisement generation system presents a time-dependent advertisement to a user based upon to a search term of a search query during a period of time via a communication network, which includes a search engine coupled to a web server for receiving the search query and for presenting a search results page.
Abstract: An advertisement generation system presents a time-dependent advertisement to a user based upon to a search term of a search query during a period of time via a communication network The system includes a search engine coupled to a web server for receiving the search query and for presenting a search results page. The search results page is accompanied with an immediate advertisement. In one embodiment, the system includes an advertisement repository to store data representing a plurality of advertisements and includes the immediate advertisement and the time-dependent advertisement. Also, the system includes a user profiler configured to communicate a plurality of search characteristics and includes a data structure to store the search characteristics. Lastly, an advertisement server is included to provide to the web server the immediate advertisement with an associated search results page and the time-dependent advertisement with a web page presented after the display of the search results page, until a period of time expires.

Journal ArticleDOI
TL;DR: Ranging from simple to complex, Web services bring the promise of flexible, open-standards-based, distributed computing to the Internet.
Abstract: Web services are a new breed of Web applications. These independent application components are published on to the Web in such a way that other Web applications can find and use them. They take the Web to its next stage of evolution, in which software components can discover other software components and conduct business transactions. Examples of Web services include a credit card service that processes credit card transactions for a given account number, a market data service that provides stock market data associated with a specified stock symbol, and an airline service that provides flight schedule, availability, and reservation functionalities. Major vendors like IBM, Microsoft, Hewlett-Packard, and Sun, among others, are investing heavily in Web services technology. Ranging from simple to complex, Web services bring the promise of flexible, open-standards-based, distributed computing to the Internet.

Journal ArticleDOI
TL;DR: It is shown that web page revisitation is a much more prevalent activity than previously reported, that most pages are visited for a surprisingly short period of time, that users maintain large (and possibly overwhelming) bookmark collections, and that there is a marked lack of commonality in the pages visited by different users.
Abstract: This paper provides an empirical characterization of user actions at the web browser. The study is based on an analysis of 4 months of logged client-side data that describes user actions with recent versions of Netscape Navigator. In particular, the logged data allow us to determine the title, URL and time of each page visit, how often they visited each page, how long they spent at each page, the growth and content of bookmark collections, as well as a variety of other aspects of user interaction with the web. The results update and extend prior empirical characterizations of web use. Among the results we show that web page revisitation is a much more prevalent activity than previously reported (approximately 81% of pages have been previously visited by the user), that most pages are visited for a surprisingly short period of time, that users maintain large (and possibly overwhelming) bookmark collections, and that there is a marked lack of commonality in the pages visited by different users. These results have implications for a wide range of web-based tools including the interface features provided by web browsers, the design of caching proxy servers, and the design of efficient web sites.

Book
01 Oct 2001
TL;DR: The Role of the Homepage, The Organization of This Book, and more: Communicating the Site's Purpose and Handling Emergencies.
Abstract: Preface. The Role of the Homepage. The Organization of This Book. Your Action Items. The Companion Web Pages. Homepage Guidelines. What's Not in These Guidelines. The Guidelines. Communicating the Site's Purpose. Communicating Information About Your Company. Content Writing. Revealing Content Through Examples. Archives and Accessing Past Content. Links. Navigation. Search. Tools and Task Shortcuts. Graphics and Animation. Graphic Design. UI Widgets. Window Titles. URLs. News and Press Releases. Popup Windows and Staging Pages. Advertising. Welcomes. Communicating Technical Problems and Handling Emergencies. Credits. Page Reload and Refresh. Customization. Gathering Customer Data. Fostering Community. Dates and Times. Stock Quotes and Displaying Numbers. Homepage Design Conventions. Homepage Design Statistics. Download Time. Basic Page Layout. Page Width. Liquid Versus Frozen Layout. Page Length. Frames. Fundamental Page Design Elements. Logo. Search. Navigation. Footer Navigation. Site Map. Routing Pages. Splash Pages. Frequent Features. Sign In. About Us. Contact Info. Privacy Policy. Job Openings. Help. Graphics and Multimedia. Pictures. ALT Text. Music. Animation. Advertising. Typography. Body Text and Background Colors. Link Formatting. Recommended Homepage Design. Sites. About http://www.about.com. Accenture http://www.accenture.com. Amazon http://www.amazon.com. The Art Institute of Chicago http://www.artic.edu. Asia Cuisine http://www.asiacuisine.com.sg. Barnes & Noble http://www.bn.com. BBC Online http://www.bbc.co.uk. Boeing http://www.boeing.com. CDNOW http://www.cdnow.com. Citigroup http://www.citigroup.com. CNET http://www.cnet.com. CNNfn cnnfn.cnn.com. Coles http://www.coles.com.au. DIRECTV http://www.directv.com. Disney http://www.disney.com. Drugstore http://www.drugstore.com. eBay http://www.ebay.com. eMagazineshop http://www.emagazineshop.com. ESPN http://www.espn.com. ExxonMobil http://www.exxon.mobil.com. FedEx http://www.fedex.com/us. Federal Highway Administration http://www.fhwa.dot.gov. Florida Department of Revenue http://www.sun6.dms.state.fl.us/dor. Ford Motor Company http://www.ford.com. Gateway http://www.gateway.com. General Electric http://www.ge.com. General Motors http://www.gm.com. Global Sources http://www.globalsources.com. IBM http://www.ibm.com. James Devaney Fuel Company http://www.jamesdevaneyfuel.com. JobMagic http://www.jobmagic.net. Learn2 http://www.learn2.com. Microsoft http://www.microsoft.com. MotherNature http://www.mothernature.com. MTV http://www.mtv.com. NewScientist http://www.newscientist.com. NewsNow http://www.newsnow.co.uk. PBS http://www.pbs.org. Petsmart http://www.petsmart.com. Philip Morris http://www.philipmorris.com. PlanetRx http://www.planetrx.com. Red Herring http://www.redherring.com. Slusser's Green Thumb http://www.slussers.com. Southwest http://www.southwest.com. Ticketmaster http://www.ticketmaster.com. Travelocity http://www.travelocity.com. USA Today http://www.usatoday.com. Victoria's Secret http://www.victoriassecret.com. Wal-Mart http://www.walmart.com. Yahoo! http://www.yahoo.com. Appendix. Index.

Patent
11 Jun 2001
TL;DR: In this article, a proxy server receives web content requests from a client and translates the content from an original format into a scalable vector representation, which is then sent to the client, where it is scaled and/or offset, enabling users to zoom or pan the web content.
Abstract: Apparatus and methods are provided for creating resolution-independent vector display of Internet content to allow it to be scaled (zoomed) larger and smaller for better viewing or to fit any resolution or screen size. According to one embodiment, novel server processing of Web content is provided that converts Web content requested by a client to a scalable vector format, such as Simple Vector Format. The vector format enables the Web content to be rendered by the client such that the rendered display substantially retains an original page layout defined in mark-up language document(s) in which the Web content is stored. In one embodiment, a proxy server receives Web content requests from a client and translates the content from an original format into a scalable vector representation. The scalable vector representation is then sent to client, where it is scaled and/or offset, enabling users to zoom and/or pan the Web content.

Patent
29 Aug 2001
TL;DR: In this paper, a technique for generating markup information to be displayed on a client computer system is described, where the client system includes memory configured to store at least one update file which comprises keyword information relating to keywords suitable for markup.
Abstract: A technique is disclosed for generating markup information to be displayed on a client computer system. The client system includes memory configured to store at least one update file which comprises keyword information relating to keywords suitable for markup. In one implementation, the update file is generated at a remote server system and downloaded to the client system. When a new document (e.g. a web page) is displayed on the client system to an end user, selected context associated from the document is analyzed for selected keywords. In a specific implementation, the selected keyword information is provided by an entity other than the end user. Using the selected keyword information, specific context in the document is selected to be marked up. According to a specific embodiment, the selection of the document context to be marked up may be performed at the client system. Markup operations are then implemented at the client system on at least a portion of the selected document context, thereby resulting in marked up document context which has a visual appearance that is different than the appearance of the initial parsed context. Additionally, a pop-up advertisement may be automatically displayed on the client system based upon the identified context in the document. When the user clicks on a portion of the marked up document context, the user may be redirected to a target URL which was not included as part of the initial parsed document context. Alternatively, a pop-up window or layer which includes a plurality of different links may be displayed to the user.

Proceedings ArticleDOI
01 Nov 2001
TL;DR: How CDNs are commonly used on the Web and a methodology to study how well they perform are defined and use of a DNS lookup in the critical path of a resource retrieval does not generally result in better server choices being made relative to client response time in either average or worst case situations.
Abstract: Content distribution networks (CDNs) are a mechanism to deliver content to end users on behalf of origin Web sites. Content distribution offloads work from origin servers by serving some or all of the contents of Web pages. We found an order of magnitude increase in the number and percentage of popular origin sites using CDNs between November 1999 and December 2000.In this paper we discuss how CDNs are commonly used on the Web and define a methodology to study how well they perform. A performance study was conducted over a period of months on a set of CDN companies employing the techniques of DNS redirection and URL rewriting to balance load among their servers. Some CDNs generally provide better results than others when we examine results from a set of clients. The performance of one CDN company clearly improved between the two testing periods in our study due to a dramatic increase in the number of distinct servers employed in its network. More generally, the results indicate that use of a DNS lookup in the critical path of a resource retrieval does not generally result in better server choices being made relative to client response time in either average or worst case situations.

Patent
22 Feb 2001
TL;DR: In this article, a system for previewing contents of Web files is presented, which includes a computing platform for executing a previewing application being designed and configured for: (a) retrieving at least a portion of the contents and/or source code of a Web file; and (b) generating a preview of said web file, said preview presenting at least the content of said Web file.
Abstract: A system for previewing contents of Web files is provided. The system includes a computing platform for executing a previewing application being designed and configured for: (a) retrieving at least a portion of the contents and/or source code of a Web file; and (b) generating a preview of the contents of said Web file, said preview presenting at least a portion of the content of said Web file.

Patent
12 Sep 2001
TL;DR: In this paper, the authors present a Web server that presents a web page including content to a plurality of users via browser applications located at each user's client site, while each user views the site, detailed activities by each user are collected in a client side data store located in each client site.
Abstract: Methods and apparatus for performing dynamic Web-based market analysis. A Web server (110) present a Web page including content to a plurality of users, via browser applications located at each user's client site. While each user views the site, detailed activities by each user are collected in a client side data store located in each client site. After a completion event occurs at each client site, such as a respective client side data store fills up, the collected activity data is sent back to the Web server (110), where its is stored in a server side data store. An analytical program (115) executed by the Web server (110) retrieves the collected response data from the server side data store and performs market analysis. The analytical program (115) produces results that reflect the success of the content presented to the Web page, and these results are used by a middleware program (112) executed on the Web server (110) to update the content presented to the Web page on a 'real-time' and automatic basis.

Proceedings ArticleDOI
01 Apr 2001
TL;DR: This paper proposes the novel concept of intelligent crawling which actually learns characteristics of the linkage structure of the world wide web while performing the crawling, and refers to this technique as intelligent crawling because of its adaptive nature in adjusting to the web page linkage structure.
Abstract: The enormous growth of the world wide web in recent years has made it important to perform resource discovery e ciently. Consequently, several new ideas have been proposed in recent years; among them a key technique is focused crawling which is able to crawl particular topical portions of the world wide web quickly without having to explore all web pages. In this paper, we propose the novel concept of intelligent crawling which actually learns characteristics of the linkage structure of the world wide web while performing the crawling. Speci cally, the intelligent crawler uses the inlinking web page content, candidate URL structure, or other behaviors of the inlinking web pages or siblings in order to estimate the probability that a candidate is useful for a given crawl. This is a much more general framework than the focused crawling technique which is based on a pre-de ned understanding of the topical structure of the web. The techniques discussed in this paper are applicable for crawling web pages which satisfy arbitrary user-de ned predicates such as topical queries, keyword queries or any combinations of the above. Unlike focused crawling, it is not necessary to provide representative topical examples, since the crawler can learn its way into the appropriate topic. We refer to this technique as intelligent crawling because of its adaptive nature in adjusting to the web page linkage structure. The learning crawler is capable of reusing the knowledge gained in a given crawl in order to provide more e cient crawling for closely related predicates.

Patent
22 May 2001
TL;DR: In this article, a web server in the device provides access to the user interface functions for the device through a device web page, such that a user of the web browser accesses the user interfaces through the web page.
Abstract: Web access functionality is embedded in a device to enable low cost widely accessible and enhanced user interface functions for the device. A web server in the device provides access to the user interface functions for the device through a device web page. A network interface in the device enables access to the web page by a web browser such that a user of the web browser accesses the user interface functions for the device through the web page.

Patent
30 Jan 2001
TL;DR: The universal Internet based telephony system as mentioned in this paper is implemented as a process that is accessible via an Internet web page and executes on the web server that hosts the web page, and/or a back-end Internet telephony server which is accessed and controlled by a web server.
Abstract: The universal Internet based telephony system is implemented as a process that is accessible via an Internet web page and executes on the web server that hosts the web page and/or a back-end Internet telephony server which is accessed and controlled by the web server. The customer accesses the universal Internet based telephony system via any existing Java Enabled Internet Browser software resident on the customer's personal computer, either as an adjunct process thereon, or as a dedicated Internet telephony process. When a customer accesses the Internet web page and clicks on the universal Internet based telephony system icon, the web server on which the web page resides executes the resident universal Internet based telephony system hyperlink script and transfers a newly opened browser session to the universal Internet based telephony system web site. The web server uploads an applet to the customer's personal computer to run on the client machine without disturbing the customer's existing web page access, by opening up a separate window on the customer's personal computer for the universal Internet based telephony system, to be seen and operated by the user. The universal Internet based telephony system web site then implements the Internet telephony communication connection between the customer's personal computer and the designated destination.

Journal ArticleDOI
TL;DR: The recent development by a consortium lead by the AICPA of the so-called “eXtensible Business Reporting Language” (XBRL) is an initiative to develop an XML-based Web-based business reporting specification that would mean that both humans and intelligent software agents could operate on financial information disseminated on the Web with a high degree of accuracy and reliability.

Proceedings ArticleDOI
01 Apr 2001
TL;DR: In this article, five methods for summarizing parts of Web pages on handheld devices, such as personal digital assistants (PDAs), or cellular phones, are introduced, which can each be hidden, partially displayed, made fully visible, or summarized.
Abstract: We introduce five methods for summarizing parts of Web pages on handheld devices, such as personal digital assistants (PDAs), or cellular phones. Each Web page is broken into text units that can each be hidden, partially displayed, made fully visible, or summarized. The methods accomplish summarization by different means. One method extracts significant keywords from the text units, another attempts to find each text unit's most significant sentence to act as a summary for the unit. We use information retrieval techniques, which we adapt to the World-Wide Web context. We tested the relative performance of our five methods by asking human subjects to accomplish single-page information search tasks using each method. We found that the combination of keywords and single-sentence summaries works best for a variety of search tasks.

Patent
14 Nov 2001
TL;DR: In this paper, a service monitor and a browser monitor determine performance metrics on both a server and a client in connection with a distributed application running on a network, and a correlated performance metric is determined by combining a server performance metric with a browser performance metric.
Abstract: A service monitor and a browser monitor determine performance metrics on both a server and a client in connection with a distributed application running on a network. While applicable to other types of distributed application data, an example is described in which a Web page is requested by a user running a browser program on the client computer. In response, the server transmits the requested Web page, along with JavaScript code that defines a browser monitor, to the client. A browser monitoring function controlled by this code determines a plurality of different performance metrics related to the access of Web pages by the client. In addition, the server monitor determines performance metrics specific to its function in providing the Web page to the client. A correlated performance metric is determined by combining a server performance metric with a browser performance metric.

Journal ArticleDOI
01 Mar 2001
TL;DR: The World Wide Web Wrapper Factory (W4F) is presented, a toolkit for the generation of wrappers for Web sources that offers an expressive language to specify the extraction of complex structures from HTML pages and a declarative mapping to various data formats like XML.
Abstract: The Web so far has been incredibly successful at delivering information to human users. So successful actually, that there is now an urgent need to go beyond a browsing human. Unfortunately, the Web is not yet a well organized repository of nicely structured documents but rather a conglomerate of volatile HTML pages. To address this problem, we present the World Wide Web Wrapper Factory (W4F), a toolkit for the generation of wrappers for Web sources, that offers: (1) an expressive language to specify the extraction of complex structures from HTML pages; (2) a declarative mapping to various data formats like XML; (3) some visual tools to make the engineering of wrappers faster and easier.