scispace - formally typeset
Search or ask a question

Showing papers on "Data management published in 1977"


01 Apr 1977
TL;DR: The author identifies the integrity problems posed by a secure military computer utility and integrity policies addressing these problems are developed and their effectiveness evaluated.
Abstract: : An integrity policy defines formal access constraints which, if effectively enforced, protect data from improper modification. The author identifies the integrity problems posed by a secure military computer utility. Integrity policies addressing these problems are developed and their effectiveness evaluated. A prototype secure computer utility, Multics, is then used as a testbed for the application of the developed access controls.

1,185 citations


Book
01 Jan 1977

126 citations


Proceedings Article
06 Oct 1977
TL;DR: An overview of several key technical problems that must be overcome in developing general purpose distributed database management systems is presented, and current r&d efforts aimed at overcoming these problems are surveyed.
Abstract: Distributed database management is a newly developed and rapidly growing sub-field of database management technology. Distributed database management is an attractive approach to solving the data management needs of many organizations because it permits the database system to act conceptually as a centralized system, while physically mirroring the geographic distribution of organizations in today's world. This paper presents an overview of several key technical problems that must be overcome in developing general purpose distributed database management systems, and surveys current r&d efforts aimed at overcoming these problems.

118 citations


Book
01 Jan 1977

76 citations


Journal ArticleDOI
TL;DR: An analytic model, based upon knowledge of data item lengths, data access costs, and user retrieval patterns, is developed to assist an analyst with this assignment problem.
Abstract: It is generally believed that 80 percent of all retrieval from a commercial database is directed at only 20 percent of the stored data items. By partitioning data items into primary and secondary record segments, storing them in physically separate files, and judiciously allocating available buffer space to the two files, it is possible to significantly reduce the average cost of information retrieval from a shared database. An analytic model, based upon knowledge of data item lengths, data access costs, and user retrieval patterns, is developed to assist an analyst with this assignment problem. A computationally tractable design algorithm is presented and results of its application are described.

68 citations


Proceedings Article
22 Aug 1977
TL;DR: An effort in the application of artificial intelligence to the access of data from a large, distributed data base over a computer network that provides access to multiple instances of a data base management system over the ARPANET in real time is discussed.
Abstract: : This report discusses an effort in the application of artificial intelligence to the access of data from a large, distributed data base over a computer network. A running system is described that provides access to multiple instances of a data base management system (DBMS) over the ARPANET in real time. The system accepts a rather wide range of natural language questions about the data, plans a sequence of appropriate queries to the data base management system to answer the question, determines on which machine to carry out the queries, establishes links to those machines over the ARPANET, monitors the prosecution of the queries and recovers from certain errors in execution, and prepares a relevant answer to the original question. In addition to the functional components that make up the demonstration system, equivalent functional components with higher levels of sophistication are discussed and proposed. Keywords: LADDER(Language Access to Distributed Data with Error Recovery).

41 citations



01 Jul 1977
TL;DR: It would seem worthwhile for those responsible for management of these mechanized information storage and retrieval data bases to attempt to use all economically feasible error-detecting and correcting schemes to reduce the error rate as much as practicable.
Abstract: : The introduction of on-line interactive literature searching systems in recent years has made it possible for information scientists to conduct bibliometric studies which might have been difficult or impractical to do by manual methods. The unconventional uses of on-line information retrieval systems are becoming more common as we learn how to search using non-subject information fields. Author's name, organizational affiliation, journal's name, year of publication, etc., can now be searched for easily. However, sometimes there are problems. It would seem worthwhile for those responsible for management of these mechanized information storage and retrieval data bases to attempt to use all economically feasible error-detecting and correcting schemes to reduce the error rate as much as practicable. Several suggestions for detecting errors are examined.

31 citations


Book
01 Jan 1977
TL;DR: In this article, the authors show the best book collections and completed collections and show how to download the book in this website lists and how to get more advantages by reading the soft file of book.
Abstract: Downloading the book in this website lists can give you more advantages. It will show you the best book collections and completed collections. So many books can be found in this website. So, this is not only this management accounting and control of data processing. However, this book is referred to read because it is an inspiring book to give you more chance to get experiences and also thoughts. This is simple, read the soft file of the book and you get it.

25 citations


Journal ArticleDOI
01 Sep 1977
TL;DR: Information Storage and Retrieval encompasses a broad scope of topics ranging from basic techniques for accessing data to sophisticated approaches for the analysis of natural language text and the deduction of information.
Abstract: Information Storage and Retrieval (IS&R) encompasses a broad scope of topics ranging from basic techniques for accessing data to sophisticated approaches for the analysis of natural language text and the deduction of information. Within the field, three general areas of investigation can be distinguished not only by their subject matter but also by the types of individuals presently interested in them:(1) Document retrieval,(2) Generalized data management, and(3) Question-answering.A functional description which applies to each of the three areas is presented together with a survey of work being conducted. The similarities and differences of the three areas of IS&R are described. Typical systems which incorporate many of the functions and techniques are described in the appendix.

22 citations


01 Feb 1977
TL;DR: This paper describes recent research in building database management systems which incorporate alerting features, and provides a framework within which this and similar research efforts can be evaluated.
Abstract: : This paper discusses the subject of 'alerters', which can be used in a database management system to provide the capability of informing a user when a specified state of the world (as reflected in the database) is reached. This paper describes recent research in building database management systems which incorporate alerting features, and provides a framework within which this and similar research efforts can be evaluated.

Book
01 Jan 1977
TL;DR: This article presents a review of the book “Introduction to Management Information Systems,” by Robert G. Murdick and Joel E. Ross.
Abstract: This article presents a review of the book “Introduction to Management Information Systems,” by Robert G. Murdick and Joel E. Ross.


Proceedings ArticleDOI
01 Jan 1977
TL;DR: This discussion represents an attempt to examine and classify several of these backend data base management machine configurations in terms of their operational parameters and application constraints.
Abstract: There has been much interest in the use of special purpose processors as the data base management component of data processing systems. The generic terms “backend” and “data management machine” have been applied to such devices. Examination of the literature reveals a broad cross section of host to backend functional distribution and interconnection methodology. This discussion represents an attempt to examine and classify several of these backend data base management machine configurations in terms of their operational parameters and application constraints. A formal taxonomy of such systems remains yet to be performed.At least three distinct classes of data management machine (DMM) are evidenced in the literature; they are the large host backend, distributed network data node and smart peripheral. The intended classes of problem that the various authors envision amenable to solution by the DMM approach exhibit overlap while the performance envelope in which each DMM architecture would provide a technically acceptable, economically sound solution to a given user requirement set varies. Some of the papers used as source for this work contained no explicit mention of either the problem classes or performance constraints that the described configuration was to address; thus liberty has been taken in interpreting the implicit application goals of these authors.


Journal ArticleDOI
TL;DR: The most common use of computers in the social sciences is for statistical analysis and related data management and the statistics and data management area is a major focus.
Abstract: The most common use of computers in the social sciences is for statistical analysis and related data management. The pervasiveness of statistical data processing has even resulted in its sometimes being considered all there is to social science computing. Many of the social science computing texts deal primarily with statistical data issues (e.g., Brier and Robinson, 1974; Harkins, Isenhour, and Jurs, 1973; Lehman and Bailey, 1968; Meyers, 1973; Rattenbury and Pelletier, 1974). Anderson’s (I974a) bibliography on social science computing lists 113 references in the statistics and data management area. The more recent and current literature is found in several periodic publications: Review of Public Data Use, a quarterly journal: Annals of Economic and Social Measurement, a journal published quarterly by the National Bureau of Economic Research; and Proceedings of the annual &dquo;Computer Science and Statistics Interface&dquo; symposia.~ Statistical program abstracts are regularly published in Behavior Research Methods and Instrumentation, Educational and Psychological Measurement, and the Journal of Marketing Research. Various articles including comparative evaluations of programs appear in American Statistician.


01 Jun 1977
TL;DR: The network database model studied in this report will be referred to as the D BTG data model and the emphasis will be on concepts rather than the detailed syntax of the DBTG language specifications.
Abstract: : This is the second of a series of reports aimed at studying the capabilities of a database computer, known as the DBC, in supporting the three major data models: hierarchical, network and relational. In the first report, DBC software requirements for handling hierarchical databases have been presented. This report will be directed towards an investigation of the software requirements for network databases. Relational database systems will be treated in a forthcoming report. The April 1971 report of the CODASYL database Task Group (DBTG) is chosen as the definitive document of network databases and systems since most commercially available network data base management systems are based on the DBTG report. Even though many of these commercial systems use a syntax that is slightly different from the DBTG specifications, the main concepts have been retained. The network database model studied in this report will be referred to as the DBTG data model. Our emphasis will be on concepts rather than the detailed syntax of the DBTG language specifications. Database computers are a recent addition to the family of computers. With the advent of large databases, there has been a growing awareness of the necessity of a computer architecture that is oriented towards storage, retrieval and manipulation of large quantities of information.


Proceedings ArticleDOI
13 Jun 1977
TL;DR: The prototype is a highly integrated, interactive, minicomputer based data management and analysis system that allows the clinical investigator to store, recall, analyze, and display his research data without resorts to computer programming.
Abstract: This paper discusses a prototype system intended for the personal use of physicians engaged in clinical research. In particular, the prototype is a highly integrated, interactive, minicomputer based data management and analysis system. The facilities offered by the system allow the clinical investigator to store, recall, analyze, and display his research data without resorting to computer programming. Modern data base techniques are available to the physician as aids in organizing, storing, and retrieving his data. The data base concepts are expressed in terms that are familiar to a clinical researcher.

Proceedings Article
06 Oct 1977
TL;DR: Conditions for data independence are described and the concepts of Extended Set Theory are summarised as a general model for expressing information systems embodying data independence for very large, distributed, backend information systems.
Abstract: Three distinct components comprise an Information System: INFORMATION MANAGEMENT, DATA MANAGEMENT, and STORAGE MANAGEMENT. Until recently, all three have been subsumed under data management. As applications become more demanding, as support criteria become more complex, and as storage capacity becomes very large the need for functional independence of these three management areas has become more apparent. Recognition of this situation has been popularized through the phrase, "data independence", or more precisely, "data independence from information" and "data independence from storage". The difficulty in achieving data independence arises through the incompatibility of a complex information space being supported by a simple storage space. The popular, but limiting approach, has been to force the information space into a restrictive record space. This achieves a deceptive compatibility allowing only the appearance of data independence at the user level. This recordoriented approach has become pervasive for small databases even though it constrains user applications, requires substantial storage overhead, and imposes inherent processing inefficiencies. As databases become very large and as distributed systems become desirable the need for inherent (not superficial) data independence becomes crucial. This paper is intended as a tutorial and will describe conditions for data independence and summarise the concepts of Extended Set Theory as a general model for expressing information systems embodying data independence. This generality will be demonstrated by considering some major problems pertinent to the design and support of very large, distributed, backend information systems. It should be emphasized that Extended Set Theory is a formalism for expressing solutions and is not a specific solution in itself. Though "redundant membership condition", "distributed membership condition", and "set-theoretic interface" may be new concepts, Extended Set Theory does not preclude any current DBMS concepts, data structures, or existing implementations. Rather, Extended Set Theory embraces them all under a unifying model.

01 Apr 1977
TL;DR: The capability of the database computer (DBC) for supporting hierarchical data models and systems is shown and it is shown that it is possible to design an interface, known as IMSI, between IMS users and the DBC.
Abstract: : This report shows the capability of the database computer (DBC) for supporting hierarchical data models and systems. Since IBM's Information Management System (IMS) is the most widely used data management system which supports a hierarchical view of data, we intend to show that the DBC can support IMS databases and IMS application programs. Furthermore, this study provides us with a case for comparing the merits of using a conventional general-purpose computer versus a special-purpose database computer (i.e., the DBC) for data management. It is shown that it is possible to design an interface, known as IMSI between IMS users and the DBC. The IMSI can faithfully execute the DL/1 calls (the data manipulation language of IMS) issued by IMS users. The design of the IMSI is considered in two phases. In the first phase, we show how to represent an IMS database utilizing the built-in (hardware) data structure of the DBC. This representation makes use of the concept of embedding symbolic identifiers into all dependent segments of an IMS database. The use of symbolic identifiers increases the degree of data independence of the stored database. Furthermore, the storage requirement for the symbolic identifiers is substantially offset by the removal of the conventional address pointers currently used in an IMS database.


Proceedings ArticleDOI
20 Jul 1977
TL;DR: A generalized high level query language has been developed for use with a network data base, as defined by the CODASYL DBTG report, and of particular interest is the capability of manipulating graphical entities with this language.
Abstract: A generalized high level query language has been developed for use with a network data base, as defined by the CODASYL DBTG report. Of particular interest is the capability of manipulating graphical entities with this language. The system was developed to supportan offshore oil lease information project where one must formulate queries pertaining to proximity of graphical entities. To facilitate such requests new relational operators were defined and implemented. Thus compound Boolean expressions can be formed which involve graphical as well as non-graphical criteria. Since one goal of the information system is to produce selectively colored maps via an FR80 microfilm recorder, the query language syntax provides the user with a comprehensive thematic cartography capability. The query language is designed to isolate the user from the underlying network structure of the data base. In order to accomplish this generalized pathfinding and accessing algorithms had to be developed. Pathfinding is facilitated through an adjacency matrix involving all record types in the data base, while accession is controlled by examining the local owner-member relation for current record types. Besides querying, the system provides the usual complement of data base management system functions, i.e. deletion, updating, reporting, etc.

Journal ArticleDOI
TL;DR: Recent conceptual and empirical work related to development of decision-oriented frameworks for management information systems design is surveyed, particularly as related to the improving the management of organizations.
Abstract: Despite the rapid growth in the use of computers in organizations, few of the resulting systems have had a significant impact on the way in which management makes decisions. Frameworks are needed which aid in understanding the structure of management information systems, toward providing focus and improving the effectiveness of systems efforts. This paper surveys recent conceptual and empirical work related to development of decision-oriented frameworks for management information systems design, particularly as related to the improving the management of organizations.

Proceedings ArticleDOI
03 Aug 1977
TL;DR: The Access Path Specification Language (APSL) is a high-level essentially nonprocedural language for specifying restructuring transformations of network databases that does so in terms of application-oriented concepts such as access strategies and selection criteria.
Abstract: The Access Path Specification Language (APSL) is a high-level essentially nonprocedural language for specifying restructuring transformations of network databases. It does so in terms of application-oriented concepts such as access strategies and selection criteria. APSL's approach to restructuring emphasizes description of the source structures from which target data is to be retrieved, rather than the operations needed to convert source constructs to target constructs. The latter approach is used by most current systems to restructure hierarchical data.APSL is based on the Relational Interface Model (RIM) of data, which permits a suitably restricted network database to be viewed simultaneously as a relational database in first normal form. Any restructuring of a RIM database which can be performed by the relational algebra operations Join, Select, Project, and Union can be described in APSL.APSL has been implemented as the restructuring specifications language in the Michigan Data Translator.


01 Jan 1977
TL;DR: The approach is to map a high-level description of user requirements in terms of individual data items and binary relations into a set of record structures and record relationships to produce a minimum number of page faults for a pre-specified set of user activities.
Abstract: : Report is concerned with the development of a methodology for partially automating the design of the logical structure of a paged data base The approach is to map a high-level description of user requirements in terms of individual data items and binary relations into a set of record structures and record relationships Subject to storage and security constraints, the goal is to produce a minimum number of page faults for a pre-specified set of user activities (Author)

Journal ArticleDOI
TL;DR: Results are presented of exploratory studies on the use of a minicomputer in conjunction with large-scale computers to perform structural design tasks, including data and program management, use of interactive graphics, and computations for structural analysis and design.