64 results on '"Boyce–Codd normal form"'
Search Results
2. Embedded Functional Dependencies and Data-completeness Tailored Database Design.
- Author
-
ZIHENG WEI and LINK, SEBASTIAN
- Subjects
- *
DATABASE design , *ELECTRONIC data processing - Abstract
We establish a principled schema design framework for data with missing values. The framework is based on the new notion of an embedded functional dependency, which is independent of the interpretation of missing values, able to express completeness and integrity requirements on application data, and capable of capturing redundant data value occurrences that may cause problems with processing data that meets the requirements. We establish axiomatic, algorithmic, and logical foundations for reasoning about embedded functional dependencies. These foundations enable us to introduce generalizations of Boyce-Codd and Third normal forms that avoid processing difficulties of any application data, or minimize these difficulties across dependency-preserving decompositions, respectively. We show how to transform any given schema into application schemata that meet given completeness and integrity requirements, and the conditions of the generalized normal forms. Data over those application schemata are therefore fit for purpose by design. Extensive experiments with benchmark schemata and data illustrate the effectiveness of our framework for the acquisition of the constraints, the schema design process, and the performance of the schema designs in terms of updates and join queries. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
3. Relational database schema design for uncertain data.
- Author
-
Link, Sebastian and Prade, Henri
- Subjects
- *
DATABASE design , *RELATIONAL databases , *ASSOCIATE degree education , *REDUNDANCY in engineering , *CERTAINTY - Abstract
Driven by the dominance of the relational model, we investigate how the requirements of applications on the certainty of functional dependencies can improve the outcomes of relational database schema design. For that purpose, we assume that tuples are assigned a degree of possibility with which they occur in a relation, and that functional dependencies are assigned a dual degree of certainty which says to which tuples they apply. A design theory is developed for functional dependencies with degrees of certainty, including efficient axiomatic and algorithmic characterizations of their implication problem. Naturally, the possibility degrees of tuples bring forward different degrees of data redundancy, caused by functional dependencies with the dual degree of certainty. Variants of the classical syntactic Boyce–Codd and Third Normal Forms are established. They are justified semantically in terms of eliminating data redundancy and update anomalies of given degrees, and minimizing data redundancy of given degrees across all dependency-preserving decompositions, respectively. As a practical outcome of our results, designers can simply fix the degree of certainty they target, and then apply classical decomposition and synthesis to the set of functional dependencies whose associated degree of certainty meets the target. Hence, by fixing the certainty degree a designer controls which integrity requirements will be enforced for the application and which data will be processed by the application. The choice of the certainty degree also balances the classical trade-off between query and update efficiency on future database instances. Our experiments confirm the effectiveness of our control parameter, and provide original insight into classical normalization strategies and their implementations. • Uncertainty in data is modeled by assigning to records a degree of possibility. • Certainty degrees of functional dependencies indicate on which records they hold. • Different degrees of data redundancy can then be defined and observed. • Refined normal forms avoid or minimize data redundancy up to a target degree. • Certainty degrees control classical trade-offs to optimize relational schema design. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
4. Some Remarks on Relational Database Schemes Having Few Minimal Keys
- Author
-
Biskup, Joachim, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Nierstrasz, Oscar, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Sudan, Madhu, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Vardi, Moshe Y., Series editor, Weikum, Gerhard, Series editor, Düsterhöft, Antje, editor, Klettke, Meike, editor, and Schewe, Klaus-Dieter, editor
- Published
- 2012
- Full Text
- View/download PDF
5. SQL schema design: foundations, normal forms, and normalization.
- Author
-
Köhler, Henning and Link, Sebastian
- Subjects
- *
SQL , *DATABASE management , *REAL-time computing , *ALGORITHMS , *INFORMATION storage & retrieval systems - Abstract
Normalization helps us find a database schema at design time that can process the most frequent updates efficiently at run time. Unfortunately, relational normalization only works for idealized database instances in which duplicates and null markers are not present. On one hand, these features occur frequently in real-world data compliant with the industry standard SQL, and especially in modern application domains. On the other hand, the features impose challenges that make it difficult to extend the existing forty year old normalization framework to SQL, and any current extensions are fairly limited. We introduce a new class of functional dependencies and show that they provide the right notion for SQL schema design. Axiomatic and linear-time algorithmic characterizations of the associated implication problem are established. These foundations enable us to propose a Boyce–Codd normal form for SQL. We justify the normal form by showing that it permits precisely those SQL instances which are free from data redundancy. Unlike the relational case, there are SQL schemata that cannot be converted into Boyce–Codd normal form. Nevertheless, for an expressive sub-class of our functional dependencies we establish a normalization algorithm that always produces a schema in Value-Redundancy free normal form. This normal form permits precisely those instances which are free from any redundant data value occurrences other than the null marker. Experiments show that our functional dependencies occur frequently in real-world data and that they are effective in eliminating redundant values from these data sets without loss of information. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
6. Data-completeness Tailored Database Design with Embedded Functional Dependencies.
- Author
-
Ziheng Wei and Link, Sebastian
- Subjects
- *
ALGORITHMS , *DATABASE design , *FUNCTIONAL dependencies , *SCHEMAS (Psychology) , *REDUNDANCY (Linguistics) - Abstract
We establish a robust schema design framework for data with missing values. The framework is based on the new notion of an embedded functional dependency, which is independent of the interpretation of missing values, able to express completeness and integrity requirements on application data, and capable of capturing many redundant data value occurrences. We establish axiomatic, algorithmic, and logical foundations for reasoning about embedded functional dependencies. These foundations allow us to establish generalizations of Boyce-Codd and Third normal forms that do not permit any redundancy in any future application data, or minimize their redundancy across dependency-preserving decompositions, respectively. We show how to transform any given schema into application schemata that meet given completeness and integrity requirements and the conditions of the generalized normal forms. Data over those application schemata are therefore fit for purpose by design. Extensive experiments with benchmark schemata and data illustrate our framework, and the effiectiveness and efficiency of our algorithms, but also provide quantified insight into database schema design trade-offs. [ABSTRACT FROM AUTHOR]
- Published
- 2019
7. A Closer Look at the Join-Equality Constraint
- Author
-
Skagestein, Gerhard, Normann, Ragnar, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Nierstrasz, Oscar, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Sudan, Madhu, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Vardi, Moshe Y., Series editor, Weikum, Gerhard, Series editor, Meersman, Robert, editor, Tari, Zahir, editor, and Herrero, Pilar, editor
- Published
- 2008
- Full Text
- View/download PDF
8. Boyce-Codd Normal Form
- Author
-
Arenas, Marcelo, Libkin, Leonid, Section Editor, Liu, Ling, editor, and Özsu, M. Tamer, editor
- Published
- 2018
- Full Text
- View/download PDF
9. Embedded functional dependencies and data-completeness tailored database design
- Author
-
Sebastian Link and Ziheng Wei
- Subjects
Interpretation (logic) ,Theoretical computer science ,Computer science ,General Engineering ,Database schema ,02 engineering and technology ,Third normal form ,Boyce–Codd normal form ,Missing data ,Database design ,020204 information systems ,Schema (psychology) ,0202 electrical engineering, electronic engineering, information engineering ,Redundancy (engineering) ,Benchmark (computing) ,020201 artificial intelligence & image processing ,Completeness (statistics) ,Functional dependency ,Information Systems - Abstract
We establish a principled schema design framework for data with missing values. The framework is based on the new notion of an embedded functional dependency, which is independent of the interpretation of missing values, able to express completeness and integrity requirements on application data, and capable of capturing redundant data value occurrences that may cause problems with processing data that meets the requirements. We establish axiomatic, algorithmic, and logical foundations for reasoning about embedded functional dependencies. These foundations enable us to introduce generalizations of Boyce-Codd and Third normal forms that avoid processing difficulties of any application data, or minimize these difficulties across dependency-preserving decompositions, respectively. We show how to transform any given schema into application schemata that meet given completeness and integrity requirements, and the conditions of the generalized normal forms. Data over those application schemata are therefore fit for purpose by design. Extensive experiments with benchmark schemata and data illustrate the effectiveness of our framework for the acquisition of the constraints, the schema design process, and the performance of the schema designs in terms of updates and join queries.
- Published
- 2019
- Full Text
- View/download PDF
10. Dealing with Dangerous Data: Part-Whole Validation for Low Incident, High Risk Data.
- Author
-
Chua, Cecil Eng Huang and Storey, Veda C.
- Subjects
DATA protection ,DATABASE security ,DATABASE administration ,FRAUD prevention ,COMMERCIAL crimes ,SECURITY systems - Abstract
In certain situations, syntactically valid, but incorrect, data entered into a database can result in nearimmediate, catastrophic financial losses for an organization. Examples include: omitting zeros in prices of goods on e-commerce sites; and financial fraud where data is directly entered into databases, bypassing application-level financial checks. Such "dangerous data" can, and should, be detected, because it deviates substantially from the statistical properties of existing data. Detection of this kind of problem requires comparing individual data items to a large amount of existing data in the database at run-time. Furthermore, the identification of errors is probabilistic, rather than deterministic, in nature. This research proposes part-whole validation as an approach to addressing the dangerous data situation. Part-whole validation addresses fundamental issues in database management, for example, integrity maintenance. Illustrative and representative examples are first defined, and analyzed. Then, an architecture for part-whole validation is presented and implemented in a prototype to illustrate the feasibility of the research. [ABSTRACT FROM AUTHOR]
- Published
- 2016
- Full Text
- View/download PDF
11. Boyce-Codd Normal Form
- Author
-
Arenas, Marcelo, LIU, LING, editor, and ÖZSU, M. TAMER, editor
- Published
- 2009
- Full Text
- View/download PDF
12. Modernization of the Second Normal Form and Boyce-Codd Normal Form for Relational Theory
- Author
-
Kseniia Ulianytska, Valerii Kolesnik, Oleksandr Amons, and Oleksandr Rolik
- Subjects
Set (abstract data type) ,Information retrieval ,Relational theory ,Relational database ,Computer science ,Database schema ,Normalization (sociology) ,Third normal form ,Second normal form ,Boyce–Codd normal form - Abstract
The task of designing relational databases has always been the subject of scientific research, as it is associated with a number of interrelated steps. The result of each step is the development of models for presenting the future database with further refinement and, finally, the creation of an adequate relational database as a set of relations with the corresponding links between them. The article focuses on the normalization of databases, as one of the steps to create a datalogical model, namely, the use of the first three Normal Forms. As a result of the analysis carried out in the article, it was concluded that the definition of the Second Normal Form can be modernized and thus achieve two goals: to ensure the correct creation of potential primary keys and, thereafter, the correct external connections between relations, even before creating the data schema in a specific relational database. Moreover, to reconsider the necessity of applying the so-called “strengthened” or a higher version of the Third Normal Form, which speaks of mutual dependencies between key and non-key attributes.
- Published
- 2020
- Full Text
- View/download PDF
13. SPECIAL ATTRIBUTES FOR DATABASE NORMAL FORMS DETERMINATION.
- Author
-
Cotelea, Vitalie
- Subjects
ALGORITHM research ,DATABASES ,SET theory ,MATHEMATICS research ,DECISION making - Abstract
The article deals with the relational schemes defined on the reduced sets of functional dependencies divided into equivalence classes. The concepts of non-essential and recoverable attributes are introduced, and the algorithms for their computing are proposed. Some properties of schemes, based on these attributes, are presented as well. Based on non-essential and recoverable attributes some conditions are introduced to make a relation be in the third or Boyce-Codd normal form. These conditions are just sufficient, because there could be a scheme that is in third or Boyce-Codd normal form but doesn't satisfy them. In addition, the article suggests a normalization algorithm that takes into account these conditions. [ABSTRACT FROM AUTHOR]
- Published
- 2012
14. Teaching Database Modeling and Design: Areas of Confusion and Helpful Hints.
- Author
-
Philip, George C.
- Subjects
- *
DATABASES , *MATHEMATICAL models , *STUDENTS , *RELATIONAL databases , *TEXTBOOKS , *TEACHING aids - Abstract
This paper identifies several areas of database modeling and design that have been problematic for students and even are likely to confuse faculty. Major contributing factors are the lack of clarity and inaccuracies that persist in the presentation of some basic database concepts in textbooks. The paper analyzes the problems and discusses ways to minimize them. Specifically, the paper discusses the practical role of normalization in database design, addresses the confusion in representing attributes with repeating values, discusses how to remove inconsistencies in defining relations and first normal form, simplifies the process of identifying candidate keys to normalize relations, clarifies the conditions under which insertion and deletion anomalies may occur, and sheds light on the confusion in defining weak entities. Normalization plays a vital role in both the theory and the practice of database design. The topdown approach popularly used in relational database design creates a conceptual schema that is represented by entity-relationship (E-R) models, and then uses mapping rules to convert the conceptual schema to relation schemas. Because E-R modeling is an intuitive process, errors could occur in identifying entities and their relationships, resulting in un-normalized relations. Unnormalized relations also could result from converting files in legacy systems and spreadsheets to relational tables. Normalization plays a key role in verifying the goodness of design of such relations and in improving the design. The concept of repeating values in relations plays a major role in defining relations and first normal form. Yet, textbooks in general do not distinguish between multi-valued and single-valued attributes in a schema. This lack of clarity may result in conflicting interpretations of the schema. The paper presents a simple solution to the problem. The lack of clarity in defining the terms tables, relations, and first normal form (1NF) in textbooks is another potential source of confusion. Some books define relation as a table with no duplicate tuples, and only atomic values. These books then redundantly define 1NF as a relation with only atomic values. Others define a relation as a table with columns and rows, and state that a relation is in 1NF if each value is atomic. These definitions fail to specify an important requirement of 1NF that there are no duplicate tuples. A third definition of 1NF that fails to include this property is that a table is in first normal form if each value is atomic. A challenging task for many students during the normalization process is checking whether a determinant is a candidate key. The standard method is to check whether every attribute of the relation is functionally dependent on the determinant. The paper presents a method that involves only the determinants, and therefore makes it easier to identify candidate keys. The paper also provides an alternate definition of Boyce-Codd Normal Form (BCNF), which is easier to apply. Discussions in textbooks and other literature on the topic of normalization often give students the impression that data redundancy in un-normalized relations leads to all three types of anomalies -- insertion, deletion, and update. The paper shows that though data redundancy generally results in insertion and deletion anomalies, that is not always the case. The conditions under which insertion/ deletion anomalies don't occur are discussed. Guidelines for mapping conceptual schema to relational tables often use the terms strong entity and weak entity to provide separate mapping rules for each. It is shown that the definition of weak entity as presented in many textbooks, however, is inaccurate. These books define weak entity using logical dependence rather than identifier dependence of entities.… [ABSTRACT FROM PUBLISHER]
- Published
- 2007
- Full Text
- View/download PDF
15. Relational database schema design for uncertain data
- Author
-
Sebastian Link, Henri Prade, Centre National de la Recherche Scientifique - CNRS (FRANCE), Institut National Polytechnique de Toulouse - Toulouse INP (FRANCE), Université Toulouse III - Paul Sabatier - UT3 (FRANCE), Université Toulouse - Jean Jaurès - UT2J (FRANCE), Université Toulouse 1 Capitole - UT1 (FRANCE), University of Auckland - UOA (NEW ZEALAND), Institut de Recherche en Informatique de Toulouse - IRIT (Toulouse, France), University of Auckland [Auckland], Argumentation, Décision, Raisonnement, Incertitude et Apprentissage (IRIT-ADRIA), Institut de recherche en informatique de Toulouse (IRIT), Université Toulouse 1 Capitole (UT1), Université Fédérale Toulouse Midi-Pyrénées-Université Fédérale Toulouse Midi-Pyrénées-Université Toulouse - Jean Jaurès (UT2J)-Université Toulouse III - Paul Sabatier (UT3), Université Fédérale Toulouse Midi-Pyrénées-Centre National de la Recherche Scientifique (CNRS)-Institut National Polytechnique (Toulouse) (Toulouse INP), Université Fédérale Toulouse Midi-Pyrénées-Université Toulouse 1 Capitole (UT1), Université Fédérale Toulouse Midi-Pyrénées, Centre National de la Recherche Scientifique (CNRS), SIGWEB, SIGIR, Mukhopadhya, Snehasis, Zhai, ChengXiang, Bertino, Elisa, Crestani, Fabio, and Mostafa, Javed
- Subjects
Theoretical computer science ,Relation (database) ,Relational database ,Computer science ,media_common.quotation_subject ,Data redundancy ,02 engineering and technology ,Third normal form ,computer.software_genre ,Boyce–Codd normal form ,[INFO.INFO-AI]Computer Science [cs]/Artificial Intelligence [cs.AI] ,020204 information systems ,Data integrity ,Functional dependency ,0202 electrical engineering, electronic engineering, information engineering ,[INFO]Computer Science [cs] ,Computer Science::Databases ,media_common ,Superkey ,Uncertain data ,Qualitative reasoning ,Certainty ,Intelligence artificielle ,Normalization ,Database schema design ,Hardware and Architecture ,Relational model ,020201 artificial intelligence & image processing ,Data mining ,Tuple ,computer ,Software ,Information Systems - Abstract
International audience; Driven by the dominance of the relational model, we investigate how the requirements of applications on the certainty of functional dependencies can improve the outcomes of relational database schema design. For that purpose, we assume that tuples are assigned a degree of possibility with which they occur in a relation, and that functional dependencies are assigned a dual degree of certainty which says to which tuples they apply. A design theory is developed for functional dependencies with degrees of certainty, including efficient axiomatic and algorithmic characterizations of their implication problem. Naturally, the possibility degrees of tuples bring forward different degrees of data redundancy, caused by functional dependencies with the dual degree of certainty. Variants of the classical syntactic Boyce–Codd and Third Normal Forms are established. They are justified semantically in terms of eliminating data redundancy and update anomalies of given degrees, and minimizing data redundancy of given degrees across all dependency-preserving decompositions, respectively. As a practical outcome of our results, designers can simply fix the degree of certainty they target, and then apply classical decomposition and synthesis to the set of functional dependencies whose associated degree of certainty meets the target. Hence, by fixing the certainty degree a designer controls which integrity requirements will be enforced for the application and which data will be processed by the application. The choice of the certainty degree also balances the classical trade-off between query and update efficiency on future database instances. Our experiments confirm the effectiveness of our control parameter, and provide original insight into classical normalization strategies and their implementations.
- Published
- 2019
- Full Text
- View/download PDF
16. Uniform probabilistic generation of relation instances satisfying a functional dependency.
- Author
-
Berens, Maximilian, Biskup, Joachim, and Preuß, Marcel
- Subjects
- *
COMBINATORICS , *ALGORITHMS , *DISTRIBUTION (Probability theory) , *CONCEPTUAL design , *COMPUTER software development - Abstract
Software engineering includes the runtime evaluation of a prototype by experiments with carefully selected sample inputs. For the development of software intended to operate on relation instances for a given relational schema with a single functional dependency, we are then challenged to generate appropriate sample instances of increasing size. Moreover, studying the impact of varying the sizes of the attribute domains might be important. We focus on seeing uniformly distributed collections of sample instances to be appropriate. If the schema is normalized (in Boyce–Codd normal form with the left-hand side of the functional dependency forming a unique key), then a straightforward heuristic insertion procedure complies with our requirements. However, for a non-normalized schema (the left-hand side not forming a key), based on a complex combinatorial analysis and exploiting an algorithm for restricted integer partitions, we develop a sophisticated probabilistic procedure of high computational complexity for generating any element of such a collection with equal probability. Moreover, we demonstrate that simpler approaches based on uniform local selections, including the heuristic insertion procedure, fail to achieve global uniformity. The conceptual design and analysis are complemented by an experimental evaluation of a prototype implementation. • Schema normalization affects the complexity of uniform random instance generation. • Naive approaches fail to achieve uniform distribution for non-normalized schemas. • Duplicate-freeness and dependency satisfaction might strongly interfere. • Achieving uniformity has to observe number-theoretic issues like integer partitions. • Cases of exponential runtimes require substantial optimization and approximation. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
17. The Problem of Equivalence for Entity-Relationship Diagrams.
- Author
-
Jajodia, Sushi, Ng, Peter A., and Springsteel, Frederick N.
- Subjects
- *
RELATIONAL databases , *ENTITY-relationship modeling , *SOFTWARE compatibility , *DATABASE design , *NORMAL forms (Mathematics) , *SOFTWARE engineering - Abstract
We investigate the question of when two entity-relationship diagrams (ERD's) should be considered equivalent, in the sense of representing the same information. This question is very important for a database design process which uses the ERD model, and can be interpreted in various ways. We give three natural and increasingly stricter criteria for developing concepts of equivalence for ERD's. We first give a notion of "domain data compatibility" which ensures that the ERD's in question represent the same universe of data in an aggregate sense. Then we define the set of functional dependencies which are naturally embedded in each ERD, and use it to develop a concept of "data dependency equivalence" which ensures that the ERD's satisfy the same constraints (functional dependencies) among the represented data. Finally, we give our strongest criterion, instance data equivalence, which requires the ERD's to have the same power to represent instances of data. We develop several alternate forms of this third notion, including some giving efficient tableaux tests for its occurrence, Indeed, for each type of equivalence, we give a polynomial-time algorithm to test for it. We also show how certain natural decompositions applied to an ERD scheme can put each relation scheme in its canonical database scheme into (at least) third normal form, without loss of the semantic information captured in fd's. These decompositions have certain other desirable properties with respect to instance data representability. [ABSTRACT FROM AUTHOR]
- Published
- 1983
18. Equivalent key problem of the relational database model
- Author
-
Kambayashi, Yahiko, Goos, G., editor, Hartmanis, J., editor, Brinch Hansen, P., editor, Gries, D., editor, Moler, C., editor, Seegmüller, G., editor, Stoer, J., editor, Wirth, N., editor, Blum, E. K., editor, Paul, M., editor, and Takasu, S., editor
- Published
- 1979
- Full Text
- View/download PDF
19. Using Fuzzy Logic to Evaluate Normalization Completeness for an Improved Database Design
- Author
-
Nayyar Iqbal, Mahaboob Sharief Shaik, and M. Rizwan Jameel Qureshi
- Subjects
FOS: Computer and information sciences ,Normalization (statistics) ,Goto ,Computer science ,Databases (cs.DB) ,computer.software_genre ,Boyce–Codd normal form ,Database design ,Fuzzy logic ,Schema transformation ,Computer Science - Databases ,Data mining ,Functional dependency ,computer - Abstract
A new approach, to measure normalization completeness for conceptual model, is introduced using quantitative fuzzy functionality in this paper. We measure the normalization completeness of the conceptual model in two steps. In the first step, different normalization techniques are analyzed up to Boyce Codd Normal Form (BCNF) to find the current normal form of the relation. In the second step, fuzzy membership values are used to scale the normal form between 0 and 1. Case studies to explain schema transformation rules and measurements. Normalization completeness is measured by considering completeness attributes, preventing attributes of the functional dependencies and total number of attributes such as if the functional dependency is non-preventing then the attributes of that functional dependency are completeness attributes. The attributes of functional dependency which prevent to go to the next normal form are called preventing attributes., Comment: 8 Pages
- Published
- 2012
- Full Text
- View/download PDF
20. Derivation of Database Keys’ Operations
- Author
-
Olusegun Folorunso, Adesina S. Sodiya, and A. T. Akinwale
- Subjects
lcsh:T58.5-58.64 ,Database ,lcsh:Information technology ,Computer science ,Relational database ,Semi-structured model ,Third normal form ,Boyce–Codd normal form ,computer.software_genre ,Database normalization ,First normal form ,Relational model ,Second normal form ,computer - Abstract
Introduction Designing database is an art process similar to building a house. Database designers always face the problems of designing a relational database that will be free of database anomalies. These anomalies bring repetition of tuples that delay processing time and occupy memory spaces. Suppose that the value of the attribute BUILDER determines values of the attribute MODEL and PRICE, (BUILDER [right arrow] MODEL, PRICE) and that the value for the attribute MODEL determines the value for PRICE, (MODEL [right arrow] PRICE). Grouping these attributes in relation HOUSE(BUILDER, MODEL, PRICE) has several undesirable properties. First the relationship between MODEL and PRICE is repeated in the relation for each BUILDER who builds a particular MODEL of home. This repetition creates difficulties if a BUILDER who happens to be the last BUILDER of a certain MODEL home is deleted from the relation, then the relationship between the MODEL and its PRICE also disappears from the relation. This is called a deletion anomaly. Similarly, if a new builder who happens to be the first BUILDER of a certain MODEL home is added then the relationship between MODEL of a home and its PRICE will also be added. This is called an insertion anomaly. Suppose that the relationship between a MODEL and its PRICE is changed e.g. the price is increased; then the MODEL and PRICE relationship should be affected for every BUILDER of the MODEL. This is called update anomaly. These anomalies are undesirable since the user is not likely to realize the consequence of the insertion, deletion or updating. The user may inadvertently affect a relationship that was not intended to be modified. Consistency, insertion, deletion and updating are not probe effecting all groupings of attributes. If the relation HOUSE(BUILDER, MODEL, PRICE) is normalized then the consistency and anomaly problems disappear. Normalization is a step by step reversible process of replacing a given collection of relations by successive collection in which the relations have a progressively simpler and more regular structure (Date & Darwen, 2000). The reversibility guarantees that the original collection of relations can be recovered and therefore no information has been lost. Codd proposed three normal forms which he called first normal form (1NF), second normal form (2NF) and third normal form (3NF). A stronger definition of 3NF was proposed by Boyce and Codd and is known as Boyce-Codd Normal Form (BCNF). All these normal forms except 1NF are based on the functional dependencies among the attributes of a relation (Elmasri & Navathe, 1994). First normal form relates to the structure of the relation. It requires that every attribute of a relation be based on a simple domain. The database designers have no problem to know if a relation violates first normal form. They can put the relation into first normal form algorithmically by replacing a non-simple domain by its constituent simple domains. In the second (2NF), third (3NF) and Boyce Codd normal form (BCNF), there is a need for the database designers to know the real meaning and application of database keys such as candidate key, primary key, super key, etc,. Problem Statement Database designers always find it difficult to determine these keys from relational database schemas. It has been difficult to motivate students and database designers to derive primary, candidate, alternative and super keys because they think this area is dry and theoretical. There are many algorithms to determine the database keys but they look abstract for students. Many database researchers indicated that relational database model to derive database keys tends to be complex for the average designers. Failure to determine the database keys at times leads to poor design that can generate database anomalies. The database key algorithms often require extensive relational algebraic backgrounds that database designers lack. …
- Published
- 2011
- Full Text
- View/download PDF
21. Normalization
- Author
-
Jan L. Harrington
- Subjects
Database normalization ,First normal form ,Theoretical computer science ,Relational database ,Computer science ,Sixth normal form ,Third normal form ,Relational algebra ,Boyce–Codd normal form ,Second normal form - Abstract
This chapter covers the process of transforming an ER diagram into a well-designed relational database. For each normal norm, the chapter presents the rules that relations must meet to reach that normal form, and the design problems that may remain. It also covers the process for transforming relations into higher normal forms, including relational algebra operations, where appropriate.
- Published
- 2016
- Full Text
- View/download PDF
22. Dependency preservation and lossless join decomposition in fuzzy normalization
- Author
-
Sharmistha Ghosh, Jaydev Mishra, and Soumitra De
- Subjects
Fuzzy classification ,Computer science ,Relational model ,Multivalued dependency ,Fuzzy set operations ,Fuzzy number ,Third normal form ,Data mining ,Type-2 fuzzy sets and systems ,Boyce–Codd normal form ,computer.software_genre ,computer - Abstract
The traditional relational database model may be extended into a fuzzy database model based on the mathematical framework of fuzzy set theory to process imprecise or uncertain information. While designing such a fuzzy relational database model that does not suffer from data redundancy and anomalies, the present authors have defined several fuzzy normal forms in ref [1]. However, as one decomposes an unnormalized relation into a desired normal form, it should also satisfy the essential properties of dependency preservation and lossless join of relation schemes which take a significant role in the design theory of a relational database. We have concentrated on these two important issues in this paper and have designed algorithms that confirm dependency preservation and lossless join decomposition of an unnormalized relation into the fuzzy third normal form or fuzzy Boyce Codd normal form. The algorithms have been tested and validated with examples.
- Published
- 2015
- Full Text
- View/download PDF
23. Normalization of Relations with Nulls in Candidate Keys
- Author
-
George C. Philip
- Subjects
Normalization (statistics) ,Entity integrity ,Relational database ,Computer science ,Third normal form ,Boyce–Codd normal form ,computer.software_genre ,Candidate key ,Hardware and Architecture ,medicine ,Data mining ,medicine.symptom ,computer ,Software ,Information Systems ,Confusion - Abstract
This paper discusses normalization of relations when the candidate keys of a relation have missing information represented by nulls. The paper shows that when the missing information is of the type “not applicable” or “does not exist,” problems and confusion can arise in normalizing relations. Candidate keys with missing information commonly are found in relations that represent information on two entities with a one-to-one relationship between them. The current definition of Boyce-Codd Normal Form (BCNF) is ineffective in identifying poor designs in such relations that may have insertion/deletion anomalies. It is shown that the above problem can be corrected by incorporating the concept of entity integrity rule into the definition of BCNF. This paper also shows that incorporating the entity integrity rule into the definition of either a relation or a candidate key does not provide a satisfactory solution to the problem.
- Published
- 2002
- Full Text
- View/download PDF
24. [Untitled]
- Author
-
V. V. Khodorovskii
- Subjects
Discrete mathematics ,Algebra ,First normal form ,Sixth normal form ,Fifth normal form ,Third normal form ,Disjunctive normal form ,Second normal form ,Boyce–Codd normal form ,Software ,Mathematics ,Fourth normal form - Abstract
A new normal form, namely, object-normal form (ONF), is introduced. It is shown that the existing definitions of the fifth normal form (5NF) are unsatisfactory. The correct definition is given for the first time. The importance of the 5NF is demonstrated. For improving data representation in a database with a 5NF schema, the notion of negating relation is introduced. It serves as a basis for the new normal form, namely, the sixth normal form. Combining requirements of the ONF and other normal forms, new normal forms are defined: the fourth object-normal form, the fifth object-normal form, and the sixth object-normal form. It is shown that the standard order of steps of the normalization procedure should be changed: first, the specific requirements of the fourth normal form should be satisfied, and only then the requirements of the second normal form, and so forth.
- Published
- 2002
- Full Text
- View/download PDF
25. Semantic foundations of 4NF in relational database design
- Author
-
Millist W. Vincent
- Subjects
Theoretical computer science ,Computer Networks and Communications ,Relational database ,Multivalued dependency ,Third normal form ,Boyce–Codd normal form ,Fourth normal form ,Database normalization ,First normal form ,Functional dependency ,Algorithm ,Software ,Information Systems ,Mathematics - Abstract
The issue of providing a formal justification for the use of fourth normal form (4NF) in relational database design is investigated. The motivation and formal definitions for three goals of database design are presented. These goals are the elimination of: redundancy, key-based update anomalies and fact-based replacement anomalies. It is then shown that, depending on the type of constraints permitted, either Boyce-Codd normal form (BCNF) or 4NF are the exact conditions needed to ensure most of the design goals. However, it is also shown that the conditions required to ensure the absence of a particular class of key-based update anomaly are new normal forms which have not previously been identified. In particular, for the case where the only constraints are functional dependencies (FDs), it is shown that the required normal form is a new normal form that is stronger than third normal form (3NF) yet weaker than BCNF. Similarly, in the more general case where both FD and multivalued dependencies (MVDs) are present, the required normal form is a new normal form that is weaker than 4NF.
- Published
- 1999
- Full Text
- View/download PDF
26. A corrected 5NF definition for relational database design
- Author
-
Millist W. Vincent
- Subjects
Discrete mathematics ,General Computer Science ,Relational database ,Normalisation ,Fifth normal form ,Third normal form ,Boyce–Codd normal form ,Database design ,Join dependencies ,Fourth normal form ,Theoretical Computer Science ,Database normalization ,First normal form ,Arithmetic ,Mathematics ,Computer Science(all) - Abstract
In this paper, the adequacy of fifth normal form (5NF) in relational database design is investigated. It is shown that 5NF is inadequate because it does not generalise fourth normal form (4NF) and because it is equivalent to the very stringent requirement that every attribute is a key, a requirement that is effectively impossible to satisfy in practical database design. By restricting the join dependencies (JDs) in the set of constraints to those that do not have superfluous components, the definition of 5NF is then changed to a new normal form, called reduced fifth normal form (5NFR), and it is shown that 5NFR generalises 4NF. It is also shown that 5NFR is a strictly weaker condition than projection-join normal form (PJ/NF), the other normal form that has been proposed for JDs.
- Published
- 1997
- Full Text
- View/download PDF
27. On keys and normal forms
- Author
-
Wai Yin Mok
- Subjects
Discrete mathematics ,Normalization (statistics) ,Algebra ,Relational database ,Signal Processing ,Relation scheme ,Third normal form ,Boyce–Codd normal form ,Computer Science Applications ,Information Systems ,Theoretical Computer Science ,Mathematics ,Fourth normal form - Abstract
In this paper, we give a necessary and sufficient condition for a Boyce-Codd Normal Form (BCNF) relation scheme to be in Fourth Normal Form (4NF). We also give a necessary and sufficient condition for a 4NF relation scheme to be in Projection-Join Normal Form (PJNF). From these results, we derive necessary and sufficient conditions for a BCNF relation scheme to be in PJNF. In addition, we give a less stringent condition for a Third Normal Form (3NF) relation scheme to be in BCNF. Finally, we show that our theorems generalize the results by Date and Fagin (1992).
- Published
- 1997
- Full Text
- View/download PDF
28. Logical design for temporal databases with multiple granularities
- Author
-
Sushil Jajodia, Claudio Bettini, X. Sean Wang, and Alexander Brodsky
- Subjects
Dependency (UML) ,Closure (mathematics) ,Computer science ,Extension (predicate logic) ,Third normal form ,Boyce–Codd normal form ,Functional dependency ,Finite set ,Algorithm ,Information Systems ,Temporal database - Abstract
The purpose of good database logical design is to eliminate data redundancy and isertion and deletion anomalies. In order to achieve this objective for temporal databases, the notions of temporal types , which formalize time granularities, and temporal functional dependencies (TFDs) are intrduced. A temporal type is a monotonic mapping from ticks of time (represented by positive integers) to time sets (represented by subsets of reals) and is used to capture various standard and user-defined calendars. A TFD is a proper extension of the traditional functional dependency and takes the form X → μ Y, meaning that there is a unique value for Y during one tick of the temporal type μ for one particular X value. An axiomatization for TFDs is given. Because a finite set TFDs usually implies an infinite number of TFDs, we introduce the notion of and give an axiomatization for a finite closure to effectively capture a finite set of implied TFDs that are essential of the logical design. Temporal normalization procedures with respect to TFDs are given. Specifically, temporal Boyce-Codd normal form (TBCNF) that avoids all data redundancies due to TFDs, and temporal third normal form (T3NF) that allows dependency preservation, are defined. Both normal forms are proper extensions of their traditional counterparts, BCNF and 3NF. Decompositition algorithms are presented that give lossless TBCNF decompositions and lossless, dependency-preserving, T3NF decompositions.
- Published
- 1997
- Full Text
- View/download PDF
29. On the Boyce - Codd normal form for relation scheme
- Author
-
Vũ Đức Thi and Lương Cao Sơn
- Subjects
Combinatorics ,Relation scheme ,Boyce–Codd normal form ,Mathematics - Published
- 2013
- Full Text
- View/download PDF
30. Some results about normal forms for functional dependency in the relational datamodel
- Author
-
János Demetrovics and Vu Duc Thi
- Subjects
Closed set ,Relation (database) ,Closure ,Boyce-Codd normal form ,Third normal form ,Boyce–Codd normal form ,Antichain ,Combinatorics ,Functional dependency ,Discrete Mathematics and Combinatorics ,Relation ,Relational datamodel ,Second normal form ,Mathematics ,Discrete mathematics ,Antikey ,Relation scheme ,Applied Mathematics ,Minimal key ,Key ,Closure (mathematics) ,Minimal generator - Abstract
In this paper we present some characterizations of relation schemes in second normal form (2NF), third normal form (3NF) and Boyce-Codd normal form (BCNF). It is known [6]that the set of minimal keys of a relation scheme is a Sperner system (an antichain) and for an arbitrary Sperner system there exists a relation scheme the set of minimal keys of which is exactly the given Sperner system. We investigate families of 2NF, 3NF and BCNF relation schemes where the sets of minimal keys are given Sperner systems. We give characterizations of these families. The minimal Armstrong relation has been investigated in the literature [3, 7, 11, 15, 18]. This paper gives new bounds on the size of minimal Armstrong relations for relation schemes. We show that given a relation scheme s such that the set of minimal keys is the Sperner system K, the number of antikeys (maximal nonkeys) of K is polynomial in the number of attributes iff so is the size of minimal Armstrong relation of s. We give a new characterization of relations and relation schemes that are uniquely determined by their minimal keys. From this characterization we give a polynomial-time algorithm deciding whether an arbitrary relation is uniquely determined by its set of all minimal keys. We present a new polynomial-time algorithm testing BCNF property of a given relation scheme.
- Published
- 1996
- Full Text
- View/download PDF
31. A normal form for preventing redundant tuples in relational databases
- Author
-
Hugh Darwen, Ronald Fagin, and C. J. Date
- Subjects
Discrete mathematics ,First normal form ,InformationSystems_DATABASEMANAGEMENT ,Fifth normal form ,Third normal form ,Disjunctive normal form ,Conjunctive normal form ,Join dependency ,Boyce–Codd normal form ,Mathematics ,Fourth normal form - Abstract
We introduce a new normal form, called essential tuple normal form (ETNF), for relations in a relational database where the constraints are given by functional dependencies and join dependencies. ETNF lies strictly between fourth normal form and fifth normal form (5NF, also known as projection-join normal form). We show that ETNF, although strictly weaker than 5NF, is exactly as effective as 5NF in eliminating redundancy of tuples. Our definition of ETNF is semantic, in that it is defined in terms of tuple redundancy. We give a syntactic characterization of ETNF, which says that a relation schema is in ETNF if and only if it is in Boyce-Codd normal form and some component of every explicitly declared join dependency of the schema is a superkey.
- Published
- 2012
- Full Text
- View/download PDF
32. Normalization of Relations with Nulls in Candidate Keys
- Author
-
George C. Philip
- Subjects
Discrete mathematics ,Normalization (statistics) ,Candidate key ,business.industry ,Pattern recognition ,Artificial intelligence ,Boyce–Codd normal form ,business ,Mathematics - Abstract
This chapter discusses normalization of relations when the candidate keys of a relation have missing information represented by nulls. The chapter shows that problems and confusion can arise in normalizing relations with nulls in candidate keys. Candidate keys with missing information commonly are found in relations that represent information on two entities with a one-to-one relationship between them. The current definition of Boyce-Codd Normal Form (BCNF) is ineffective in identifying poor designs in such relations that may have insertion/deletion anomalies. Domain Key Normal Form (DKNF) also suffers from the same problem. It is shown that the above problem can be corrected by incorporating the concept of entity integrity rule into the definitions of BCNF and DKNF. This chapter also shows that incorporating the entity integrity rule into the definition of either a relation or a candidate key does not provide a satisfactory solution to the problem.
- Published
- 2011
- Full Text
- View/download PDF
33. REDUNDANCY AND THE JUSTIFICATION FOR FOURTH NORMAL FORM IN RELATIONAL DATABASES
- Author
-
Millist W. Vincent and Bala Srinivasan
- Subjects
Dependency theory (database theory) ,Discrete mathematics ,Computer Science (miscellaneous) ,Redundancy (engineering) ,Multivalued dependency ,Join dependency ,Functional dependency ,Boyce–Codd normal form ,Fourth normal form ,Mathematics ,Superkey - Abstract
The relationship between the absence of redundancy in relational databases and fourth normal form (4NF) is investigated. A relation scheme is defined to be redundant if there exists a legal relation defined over it which has at least two tuples that are identical on the attributes in a functional dependency (FD) or multivalued dependency (MVD) constraint. Depending on whether the dependencies in a set of constraints or the dependencies in the closure of the set is used, two different types of redundancy are defined. It is shown that the two types of redundancy are equivalent and their absence in a relation scheme is equivalent to the 4NF condition.
- Published
- 1993
- Full Text
- View/download PDF
34. Simple conditions for guaranteeing higher normal forms in relational databases
- Author
-
C. J. Date and Ronald Fagin
- Subjects
Algebra ,First normal form ,Computer science ,Multivalued dependency ,Fifth normal form ,Third normal form ,Disjunctive normal form ,Boyce–Codd normal form ,Functional dependency ,Algorithm ,Information Systems ,Fourth normal form - Abstract
A key is simple if it consists of a single attribute. It is shown that if a relation schema is in third normal form and every key is simple, then it is in projection-join normal form (sometimes called fifth normal form), the ultimate normal form with respect to projections and joins. Furthermore, it is shown that if a relation schema is in Boyce-Codd normal form and some key is simple, then it is in fourth normal form (but not necessarily projection-join normal form). These results give the database designer simple sufficient conditions, defined in terms of functional dependencies alone, that guarantee that the schema being designed is automatically in higher normal forms.
- Published
- 1992
- Full Text
- View/download PDF
35. An ameliorated methodology to design normalized relations
- Author
-
Shivanand M. Handigund and Ajeet A. Chikkamannur
- Subjects
Dependency theory (database theory) ,Theoretical computer science ,Relation (database) ,Computer science ,Relational database ,Software requirements specification ,Functional requirement ,Data mining ,Functional dependency ,computer.software_genre ,Boyce–Codd normal form ,computer ,Superkey - Abstract
In Database design, the attributes and their functional dependencies are abstracted from the software requirements specification (SRS) by the forward engineering process. The abstracted attributes from a forward engineering process are structured depending on the functional requirements. Then, the database relations are composed, by the related attributes and their functional dependencies, based on the structures defined by Codd. Any random structural composition of relations leads to INSERT, DELETE and UPDATE anomalies This paper presents a simple methodology that blends the analytical approach and the synthetic approach, to constitute the relations from a set of attributes and minimally covered functional dependencies, through the use of a dependency matrix to get the desired manipulation. The distinct attribute(s) from a set of functional dependencies is identified for a separate relation. All the dependencies are preserved and the lossless join is ensured by the framed algorithm. Further, the Boyce-Codd Normal Form (BCNF) is persuaded on each relation by revamping the determinant attributes to candidatekey.
- Published
- 2009
- Full Text
- View/download PDF
36. Evaluation of Learning Costs of Rule Evaluation Models Based on Objective Indices to Predict Human Hypothesis Construction Phases
- Author
-
Hidenao Abe, Shusaku Tsumoto, Miho Ohsaki, Hideto Yokoi, and Takahira Yamaguchi
- Subjects
Database normalization ,Algebra ,Relational database ,Dominance-based rough set approach ,Relational model ,Conjunctive query ,Rough set ,Set theory ,Boyce–Codd normal form ,Algorithm ,Mathematics - Abstract
In this paper a new method to judge the grade of normal form for relational database is proposed based on rough sets theory. First, some concepts about INF, 2NF, 3NF and BCNF are given and the principles of rough set theory are discussed. Second, the method to judge the grade of normal forms for a given relation is analyzed using rough sets theory, and some properties of a relation satisfying some grade of normal form are obtained. The study in this paper is a new application of rough sets theory.
- Published
- 2007
- Full Text
- View/download PDF
37. Finding Faithful Boyce-Codd Normal Form Decompositions
- Author
-
Henning Koehler
- Subjects
Set (abstract data type) ,Discrete mathematics ,Dependency (UML) ,Relational database ,Computer science ,Algorithmics ,Schema (psychology) ,InformationSystems_DATABASEMANAGEMENT ,Decision problem ,Functional dependency ,Boyce–Codd normal form ,Exponential function - Abstract
It is well known that faithful (i.e. dependency preserving) decompositions of relational database schemas into Boyce-Codd Normal Form (BCNF) do not always exist, depending on the set of functional dependencies given, and that the corresponding decision problem is NP-hard. The only algorithm to guarantee both faithfulness and BCNF (if possible) proposed so far in [Os79] is a brute-force approach which always requires exponential time. To be useful in practice, e.g. in automated design tools, we require more efficient means. In this paper we present an algorithm which always finds a faithful BCNF decomposition if one exists, and which is usually efficient, and exponential only in notorious cases.
- Published
- 2006
- Full Text
- View/download PDF
38. An extended Boyce-Codd normal form in fuzzy relational databases
- Author
-
J. Vandenbulcke, Etienne Kerre, and Guoqing Chen
- Subjects
Discrete mathematics ,Transitive relation ,Relation (database) ,Relational database ,Fuzzy set ,Database theory ,Anomaly (physics) ,Boyce–Codd normal form ,Fuzzy logic ,Mathematics - Abstract
If the relation schemes of any fuzzy database are not properly designed certain anomaly problems may occur when the fuzzy database is updated or maintained. This is because some undesired relationships exist among attributes. Based on the notions of fuzzy functional dependency (FFD) and /spl theta/-keys, the Boyce-Codd normal form is extended to cope with the anomaly problems caused by partial and transitive FFDs of /spl theta/-prime on some /spl theta/-keys, thus resulting in /spl theta/-fuzzy Boyce-Codd normal form (/spl theta/-FBCNF). Furthermore, the relationship between /spl theta/-FBCNF and other existing fuzzy normal forms is analyzed, showing that /spl theta/-FBCNF reflects the strongest restriction for attributes. Finally an algorithm is provided that can decompose, in a lossless-join manner, any scheme into a number of /spl theta/-FBCNF schemes.
- Published
- 2002
- Full Text
- View/download PDF
39. A note on relation schemes which are in 3NF but not in BCNF
- Author
-
Bala Srinivasan and Millist W. Vincent
- Subjects
Combinatorics ,Candidate key ,Relation (database) ,Relational database ,Signal Processing ,Relation scheme ,Third normal form ,Boyce–Codd normal form ,Algorithm ,Computer Science Applications ,Information Systems ,Theoretical Computer Science ,Mathematics - Abstract
We investigate the properties of relation schemes which are in third normal form (3NF) but not in Boyce-Codd normal form (BCNF) and prove that such a relation scheme must have a pair of candidate keys which overlap.
- Published
- 1993
- Full Text
- View/download PDF
40. Method to determine existence of perfect BCNF decomposition
- Author
-
W.-Y. Liu
- Subjects
Lossless compression ,Computer science ,Relational database ,Decomposition (computer science) ,Boyce–Codd normal form ,Functional dependency ,Algorithm ,Software ,Computer Science Applications ,Information Systems - Abstract
Boyce-Codd Normal Form (BCNF) is an important normal form in the theory of relational databases. Unfortunately, it is not always possible to obtain a lossless functional-dependency-preserving BCNF decomposition. (This decomposition is called a perfect BCNF decomposition for short.) The necessary condition of the existence of a perfect BCNF decomposition is discussed. Three propositions about a perfect BCNF decomposition are obtained. The concrete determining algorithms are given.
- Published
- 1992
- Full Text
- View/download PDF
41. Decomposition of four component items
- Author
-
John K. Debenham
- Subjects
Algebra ,Formalism (philosophy of mathematics) ,First normal form ,Four component ,Computer science ,Relational database ,Boyce–Codd normal form - Abstract
Items are introduced as a universal formalism for describing data (ie individual things), information (ie relations) and knowledge (ie rules). A single rule for the normalisation of items is defined. The five normal forms for relational database together with the Boyce-Codd normal form provide a complete characterisation of the normalisation of 3-adic information items. New normal forms may be derived by applying our single rule of normalisation to n-adic items for n>3.
- Published
- 1993
- Full Text
- View/download PDF
42. Removing redundancy and updating databases
- Author
-
De Bra, P.M.E., Paredaens, J., Abiteboul, S., Kanellakis, P. C., and Mathematics and Computer Science
- Subjects
Database normalization ,Redundancy (information theory) ,Database ,Fragment (logic) ,Computer science ,Third normal form ,computer.software_genre ,Functional dependency ,Boyce–Codd normal form ,computer ,Database design ,Time complexity - Abstract
In order to eliminate redundant information in a database one usually uses decomposition or special database design methods to achieve Third Normal Form (or some other normal form like Boyce-Codd Normal Form). This method of database design assumes that the real world satisfies very strong constraints like functional dependencies (fds), without any exceptions. Most databases in "Whatever" Normal Form still contain a large amount of duplicated information, because the real world satisfies very few constraints, but a lot of "almost" constrains. Horizontal Fragments (resulting from Horizontal Decomposition) can be used to separate information with different properties. The different constraints that hold in these fragments can be used to eliminate redundant information in the fragments. We focus on the horizontal decomposition for eliminating exceptions to fds. We show how to update such horizontally decomposed databases in an efficient way. Generating separate horizontal fragments for all desired fds that do not hold in the real world may not be possible (although for every fd there will be a fragment for its exceptions, but the fragments for a number of fds may coincide). The number of fds for which a separate exception fragment can be generated depends on the choice of fds and the order in which they are used. We show that it is possible to find (in polynomial time) the order which leads to the "optimal" decomposition of a relation into horizontal fragments.
- Published
- 1990
- Full Text
- View/download PDF
43. Response to 'Remarks on two new theorems of Date and Fagin'
- Author
-
C. J. Date and Ronald Fagin
- Subjects
Discrete mathematics ,Computer science ,Relational database ,InformationSystems_DATABASEMANAGEMENT ,Third normal form ,Disjunctive normal form ,Boyce–Codd normal form ,Fourth normal form ,Database normalization ,First normal form ,Canonical form ,Fifth normal form ,Conjunctive normal form ,Functional dependency ,Second normal form ,Algorithm ,Software ,Information Systems - Abstract
In [DF92], we present simple conditions, which we now describe, for guaranteeing higher normal forms for relational databases. A key is simple if it consists of a single attribute. We show in [DF92] that if a relation schema is in third normal form (3NF) and every key is simple, then it is in projection-join normal form (sometimes called fifth normal form), the ultimate normal form with respect to projections and joins. We also show in [DF92] that if a relation schema is in Boyce-Codd normal form (BCNF) and some key is simple, then it is in fourth normal form (4NF). These results give the database designer simple sufficient conditions, defined in terms of functional dependencies alone, that guarantee that the schema being designed is automatically in higher normal forms.
- Published
- 1993
- Full Text
- View/download PDF
44. Entity-relationship diagrams which are in BCNF
- Author
-
Jajodia, Sushil, Ng, Peter A., and Springsteel, Frederick N.
- Published
- 1983
- Full Text
- View/download PDF
45. Characterizations for functional dependency and Boyce-Codd normal form families
- Author
-
Seymour Ginsburg and Richard Hull
- Subjects
Discrete mathematics ,General Computer Science ,Closed set ,Intersection (set theory) ,Boyce–Codd normal form ,Theoretical Computer Science ,Decidability ,Combinatorics ,Set (abstract data type) ,Algebraic number ,Functional dependency ,Projection (set theory) ,Computer Science(all) ,Mathematics - Abstract
A functional dependency (fd) family was recently defined [20] as the set of all instances satisfying some set of functional dependencies. A Boyce-Codd normal form, abbreviated BCNF, family is defined here as an fd-family specified by some BCNF set of functional dependencies. The purpose of this paper is to present set-theoretic/algebraic characterizations relating to both types of families. Two characterizations of F ( I ), the smallest fd-family containing the family I of instances, are established. The first involves the notion of agreement, a concept related to that of a closed set of attributes. The second describes F ( I ) as the smallest family of instances containing I and closed under four specific operations on instances. Companion results are also given for BCNF- families. The remaining results concern characterizations involving the well-known operations of projection, join and union. Two characterizations for when the projection of an fd-family is again an fd-family are given. Several corollaries are obtained, including the effective decidability of whether a projection of an fd-family is an fd-family. The problem for BCNF-families disappears since it is shown that the projection of a BCNF-family is always a BCNF-family. Analogous to results for fd-families presented in [20], characterizations of when the join and union of BCNF-families are BCNF-families are given. Finally, the collections of all fd-families and all BCNF-families are characterized in terms of inverse projection operations and intersection.
- Published
- 1983
- Full Text
- View/download PDF
46. Multivalued dependencies and a new normal form for relational databases
- Author
-
Ronald Fagin
- Subjects
Algebra ,First normal form ,Computer science ,Multivalued dependency ,Canonical form ,Third normal form ,Functional dependency ,Second normal form ,Boyce–Codd normal form ,Algorithm ,Information Systems ,Fourth normal form - Abstract
A new type of dependency, which includes the well-known functional dependencies as a special case, is defined for relational databases. By using this concept, a new (“fourth”) normal form for relation schemata is defined. This fourth normal form is strictly stronger than Codd's “improved third normal form” (or “Boyce-Codd normal form”). It is shown that every relation schema can be decomposed into a family of relation schemata in fourth normal form without loss of information (that is, the original relation can be obtained from the new relations by taking joins).
- Published
- 1977
- Full Text
- View/download PDF
47. Independent components of relations
- Author
-
Jorma Rissanen
- Subjects
Pure mathematics ,Component (thermodynamics) ,Computer science ,Third normal form ,Cartesian product ,Boyce–Codd normal form ,Fourth normal form ,symbols.namesake ,Operator (computer programming) ,symbols ,Functional dependency ,Algorithm ,Independence (probability theory) ,Information Systems - Abstract
In a multiattribute relation or, equivalently, a multicolumn table a certain collection of the projections can be shown to be independent in much the same way as the factors in a Cartesian product or orthogonal components of a vector. A precise notion of independence for relations is defined and studied. The main result states that the operator which reconstructs the original relation from its independent components is the natural join, and that independent components split the full family of functional dependencies into corresponding component families. These give an easy-to-check criterion for independence.
- Published
- 1977
- Full Text
- View/download PDF
48. A normal form for relational databases that is based on domains and keys
- Author
-
Ronald Fagin
- Subjects
Combinatorics ,Relational database ,Computer science ,Bounded function ,Multivalued dependency ,Third normal form ,Join dependency ,Functional dependency ,Boyce–Codd normal form ,Algorithm ,Information Systems ,Fourth normal form - Abstract
A new normal form for relational databases, called domain-key normal form (DK/NF), is defined. Also, formal definitions of insertion anomaly and deletion anomaly are presented. It is shown that a schema is in DK/NF if and only if it has no insertion or deletion anomalies. Unlike previously defined normal forms, DK/NF is not defined in terms of traditional dependencies (functional, multivalued, or join). Instead, it is defined in terms of the more primitive concepts of domain and key, along with the general concept of a “constraint.” We also consider how the definitions of traditional normal forms might be modified by taking into consideration, for the first time, the combinatorial consequences of bounded domain sizes. It is shown that after this modification, these traditional normal forms are all implied by DK/NF. In particular, if all domains are infinite, then these traditional normal forms are all implied by DK/NF.
- Published
- 1981
- Full Text
- View/download PDF
49. Decomposition of a relation scheme into Boyce-Codd Normal Form
- Author
-
Patrick C. Fischer and Don-Min Tsou
- Subjects
Discrete mathematics ,Lossless compression ,Multidisciplinary ,Relational database ,Decomposition (computer science) ,InformationSystems_DATABASEMANAGEMENT ,Extension (predicate logic) ,Join dependency ,Lossless-Join Decomposition ,Functional dependency ,Boyce–Codd normal form ,Algorithm ,Mathematics - Abstract
Decomposition into Boyce-Codd Normal Form (BCNF) with a lossless join and preservation of dependencies is desired in the design of a relational database scheme. However, there may be no decomposition of a relation scheme into BCNF that is dependency preserving, and the known algorithms for lossless join decomposition into BCNF require exponential time and space. In this paper we give an efficient algorithm for lossless join decomposition and show that the problem of deciding whether a relation scheme has a dependency-preserving decomposition into BCNF is NP-hard. The algorithm and the proof assume that all data dependencies are functional. We then discuss the extension of our techniques to the case where data dependencies are multivalued.
- Published
- 1982
- Full Text
- View/download PDF
50. The complexity of recognizing 3NF relation schemes
- Author
-
Jiann H. Jou and Patrick C. Fischer
- Subjects
Theoretical computer science ,Relation (database) ,Relational database ,Third normal form ,Relational algebra ,Boyce–Codd normal form ,Topology ,Computer Science Applications ,Theoretical Computer Science ,Database normalization ,First normal form ,Signal Processing ,Relational model ,Information Systems ,Mathematics - Published
- 1982
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.