Higher degrees of normalization tend to involve the creation of more tables, and more joins when retrieving functional datasets.
Data anomaly types addressed by normalization
- Update Anomaly - similar data is represented in multiple areas. When data is updated, it may not capture each specific source. When queried, this results in multiple records and inconsistent results.
- Insertion Anomaly - When combined with the incorrect combination of fields, an insert statement into the table may be incorrectly refused.
- Deletion Anomaly - Similar to insertion, when combined incorrectly on a table, the last instance of a record may be incorrectly deleted.
The normal forms of a database determine the degree of vulnerability to data inconsistencies. The higher the form, the less vulnerable the table is.
- First Normal Form (1NF) - The table must faithfully represent a relation. One requirement for a relation being that there is only one value for each attribute or "no repeating groups"
- Second Normal Form (2NF) -
- Third Normal Form (3NF) -
Purposeful denormalization can be used to represent by read-only applications or Business Intelligence applications. These are characterized as OLAP (Online Analytical Processing). An implementation of this is the Amazon database previously written about, allowing data set in a field, more like hash table/key value pairs.
No comments:
Post a Comment