Deletion anomalies-data redundancy, Database Management System

Assignment Help:

Deletion Anomalies: Loss of important Information: In some cases, useful information may be lost when a tuple is deleted. For instance, if we delete the tuple corresponding to student 050111341 enrolled for MCS-014, we will misplace relevant information about the student by enrolment number, address and name of this student.  Likewise deletion of tuple having Sname "Rahul" and Cno 'MCS-012" will result in failure of information that MCS-012 is named computer organisation having an instructor "Anurag Sharma", whose office number is 105. This is known as deletion anomaly.

The anomalies arise primarily as the relation STUDENT has information about students as well as subjects. One solution to the troubles is to decompose the relation into two or more smaller relations. But what should be the basis of this decomposition? To solution the questions let us attempt to formulate how data is related in the relation with the help of the following Figure:

                              2336_Deletion Anomalies.png

 

                                                       Figure :  The dependencies of relation

Please note that the arrows in Figure are defines data inter-relationship. For instance, enrolment no column is unique for a student so if we identify the enrolment no of a student we can uniquely describe his/her name and address. Likewise, the course code (Cno) uniquely defines course name (Cname) and Instructor (we are assuming that a course is taught by only single instructor). Please also note one vital interrelationship in Figure that is, the Office (address) of an instructor is relying on Instructor (name), assuming unique instructor names. The root cause of the being there of anomalies in a relation is determination of data by the components of the non-key and key attributes.

Normalisation includes decomposition of a relation into minor relations based on the concept of functional dependence to come over undesirable anomalies.

Normalisation few times can affect performance. As it results in decomposition of tables, few queries desire to join these tables to create the data once again. But such performance overheads are smallest as Normalisation results in minimisation of data redundancy and may result in minor relation sizes. Also DBMSs executes optimised algorithms for joining of relations and many indexing schemes that decrease the load on joining of relations. In any case the benefits of normalization normally overweigh the performance constraints. Normalisation does lead to more well-organized updates since an update that might have needs various tuples to be updated, while normalised relations, in general, need the information updating at only one place.

A relation that requires to be normalised may have a very large number of attributes. In such relations, it is almost impossible for someone to conceptualise all the information and recommend a suitable decomposition to overcome the troubles. Such relations require an algorithmic approach of searching if there are troubles in a proposed database design and how to remove them if they exist. The discussions of these algorithms are beyond the scope of this part, except, we will initial introduce you to the basic concept that supports the process of Normalisation of big databases. So let us first describes the concept of functional dependence in the subsequent part and follow it up with the thoughts of normalisation.


Related Discussions:- Deletion anomalies-data redundancy

In which format database schema is written, In which format database schema...

In which format database schema is written ? Data base schema is written in the DDL format.

What is called as an n-way merge, What is called as an N-way merge? The...

What is called as an N-way merge? The merge operation is a generalization of the two-way merge used by the standard in-memory sort-merge algorithm. It merges N runs, so it is k

Which sorting technique is used to sort databases, Which sorting technique ...

Which sorting technique is used to sort databases, whose sizes are very big?  Give one such algorithm.  Why do sorting techniques like quicksort, insertion sort, etc.  not applied

State the definition of association, State the definition of association ...

State the definition of association An association is a structural relationship which specifies that objects of one thing are connected with objects of another. In UML, asso

What is a records management consultation, What is a records management con...

What is a records management consultation? Our organization is an authority for disseminating your specific requirements and giving you with a solution to effectively manage yo

Which data manipulation command combines the records, Which data manipulati...

Which data manipulation command combines the records from one or more tables ? JOIN data manipulation command combines the records from one or more tables.

Set concept is used in which type of modelling, SET concept is used in whic...

SET concept is used in which type of modelling? SET concept is used in Network Model.

Support by adding indexes, Support by Adding Indexes Multiple indexes c...

Support by Adding Indexes Multiple indexes can be used to access a data file by multiple access paths. In such a scheme only single copy of the data is consider, only the numbe

What are the acid properties?, Question 1 What are the advantages of DBMS ...

Question 1 What are the advantages of DBMS approach in managing data? Question 2 What are the services provided by a database system Question 3 List and briefly describe

Write Your Message!

Captcha
Free Assignment Quote

Assured A++ Grade

Get guaranteed satisfaction & time on delivery in every assignment order you paid with us! We ensure premium quality solution document along with free turntin report!

All rights reserved! Copyrights ©2019-2020 ExpertsMind IT Educational Pvt Ltd