Deletion anomalies-data redundancy, Database Management System

Assignment Help:

Deletion Anomalies: Loss of important Information: In some cases, useful information may be lost when a tuple is deleted. For instance, if we delete the tuple corresponding to student 050111341 enrolled for MCS-014, we will misplace relevant information about the student by enrolment number, address and name of this student.  Likewise deletion of tuple having Sname "Rahul" and Cno 'MCS-012" will result in failure of information that MCS-012 is named computer organisation having an instructor "Anurag Sharma", whose office number is 105. This is known as deletion anomaly.

The anomalies arise primarily as the relation STUDENT has information about students as well as subjects. One solution to the troubles is to decompose the relation into two or more smaller relations. But what should be the basis of this decomposition? To solution the questions let us attempt to formulate how data is related in the relation with the help of the following Figure:

                              2336_Deletion Anomalies.png

 

                                                       Figure :  The dependencies of relation

Please note that the arrows in Figure are defines data inter-relationship. For instance, enrolment no column is unique for a student so if we identify the enrolment no of a student we can uniquely describe his/her name and address. Likewise, the course code (Cno) uniquely defines course name (Cname) and Instructor (we are assuming that a course is taught by only single instructor). Please also note one vital interrelationship in Figure that is, the Office (address) of an instructor is relying on Instructor (name), assuming unique instructor names. The root cause of the being there of anomalies in a relation is determination of data by the components of the non-key and key attributes.

Normalisation includes decomposition of a relation into minor relations based on the concept of functional dependence to come over undesirable anomalies.

Normalisation few times can affect performance. As it results in decomposition of tables, few queries desire to join these tables to create the data once again. But such performance overheads are smallest as Normalisation results in minimisation of data redundancy and may result in minor relation sizes. Also DBMSs executes optimised algorithms for joining of relations and many indexing schemes that decrease the load on joining of relations. In any case the benefits of normalization normally overweigh the performance constraints. Normalisation does lead to more well-organized updates since an update that might have needs various tuples to be updated, while normalised relations, in general, need the information updating at only one place.

A relation that requires to be normalised may have a very large number of attributes. In such relations, it is almost impossible for someone to conceptualise all the information and recommend a suitable decomposition to overcome the troubles. Such relations require an algorithmic approach of searching if there are troubles in a proposed database design and how to remove them if they exist. The discussions of these algorithms are beyond the scope of this part, except, we will initial introduce you to the basic concept that supports the process of Normalisation of big databases. So let us first describes the concept of functional dependence in the subsequent part and follow it up with the thoughts of normalisation.


Related Discussions:- Deletion anomalies-data redundancy

Explain the 10 reasons to use mysql, Question 1 Explain the 10 reasons to ...

Question 1 Explain the 10 reasons to use MySQL Question 2 Write the SQL statements to demonstrate the following using SELECT command Expression Evaluation ORDER BY

How many types of locks are there, Depending upon the rules we have found, ...

Depending upon the rules we have found, we can classify the locks into two types. Shared Lock: A transaction might be acquiring shared lock on a data item in order to read it

Explain the concept of modeling with objects, Explain the concept of Modeli...

Explain the concept of Modeling With Objects A model is a conceptual representation of a specification, design or system from a particular point of view. A modelling language

Describe physical data independence, Describe Physical Data independence ? ...

Describe Physical Data independence ? Physical Data Independence - It is capacity to modify the internal schema without having to change conceptual schema. Therefore, the exter

Stock market, need requirement analysis for database management of stock ma...

need requirement analysis for database management of stock market

Why containment is important in oosystems, Why containment is important in ...

Why containment is important in oosystems? Containment is an significant concept in oosystems because it allows dissimilar users to view data at different granularities.

Data aggregation, Q.   Explain  data aggregatio n and discuss differen...

Q.   Explain  data aggregatio n and discuss different design constraints. Sol. Aggregation   One limitation of the E-R model is that it cannot express relationships amon

Define granularity and hierarchy of granularity of locks, Define granularit...

Define granularity, hierarchy of granularity of locks & multiple granularity locking. Explain the modified two phase locking along with multiple granularity locking. Ans: The s

Assignment, Gyan Gurukul is a software company that provides software relat...

Gyan Gurukul is a software company that provides software related solution. Ashutosh Dubey who is the director of Gyan Gurukul want a three-level schema architecture for the Gyan G

Write Your Message!

Captcha
Free Assignment Quote

Assured A++ Grade

Get guaranteed satisfaction & time on delivery in every assignment order you paid with us! We ensure premium quality solution document along with free turntin report!

All rights reserved! Copyrights ©2019-2020 ExpertsMind IT Educational Pvt Ltd