Deletion anomalies-data redundancy, Database Management System

Assignment Help:

Deletion Anomalies: Loss of important Information: In some cases, useful information may be lost when a tuple is deleted. For instance, if we delete the tuple corresponding to student 050111341 enrolled for MCS-014, we will misplace relevant information about the student by enrolment number, address and name of this student.  Likewise deletion of tuple having Sname "Rahul" and Cno 'MCS-012" will result in failure of information that MCS-012 is named computer organisation having an instructor "Anurag Sharma", whose office number is 105. This is known as deletion anomaly.

The anomalies arise primarily as the relation STUDENT has information about students as well as subjects. One solution to the troubles is to decompose the relation into two or more smaller relations. But what should be the basis of this decomposition? To solution the questions let us attempt to formulate how data is related in the relation with the help of the following Figure:

                              2336_Deletion Anomalies.png

 

                                                       Figure :  The dependencies of relation

Please note that the arrows in Figure are defines data inter-relationship. For instance, enrolment no column is unique for a student so if we identify the enrolment no of a student we can uniquely describe his/her name and address. Likewise, the course code (Cno) uniquely defines course name (Cname) and Instructor (we are assuming that a course is taught by only single instructor). Please also note one vital interrelationship in Figure that is, the Office (address) of an instructor is relying on Instructor (name), assuming unique instructor names. The root cause of the being there of anomalies in a relation is determination of data by the components of the non-key and key attributes.

Normalisation includes decomposition of a relation into minor relations based on the concept of functional dependence to come over undesirable anomalies.

Normalisation few times can affect performance. As it results in decomposition of tables, few queries desire to join these tables to create the data once again. But such performance overheads are smallest as Normalisation results in minimisation of data redundancy and may result in minor relation sizes. Also DBMSs executes optimised algorithms for joining of relations and many indexing schemes that decrease the load on joining of relations. In any case the benefits of normalization normally overweigh the performance constraints. Normalisation does lead to more well-organized updates since an update that might have needs various tuples to be updated, while normalised relations, in general, need the information updating at only one place.

A relation that requires to be normalised may have a very large number of attributes. In such relations, it is almost impossible for someone to conceptualise all the information and recommend a suitable decomposition to overcome the troubles. Such relations require an algorithmic approach of searching if there are troubles in a proposed database design and how to remove them if they exist. The discussions of these algorithms are beyond the scope of this part, except, we will initial introduce you to the basic concept that supports the process of Normalisation of big databases. So let us first describes the concept of functional dependence in the subsequent part and follow it up with the thoughts of normalisation.


Related Discussions:- Deletion anomalies-data redundancy

Mapping, external and internal mapping with examples

external and internal mapping with examples

Define average latency time, Define average latency time. The average l...

Define average latency time. The average latency time of the disk is one-half the time for a full rotation of the disk.

What is hash file organization, What is hash file organization? What are t...

What is hash file organization? What are the causes of bucket overflow within a hash file organization? What can be completed to decrease the occurrence of bucket overflow?

Define tables students and teachers as sub tables of people, Write a query ...

Write a query to define tables students and teachers as sub tables of people? Create table students of student under people Create table teachers of teacher under people

What are the two integrity rules, What are the two integrity rules?  An...

What are the two integrity rules?  Ans:  The two integrity rules are: Entity Integrity Rule & Referential Integrity Rule.   Entity Integrity Rule - If the attribute A of rel

Ideal data dictionary, An ideal data dictionary should have everything a DB...

An ideal data dictionary should have everything a DBA wants to know about the database. 1 Conceptual, External, and internal database descriptions. 2.    Descriptions of attr

Define null value, Define NULL value NULL value NULL means something is...

Define NULL value NULL value NULL means something is unknown. It does NOT mean null (the digit 0). Null is as well employed as attribute value for a particular entity in which

Components of a typical database environment, Question: (a) List four d...

Question: (a) List four disadvantages of the File Processing System as compared to the Database Systems. (b) The integrity of a database is enforced by integrity rules. One

Define the concept of aggregation, Define the concept of aggregation.  Give...

Define the concept of aggregation.  Give two examples where this concept is useful. Ans: Aggregation transforms a relationship among the objects into a higher-level object. A

Write Your Message!

Captcha
Free Assignment Quote

Assured A++ Grade

Get guaranteed satisfaction & time on delivery in every assignment order you paid with us! We ensure premium quality solution document along with free turntin report!

All rights reserved! Copyrights ©2019-2020 ExpertsMind IT Educational Pvt Ltd