Arbitrary categorisation - learning decision trees, Computer Engineering

Assignment Help:

Arbitrary categorisation - learning decision trees:

Through visualising  a set of boxes with some balls in. There if all the balls were in a single box so this would be nicely ordered but it would be extremely easy to find a particular ball. Moreover If the balls were distributed amongst the boxes then this would not be so nicely ordered but it might take rather a whereas to find a particular ball. It means if we were going to define a measure based at this notion of purity then we would want to be able to calculate a value for each box based on the number of balls in it so then take the sum of these as the overall measure. Thus we would want to reward two situations: nearly empty boxes as very neat and boxes just with nearly all the balls in as also very neat. However this is the basis for the general entropy measure that is defined follows like: 

Now next here instantly an arbitrary categorisation like C into categories c1, ..., cn and a set of examples, S, for that the proportion of examples in ci is pi, then the entropy of S is as: 

198_Arbitrary categorisation - learning decision trees.png

Here measure satisfies our criteria that is of the -p*log2(p) construction: where p gets close to zero that is the category has only a few examples in it so then the  log(p) becomes a big negative number and the  p  part dominates the calculation then the entropy works out to be nearly zero. However make it sure that entropy calculates the disorder in the data in this low score is good and as it reflects our desire to reward categories with few examples in. Such of similarly if p gets close to 1 then that's the category has most of the examples in so then the  log(p) part gets very close to zero but it  is this that dominates the calculation thus the overall value gets close to zero. Thus we see that both where the category is nearly  -  or completely  -  empty and when the category nearly contains as - or completely contains as  - all the examples and the score for the category gets close to zero that models what we wanted it to. But note that 0*ln(0) is taken to be zero by convention them.


Related Discussions:- Arbitrary categorisation - learning decision trees

Example on public divisor, Q. Example on PUBLIC DIVISOR? Linker appends...

Q. Example on PUBLIC DIVISOR? Linker appends all segments having the same name and PUBLIC directive with segment name into one segment. Their contents are pulled together in co

Assembly program, where do i get some sample assembly projects (with coding...

where do i get some sample assembly projects (with coding included)?

Average waiting time, Customer arrivals at a gas station can be characteriz...

Customer arrivals at a gas station can be characterized by exponential distribution with mean of 10 minutes. The amount of time they take to pump gas can be characterized by an ex

Define the message queues, The message queue provides the information about...

The message queue provides the information about sizes of queues under utilization of various processors. It points to size of every processor incoming message queue that would be

State about the object oriented analysis design, State about the Object ori...

State about the Object oriented analysis design Object oriented analysis design (OOAD) is a bottom up approach which supports viewing system as a set of components (objects) w

Appropriate problems for ann learning , Appropriate Problems for ANN learni...

Appropriate Problems for ANN learning - artificial intelligence-  As we did for decision trees, it is essential to know when ANNs are the correct representation scheme for the

Depth first search - artificial intelligence, Depth First Search - artifici...

Depth First Search - artificial intelligence: Depth first search is very similar to breadth first, except for that the things are added to the top of this agenda rather than o

Which electro mechanical switch had fewer moving parts, The             el...

The             electro mechanical switch (developed in 1938) had fewer moving parts than earlier switches. (A)  No. 1ESS                                 (B)  Strowger (

Component fulfil this requirement, A component has the time to failure dist...

A component has the time to failure distribution that is modelled as the Weibull distribution with shape parameter 3 and scale parameter of 36 months. This component is planned to

DATA DICTIONARY, DATA DICTIONARY ON ONLINE QUIZ SYSTEM

DATA DICTIONARY ON ONLINE QUIZ SYSTEM

Write Your Message!

Captcha
Free Assignment Quote

Assured A++ Grade

Get guaranteed satisfaction & time on delivery in every assignment order you paid with us! We ensure premium quality solution document along with free turntin report!

All rights reserved! Copyrights ©2019-2020 ExpertsMind IT Educational Pvt Ltd