Chapter 27: Problem 12
How does clustering differ from classification?
Chapter 27: Problem 12
How does clustering differ from classification?
All the tools & learning materials you need for study success - in one app.
Get started for freeWhat are the difficulties of mining association rules from large databases?
What is entropy and how is it used in building decision trees?
Apply the Apriori algorithm to the following data set. $$\begin{array}{ll} \text { Trans ID } & \text { Items Purchased } \\ \hline 101 & \text { milk, bread, eggs } \\ 102 & \text { milk, juice } \\ 103 & \text { juice, butter } \\ 104 & \text { milk, bread, eggs } \\ 105 & \text { coffee, eggs } \\ 106 & \text { coffee } \\ 107 & \text { coffee, juice } \\ 108 & \text { milk, bread, cookies, eggs } \\ 109 & \text { cookies, butter } \\ 110 & \text { milk, bread } \end{array}$$ The set of items is \(\\{\text { milk, bread, cookies, eggs, butter, coffee, juice }\\}\). Use 0.2 for the minimum support value.
The K-means algorithm uses a similarity metric of distance between a record and a cluster centroid. If the attributes of the records are not quantitative but categorical in nature, such as Income Level with values \\{low, medium, hight or Married with values \\{Yes, Nof or State of Residence with values \\{Alabama, Alaska, \(\ldots,\) Wyoming then the distance metric is not meaningful. Define a more suitable similarity metric that can be used for clustering data records that contain categorical data.
What are classification rules and how are decision trees related to them?
What do you think about this solution?
We value your feedback to improve our textbook solutions.