MachineX: Two Parts of Association Rule Learning

DZone 's Guide to

MachineX: Two Parts of Association Rule Learning

Decouple the support and confidence requirements for Association Rule Learning in this article.

· AI Zone ·
Free Resource

In our previous blog, MachineX: Layman's Guide to Association Rule Learning, we discussed what Association rule learning is all about, and as you can already tell, with a large dataset, which almost every market has, finding association rules isn't very easy. For these purposes, we introduced measures of interestingness, which were support, confidence, and lift. Support tells us how frequent an itemset is in a given dataset and confidence tells us about the reliability of that rule. Due to large datasets, it is computationally very expensive to find both support and confidence for it.

An initial step towards improving the performance of association rule mining algorithms is to decouple the support and confidence requirements. Let's look at it more closely.

Suppose a dataset exists such as the one below:

For the itemset {Beer, Diapers, Milk}, the following rules exist:

Support for all the above rules is identical since the rules involve items from the same item set. Now, without any optimization, our next step would have been to calculate their confidence values, but, supposing this itemset to be infrequent, that is with a low support value, rules generated by this itemset cannot be of any interest to us. So, we don't need to calculate the confidence to tell that we do not require these rules. So, just on the basis of its support, we can immediately prune all the rules generated by it. This way, we can achieve some degree of optimization.

For these purposes, association rule learning is divided into two parts:

Frequent Itemset Generation

In this step, all the itemsets with support higher than the minimum support threshold are selected. This step gives us all the frequent itemsets in the dataset.

Rule Generation

In this step, high-confidence rules are extracted from the frequent itemsets that were obtained from the previous step. These rules are known as strong rules.

The computational requirements for frequent itemset generation are generally more expensive than those of rule generation.

In our next blog, we will be discussing the apriori algorithm for frequent itemset generation, why is it not used much, and it's alternative. Stay tuned!

data science ,artificial intelligence

Published at DZone with permission of Akshansh Jain , DZone MVB. See the original article here.

Opinions expressed by DZone contributors are their own.

{{ parent.title || parent.header.title}}

{{ parent.tldr }}

{{ parent.urlSource.name }}