The document discusses the FP-Growth algorithm for frequent pattern mining. It improves upon the Apriori algorithm by not requiring candidate generation and only requiring two scans of the database. FP-Growth works by first building a compact FP-tree structure using two passes over the data, then extracting frequent itemsets directly from the FP-tree. An example is provided where an FP-tree is constructed from a sample transaction database and frequent patterns are generated from the tree. Advantages of FP-Growth include only needing two scans of data and faster runtime than Apriori, while a disadvantage is the FP-tree may not fit in memory.
FP GrowthStands for frequent pattern growth It is a scalable technique for mining frequent pattern in a database
3.
FP growthimproves Apriority to a big extent Frequent Item set Mining is possible without candidate generation Only “two scan” to the database is needed BUT HOW?
4.
Simply atwo step procedure – Step 1: Build a compact data structure called the FP-tree • Built using 2 passes over the data-set. – Step 2: Extracts frequent item sets directly from the FP- tree
5.
Now LetsConsider the following transaction table TID List of item IDs T100 I1,I2,I3 T200 I2,I4 T300 I2,I3 T400 I1,I2,I4 T500 I1,I3 T600 I2,I3 T700 I1,I3 T800 I1,I2,I3,I5 T900 I1,I2,I3
6.
Now wewill build a FP tree of that database Item sets are considered in order of their descending value of support count.
I2 7 I1 6 I36 I4 2 I5 2 null I2:7 I1:4 I3:2 I3:2 I5:1 I4:1 I4:1 To facilitate tree traversal, an item header table is built so that each item points to its occurrences in the tree via a chain of node-links. I1:2 I3:2 I5:1 FP Tree Construction Over!! Now we need to find conditional pattern base and Conditional FP Tree for each item