Title of Invention

A METHOD AND APPARATUS FOR MINING A LARGE DATA BASE

Abstract A computer method of online mining of quantitative association rules consisting of two stages, a preprocessing stage followed by an online rule generation stage. The required computational effort is reduced by the pre¬processing stage, defined by pre-processing data to organize the relationship between antecedent attributes to create a heirarchially arranged multidimensional indexing structure. The resulting structure facilitates the performance of the second stage, online processing, which involves the generation of quantitative association rules. The second stage, online rule generation, utilizes the multidimensional index structure created by the preprocessing stage by first finding the areas in the data which correspond to the rules and then uses a merging step to create a merged tree in order to carefully combine interesting regions in order to give a heirarchical representation of the rule set. The merged tree is then used in order to actually generate the rules.
Full Text The invention relates to a method of mining a large data base. It relates generally to online searching for data dependencies in large databases and more particularly to an online method of data mining of data items to find quantitative association rules, where the data items comprise various kinds of quantitative and categorical attributes.
Discussion of the Prior Art
Data mining, also known as knowledge discovery in databases, has been recognized as a new area for database research. The volume of data stored in electronic format has increased dramatically over the past two decades. The increase in use of electronic data gathering devices such as point-of-sale or remote sensing devices has contributed to this explosion of available data. Data storage is becoming easier and more attractive to the business community as the availability of large amounts of computing power and data storage resources are being made available at increasingly reduced costs.

with much attention focused on the accumulation of data, there arose a complimentary need to focus on hpw this valuable resource could be utilized. Businesses soon recognized that valuable insights could be gleaned by decision-makers who could make use of the stored data. By using data from bar code companies, or sales data from catalog companies, it is possible to gain valueible information about customer buying behavior. The derived information might be used, for example, by retailers in deciding which items to shelve in a supermarket, or for designing a well targeted marketing progreim, among others. Numerous meaningful insights can be unearthed from the data utilizing proper analysis techniques. In the most general sense, data mining is concerned with the analysis of data and the use of software techniques for finding patterns and regularities in sets of data. The objective of data mining is to source out discernible patterns and trends in data and infer association rules from these patterns.
Data mining technologies are characterized by intensive computations on large volumes of data. Large databases are definable as consisting of a million records or more. In a typical application, end users will test association rules such as; "75% of customers who buy Cola also buy corn chips", where 75% refers to the rule's confidence factor. The support
VOQ07 237

of the rule is the percentage of transactions that contain both Cola and corn chips.
To date the prior art has not addressed the issue of online mining but has instead focused on an itemset approach. IBM's Almaden's project called Quest is based upon this method, A significant drawback of the itemset approach is that as the user tests the database for association rules at differing values of support and confidence, multiple passes have to be made over the database, which could be of the order of Gigabytes. For very large databases, this may involve a considerable amount of I/O and in some situations, it may lead to unacceptable response times for online queries. A user must make multiple tjueries on a database because it is difficult to guess apriori, how many rules might satisfy a given level of support and confidence. Typically one may be interested in only a few rules. This makes the problem all the more difficult, since a user may need to run the query multiple times in order to find appropriate levels of minimum support and minimum confidence in order to mine the rules. In other words, the problem of mining association rules may require considerable manual parameter tuning by repeated queries, before useful business information can be gleaned from the transaction database. The processing methods of mining described heretofore are therefore unsuitable to repeated online queries as a result of the
VBPPi' 337

extensive disk I/O or computation leading to unacceptable response times. The need for expanding, the capabilities of data mining to the internet requires dynamic online methods rather than the batch oriented method of the itemset approach. It is therefore a primary object of the invention to provide a computationally efficient method for making online queries on a dateJsase to evaluate the strength of association rules utilizing user supplied levels of support and confidence as predictors.
It is a further object object of the invention to discover quantitative association rules.
SUMMRRY Qg THE INVEKTION
The present invention is directed to a method for efficiently performing online mining of quantitative association rules. An association rule can be generally defined as a conditional statement that suggests that there exists some correlation between its two component parts, antecedent and consequent. In a quantitative association rule both the antecedent and consequent are composed from some user specified combination of quantitative and categorical attributes. Along with the proposed rule, the user would provide three additional inputs representing the confidence and support level of interest to the user and a value referred to as interest level. These

inputs provide an indication of the strength of the rule proposed by the user (the user query). In other words the strength of the suggested correlation between antecedent and consequent defined by the user query.
In order to carry out the object of the present invention, there is disclosed, a method for preprocessing the raw data by utilizing the antecedent attributes to partition the data so as to create a multidemensional indexing structure,followed by an on¬line rule generation step. By effectively pre¬processing the data into an indexing structure it is placed in a form suitable to answer repeated online queries with practically instantaneous response times. Once created, the indexing structure obviates the need to make multiple passes over the database. The indexing structure creates significant performance advantages over previous techniques. The indexing structure (pre-processed data) is stored in such a way that online processing may be done by applying a graph theoretic search algorithm whose complexity is proportional to the size of the output. This results in an online algorithm which is practically instantaneous in terms of response time, minimizing excessive amounts of I/O or computation.

Accordingly the present invention provides a method of mining of a large database having a plurality of records, each record having a plurality of quantitative and categorical items for providing quantitative association rules, comprising the steps of:
a) receiving a user defined value of minimum confidence, a user defined value of minimum support, and a user query comprising antecedents and consequent attributes expressed in terms of said quantative and/or categorical items;
b) organizing the relationship between antecedent and consequent attributes by pre-storing antecedent data hierarchically into an index tree comprising a multiplicity of index nodes, each index node having first and second values representing die actual support and the confidence for each user query consequent attribute; and
c) deriving an answer from said pre-stored data in response to said user query by searching all the index nodes of said index, tree to isolate those nodes whose antecedent attribute range corresponds to said user query antecedent a attribute range and which have confidence at least equal to said user defined value of minimum confidence and a value of support at least equal to said user defmed minimum value of support.
With reference to the accompanying drawings:

BRIEF DESCRIPTIOK QP THE DRAWINGS
FIG0RE 1 is an overall description of the computer network in which this invention operates.
FIGURE 2 is an overall description of the method performed by the invention. It consists of two stages described by Figures 2(a) and 2(b). Figure 2(a) is a description of the preprocessing stage. Figure 2(b) is a description of the on-line stage of the algorithm.
FIGURE 3 is a detailed description of how the index tree is constructed using the antecedent set. It can be considered an expansion of step 75 of Figure 2(a).
FIGURE 4 is a detailed description of how the unmerged rule tree is generated from the index tree. It can be considered an expansion of step 100 of Figure 2(b),
FIGURE 5 is a description of how the merged rule tree is built from the unmerged rule tree.
FIGURE 6 is a description of how the quantitative association rules are generated from the merged rule tree at some user specified interest level r.

DESCRIPTION OF THE PREFERRED EMBODIMENTS
The present invention is directed to a method for online data mining of quantitative association rules. Traditional database queries consisting of simple questions such as "what were the sales of orange juice in January 1995 for the Long Island area?". Data mining, by contrast, attempts to source out discernible patterns and trends in the data and infers rules from these patterns. With these rules the user is then able to support, review and examine decisions in some related business or scientific area. Consider, for example, a supermarket with a large collection of items. Typical business decisions associated with the operation concern what to put on sale, how to design coupons, and how to place merchandise on shelves in order to maximize profit, etc. Analysis of past transaction data is a commonly used approach in order to improve the quality of such decisions. Modern technology has made it possible to store the so called basket data that stores items purchased on a per-transaction basis. Organizations collect massive cutiounts of such data. The problem becomes one of "mining" a large collection of basket data type transactions for association rules between sets of items with some minimum specified confidence. Given a set of transactions, where each transaction is a set of items, an association rule is an expression of the form X => Y, where X and Y are sets of items.

An example of an association rule is: 30% of transactions that contain beer also contain diapers;. 2% of all transactions contain both of these items". Here 30% is called the confidence of the rule, and 2% the support of the rule.
Another example of such an association rule is the statement that 90% of customer transactions that purchase bread and butter also purchase milk. The antecedent of this rule, X, consists of bread and butter and the consequent, Y, consists of milk alone. Ninety percent is the confidence factor of the rule. It may be desirable, for instance to find all rules that have 'bagels' in the antecedent which may help determine what products (the consequent) may be impacted if the store discontinues selling bagels.
Given a set of raw transactions,!?, the problem of mining association rules is to find all rules that have support and confidence greater than the user-specified minimum support (minsupport s} and minimum confidence (minconfidence c). Generally, the support of a rule X =>Y is the percentage of customer transactions, or tuples in a generalized database, which contain both X and Y itemsets. In more formal mathematical terminology, the rule X=> Y has support s in the transaction set D if s% of transactions in D ^^^T^^-a^■n 5f imion Y, X V Y. The confidence of a rule

X => Y is defined as the percentage of transactions that contain X which also contain Y. Or more fonnally, the rule X=> Y has confidence c in the transaction set D if c% of transactions in D that contain X also contain Y. Thus if a rule has 90% confidence then it means that 90% of the transactions containing X also contain Y.
As previously stated, an association rule is an expression of the form X => Y. For example if the itemsets X and Y were defined to be
X= [milk & cheese & butter] Y= [eggs & ham] respectively
The rule may be interpreted as:
RULE : X=> Y , implies that given the occurrence of milk, cheese and butter in a transaction, what is the likelihood of eggs and ham appearing in that same transaction to within some defined support and confidence level.
The support and confidence of the rule collectively define the strength of the rule. There are a number of ways in which a user may pose a rule to such a system in order to test its strength, A non-inclusive

yet representative list of the kinds of online queries that such a system can support include ;
(1) Find all association rules above a certain level of minsupport and minconfidence.
(2) At a certain level of minsupport and minconfidence, find all association rules that have the set of items X in the
antecedent.
(3) At a certain level of minsupport and
minconfidence, find all association rules
that have the set of items Y in the
consequent.
(4) At a certain level of minsupport and minconfidence, find all association rules that have the set of items y either in the antecedent or consequent or distributed between the antecedent and consequent.
(5) Find the number of association rules/itemsets in any of the cases (1), (2), (3), (4) above.

(6) At what level of minsupport do exactly k itemsets exist containing the set of items. z.
The present method particularizes the method of discovering general association rules to finding quantitative rules from a large database consisting of a set of raw transactions, D, defined by various quantitative and categorical attributes.
For example, a typical quantitative/categorical database for a general marketing survey would consist of a series of records where each record reflects some combination of consumer characteristics and preferences;
Record (1) = age=21, sex=male, homeowner=no Record (2) = age=43, sex=male, homeowner=yes Record (3) = age=55, sex=female, homeowner=no
In general, a quantitative association rule is a condition of the form;
GENERAL RULE : Xl[ll.,ul],x2[12..u21...Xk(lk..ul(] Yl=cl. Y2Bc2..Yr=cr -> Z1=Z1,Z2=E2
where Xl,X2,.,Xk correspond to quantitative antecedent attributes, and Yl,Y2,..Y'r, and C correspond to

categorical antecedent attributes. Here [11..ul], [12..u2], ...[lk...uk] correspond to the ranges for the various quantitative attributes. Zl and Z2 correspond to a multiple consequent condition.
The present method requires that a user supply three inputs, a proposed rule, otherwise referred to as the user query, in the form of an antecedent/consequent pair. In addition to the proposed rule a user would supply values for minimum required confidence (minconfidence= c), and minimum required support, (minsupport= s), to test the strength of the proposed rule (user query).
Both the minimum confidence and and minimum support are as relevant to the discovery of quantitative association rules as they are to the discovery of general association rules. An exan^le of a typical user input might be;
BXAHPLB A ; TYPICAL USER INPUT
1. User supplies a proposed Rule to be tested (query)
ANTECEDEHT CONOrTIOM CONSEQUENT COHPrTION
Age[20-40].Salary[100k-2ODk], Sex=Female => Cars=2
2. User supplies a confidence value for the proposed rule,
referred to as minconfidence, c.
Minconfidence = 50%

3. User supplies a support value for the proposed rule, Minsupport, s.
Minsupport = 10%
Figure 1 is an overall description of the architecture of the present method. There are assumed to be multiple clients 40 which can access the preprocessed data over the network 35. The preprocessed data resides at the server 5. There may be a cache 25 at the server end. along with the preprocessed data 20. The preprocessing as well as the online processing takes place in the CPU 10, In addition, a disk 15 is present in the event that the data is stored on disk.
The present method comprises two stages, a pre¬processing stage followed by an online processing stage. Fig. 2(a) shows an overall description of the preprocessing step as well as the online processing {rule generation steps) for the algorithm. The pre¬processing stage involves the construction of a binary index tree structure, see step 75 of FIG. 2 and the associated detailed description of FIG. 3(a). The use of an index tree structure is a well known spatial data structure in the art which is used as a means to index on multidimensional data. Related work in prior art may be found in Guttman. ft. ., h- dY"ainic Jndfi-x structure for Snatial Searching. Proceedings of the

ftCM STflMOD Conference. In the present method a variation on this index tree structure is employed in order to perform the on-line queries. Antecedent attributes are utilized to partition the data so as to create a multidimensional indexing structure. The indexing structure is a two-level structure where the higher level nodes are associated with at most two successor nodes and lower level nodes may have more than two successor nodes. The construction of the indexing structure is crucial to performing effective online data mining. The key advantage resides in minimizing the amount of disk I/O required to respond to user queries.
A graphical analogue of the indexing structure, stored in computer memory, is shown shown in FIG. 3(b) in the form of an index tree. An index tree is a well known spatial data structure which is used in order to index on multi-dimensional data. A separate index structure will be created in computer memory for each dimension, defined by a particular quantitative attribute, specified by the user in the online query. Figure 3(b) is a specific example of an index tree structure which represents the antecedent condition, "Age" and its associated consequent condition, "FirstTimeBuyer*. To further clarify the concept of an index tree. Pig. 3(b) could have represented the "Age" dimension in the example below;

EXAMPLE B : SAKFLE USER QDERY
AMTECEDEHT COHDITIOH COHSROHECT CONDTTIQN
Salary[40k-85k],Age[0-100],Sex => FirstTimeBuyer
In general there are no restrictions with respect to the quantity or combination of quantitative and categorical attributes which comprise the antecedent and consequent conditions.
In Figure 3(b) the root node of the index tree structure defines the user specified quantitative attribute, AgetO-100]. Each of the successive nodes of the tree also represent the quantitative attribute. Age, with increasingly narrower range limits from the top to the bottom of the tree heirarchy. For example, the binary successors to the root node for age[0-100] are Age[0-45] and Age[45-100]. The present method stores two pieces of data at each node of the index tree representing the confidence and support levels of Interest, For example, with reference to Figure 3{b), at the root node, two pieces of data are stored consisting of;
1. confidence level = 50%
2. support level = function of data input to
the raw database

defining the confidence and support for the user query, (antecedent/consequent pair),
age[0-100] =>Firs t timebuyer at the root node,
FIG. 3{a) is the detailed flowchart of the preprocessing stage of the algorithm, illustrated in PIG.2 as element 100. The process steps of this stage involve generating the binary index tree structure and storing the support and confidence levels for the consequent attribute at each node of the structure, followed by utilizing a compression algorithm on the lower levels of the structure to ensure that the index tree fits into the available memory. Step 300 is the point of entry into the preprocessing stage. Step 310 represents the software to implement the process step of using a binarization algorithm to generate a binary index tree. The binarization step has been discussed in the prior art in Rggarwal r. C Wolf J.. Yu p. s.. and Eoelman M. A. The S-Tree: An effic^ienti index tree for multidimensional index trp.es. Swmosium of Spatjal Databases. 1997. However, the present method diverges from the prior art in at least one aspect. At Step 315, the way in which the entries of an index

node are organized is unique in that both the support level and the confidence level for each value of the. consequent attribute are stored at each node in the structure. Step 320 represents the software to implement the process step of utilizing a compression algorithm to compress the lower level index nodes into a single node.
PIG. 4(a> is the detailed flowchart of the primary search algorithm which is used in order to generate the unmerged rule tree from the index tree, illustrated in PIG, 2(b) as element 100. The algorithm requires as input, user specified values for minconfidence c, minsupport s, and a user query which consists of a Ouerybox Q and one or more right hand side values, Zl^zl, Z2=z2. The Querybox is merely a descriptive term to denote the lefthand or antecedent portion of the user query. To further clarify the meaning of Querybox, Excimple C below describes what is recjuired of an online user as input in the present method;
1BXMII>LE C I TYPICAL USER INPOT
The user would specify :
(1.) a minimum confidence value, I minconfidence, c]
(2.) a minimum support value, [ minsupport, s ]

An online user would, in addition be required to input a user query (proposed rule) in Che form of an an(antecedent/consequent) pair, items 384.
(3.) a Querybox, 'Q- [the antecedent] (4.) Zl=zl, Z2=z2, etc.. ta consequent]
Item three, the Querybox, is further explained by the following examples, and can generally consist of any combination of quantitative and categorical attributes. Item four, the consequent attribute, can consist of one or more categorical attributes.
[Exaniple 1]: This user specified query consists of an antecedent condition, querybojc, with two dimensions. Age and Lefthandedneas, and a single categorical consequent condition, asmoker.
Age [0-241.Lefthanded ==> asmoker
(Example 2]: This user specified query consists of an antecedent condition, querybox, with two dimensions, Height and Income and a multiple consequent condition.
Oijerv^iox
Height(5-7],Income[10k-40k] ==> ownsahome.ovmsacar

[Example 3]I The user specified query consists of a single antecedent condition, querybox, with a single dimension. Age, and a single consequent condition.
Ouervbox AgeflO-43] *=> aamoker
Example C above, describes in general terms what a user supplies as input to the method. Example D below provides a representative example. Using the user query in example 2 above, a typical input/output result could look as follows:
Example D :
user specifies as input::
1. minconfidence = .50
2. minsupport = .4
3. querybox (antecedent condition) = Height [5-6) ,Inconie(103t-40k]
4. consequent condition of interest = ownsahome=l,ownsacar=l
user query fonaed from items (3&4} ,-
Height (5-71,Income[10k-4Ok] ==> ownsahome,ownaacar
[leBUlti"Q output : generated rule
heightl5.5 ■ 6,2],lncomeI13k - 27.4k] ==> ownsahorae=l,ownsacar=l
In general, the output can conceivably generate no rules, one rule, or multiple rules. A single rule was

generated in the example above. The generated rule is said to satisfy the user query,(antecedent/consequent pair), at the user specified confidence and support level, .5 and .4 respectively.
The algorithm for generating the unmerged rule tree from the index tree, defined by Figure 4(a), proceeds by searching all the nodes in the index tree one by one. Step 400 is the point of entry into the primary search algorithm. Step 410 represents the software to implement the process step of setting a pointer, Currentnode to point to the root node of the index tree. Pointer CurrenCWode will always point to the particular node of the index tree which the algorithm is presently searching. Step 420 defines LIST as a set of nodes which are considered to be eligible nodes to be scanned by the search algorithm. LIST is initialized to contain only the root node in step 420. Step 430 represents the software to implement the process step of adding all the child nodes of the node pointed to by Currentnode to LIST which intersect with Querybox 0, and have support at least equal to the user supplied input value, minsupport, s. A child node is said to intersect with Querybox Q, when all of the antecedent conditions associated with the child node are wholly contained within the antecedent condition defined by the Querybox. Step 440 is a

decision step which determines whether the individual data records contained in CurrentWcjde satisfy the consequent condition, Zl=2l and Z2=z2 at least c percent of the time. If the condition of step 440 is satisfied then the algorithm proceeds to step 445. Step 445 generates the rule corresponding to the set of attributes on the right hand side, the consequent condition. Step 450 follows steps 440 and 445 and represents the software to implement the process step of deleting the node presently pointed to by Currentnode from LIST and setting the pointer Currentnode to the next node contained in LIST. Step 460 determines whether LIST is empty and terminates the algorithm when the condition is met, see Step 470. Otherwise, the algorithm returns to step 430 and repeats the steps for the node currently pointed to by the pointer CurrentWode. Upon termination of the algorithm, an unmerged rule tree is output which consists of all nodes in the input index tree which satisfy the user specified minimum support, minsupport s.
FIG. 5(a) is the detailed flowchart which describes the process of constructing the merged rule tree from the unmerged rule tree. The algorithm described by the flowchart compresses the unmerged rule tree to obtain a hierarchical representation of the rules. The unmerged rule tree is traversed in depth first

search order where at each node a determination is made as to whether that node is meaningful. A meaningful node is defined to be a node which has a rule associated with it. A rule may or may not have been associated with a node when the unmerged rule tree was created. To further clarify the distinction between meaningful and nonmeaningful nodes, refer back to Fig. 4(b), the unmerged rule tree, where meaningful nodes correspond to nodes 1,2, and 4. All meaningful nodes are preserved in the merged rule tree. If a node is determined not to be meaningful then the algorithm either eliminates that node, or merges multiple child nodes into a single node when certain conditions are met.
Step 500 represents the point of entry into the algorithm. Step 510 represents the software to implement the process step of insuring that the unmerged rule tree is traversed in depth first search order. Step 515 represents the step of proceeding to the next node in the unmerged rule tree in the depth first traversal. Step 520 represents a decision step which determines whether the current rule node is a meaningful node. A branch is made to step 530 when the current node is determined to be meaningful. Otherwise the algorithm branches to step 540 thereby classifying the node as nonmeaningful. Step 540 is a decision step which determines whether the

nonmeaningful node has a child node. If the nonrneaningful node does have a child node a branch is taken to step 550. Step 550 represents the software to implement the process step of deleting the current nonmeaningful node. Otherwise, if it is determined in step 540 that the current node does not have a child node, a branch will be taken to step 560. Step 560 is a decision step for the purpose of determining whether the current nonmeaningful node has one or more than one child nodes. If the current node has only a single child node then a branch is taken to step 570. Step 570 represents the software to implement the process step of deleting the current node and directly connecting the parent and child nodes of the deleted nonmeaningful node together in the index tree. Otherwise, in the case where the current node is found to have multiple child nodes a branch is taken to step 580. Step 580 is a decision step which determines whether the minimum bounding rectangle of the two child nodes are more than that of the nonmeaningful parent node. The minimum bounding rectangle is defined by the upper and lower bounds (the range) of the quantitative attribute for each child node. When the ranges of the child nodes are combined and found to be broader than the range of the parent node, a merger occurs. For example, if the child nodes were defined as;

child node 1 - age [10-201 child node 2 - age [30-40]
and the corresponding parent node were defined as;
parent node - age [10-30]
then a merger would occur in this example, since the combination of the child attribute ranges yields a combined range of [10-40] which is broader than than range specified by the parent node,[10-30] .
If the confidence of the minimum bounding rectangle of the two child nodes exceeds that of the parent node, a branch will occur to step 590. Step 590 represents the software to perform the process step of adjusting the minimum bounding rectangle of the parent to be the minimum bounding rectangle of the two child nodes. A branch to decision step 600 determines whether there are any more nodes to traverse in the tree. A branch to termination step 610 occurs if there are no more nodes to traverse, otherwise process steps 490-515 are repeated for the remaining index nodes.
PIG. 6 is the detailed flowchart which describes the process of using the merged rule tree as input to define the rules at the user specified interest level r. The merged rule tree is traversed in depth first

order. Step 616 is the point of entry into the flowchart. A user would specify an input value for r, representing the interest level. Step 618 represents the software to select the next node in the merged rule tree in depth first order. Step 620 is a decision step which represents the software which looks at all ancestral nodes of the current node of interest to determine whether any of them has a confidence value at least egual to 1/r of the current node. A branch to Step 630 will be taken when condition is true. Step 630 represents the software to prune the rule associated with the current node. If the condition is not met, a branch to step 640 is taken. Step 640 is a decision step which determines whether there are any remaining nodes to be evaluated in the merged rule tree. The process steps will be repeated if there are additional nodes to be evaluated, otherwise the process terminates at this point.
While the invention has beer, particularly shown and described with respect to illustrative and preformed embodiments thereof, it will be understood by those
skilled in the art that the foregoing and other changes in form and details nay be made therein without departing from the spirit and scope of the invention which should be linitec only by the scope of the appended claims.


WE CLAIM :
1. A method of mining of a large database having a plurality of records, each
record having a plurality of quantitative and categorical items for providing
quantitative association rules, comprising the steps of:
a) receiving a user defined value of minimum confidence, a user defined value of minimum support, and a user query comprising antecedents and consequent attributes expressed in terms of said quantative and/or categorical items;
b) organizing the relationship between antecedent and consequent attributes by pre-storing antecedent data hierarchically into an index tree comprising a multiplicity of index nodes, each index node having first and second values representing the actual support and the confidence for each user query consequent attribute; and
c) deriving an answer from said pre-stored data in response to said user query by searching all the index nodes of said index, tree to isolate those nodes whose antecedent attribute range corresponds to said user query antecedent a attribute range and which have confidence at least equal to said user defined value of minimum confidence and a value of support at least equal to said user defined minimum value of support,
2. The method as claimed in claim 1, wherein said answer comprises one or
more quantitative association rules, an actual confidence value associated
with each rule, and an actual support value associated with each rule.

3. The method as claimed in claim 1 or 2, wherein a user defined interest level is provided in said receiving step, and said answer includes an interest level associated with each rule, whereby said one or more quantitative association rules consist of only those rules whose computed interest level is at least equal to said user defined interest value.
4. The method as claimed in claim 3, wherein said interest level is defined as the minimum of a first and a second computed ratio, wherein said first ratio is defined as the actual confidence divided by an expected confidence and a second ratio is defined as the actual support divided by an expected support, wherein said expected confidence and support are computed values based on a presumption of statistical independence.
5. The method as claimed in any one of the preceding claims, wherein said antecedent attributes are comprised of categorical and quantitative attributes.
6. The method as claimed in claim 5, wherein said quantitative attributes are further defined by a range consisting of a lower and upper bound.
7. The method of any preceding claim, wherein said deriving step comprises building a merge tree by deleting meaningless nodes and combining other nodes, wherein a meaningless node is a node which does not have a corresponding calculated value of confidence at least equal to said user defined value of minimum confidence.
8. The method as claimed in claim 7, wherein the merge tree may be built either for a single or for multiple consequent attributes.

9. The method as claimed in claim I, wherein:
said receiving step comprising inputting to a computet data including a user defined value of minimum support, a user defined value of minimum confidence, a user defined value of interest, and a user query comprising an antecedent and consequent condition, where said antecedent and consequent condition comprise a plurality of quantitative and categorical attributes;
said organizing step comprises constructing in memory an index tree comprised of one or more dimensions, where each dimension is defined by one of the user supplied quantitative attributes contained in said antecedent condition, said index tree consisting of a plurality of index nodes where said index nodes consist of a plurality of data records;
and said deriving step comprises constructing in memory an unmerged rule tree from said index tree, and a merged rule tree from said unmerged rule tree and generating one or more quantitative association rules fi"om those index nodes that satisfy said user query and whose support is at least equal to said minimum support, and whose confidence is at least equal to said minimum confidence;
said method comprising the step of displaying to a user output data consisting of: said quantitative association rules fi-om the generating step; a value of actual confidence associated with each generated quantitative association rule; a value of support associated with each generated quantitative association rule; and a value of interest level associated with each generated quantitative association rule.

10. The method as claimed in claim 9, wherein the step of generating one or more quantitative association rule is repeated as that said user query is interactively modified to further define said association rules.
11. The method as claimed in claim 9 or 10, wherein the step of constructing an
index tree comprises the step of:
constructing a binary index tree of one or more dimensions, where each dimension is defined by one of said user supplied quantitative antecedent attributes; and
storing at each index node said support level and confidence level.
12. The method as claimed in claim 9, 10 or 11, wherein the step of
constructing an unmerged rule tree comprises the step of:
searching each node of said index tree; and
selecting these nodes which contain rules which satisfy the user specified consequent condition and have confidence at least equal to said user defined value of minimum confidence, and a value of support at least equal to said user defined value of minimum support.
13.The method as claimed in claim 12, wherein the step of selecting those nodes which contain rules which satisfy the user specified consequent condition comprises:
constructing a pointer;
equating said pointer to a root node in said index tree;
adding said node associated with said pointer to a list;
adding to the list all children of the node pointed to by said pointer with antecedent attribute wholly contained within the parameters of said

user specified antecedent attribute and with a minimum support value at least equal to said user defined minimum support;
determining whether the data records stored at the node pointed to by said pointer at least equal the user specified consequent condition and have a confidence at least equal to said user defined minimum confidence; generating a quanthative association rule associated with said consequent condition;
deleting said node from said list when the conditions of the previous step are not satisfied;
determing whether said list is empty; and
terminating when said list is empty, otherwise equating said pointer to the next node of said index tree, and repeating the above steps from said step of adding the node associated with said pointer to the list onwards.
14. The method as claimed in any one of claims 9 to 13, 12, wherein the step of building a merged rule tree comprises :
a) traversing each node of the unmerged rule tree in post order;
b) evaluating each traversed node for inclusion or exclusion in the unmerged rule tree, by:
i) determining whether each said user defined consequent attribute
value is greater than the consequent attribute value stored at said
node; ii) preserving said node in said merged rule tree when the condition
of (i) is satisfied; iii) deleting said node from said merged rule tree when the condition
of (i) is not satisfied and said node has no associated child
nodes; iv) deleting said node from said merged rule tree and directly
associating an ancestor node and child node of said deleted node

when the condition of (i) is not satisfied and said node has one
child node; and v) adjusting the range of said consequent attribute when the
condition of (i) is not satisfied; wherein said evaluating step is repeated until all nodes have been traversed in post order.
15. Apparatus for mining of a large database having a plurality of records, each record having a plurality of quantitative and categorical items for providing quantitative association rules, comprising;
means for receiving a user defined value of minimum confidence, a user defined value of minimum support, and a user query comprising antecedent and consequent attributes expressed in terms of said quantitative and/or categorical items;
means for organizing the relationship between antecedent and consequent attributes by pre-storing antecedent data hierarchically into an index tree comprising a multiplicity of index nodes, each index node having first and second value representing the actual support and the confidence for each user query consequent attribute; and
means for deriving an answer from said pre-stored data in response to said user query by searching all the index nodes of said index tree to isolate those nodes whose antecedent attribute range corresponds to said user query antecedent attribute range and which have confidence at least equal to said user defined value of minimum confidence and a value of support at least equal to said user defined minimum value of support.




Documents:

2315-mas-1998 abstract-duplicate.pdf

2315-mas-1998 abstract.pdf

2315-mas-1998 assignment.pdf

2315-mas-1998 claims-duplicate.pdf

2315-mas-1998 claims.pdf

2315-mas-1998 correspondence-others.pdf

2315-mas-1998 correspondence-po.pdf

2315-mas-1998 description (complete)-duplicate.pdf

2315-mas-1998 description (complete).pdf

2315-mas-1998 drawings-duplicate.pdf

2315-mas-1998 drawings.pdf

2315-mas-1998 form-19.pdf

2315-mas-1998 form-2.pdf

2315-mas-1998 form-26.pdf

2315-mas-1998 form-4.pdf

2315-mas-1998 form-6.pdf

2315-mas-1998 others.pdf

2315-mas-1998 petition.pdf


Patent Number 200751
Indian Patent Application Number 2315/MAS/1998
PG Journal Number 30/2009
Publication Date 24-Jul-2009
Grant Date
Date of Filing 15-Oct-1998
Name of Patentee INTERNATIONAL BUSINESS MACHINE CORPORATION
Applicant Address NEW YORK 10504
Inventors:
# Inventor's Name Inventor's Address
1 CHARU CHANDRA AGGARWAL APT. C-2-5, 38 (1/2) WOLDEN ROAD OSSINING, NEW YORK 10562
2 PHILIP SHI-LUNG YU 18 STORNOWAYE, CHAPPAQUA, NEW YORK 10514
PCT International Classification Number G06F17/30
PCT International Application Number N/A
PCT International Filing date
PCT Conventions:
# PCT Application Number Date of Convention Priority Country
1 08/964,064 1997-11-04 U.S.A.