rule generation in a big dataset


Given a dataset with 30 fields and 25000 instances,

1) what are your suggestions for novel methods of rule extraction?

2) Can I use association rule mining in addition to sequential rule mining?

3) which method can be more appropriate for such a big dataset?Apriori-based like SPIRIT, SPADE, SPAM, IBM or pattern growth ones like FreeSpan, PrefixSpan, SLPMiner?

*the output field (risk) is labelled as very low, low, medium, high, very high. Also, there is a temporal field (date) for each instance. An example is given below.

$$ \begin{array}{|c|c|c|c|} \hline \mathbf{date}& \mathbf{temperature} & \mathbf{density} & \mathbf{risk} \\ \hline 2018/1/2 & 15 & 100 & \text{very high}\\ \hline & & &\\ \end{array} $$


Posted 2019-04-01T14:56:02.463

Reputation: 11

Hi Anna, What do you mean by novel? There are methods to extract rules, yes. Without any given target variable you could use both methods to find correlations in the dataset. – S van Balen – 2019-04-01T22:15:39.103

Hi Balen, by the novel, I meant, for example, using deep neural networks or those that can work with my large dataset. – anna – 2019-04-02T17:08:15.103

No answers