RapidMiner52.7 - Data Mining and RapidMiner
Machine LearningMachine learning algorithms: RapidMiner offers a huge number of learning schemes for:support vector machines (SVM), decision treerule learnerslazy learners,Bayesian learnersLogistic learners. association rule mining and clustering meta learning schemes including Bayesian Boosting.
Machine LearningDecision Trees: This operator learns decision trees from both nominal and numerical data. Decision trees are powerful classification methods which often can also easily be understood. This decision tree learner works similar to Quinlan‘s C4.5 or CART. The actual type of the tree is determined by the criterion, e.g. using gain ratio or Gini for CART / C4.5.
Machine LearningNeural Net: T his operator learns a model by means of a feed-forward neural network. The learning is done via back-propagation. The user can define the structure of the neural network with the parameter list “hidden layer types“. Each list entry describes a new hidden layer. The key of each entry must correspond to the layer type which must be one out ofˆ linearˆ sigmoid (default)ˆ tanhˆ sineˆ logarithmicˆ gaussian
Machine LearningBayesian Boosting: This operator trains an ensemble of classifiers for boolean target attributes. In each iteration the training set is reweighted, so that previously discovered patterns and other kinds of prior knowledge are sampled out.  An inner classifier, typically a rule or decision tree induction algorithm, is sequentially applied several times, and the models are combined to a single global model. The number of models to be trained maximally are specified by the parameter iterations.
Meta LearningBootstrap aggregating (bagging): is a machine learning ensemble meta-algorithm to improve machine learning of classification and regression models in terms of stability and classification accuracy. It also reduces variance and helps to avoid over fitting. Although it is usually applied to decision tree models, it can be used with any type of model. Bagging is a special case of the model averaging approach.
PreprocessingFeature Selection: Assume that we have a dataset with numerous attributes. We would like to test, whether all of these attributes are really relevant, or whether we can get a better model by omitting some of the original attributes. This task is called feature selection and the backward elimination algorithm is an approach that can solve it.
PreprocessingBackward Elimination in RapidMiner5:	Enclose the cross-validation chain by a FeatureSelection operator.	 This operator repeatedly applies the cross-validation chain, which now is its inner operator, until the specified stopping criterion is complied with. The backward elimination approach iteratively removes the attribute whose removal yields the largest performance improvement.
PreprocessingUserBasedDiscretization:This operator discretizes a numerical attribute to either a nominal or an ordinal attribute. The numerical values are mapped to the classes according to the thresholds specified by the user. The user can define the classes by specifying the upper limits of each class.
PreprocessingNormalization: This operator performs a normalization. This can be done between a user defined minimum and maximum value or by a z-transformation, i.e. on mean 0 and variance 1. or by a proportional transformation as proportion of the total sum of the respective attribute.
PreprocessingSampling: This operator performs a random sampling of a given fraction. For example, if the input example set contains 5000 examples and the sample ratio is set to 0.1, the result will have approximately 500 examples.
Genetic AlgorithmGenetic Algorithm: A genetic algorithm for feature selection (mutation=switch features on and off, crossover=interchange used features). Selection is done by roulette wheel. Genetic algorithms are general purpose optimization / search algorithms that are suitable in case of no or little problem knowledge.
ValidationA Simple Validation randomly splits up the example set into a training and test set and evaluates the model.
More Questions?Reach us at support@dataminingtools.netVisit: www.dataminingtools.net

RapidMiner: Data Mining And Rapid Miner

  • 1.
    RapidMiner52.7 - DataMining and RapidMiner
  • 2.
    Machine LearningMachine learningalgorithms: RapidMiner offers a huge number of learning schemes for:support vector machines (SVM), decision treerule learnerslazy learners,Bayesian learnersLogistic learners. association rule mining and clustering meta learning schemes including Bayesian Boosting.
  • 3.
    Machine LearningDecision Trees:This operator learns decision trees from both nominal and numerical data. Decision trees are powerful classification methods which often can also easily be understood. This decision tree learner works similar to Quinlan‘s C4.5 or CART. The actual type of the tree is determined by the criterion, e.g. using gain ratio or Gini for CART / C4.5.
  • 4.
    Machine LearningNeural Net:T his operator learns a model by means of a feed-forward neural network. The learning is done via back-propagation. The user can define the structure of the neural network with the parameter list “hidden layer types“. Each list entry describes a new hidden layer. The key of each entry must correspond to the layer type which must be one out ofˆ linearˆ sigmoid (default)ˆ tanhˆ sineˆ logarithmicˆ gaussian
  • 5.
    Machine LearningBayesian Boosting:This operator trains an ensemble of classifiers for boolean target attributes. In each iteration the training set is reweighted, so that previously discovered patterns and other kinds of prior knowledge are sampled out. An inner classifier, typically a rule or decision tree induction algorithm, is sequentially applied several times, and the models are combined to a single global model. The number of models to be trained maximally are specified by the parameter iterations.
  • 6.
    Meta LearningBootstrap aggregating (bagging):is a machine learning ensemble meta-algorithm to improve machine learning of classification and regression models in terms of stability and classification accuracy. It also reduces variance and helps to avoid over fitting. Although it is usually applied to decision tree models, it can be used with any type of model. Bagging is a special case of the model averaging approach.
  • 7.
    PreprocessingFeature Selection: Assumethat we have a dataset with numerous attributes. We would like to test, whether all of these attributes are really relevant, or whether we can get a better model by omitting some of the original attributes. This task is called feature selection and the backward elimination algorithm is an approach that can solve it.
  • 8.
    PreprocessingBackward Elimination inRapidMiner5: Enclose the cross-validation chain by a FeatureSelection operator. This operator repeatedly applies the cross-validation chain, which now is its inner operator, until the specified stopping criterion is complied with. The backward elimination approach iteratively removes the attribute whose removal yields the largest performance improvement.
  • 9.
    PreprocessingUserBasedDiscretization:This operator discretizesa numerical attribute to either a nominal or an ordinal attribute. The numerical values are mapped to the classes according to the thresholds specified by the user. The user can define the classes by specifying the upper limits of each class.
  • 10.
    PreprocessingNormalization: This operatorperforms a normalization. This can be done between a user defined minimum and maximum value or by a z-transformation, i.e. on mean 0 and variance 1. or by a proportional transformation as proportion of the total sum of the respective attribute.
  • 11.
    PreprocessingSampling: This operatorperforms a random sampling of a given fraction. For example, if the input example set contains 5000 examples and the sample ratio is set to 0.1, the result will have approximately 500 examples.
  • 12.
    Genetic AlgorithmGenetic Algorithm:A genetic algorithm for feature selection (mutation=switch features on and off, crossover=interchange used features). Selection is done by roulette wheel. Genetic algorithms are general purpose optimization / search algorithms that are suitable in case of no or little problem knowledge.
  • 13.
    ValidationA Simple Validationrandomly splits up the example set into a training and test set and evaluates the model.
  • 14.
    More Questions?Reach usat [email protected]: www.dataminingtools.net