use of org.apache.spark.ml.Pipeline in project mmtf-spark by sbl-sdsc.
the class SparkMultiClassClassifier method fit.
/**
* Dataset must at least contain the following two columns:
* label: the class labels
* features: feature vector
* @param data
* @return map with metrics
*/
public Map<String, String> fit(Dataset<Row> data) {
int classCount = (int) data.select(label).distinct().count();
StringIndexerModel labelIndexer = new StringIndexer().setInputCol(label).setOutputCol("indexedLabel").fit(data);
// Split the data into training and test sets (30% held out for testing)
Dataset<Row>[] splits = data.randomSplit(new double[] { 1.0 - testFraction, testFraction }, seed);
Dataset<Row> trainingData = splits[0];
Dataset<Row> testData = splits[1];
String[] labels = labelIndexer.labels();
System.out.println();
System.out.println("Class\tTrain\tTest");
for (String l : labels) {
System.out.println(l + "\t" + trainingData.select(label).filter(label + " = '" + l + "'").count() + "\t" + testData.select(label).filter(label + " = '" + l + "'").count());
}
// Set input columns
predictor.setLabelCol("indexedLabel").setFeaturesCol("features");
// Convert indexed labels back to original labels.
IndexToString labelConverter = new IndexToString().setInputCol("prediction").setOutputCol("predictedLabel").setLabels(labelIndexer.labels());
// Chain indexers and forest in a Pipeline
Pipeline pipeline = new Pipeline().setStages(new PipelineStage[] { labelIndexer, predictor, labelConverter });
// Train model. This also runs the indexers.
PipelineModel model = pipeline.fit(trainingData);
// Make predictions.
Dataset<Row> predictions = model.transform(testData).cache();
// Display some sample predictions
System.out.println();
System.out.println("Sample predictions: " + predictor.getClass().getSimpleName());
predictions.sample(false, 0.1, seed).show(25);
predictions = predictions.withColumnRenamed(label, "stringLabel");
predictions = predictions.withColumnRenamed("indexedLabel", label);
// collect metrics
Dataset<Row> pred = predictions.select("prediction", label);
Map<String, String> metrics = new LinkedHashMap<>();
metrics.put("Method", predictor.getClass().getSimpleName());
if (classCount == 2) {
BinaryClassificationMetrics b = new BinaryClassificationMetrics(pred);
metrics.put("AUC", Float.toString((float) b.areaUnderROC()));
}
MulticlassMetrics m = new MulticlassMetrics(pred);
metrics.put("F", Float.toString((float) m.weightedFMeasure()));
metrics.put("Accuracy", Float.toString((float) m.accuracy()));
metrics.put("Precision", Float.toString((float) m.weightedPrecision()));
metrics.put("Recall", Float.toString((float) m.weightedRecall()));
metrics.put("False Positive Rate", Float.toString((float) m.weightedFalsePositiveRate()));
metrics.put("True Positive Rate", Float.toString((float) m.weightedTruePositiveRate()));
metrics.put("", "\nConfusion Matrix\n" + Arrays.toString(labels) + "\n" + m.confusionMatrix().toString());
return metrics;
}
use of org.apache.spark.ml.Pipeline in project mmtf-spark by sbl-sdsc.
the class SparkRegressor method fit.
/**
* Dataset must at least contain the following two columns:
* label: the class labels
* features: feature vector
* @param data
* @return map with metrics
*/
public Map<String, String> fit(Dataset<Row> data) {
// Split the data into training and test sets (30% held out for testing)
Dataset<Row>[] splits = data.randomSplit(new double[] { 1.0 - testFraction, testFraction }, seed);
Dataset<Row> trainingData = splits[0];
Dataset<Row> testData = splits[1];
// Train a RandomForest model.
predictor.setLabelCol(label).setFeaturesCol("features");
// Chain indexer and forest in a Pipeline
Pipeline pipeline = new Pipeline().setStages(new PipelineStage[] { predictor });
// Train model. This also runs the indexer.
PipelineModel model = pipeline.fit(trainingData);
// Make predictions.
Dataset<Row> predictions = model.transform(testData);
// Display some sample predictions
System.out.println("Sample predictions: " + predictor.getClass().getSimpleName());
String primaryKey = predictions.columns()[0];
predictions.select(primaryKey, label, "prediction").sample(false, 0.1, seed).show(50);
Map<String, String> metrics = new LinkedHashMap<>();
metrics.put("Method", predictor.getClass().getSimpleName());
// Select (prediction, true label) and compute test error
RegressionEvaluator evaluator = new RegressionEvaluator().setLabelCol(label).setPredictionCol("prediction").setMetricName("rmse");
metrics.put("rmse", Double.toString(evaluator.evaluate(predictions)));
return metrics;
}
use of org.apache.spark.ml.Pipeline in project mm-dev by sbl-sdsc.
the class CathClassificationDataset method sequenceToFeatureVector.
private static Dataset<Row> sequenceToFeatureVector(Dataset<Row> data, int n, int windowSize, int vectorSize) {
// split sequence into an array of one-letter codes (1-grams)
// e.g. IDCGHVDSL => [i, d, c, g, h, v...
RegexTokenizer tokenizer = new RegexTokenizer().setInputCol("sequence").setOutputCol("1gram").setPattern("(?!^)");
// create n-grams out of the sequence
// e.g., 2-gram [i, d, c, g, h, v... => [i d, d c, c g, g...
NGram ngrammer = new NGram().setN(n).setInputCol("1gram").setOutputCol("ngram");
// convert n-grams to W2V feature vector
// [i d, d c, c g, g... => [0.1234, 0.23948, ...]
Word2Vec word2Vec = new Word2Vec().setInputCol("ngram").setOutputCol("features").setWindowSize(windowSize).setVectorSize(vectorSize).setMinCount(0);
Pipeline pipeline = new Pipeline().setStages(new PipelineStage[] { tokenizer, ngrammer, word2Vec });
// .setStages(new PipelineStage[] {tokenizer, word2Vec});
PipelineModel model = pipeline.fit(data);
data = model.transform(data);
return data;
}
Aggregations