Exploratory Data Analysis and Manipulation of Multi-Label Data Sets


[Up] [Top]

Documentation for package ‘mldr’ version 0.4.3

Help Pages

+.mldr Generates a new mldr object joining the rows in the two mldrs given as input
==.mldr Checks if two mldr objects have the same structure
accuracy Multi-label averaged evaluation metrics
Averaged metrics Multi-label averaged evaluation metrics
average_precision Multi-label ranking-based evaluation metrics
Basic metrics Multi-label evaluation metrics
birds birds
concurrenceReport Generates a label concurrence report
coverage Multi-label ranking-based evaluation metrics
emotions emotions
example_auc Multi-label ranking-based evaluation metrics
fmeasure Multi-label averaged evaluation metrics
genbase genbase
hamming_loss Multi-label evaluation metrics
labelInteractions Provides data about interactions between labels
macro_auc Multi-label ranking-based evaluation metrics
macro_fmeasure Multi-label averaged evaluation metrics
macro_precision Multi-label averaged evaluation metrics
macro_recall Multi-label averaged evaluation metrics
micro_auc Multi-label ranking-based evaluation metrics
micro_fmeasure Multi-label averaged evaluation metrics
micro_precision Multi-label averaged evaluation metrics
micro_recall Multi-label averaged evaluation metrics
mldr Creates an object representing a multilabel dataset
mldrGUI Launches the web-based GUI for mldr
mldr_evaluate Evaluate predictions made by a multilabel classifier
mldr_from_dataframe Generates an mldr object from a data.frame and a vector with label indices
mldr_to_labels Label matrix of an MLD
mldr_transform Transformns an MLDR into binary or multiclass datasets
one_error Multi-label ranking-based evaluation metrics
plot.mldr Generates graphic representations of an mldr object
precision Multi-label averaged evaluation metrics
print.mldr Prints the mldr content
Ranking-based metrics Multi-label ranking-based evaluation metrics
ranking_loss Multi-label ranking-based evaluation metrics
read.arff Read an ARFF file
recall Multi-label averaged evaluation metrics
remedial Decouples highly imbalanced labels
roc ROC curve
roc.mldr ROC curve
subset_accuracy Multi-label evaluation metrics
summary.mldr Provides a summary of measures about the mldr
write_arff Write an '"mldr"' object to a file
[.mldr Filter rows in a'mldr' returning a new 'mldr'