+.mldr |
Generates a new mldr object joining the rows in the two mldrs given as input |
==.mldr |
Checks if two mldr objects have the same structure |
accuracy |
Multi-label averaged evaluation metrics |
Averaged metrics |
Multi-label averaged evaluation metrics |
average_precision |
Multi-label ranking-based evaluation metrics |
Basic metrics |
Multi-label evaluation metrics |
birds |
birds |
concurrenceReport |
Generates a label concurrence report |
coverage |
Multi-label ranking-based evaluation metrics |
emotions |
emotions |
example_auc |
Multi-label ranking-based evaluation metrics |
fmeasure |
Multi-label averaged evaluation metrics |
genbase |
genbase |
hamming_loss |
Multi-label evaluation metrics |
labelInteractions |
Provides data about interactions between labels |
macro_auc |
Multi-label ranking-based evaluation metrics |
macro_fmeasure |
Multi-label averaged evaluation metrics |
macro_precision |
Multi-label averaged evaluation metrics |
macro_recall |
Multi-label averaged evaluation metrics |
micro_auc |
Multi-label ranking-based evaluation metrics |
micro_fmeasure |
Multi-label averaged evaluation metrics |
micro_precision |
Multi-label averaged evaluation metrics |
micro_recall |
Multi-label averaged evaluation metrics |
mldr |
Creates an object representing a multilabel dataset |
mldrGUI |
Launches the web-based GUI for mldr |
mldr_evaluate |
Evaluate predictions made by a multilabel classifier |
mldr_from_dataframe |
Generates an mldr object from a data.frame and a vector with label indices |
mldr_to_labels |
Label matrix of an MLD |
mldr_transform |
Transformns an MLDR into binary or multiclass datasets |
one_error |
Multi-label ranking-based evaluation metrics |
plot.mldr |
Generates graphic representations of an mldr object |
precision |
Multi-label averaged evaluation metrics |
print.mldr |
Prints the mldr content |
Ranking-based metrics |
Multi-label ranking-based evaluation metrics |
ranking_loss |
Multi-label ranking-based evaluation metrics |
read.arff |
Read an ARFF file |
recall |
Multi-label averaged evaluation metrics |
remedial |
Decouples highly imbalanced labels |
roc |
ROC curve |
roc.mldr |
ROC curve |
subset_accuracy |
Multi-label evaluation metrics |
summary.mldr |
Provides a summary of measures about the mldr |
write_arff |
Write an '"mldr"' object to a file |
[.mldr |
Filter rows in a'mldr' returning a new 'mldr' |