Okay so just trying to understand please, in general…
If we work on a paper, and create different models there as part of my work, then:
To compare my models, it’s okay to just use machine learning metrics (the same metric), for example, if I have 4 models with say accuracy of 70%, 65%, 85%, 95%, then the best one will be the one with 95%. As we know accuracy is generally not the ideal metric due to possible class imbalance etc in general, so then, what would your advice be for: which metrics to choose, how many metrics to choose, and then, if we have multiple metrics, how do you say which is the “best model” from your study in general for your paper? Just would like to get your expert advice on these aspects please?
For example, I’ve read papers and seen authors choose metrics (they almost always include accuracy even when not ideal) and then for example, they’ll say model X is the best model in their study, from all the models they tried, because model X had the highest performance value for say, 3 of the 4 machine learning metrics.
What is your advice on all of this?
Also, one should always report in one’s paper, what other authors have done in their study, I mean, the performance of their models, however, what if the dataset those authors use were different? So does one report other authors work, best models from their published study, only if they used the exact same dataset you are also using? What if you create a new dataset in a different way, so there won’t be previous authors papers to report about in terms of comparing their models, to yours in your paper, although there may have been work done in the same field before… Then is it okay to just speak about the other authors model results in general in the discussion section of your paper, as in the Results part, there won’t be anything to compare against due to the aspect I mentioned above, re: the dataset is newly curated and hasn’t been used by others before.