I obviously get accuracy/predictions that have 4 outputs: A, B, C, and None. It’s pretty accurate too, around 90%. However, my question is as follows: Is there a way to force tabular learner to only pick A or B? This works if I remove all C and None examples from the data, but I’d like to leave them in. They have value since this is a demographics based dataset, I just specifically want to avoid the scenario of the model predicting C or None.
Edit: I’ve got a lot of learning to do, so, I’m sorry if this is a dumb question!
You do it just how you described. If you wanted a case for c then you do rest vs C. A “tree” ish design to a bunch of tabular models. I’ve done this with my research. Sometimes it gives quite good results.
Kinda halfway related to this topic, do you know what happens to procs when you export a model and try single inference?
psuedocode:
newly_created_row = some_df_values #from a form input or something
learn = load_learner(some_exported_model.pkl)
learn.predict(newly_created_row)
edit: the relevant and not-clear-to-me part from the docs is: “As in the other applications, we just have to type learn.export() to save everything we’ll need for inference (here it includes the inner state of each processor)”
Nope! Just learn.predict() with some dataframe row think of procs the same as our image’s transforms (or processor) So they’re stored in our learner already
damn, I was hoping you wouldn’t have said that. I’m just getting really crappy predictions on things that are intuitively true and was hoping it was related to procs
How similar is your test data to your training data? Perhaps Maybe there’s data missing in columns that isn’t missing in your training data? Few things to try