-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
compare_classifiers_performance_from_pred - hdf5 is not supported for ground truth file #3550
Comments
Hi @iflow whenever you can confirm that this fixes the issue, we're good to merge it our fix in! |
Thanks for the quick fix! Unfortunately I couldn't try it yet because I installed the library with Google Colab. So I have to set everything up on my local machine, which might take some time. |
Hey @iflow, in Collab you can test out the branch by installing Ludwig like this:
|
Thank you @tgaddair, I did not know about this awesome command :) With using the fixed version the error "hd5 is not supported..." does not show up anymore 👍 However a different error is given: I guess it is not related to this issue? Full trace:
|
Hey @iflow, for the |
Describe the bug
The visualization compare_classifiers_performance_from_pred does not work, because the following error is raised:
ValueError: hdf5 is not supported for ground truth file, valid types are {'stata', 'dataframe', <class 'dask.dataframe.core.DataFrame'>, 'html', 'df', 'tsv', 'json', 'jsonl', <class 'pandas.core.frame.DataFrame'>, 'orc', 'parquet', 'sas', 'fwf', 'feather', 'csv', 'spss', 'excel', 'pickle'}
According to the documentation the parameter ground_truth should be the name of the HDF5 file obtained during training preprocessing.
Documentation: https://ludwig.ai/latest/user_guide/visualizations/#compare_classifiers_performance_from_pred
To Reproduce
Steps to reproduce the behavior:
Expected behavior
Some plots like in the documentation.
Environment:
Additional context
I tried to find a bug in ludwig/utils/data_utils.py, but it looks good. I have also tried to call directly from a Jupyter Notebook (compare_classifiers_performance_from_pred_cli), but the same error raises there.
The text was updated successfully, but these errors were encountered: