-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Problems in the training process #2
Comments
The cluster centres are used only to select seed samples. As in credanno/eval_linear_joint_recycle.py Lines 142 to 143 in ff63a82
I am sorry that I do not get what you mean by "clustering in advance and train it directly as indicators". Could you elaborate? |
Thank you for your reply, I mean “If I take this code out alone and run it, the clustering center is indicators = [1414, ..., 500].”,If I take this code out alone and run it, the clustering center is indicators = [1414, ..., 500]. Now that we have the cluster center, in order to simplify our subsequent code running, I will directly assign these recorded values to indices (that is, cancel the code extract_features), which will eventually affect the accuracy. |
thats mean I created a new python file to run the extract_feature part of your code, and I'm sure that the output of indices in the source code is the same as that in the new python file. Then, after annotating the relevant part of "extract_features" in the source code, directly assign the values of indices (the cluster centers extracted in advance in the newly-built python file), which will affect the accuracy, even the difference is 2%. |
I think the extract_features function in line 141 of the code may have affected the parameters in the model, but I saw that you set "model.eval ()", so I'm not sure why it has the influence of precision. |
I have to admit that I have not tried this approach myself. To locate the problem, I would suggest trying Randomness is a known issue in this problem. Different random seeds can certainly lead to fluctuation in performance. It actually makes more sense to run several it several times and compare the statistical mean and variance in performance. |
|
Hello author, thank you for such a great article. But I found some problems when I reproduced the results. The function (extract_features) seems to have an influence on the result when the code extracts the cluster center. I tried to get the center by clustering in advance and train it directly as indicators, which would be different from calling the function to extract features in the program and then getting the clustering center to train (the clustering center is exactly the same).
The text was updated successfully, but these errors were encountered: