Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Interpretation of Baseline Overview acc on Twibot-22 f1 on Twibot-22 #31

Open
ohauglend opened this issue Apr 11, 2023 · 5 comments
Open

Comments

@ohauglend
Copy link

Hi, as far as I understand the Baseline Overview acc and f1 are training measures where the model associated with the paper have been fitted on the whole data and then the measures are produced. I am wondering if I can compare these measures directly to a model trained on the Twibot-22 data, and if not then why so?

The word re-implemented is used consistently in the paper and on git. I understand this as re-trained, please let me know if I have misunderstood?

@BunsenFeng
Copy link
Contributor

Hi, thank you for your interest in our work. Yes, you can compare the accuracy and F1-score numbers directly to a model trained on the TwiBot-22 data. By "reimplemented" we meant that not all baselines have available code, so we implemented baselines when necessary. All results in the "TwiBot-22" column indicate that the model is trained on the TwiBot-22 training set and evaluated performance on the TwiBot-22 test set.

Please let me know if you have more questions.

@ohauglend
Copy link
Author

Thanks, this is really helpful, is there anywhere to access the Twibot-22 test set either through git or the drive folder?

@whr000001
Copy link
Collaborator

Hi, thank you for your interest in our work. The test set can be found in the 'split.csv' file. The first column denotes the user id and the second denotes this user belongs to train/val/test set.

@ohauglend
Copy link
Author

Great, is it anyway to see what data is linked to Test 1 Test 2 and Test 3?

@TamSiuhin
Copy link
Contributor

Hi @ohauglend! Test1 contains 1000 expert annotated users described in section 3.2. Test2 denotes 500 users in the annotation quality study (section A.5). Test 3 is the merge of test1 and test2.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants