Replies: 1 comment 1 reply
-
First of all, make sure you know what arrives in the model, because this looks like you're using two system prompts with what you're putting into the Second, follow the suggested template. You may change the instructions, but make sure to keep the rest as suggested, including whitespace. It may be easier to experiment in the chat application first, by the way. You're doing a strange mix of template + input in your Next, you may want to try different values as parameters; top-p, top-K and especially temperature. The FAQ on the wiki has an entry about model settings with a useful link to get a better idea how these impact the generation. For structured output like JSON you may be better off if you lower those a bit. And finally, all of this still doesn't guarantee that it'll behave correctly in every case. All LLMs can hallucinate, and smaller ones are typically harder to steer than larger ones. You should definitely also look into how to make better prompts. |
Beta Was this translation helpful? Give feedback.
-
Hi all,
I am use Phi3 to do an address parser
Am i find sometime the response will contain some noise.
For example:
Given the input below:
It may output:
This just contain explanation, i don't want it but i think it may be fine.
Sometime it would be worse, it just contain a lot noise (you can even see in the end of response, it complete unrelated indeed.
What i am expected:
Right now i am doing a workaround by add logic before it response. something like below:
Did it have any better way ? May be i am doing something wrong on using GPT4All SDK ?
here is how i using the SDK
Beta Was this translation helpful? Give feedback.
All reactions