Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Obs AI Assistant] Boost user prompt in recall #184933

Merged
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -72,10 +72,6 @@ export function createService({
return of(
createFunctionRequestMessage({
name: 'context',
args: {
queries: [],
categories: [],
},
}),
createFunctionResponseMessage({
name: 'context',
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -40,34 +40,10 @@ export function registerContextFunction({
description:
'This function provides context as to what the user is looking at on their screen, and recalled documents from the knowledge base that matches their query',
visibility: FunctionVisibility.Internal,
parameters: {
type: 'object',
properties: {
queries: {
type: 'array',
description: 'The query for the semantic search',
items: {
type: 'string',
},
},
categories: {
type: 'array',
description:
'Categories of internal documentation that you want to search for. By default internal documentation will be excluded. Use `apm` to get internal APM documentation, `lens` to get internal Lens documentation, or both.',
items: {
type: 'string',
enum: ['apm', 'lens'],
},
},
},
required: ['queries', 'categories'],
} as const,
},
async ({ arguments: args, messages, screenContexts, chat }, signal) => {
async ({ messages, screenContexts, chat }, signal) => {
const { analytics } = (await resources.context.core).coreStart;

const { queries, categories } = args;

async function getContext() {
const screenDescription = compact(
screenContexts.map((context) => context.screenDescription)
Expand All @@ -94,38 +70,29 @@ export function registerContextFunction({
messages.filter((message) => message.message.role === MessageRole.User)
);

const nonEmptyQueries = compact(queries);

const queriesOrUserPrompt = nonEmptyQueries.length
? nonEmptyQueries
: compact([userMessage?.message.content]);

queriesOrUserPrompt.push(screenDescription);

const suggestions = await retrieveSuggestions({
client,
categories,
queries: queriesOrUserPrompt,
});
const userPrompt = userMessage?.message.content;
const queries = [{ text: userPrompt, boost: 3 }, { text: screenDescription }].filter(
({ text }) => text
) as Array<{ text: string; boost?: number }>;

const suggestions = await retrieveSuggestions({ client, queries });
if (suggestions.length === 0) {
return {
content,
};
return { content };
}

try {
const { relevantDocuments, scores } = await scoreSuggestions({
suggestions,
queries: queriesOrUserPrompt,
screenDescription,
userPrompt,
Comment on lines -120 to +87
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@dgieselaar Per your comment we now use queries that includes boost parameter in recall but for scoring suggestions I am getting better results by telling the LLM what each query is (user prompt and screen description).

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actually that's a great idea!

messages,
chat,
signal,
logger: resources.logger,
});

analytics.reportEvent<RecallRanking>(RecallRankingEventType, {
prompt: queriesOrUserPrompt.join('|'),
prompt: queries.map((query) => query.text).join('|'),
scoredDocuments: suggestions.map((suggestion) => {
const llmScore = scores.find((score) => score.id === suggestion.id);
return {
Expand Down Expand Up @@ -178,15 +145,12 @@ export function registerContextFunction({
async function retrieveSuggestions({
queries,
client,
categories,
}: {
queries: string[];
queries: Array<{ text: string; boost?: number }>;
client: ObservabilityAIAssistantClient;
categories: Array<'apm' | 'lens'>;
}) {
const recallResponse = await client.recall({
queries,
categories,
});

return recallResponse.entries.map((entry) => omit(entry, 'labels', 'is_correction'));
Expand All @@ -208,14 +172,16 @@ const scoreFunctionArgumentsRt = t.type({
async function scoreSuggestions({
suggestions,
messages,
queries,
userPrompt,
screenDescription,
chat,
signal,
logger,
}: {
suggestions: Awaited<ReturnType<typeof retrieveSuggestions>>;
messages: Message[];
queries: string[];
userPrompt: string | undefined;
screenDescription: string;
chat: FunctionCallChatFunction;
signal: AbortSignal;
logger: Logger;
Expand All @@ -237,7 +203,10 @@ async function scoreSuggestions({
- The document contains new information not mentioned before in the conversation

Question:
${queries.join('\n')}
${userPrompt}

Screen description:
${screenDescription}

Documents:
${JSON.stringify(indexedSuggestions, null, 2)}`);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,16 @@ const functionRecallRoute = createObservabilityAIAssistantServerRoute({
params: t.type({
body: t.intersection([
t.type({
queries: t.array(nonEmptyStringRt),
queries: t.array(
t.intersection([
t.type({
text: t.string,
}),
t.partial({
boost: t.number,
}),
])
),
}),
t.partial({
categories: t.array(t.string),
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,9 +28,5 @@ export function getContextFunctionRequestIfNeeded(

return createFunctionRequestMessage({
name: CONTEXT_FUNCTION_NAME,
args: {
queries: [],
categories: [],
},
});
}
Original file line number Diff line number Diff line change
Expand Up @@ -1232,7 +1232,6 @@ describe('Observability AI Assistant client', () => {
role: MessageRole.Assistant,
function_call: {
name: CONTEXT_FUNCTION_NAME,
arguments: JSON.stringify({ queries: [], categories: [] }),
trigger: MessageRole.Assistant,
},
},
Expand Down Expand Up @@ -1456,7 +1455,6 @@ describe('Observability AI Assistant client', () => {
role: MessageRole.Assistant,
function_call: {
name: CONTEXT_FUNCTION_NAME,
arguments: JSON.stringify({ queries: [], categories: [] }),
trigger: MessageRole.Assistant,
},
},
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -694,7 +694,7 @@ export class ObservabilityAIAssistantClient {
queries,
categories,
}: {
queries: string[];
queries: Array<{ text: string; boost?: number }>;
categories?: string[];
}): Promise<{ entries: RecalledEntry[] }> => {
return this.dependencies.knowledgeBaseService.recall({
Expand Down Expand Up @@ -757,11 +757,9 @@ export class ObservabilityAIAssistantClient {
};

fetchUserInstructions = async () => {
const userInstructions = await this.dependencies.knowledgeBaseService.getUserInstructions(
return this.dependencies.knowledgeBaseService.getUserInstructions(
this.dependencies.namespace,
this.dependencies.user
);

return userInstructions;
};
}
Original file line number Diff line number Diff line change
Expand Up @@ -303,19 +303,20 @@ export class KnowledgeBaseService {
user,
modelId,
}: {
queries: string[];
queries: Array<{ text: string; boost?: number }>;
categories?: string[];
namespace: string;
user?: { name: string };
modelId: string;
}): Promise<RecalledEntry[]> {
const query = {
bool: {
should: queries.map((text) => ({
should: queries.map(({ text, boost = 1 }) => ({
text_expansion: {
'ml.tokens': {
model_text: text,
model_id: modelId,
boost,
},
},
})),
Expand Down Expand Up @@ -385,7 +386,7 @@ export class KnowledgeBaseService {
uiSettingsClient,
modelId,
}: {
queries: string[];
queries: Array<{ text: string; boost?: number }>;
asCurrentUser: ElasticsearchClient;
uiSettingsClient: IUiSettingsClient;
modelId: string;
Expand Down Expand Up @@ -414,15 +415,16 @@ export class KnowledgeBaseService {
const vectorField = `${ML_INFERENCE_PREFIX}${field}_expanded.predicted_value`;
const modelField = `${ML_INFERENCE_PREFIX}${field}_expanded.model_id`;

return queries.map((query) => {
return queries.map(({ text, boost = 1 }) => {
return {
bool: {
should: [
{
text_expansion: {
[vectorField]: {
model_text: query,
model_text: text,
model_id: modelId,
boost,
},
},
},
Expand Down Expand Up @@ -470,7 +472,7 @@ export class KnowledgeBaseService {
asCurrentUser,
uiSettingsClient,
}: {
queries: string[];
queries: Array<{ text: string; boost?: number }>;
categories?: string[];
user?: { name: string };
namespace: string;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@ describe('<ChatBody>', () => {
role: 'assistant',
function_call: {
name: CONTEXT_FUNCTION_NAME,
arguments: '{"queries":[],"categories":[]}',
arguments: '{}',
trigger: 'assistant',
},
content: '',
Expand Down Expand Up @@ -88,7 +88,7 @@ describe('<ChatBody>', () => {
role: 'assistant',
function_call: {
name: CONTEXT_FUNCTION_NAME,
arguments: '{"queries":[],"categories":[]}',
arguments: '{}',
trigger: 'assistant',
},
content: '',
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -193,7 +193,6 @@ export default function ApiTest({ getService }: FtrProviderContext) {
role: MessageRole.Assistant,
function_call: {
name: 'context',
arguments: JSON.stringify({ queries: [], categories: [] }),
trigger: MessageRole.Assistant,
},
},
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -72,6 +72,7 @@ export default function ApiTest({ getService }: FtrProviderContext) {
format,
})
.set('kbn-xsrf', 'foo')
.set('elastic-api-version', '2023-10-31')
.send({
messages,
connectorId,
Expand All @@ -83,13 +84,20 @@ export default function ApiTest({ getService }: FtrProviderContext) {
if (err) {
return reject(err);
}
if (response.status !== 200) {
return reject(new Error(`${response.status}: ${JSON.stringify(response.body)}`));
}
return resolve(response);
});
});

const [conversationSimulator, titleSimulator] = await Promise.all([
conversationInterceptor.waitForIntercept(),
titleInterceptor.waitForIntercept(),
const [conversationSimulator, titleSimulator] = await Promise.race([
Promise.all([
conversationInterceptor.waitForIntercept(),
titleInterceptor.waitForIntercept(),
]),
// make sure any request failures (like 400s) are properly propagated
responsePromise.then(() => []),
]);

await titleSimulator.status(200);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -94,7 +94,7 @@ export default function ApiTest({ getService, getPageObjects }: FtrProviderConte
content: '',
function_call: {
name: 'context',
arguments: '{"queries":[],"categories":[]}',
arguments: '{}',
trigger: MessageRole.Assistant,
},
},
Expand Down Expand Up @@ -290,7 +290,6 @@ export default function ApiTest({ getService, getPageObjects }: FtrProviderConte

expect(pick(contextRequest.function_call, 'name', 'arguments')).to.eql({
name: 'context',
arguments: JSON.stringify({ queries: [], categories: [] }),
});

expect(contextResponse.name).to.eql('context');
Expand Down Expand Up @@ -354,7 +353,6 @@ export default function ApiTest({ getService, getPageObjects }: FtrProviderConte

expect(pick(contextRequest.function_call, 'name', 'arguments')).to.eql({
name: 'context',
arguments: JSON.stringify({ queries: [], categories: [] }),
});

expect(contextResponse.name).to.eql('context');
Expand Down
Loading