All files / kernel-test-local/src/vats lms-chat-vat.ts

0% Statements 0/7
0% Branches 0/2
0% Functions 0/3
0% Lines 0/7

Press n or j to go to the next uncovered block, b, p or k for the previous block.

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42                                                                                   
import type { ERef } from '@endo/eventual-send';
import { makeDefaultExo } from '@metamask/kernel-utils/exo';
import type { Logger } from '@metamask/logger';
import { makeChatClient } from '@ocap/kernel-language-model-service';
import type { ChatService } from '@ocap/kernel-language-model-service';
 
/**
 * A vat that uses a kernel language model service to perform a chat completion
 * and logs the response. Used by lms-chat.test.ts and lms-chat.e2e.test.ts to verify the full
 * kernel → LMS service → Ollama round-trip.
 *
 * @param vatPowers - Vat powers, expected to include a logger.
 * @param parameters - Vat parameters.
 * @param parameters.model - The model to use for chat completion.
 * @returns A default Exo instance.
 */
// eslint-disable-next-line @typescript-eslint/explicit-function-return-type
export function buildRootObject(
  vatPowers: Record<string, unknown>,
  { model }: { model: string },
) {
  const logger = vatPowers.logger as Logger;
  const tlog = (message: string): void => {
    logger.subLogger({ tags: ['test', 'lms-chat'] }).log(message);
  };
 
  return makeDefaultExo('root', {
    async bootstrap(
      _roots: unknown,
      { languageModelService }: { languageModelService: ERef<ChatService> },
    ) {
      const client = makeChatClient(languageModelService, model);
      const result = await client.chat.completions.create({
        messages: [
          { role: 'user', content: 'Reply with exactly one word: hello.' },
        ],
      });
      tlog(`lms-chat response: ${result.choices[0]?.message.content ?? ''}`);
    },
  });
}