prompt method
Generates a stream of responses based on the provided list of chat messages.
This method takes a list of LlamaMessage objects and returns a Stream of strings, where each string represents a response generated from the chat messages.
The stream allows for asynchronous processing of the chat messages, enabling real-time or batched responses.
- Parameter messages: A list of LlamaMessage objects that represent the chat history.
- Returns: A Stream of strings, where each string is a generated response.
Implementation
Stream<String> prompt(List<LlamaMessage> messages) async* {
throw LlamaException('Web not supported');
}