Awesome
Ollama for Dart
A Dart client for interacting with the Ollama API. This library provides an easy-to-use interface for generating text completions, chat responses, and embeddings using Ollama inference engine.
Features
- Generate text completions
- Generate chat responses
- Generate embeddings
- Support for streaming responses
- Customizable model parameters
Installation
Run the following command to install the package:
dart pub add ollama
Usage
Initializing the client
import 'package:ollama/ollama.dart';
final ollama = Ollama();
// Or with a custom base URL:
// final ollama = Ollama(baseUrl: Uri.parse('http://your-ollama-server:11434'));
Generating text completions
final stream = ollama.generate(
'Tell me a joke about programming',
model: 'llama3',
);
await for (final chunk in stream) {
print(chunk.response);
}
Generating chat responses
final messages = [
ChatMessage(role: 'user', content: 'Hello, how are you?'),
];
final stream = ollama.chat(
messages,
model: 'llama3',
);
await for (final chunk in stream) {
print(chunk.message?.content);
}
Generating embeddings
final embeddings = await ollama.embeddings(
'Here is an article about llamas...',
model: 'llama3',
);
print(embeddings);
Contributing
Contributions are welcome! Please feel free to submit a Pull Request.