| 123456789101112131415161718192021222324252627282930313233343536 |
- # List of ongoing issues
- @bug
- Feature: Issues
- # Issue #5655
- Scenario: Multi users embeddings
- Given a server listening on localhost:8080
- And a model file stories260K.gguf
- And a model alias tinyllama-2
- And 42 as server seed
- And 64 KV cache size
- And 2 slots
- And continuous batching
- And embeddings extraction
- Then the server is starting
- Then the server is healthy
- Given a prompt:
- """
- Write a very long story about AI.
- """
- And a prompt:
- """
- Write another very long music lyrics.
- """
- And a prompt:
- """
- Write a very long poem.
- """
- And a prompt:
- """
- Write a very long joke.
- """
- Given concurrent embedding requests
- Then the server is busy
- Then the server is idle
- Then all embeddings are generated
|