|
| 1 | +import itertools |
1 | 2 | from abc import ABC, abstractmethod
|
2 | 3 | from typing import Generic, TypeVar
|
3 | 4 |
|
@@ -29,24 +30,49 @@ def push_response(self, response: ResponseT) -> None: ...
|
29 | 30 | def complete(self) -> bool: ...
|
30 | 31 |
|
31 | 32 |
|
32 |
| -# TODO: Implement multiturn support |
| 33 | +# FIXME: Bad implementation. Can only handle string requests |
33 | 34 | class GenerativeRequestSession(RequestSession[GenerationRequest, ResponseSummary]):
|
34 |
| - def __init__(self, request: GenerationRequest) -> None: |
35 |
| - self.request = request |
36 |
| - self._complete = False |
| 35 | + def __init__(self, prompts: list[GenerationRequest]) -> None: |
| 36 | + if not prompts: |
| 37 | + raise ValueError("Prompts cannot be empty") |
| 38 | + |
| 39 | + self.prompts = prompts |
| 40 | + self.responses: list[str] = [] |
37 | 41 |
|
38 | 42 | def __len__(self) -> int:
|
39 |
| - return 1 |
| 43 | + return len(self.prompts) |
40 | 44 |
|
41 | 45 | def get_next_request(self) -> GenerationRequest:
|
42 |
| - return self.request |
| 46 | + completed_responses = len(self.responses) |
| 47 | + base_request = self.prompts[completed_responses].model_copy(deep=True) |
| 48 | + base_request.content = "".join( |
| 49 | + itertools.chain.from_iterable( |
| 50 | + zip((x.content for x in self.prompts), self.responses + [""]) |
| 51 | + ) |
| 52 | + ) |
| 53 | + base_request.stats["prompt_tokens"] = sum( |
| 54 | + x.stats["prompt_tokens"] for x in self.prompts[: completed_responses + 1] |
| 55 | + ) |
| 56 | + base_request.constraints["output_tokens"] = sum( |
| 57 | + x.constraints["output_tokens"] |
| 58 | + for x in self.prompts[: completed_responses + 1] |
| 59 | + ) |
| 60 | + |
| 61 | + return base_request |
43 | 62 |
|
44 | 63 | def get_next_delay(self) -> float:
|
45 | 64 | return 0.0
|
46 | 65 |
|
47 |
| - def push_response(self, response: ResponseSummary) -> None: # noqa: ARG002 |
48 |
| - self._complete = True |
| 66 | + def push_response(self, response: ResponseSummary) -> None: |
| 67 | + if len(self.responses) < len(self.prompts): |
| 68 | + if response.response_output_tokens is not None: |
| 69 | + self.prompts[len(self.responses)].constraints["output_tokens"] = ( |
| 70 | + response.response_output_tokens |
| 71 | + ) |
| 72 | + self.responses.append(response.value) |
| 73 | + else: |
| 74 | + raise ValueError("Response list full") |
49 | 75 |
|
50 | 76 | @property
|
51 | 77 | def complete(self) -> bool:
|
52 |
| - return self._complete |
| 78 | + return len(self.responses) >= len(self.prompts) |
0 commit comments