Skip to content

Commit 97671f1

Browse files
committed
refactor: rename prompt_prefix_cache to prefix_cache for clarity
- Updated the variable name from prompt_prefix_cache to prefix_cache in OpenAICompatibleTinkerServer for improved readability. - Adjusted related cache lookup and insertion methods to reflect the new naming convention.
1 parent 7baebb2 commit 97671f1

1 file changed

Lines changed: 3 additions & 3 deletions

File tree

src/art/tinker/server.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ class OpenAICompatibleTinkerServer:
3131
sampling_clients_and_renderers: dict[
3232
str, tuple[tinker.SamplingClient, renderers.Renderer]
3333
] = field(default_factory=dict)
34-
prompt_prefix_cache: LRUTrieCache = field(
34+
prefix_cache: LRUTrieCache = field(
3535
default_factory=lambda: LRUTrieCache(max_entries=1000)
3636
)
3737
_task: asyncio.Task[None] | None = None
@@ -98,7 +98,7 @@ async def chat_completions(
9898
),
9999
)
100100
prompt_tokens = rendered_prompt_tokens
101-
prefix_entry = self.prompt_prefix_cache.lookup(rendered_prompt_tokens)
101+
prefix_entry = self.prefix_cache.lookup(rendered_prompt_tokens)
102102
if prefix_entry is not None and prefix_entry.rendered_len <= len(
103103
rendered_prompt_tokens
104104
):
@@ -140,7 +140,7 @@ async def chat_completions(
140140
renderer.tokenizer.decode(sequence.tokens)
141141
)
142142
if rendered_response_tokens != sequence.tokens:
143-
self.prompt_prefix_cache.insert(
143+
self.prefix_cache.insert(
144144
rendered_prompt_tokens + rendered_response_tokens,
145145
prompt_tokens + sequence.tokens,
146146
)

0 commit comments

Comments
 (0)