github-actions[bot]
Auto-sync from demo at Wed Oct 29 11:25:28 UTC 2025
d02622b
raw
history blame
990 Bytes
from typing import Any, List, Optional
from graphgen.bases import BaseLLMWrapper
from graphgen.bases.datatypes import Token
# TODO: implement TGIWrapper methods
class TGIWrapper(BaseLLMWrapper):
"""
Async inference backend based on TGI (Text-Generation-Inference)
"""
def __init__(
self,
model_url: str, # e.g. "http://localhost:8080"
temperature: float = 0.0,
top_p: float = 1.0,
topk: int = 5,
**kwargs: Any
):
super().__init__(temperature=temperature, top_p=top_p, **kwargs)
async def generate_answer(
self, text: str, history: Optional[List[str]] = None, **extra: Any
) -> str:
pass
async def generate_topk_per_token(
self, text: str, history: Optional[List[str]] = None, **extra: Any
) -> List[Token]:
pass
async def generate_inputs_prob(
self, text: str, history: Optional[List[str]] = None, **extra: Any
) -> List[Token]:
pass