langchain_core.language_models.llms
.get_prompts¶
- langchain_core.language_models.llms.get_prompts(params: Dict[str, Any], prompts: List[str], cache: Optional[Union[BaseCache, bool]] = None) Tuple[Dict[int, List], str, List[int], List[str]] [source]¶
è·ćć·ČçŒćçæç€șă
- Parameters
params (Dict[str, Any]) â
prompts (List[str]) â
cache (Optional[Union[BaseCache, bool]]) â
- Return type
Tuple[Dict[int, List], str, List[int], List[str]]