langchain.agents.structured_chat.base
.StructuredChatAgentΒΆ
- class langchain.agents.structured_chat.base.StructuredChatAgent[source]ΒΆ
Bases:
Agent
[Deprecated] Structured Chat Agent.
Notes
Deprecated since version langchain==0.1.0: Use create_structured_chat_agent instead.
Create a new model by parsing and validating input data from keyword arguments.
Raises ValidationError if the input data cannot be parsed to form a valid model.
- param allowed_tools: Optional[List[str]] = NoneΒΆ
- param output_parser: AgentOutputParser [Optional]ΒΆ
Output parser for the agent.
- async aplan(intermediate_steps: List[Tuple[AgentAction, str]], callbacks: Optional[Union[List[BaseCallbackHandler], BaseCallbackManager]] = None, **kwargs: Any) Union[AgentAction, AgentFinish] ΒΆ
Given input, decided what to do.
- Parameters
intermediate_steps (List[Tuple[AgentAction, str]]) β Steps the LLM has taken to date, along with observations
callbacks (Optional[Union[List[BaseCallbackHandler], BaseCallbackManager]]) β Callbacks to run.
**kwargs (Any) β User inputs.
- Returns
Action specifying what tool to use.
- Return type
Union[AgentAction, AgentFinish]
- classmethod construct(_fields_set: Optional[SetStr] = None, **values: Any) Model ΒΆ
Creates a new model setting __dict__ and __fields_set__ from trusted or pre-validated data. Default values are respected, but no other validation is performed. Behaves as if Config.extra = βallowβ was set since it adds all passed values
- Parameters
_fields_set (Optional[SetStr]) β
values (Any) β
- Return type
Model
- copy(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, update: Optional[DictStrAny] = None, deep: bool = False) Model ΒΆ
Duplicate a model, optionally choose which fields to include, exclude and change.
- Parameters
include (Optional[Union[AbstractSetIntStr, MappingIntStrAny]]) β fields to include in new model
exclude (Optional[Union[AbstractSetIntStr, MappingIntStrAny]]) β fields to exclude from new model, as with values this takes precedence over include
update (Optional[DictStrAny]) β values to change/add in the new model. Note: the data is not validated before creating the new model: you should trust this data
deep (bool) β set to True to make a deep copy of the model
self (Model) β
- Returns
new model instance
- Return type
Model
- classmethod create_prompt(tools: Sequence[BaseTool], prefix: str = 'Respond to the human as helpfully and accurately as possible. You have access to the following tools:', suffix: str = 'Begin! Reminder to ALWAYS respond with a valid json blob of a single action. Use tools if necessary. Respond directly if appropriate. Format is Action:```$JSON_BLOB```then Observation:.\nThought:', human_message_template: str = '{input}\n\n{agent_scratchpad}', format_instructions: str = 'Use a json blob to specify a tool by providing an action key (tool name) and an action_input key (tool input).\n\nValid "action" values: "Final Answer" or {tool_names}\n\nProvide only ONE action per $JSON_BLOB, as shown:\n\n```\n{{{{\n "action": $TOOL_NAME,\n "action_input": $INPUT\n}}}}\n```\n\nFollow this format:\n\nQuestion: input question to answer\nThought: consider previous and subsequent steps\nAction:\n```\n$JSON_BLOB\n```\nObservation: action result\n... (repeat Thought/Action/Observation N times)\nThought: I know what to respond\nAction:\n```\n{{{{\n "action": "Final Answer",\n "action_input": "Final response to human"\n}}}}\n```', input_variables: Optional[List[str]] = None, memory_prompts: Optional[List[BasePromptTemplate]] = None) BasePromptTemplate [source]ΒΆ
Create a prompt for this class.
- Parameters
tools (Sequence[BaseTool]) β
prefix (str) β
suffix (str) β
human_message_template (str) β
format_instructions (str) β
input_variables (Optional[List[str]]) β
memory_prompts (Optional[List[BasePromptTemplate]]) β
- Return type
- dict(**kwargs: Any) Dict ΒΆ
Return dictionary representation of agent.
- Parameters
kwargs (Any) β
- Return type
Dict
- classmethod from_llm_and_tools(llm: BaseLanguageModel, tools: Sequence[BaseTool], callback_manager: Optional[BaseCallbackManager] = None, output_parser: Optional[AgentOutputParser] = None, prefix: str = 'Respond to the human as helpfully and accurately as possible. You have access to the following tools:', suffix: str = 'Begin! Reminder to ALWAYS respond with a valid json blob of a single action. Use tools if necessary. Respond directly if appropriate. Format is Action:```$JSON_BLOB```then Observation:.\nThought:', human_message_template: str = '{input}\n\n{agent_scratchpad}', format_instructions: str = 'Use a json blob to specify a tool by providing an action key (tool name) and an action_input key (tool input).\n\nValid "action" values: "Final Answer" or {tool_names}\n\nProvide only ONE action per $JSON_BLOB, as shown:\n\n```\n{{{{\n "action": $TOOL_NAME,\n "action_input": $INPUT\n}}}}\n```\n\nFollow this format:\n\nQuestion: input question to answer\nThought: consider previous and subsequent steps\nAction:\n```\n$JSON_BLOB\n```\nObservation: action result\n... (repeat Thought/Action/Observation N times)\nThought: I know what to respond\nAction:\n```\n{{{{\n "action": "Final Answer",\n "action_input": "Final response to human"\n}}}}\n```', input_variables: Optional[List[str]] = None, memory_prompts: Optional[List[BasePromptTemplate]] = None, **kwargs: Any) Agent [source]ΒΆ
Construct an agent from an LLM and tools.
- Parameters
llm (BaseLanguageModel) β
tools (Sequence[BaseTool]) β
callback_manager (Optional[BaseCallbackManager]) β
output_parser (Optional[AgentOutputParser]) β
prefix (str) β
suffix (str) β
human_message_template (str) β
format_instructions (str) β
input_variables (Optional[List[str]]) β
memory_prompts (Optional[List[BasePromptTemplate]]) β
kwargs (Any) β
- Return type
- classmethod from_orm(obj: Any) Model ΒΆ
- Parameters
obj (Any) β
- Return type
Model
- get_allowed_tools() Optional[List[str]] ΒΆ
- Return type
Optional[List[str]]
- get_full_inputs(intermediate_steps: List[Tuple[AgentAction, str]], **kwargs: Any) Dict[str, Any] ΒΆ
Create the full inputs for the LLMChain from intermediate steps.
- Parameters
intermediate_steps (List[Tuple[AgentAction, str]]) β
kwargs (Any) β
- Return type
Dict[str, Any]
- json(*, include: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, exclude: Optional[Union[AbstractSetIntStr, MappingIntStrAny]] = None, by_alias: bool = False, skip_defaults: Optional[bool] = None, exclude_unset: bool = False, exclude_defaults: bool = False, exclude_none: bool = False, encoder: Optional[Callable[[Any], Any]] = None, models_as_dict: bool = True, **dumps_kwargs: Any) unicode ΒΆ
Generate a JSON representation of the model, include and exclude arguments as per dict().
encoder is an optional function to supply as default to json.dumps(), other arguments as per json.dumps().
- Parameters
include (Optional[Union[AbstractSetIntStr, MappingIntStrAny]]) β
exclude (Optional[Union[AbstractSetIntStr, MappingIntStrAny]]) β
by_alias (bool) β
skip_defaults (Optional[bool]) β
exclude_unset (bool) β
exclude_defaults (bool) β
exclude_none (bool) β
encoder (Optional[Callable[[Any], Any]]) β
models_as_dict (bool) β
dumps_kwargs (Any) β
- Return type
unicode
- classmethod parse_file(path: Union[str, Path], *, content_type: unicode = None, encoding: unicode = 'utf8', proto: Protocol = None, allow_pickle: bool = False) Model ΒΆ
- Parameters
path (Union[str, Path]) β
content_type (unicode) β
encoding (unicode) β
proto (Protocol) β
allow_pickle (bool) β
- Return type
Model
- classmethod parse_obj(obj: Any) Model ΒΆ
- Parameters
obj (Any) β
- Return type
Model
- classmethod parse_raw(b: Union[str, bytes], *, content_type: unicode = None, encoding: unicode = 'utf8', proto: Protocol = None, allow_pickle: bool = False) Model ΒΆ
- Parameters
b (Union[str, bytes]) β
content_type (unicode) β
encoding (unicode) β
proto (Protocol) β
allow_pickle (bool) β
- Return type
Model
- plan(intermediate_steps: List[Tuple[AgentAction, str]], callbacks: Optional[Union[List[BaseCallbackHandler], BaseCallbackManager]] = None, **kwargs: Any) Union[AgentAction, AgentFinish] ΒΆ
Given input, decided what to do.
- Parameters
intermediate_steps (List[Tuple[AgentAction, str]]) β Steps the LLM has taken to date, along with observations
callbacks (Optional[Union[List[BaseCallbackHandler], BaseCallbackManager]]) β Callbacks to run.
**kwargs (Any) β User inputs.
- Returns
Action specifying what tool to use.
- Return type
Union[AgentAction, AgentFinish]
- return_stopped_response(early_stopping_method: str, intermediate_steps: List[Tuple[AgentAction, str]], **kwargs: Any) AgentFinish ΒΆ
Return response when agent has been stopped due to max iterations.
- Parameters
early_stopping_method (str) β
intermediate_steps (List[Tuple[AgentAction, str]]) β
kwargs (Any) β
- Return type
- save(file_path: Union[Path, str]) None ΒΆ
Save the agent.
- Parameters
file_path (Union[Path, str]) β Path to file to save the agent to.
- Return type
None
Example: .. code-block:: python
# If working with agent executor agent.agent.save(file_path=βpath/agent.yamlβ)
- classmethod schema(by_alias: bool = True, ref_template: unicode = '#/definitions/{model}') DictStrAny ΒΆ
- Parameters
by_alias (bool) β
ref_template (unicode) β
- Return type
DictStrAny
- classmethod schema_json(*, by_alias: bool = True, ref_template: unicode = '#/definitions/{model}', **dumps_kwargs: Any) unicode ΒΆ
- Parameters
by_alias (bool) β
ref_template (unicode) β
dumps_kwargs (Any) β
- Return type
unicode
- tool_run_logging_kwargs() Dict ΒΆ
- Return type
Dict
- classmethod update_forward_refs(**localns: Any) None ΒΆ
Try to update ForwardRefs on fields based on this Model, globalns and localns.
- Parameters
localns (Any) β
- Return type
None
- classmethod validate(value: Any) Model ΒΆ
- Parameters
value (Any) β
- Return type
Model
- property llm_prefix: strΒΆ
Prefix to append the llm call with.
- property observation_prefix: strΒΆ
Prefix to append the observation with.
- property return_values: List[str]ΒΆ
Return values of the agent.