top of page

What's an AI Assistant Made of?



In order to understand the full potential and uses of Large Language Models (LLM) agents, which serve as the foundation of any text-based AI Assistant (VCSA), it is crucial to comprehend their fundamental elements and how they operate within the wider scope of artificial intelligence.


Outlined below are the key components that characterise LLM agents and their operational dynamics:


'Brain'

The 'brain' of an LLM agent is essentially the underlying language model itself.  This integral element handles input processing, response generation, and decision-making through contextual comprehension. The sophistication of this brain allows LLM agents to participate in dialogues, provide answers, and carry out tasks with a fluency that closely resembles human interaction.


Memory

Memory plays a crucial role in how LLM agents operate. It can be categorised into two types:


Short-term Memory: This allows agents to keep track of context within a single interaction, enabling them to follow conversations and respond appropriately.

 

Long-term Memory: Some advanced LLM agents are equipped with features that enable them to retain information for extended periods and retrieve it during subsequent interactions. This capability enhances personalisation and improves user experience.


Tools

LLM agents come with a range of tools that empower them to engage with external systems and APIs. These tools encompass databases, web services, and other software applications, enabling the agent to seamlessly carry out tasks like fetching information, running commands, and integrating with different technologies.


Planning

An essential aspect of LLM agents is their ability to plan tasks effectively. This involves breaking down complex objectives into manageable steps and determining the best course of action to achieve desired outcomes. Planning capabilities enhance the agent’s efficiency in executing multi-step processes, making them valuable in various applications ranging from project management to customer support. 


This involves two critical steps: Plan Formulation and Plan Reflection.


Plan Formulation

At this phase, agents divide extensive tasks into smaller subtasks. There are different methods for task decomposition available:


Some techniques propose creating a comprehensive plan in one go and then following it step by step.

Others, like the Chain of Thought (CoT) approach, advocate for a more adaptable strategy where agents address sub-tasks individually, providing more flexibility.


The Tree of Thought (ToT) method extends the CoT concept by exploring multiple problem-solving paths. It dissects the issue into several stages, generating diverse ideas at each step and structuring them akin to branches on a tree.


Hierarchical structures or decision trees can also be utilised, evaluating all potential choices before finalising a plan.


While LLM-based agents typically possess broad knowledge, they may encounter challenges with tasks necessitating specialised expertise. Integrating these agents with domain-specific planners can significantly boost their performance.


Plan Reflection

Once a plan is developed, it is essential for agents to evaluate its efficacy. Agents relying on LLMs employ internal feedback systems that leverage established models to enhance their strategies. They have the capability to engage with humans to adapt their plans according to feedback and preferences. Moreover, agents collect insights from their surroundings, whether physical or digital, utilising results and observations to continuously improve their plans.


There are two effective methods for incorporating feedback in planning - ReAct and Reflexion.


ReAct helps an LLM solve complex tasks by cycling through a sequence of thought, action, and observation, repeating these steps as needed. It takes in feedback from the environment—including observations and input from humans or other models—allowing the LLM to adjust its approach based on real-time feedback.


Reflexion focuses on self-assessment, enabling the agent to evaluate its previous actions against desired outcomes, thereby fostering continuous improvement.


Having a grasp of these fundamental elements lays a strong groundwork for delving into the various LLM agents currently in use.




Note: the article is based on the information provided at AI-Pro platform.


Comments


bottom of page