This information dives into constructing a customized conversational agent with LangChain, a robust framework that integrates Massive Language Fashions (LLMs) with a spread of instruments and APIs. Designed for versatility, the agent can sort out duties like producing random numbers, sharing philosophical insights, and dynamically fetching and extracting content material from webpages. By combining pre-built instruments with customized options, we create an agent able to delivering real-time, informative, and context-aware responses.
Studying Aims
- Achieve information of the LangChain framework and its integration with Massive Language Fashions and exterior instruments.
- Study to create and implement customized instruments for specialised duties inside a conversational agent.
- Purchase expertise in fetching and processing dwell knowledge from the online for correct responses.
- Develop a conversational agent that maintains context for coherent and related interactions.
This text was revealed as part of the Information Science Blogathon.
Why Combine LangChain, OpenAI, and DuckDuckGo?
Integrating LangChain with OpenAI and DuckDuckGo empowers us to construct superior conversational AI purposes that may carry out each structured and unstructured searches. OpenAI’s language fashions allow pure language understanding, whereas DuckDuckGo supplies a dependable, privacy-respecting search API. Collectively, they permit our AI to not solely generate coherent and contextual responses but additionally fetch real-time data, enhancing its versatility and relevance. This setup is good for creating clever, responsive chatbots or digital assistants that may deal with a variety of person inquiries, making it a robust toolkit for builders in conversational AI.
Putting in Important Packages
First, you’ll want to put in the required Python packages, together with langchain and openai, and the DuckDuckGo search instrument. You’ll be able to simply do that utilizing pip:
!pip -q set up langchain==0.3.4 openai
pip set up langchain
!pip -q set up duckduckgo-search
As soon as put in, affirm the set up of LangChain to make sure that it’s arrange appropriately:
!pip present langchain
Configuring API Entry
To make the perfect use of OpenAI’s fashions, you’ll must arrange an API key. For this, you should utilize the os library to load your API keys into the atmosphere variables:
import os
os.environ["OPENAI_API_KEY"] = "your_openai_key_here"
Be sure that to exchange “your_openai_key_here” together with your precise OpenAI API key. This key can be required for interacting with the GPT-3.5-turbo mannequin.
Connecting OpenAI Fashions to LangChain
We are going to now arrange a connection to OpenAI’s mannequin utilizing LangChain. This mannequin supplies the pliability to deal with a variety of language duties, from primary conversations to extra superior queries:
from langchain import OpenAI
from langchain.chat_models import ChatOpenAI
from langchain.chains.dialog.reminiscence import ConversationBufferWindowMemory
# Arrange the turbo LLM
turbo_llm = ChatOpenAI(
temperature=0,
model_name="gpt-4o"
)
On this case, we’ve configured the GPT-4o mannequin with a low temperature (temperature=0) to make sure constant responses.
To make your agent extra versatile, you’ll need to combine instruments that allow it to entry exterior knowledge. On this case, we are going to combine the DuckDuckGo search instrument, which permits the agent to carry out internet searches:
from langchain.instruments import DuckDuckGoSearchTool
from langchain.brokers import Instrument
from langchain.instruments import BaseTool
search = DuckDuckGoSearchTool()
# Defining a single instrument
instruments = [
Tool(
name = "search",
func=search.run,
description="useful for when you need to answer questions about current events. You should ask targeted questions"
)
]
The DuckDuckGoSearchTool is about up right here as a Instrument object, with its description indicating that it’s supreme for answering questions associated to present occasions.
Creating Customized Functionalities
Along with the usual instruments offered by LangChain, it’s also possible to create customized instruments to reinforce your agent’s skills. Under, we reveal how you can create a few easy but illustrative instruments: one which returns the that means of life, and one other that generates random numbers.
Customized Instrument: Which means of Life
The primary instrument is a straightforward perform that gives a solution to the age-old query, “What’s the that means of life?”. On this case, it returns a humorous response with a slight twist:
def meaning_of_life(enter=""):
return 'The that means of life is 42 if rounded however is definitely 42.17658'
life_tool = Instrument(
identify="Which means of Life",
func= meaning_of_life,
description="Helpful for when you want to reply questions concerning the that means of life. enter needs to be MOL "
)
This instrument may be built-in into your agent to supply some enjoyable or philosophical insights throughout a dialog.
Customized Instrument: Random Quantity Generator
The second instrument generates a random integer between 0 and 5. This might be helpful in conditions the place randomness or decision-making is required:
import random
def random_num(enter=""):
return random.randint(0,5)
random_tool = Instrument(
identify="Random quantity",
func= random_num,
description="Helpful for when you want to get a random quantity. enter needs to be 'random'"
)
By integrating this instrument into your LangChain agent, it turns into able to producing random numbers when wanted, including a contact of unpredictability or enjoyable to the interactions.
Making a conversational agent with customized instruments opens up infinite potentialities for tailoring interactions to particular wants. By integrating distinctive functionalities, we will construct a dynamic, responsive assistant that goes past customary responses, delivering customized and contextually related data.
Initializing the Agent
We begin by importing the initialize_agent perform from LangChain and defining the instruments we’ve created earlier—DuckDuckGo search, random quantity generator, and the “Which means of Life” instrument:
from langchain.brokers import initialize_agent
instruments = [search, random_tool, life_tool]
Including Reminiscence to the Agent
To allow the agent to recollect current elements of the dialog, we incorporate reminiscence. The ConversationBufferWindowMemory characteristic permits the agent to maintain a rolling reminiscence of the previous couple of messages (set to three on this case):
from langchain.chains.dialog.reminiscence import ConversationBufferWindowMemory
reminiscence = ConversationBufferWindowMemory(
memory_key='chat_history',
ok=3,
return_messages=True
)
This reminiscence will permit the agent to keep up context over brief conversations, which is particularly helpful for a extra pure dialogue move.
Constructing the Conversational Agent
We initialize the agent utilizing initialize_agent. The important thing parameters embrace the next:
- agent: Specifies the kind of agent. On this case, we’re utilizing the conversational ReAct-style agent (chat-conversational-react-description).
- instruments: The instruments we’ve outlined are handed right here.
- llm: The OpenAI GPT-3.5-turbo mannequin.
- verbose: Set to True in order that we will see detailed output of the agent’s conduct.
- reminiscence: The reminiscence object that permits the agent to recall dialog historical past.
- max_iterations: These are set to cease the agent after a most of three iterations or when it generates a legitimate reply.
conversational_agent = initialize_agent(
agent="chat-conversational-react-description",
instruments=instruments,
llm=turbo_llm,
verbose=True,
max_iterations=3,
early_stopping_method='generate',
reminiscence=reminiscence
)
Testing the Agent
Now you can work together with the agent utilizing pure language queries. Listed below are some examples of questions you’ll be able to ask the agent:
Ask for the present time in a metropolis
conversational_agent("What time is it in London?")
- The agent will use the DuckDuckGo search instrument to search out the present time in London.
Request a random quantity:
conversational_agent("Are you able to give me a random quantity?")
- The agent will generate a random quantity utilizing the customized random_tool.
Ask for the that means of life
conversational_agent("What's the that means of life?")
The agent will reply this query by using the customized life_tool.
Customizing the System Immediate
To information the agent’s conduct extra successfully, we will regulate the system immediate it follows. By default, the agent supplies normal help throughout quite a lot of subjects. Nonetheless, we will tailor this by explicitly telling it to make use of instruments for particular duties, akin to answering questions on random numbers or the that means of life.
Right here’s how we modify the system immediate:
# system immediate
conversational_agent.agent.llm_chain.immediate.messages[0].immediate.template
fixed_prompt=""'Assistant is a big language mannequin skilled by OpenAI.
Assistant is designed to have the ability to help with a variety of duties, from answering
easy inquiries to offering in-depth explanations and discussions on a variety of
subjects. As a language mannequin, Assistant is ready to generate human-like textual content based mostly on the enter it receives, permitting it to have interaction in natural-sounding conversations and supply responses which are coherent and related to the subject at hand.
Assistant would not know something about random numbers or something associated to the that means
of life and may use a instrument for questions on these subjects.
Assistant is continually studying and enhancing, and its capabilities are always
evolving. It is ready to course of and perceive giant quantities of textual content, and may use
this data to supply correct and informative responses to a variety of
questions. Moreover, Assistant is ready to generate its personal textual content based mostly on the
enter it receives, permitting it to have interaction in discussions and supply explanations and
descriptions on a variety of subjects.
Total, Assistant is a robust system that may assist with a variety of duties and
present precious insights and knowledge on a variety of subjects. Whether or not you want
assist with a selected query or simply need to have a dialog a few specific
subject, Assistant is right here to help.'''
We then apply the modified immediate to the agent:
conversational_agent.agent.llm_chain.immediate.messages[0].immediate.template = fixed_prompt
Now, when the agent is requested about random numbers or the that means of life, it should make sure to seek the advice of the suitable instrument as a substitute of trying to generate a solution by itself.
Testing with the New Immediate
Let’s check the agent once more with the up to date immediate:
conversational_agent("What's the that means of life?")
This time, the agent ought to name the life_tool to reply the query as a substitute of producing a response from its personal information.
On this part, we’ll create a customized instrument that may strip HTML tags from a webpage and return the plain textual content content material. This may be significantly helpful for duties like summarizing articles, extracting particular data, or just changing HTML content material right into a readable format.
Constructing the Net Scraper Instrument
We are going to use the requests library to fetch the webpage and BeautifulSoup from the bs4 package deal to parse and clear the HTML content material. The perform under retrieves a webpage, removes all HTML tags, and returns the primary 4000 characters of the stripped textual content:
from bs4 import BeautifulSoup
import requests
def stripped_webpage(webpage):
# Fetch the content material of the webpage
response = requests.get(webpage)
html_content = response.textual content
# Operate to strip HTML tags from the content material
def strip_html_tags(html_content):
soup = BeautifulSoup(html_content, "html.parser")
stripped_text = soup.get_text()
return stripped_text
# Strip the HTML tags
stripped_content = strip_html_tags(html_content)
# Restrict the content material to 4000 characters to keep away from lengthy outputs
if len(stripped_content) > 4000:
stripped_content = stripped_content[:4000]
return stripped_content
# Testing the perform with a pattern webpage
stripped_webpage('https://www.google.com')
How It Works
- Fetching the webpage: The perform makes use of the requests library to get the uncooked HTML content material from a given URL.
- Stripping HTML tags: Utilizing BeautifulSoup, we parse the HTML content material and extract solely the textual content, eradicating all HTML tags.
- Character restrict: To keep away from overly lengthy outputs, the textual content is proscribed to the primary 4000 characters. This retains the content material concise and manageable, particularly to be used in a conversational agent.
Testing the Instrument
You’ll be able to check this instrument by passing any URL to the stripped_webpage perform. For instance, when used with https://www.google.com, it should return a text-only model of Google’s homepage (restricted to the primary 4000 characters).
Integrating into LangChain as a Instrument
Now you can wrap this perform inside a LangChain Instrument and combine it into your agent for dynamic internet scraping:
from langchain.brokers import Instrument
# Create the instrument for internet scraping
web_scraper_tool = Instrument(
identify="Net Scraper",
func=stripped_webpage,
description="Fetches a webpage, strips HTML tags, and returns the plain textual content content material (restricted to 4000 characters)."
)
Now, this internet scraper instrument may be a part of the agent’s toolbox and used to fetch and strip content material from any webpage in actual time, including additional utility to the conversational agent you’ve constructed.
On this part, we’re constructing a customized WebPageTool class that can permit our conversational agent to fetch and strip HTML tags from a webpage, returning the textual content content material. This instrument is beneficial for extracting data from web sites dynamically, including a brand new layer of performance to the agent.
Defining the WebPageTool Class
We begin by defining the WebPageTool class, which inherits from BaseTool. This class has two strategies: _run for synchronous execution (the default for this instrument) and _arun which raises an error since asynchronous operations should not supported right here.
from langchain.instruments import BaseTool
from bs4 import BeautifulSoup
import requests
class WebPageTool(BaseTool):
identify = "Get Webpage"
description = "Helpful for when you want to get the content material from a selected webpage"
# Synchronous run technique to fetch webpage content material
def _run(self, webpage: str):
# Fetch webpage content material
response = requests.get(webpage)
html_content = response.textual content
# Strip HTML tags from the content material
def strip_html_tags(html_content):
soup = BeautifulSoup(html_content, "html.parser")
stripped_text = soup.get_text()
return stripped_text
# Get the plain textual content and restrict it to 4000 characters
stripped_content = strip_html_tags(html_content)
if len(stripped_content) > 4000:
stripped_content = stripped_content[:4000]
return stripped_content
# Async run technique (not carried out)
def _arun(self, webpage: str):
elevate NotImplementedError("This instrument doesn't assist async")
# Instantiate the instrument
page_getter = WebPageTool()
How the WebPageTool Works
- Fetching Webpage Content material: The _run technique makes use of requests.get to fetch the webpage’s uncooked HTML content material.
- Stripping HTML Tags: Utilizing BeautifulSoup, the HTML tags are eliminated, leaving solely the plain textual content content material.
- Limiting Output Measurement: The content material is truncated to 4000 characters to keep away from extreme output, making the agent’s response concise and manageable.
Reinitializing the Conversational Agent
Subsequent, we have to combine the brand new WebPageTool into our conversational agent. This entails including the instrument to the agent’s toolbox and updating the system immediate to instruct the agent to make use of this instrument for fetching webpage content material.
Updating the System Immediate
We modify the system immediate to explicitly instruct the assistant to at all times verify webpages when requested for content material from a selected URL:
fixed_prompt=""'Assistant is a big language mannequin skilled by OpenAI.
Assistant is designed to have the ability to help with a variety of duties, from answering
easy inquiries to offering in-depth explanations and discussions on a variety of
subjects. As a language mannequin, Assistant is ready to generate human-like textual content based mostly on the
enter it receives, permitting it to have interaction in natural-sounding conversations and supply
responses which are coherent and related to the subject at hand.
Assistant would not know something about random numbers or something associated to the that means
of life and may use a instrument for questions on these subjects.
Assistant additionally would not know details about content material on webpages and may at all times
verify if requested.
Total, Assistant is a robust system that may assist with a variety of duties and
present precious insights and knowledge on a variety of subjects. Whether or not you want
assist with a selected query or simply need to have a dialog a few specific
subject, Assistant is right here to help.'''
This immediate clearly instructs the assistant to depend on instruments for sure subjects, akin to random numbers, the that means of life, and webpage content material.
Reinitializing the Agent with the New Instrument
Now we will reinitialize the conversational agent with the brand new WebPageTool included:
from langchain.brokers import initialize_agent
instruments = [search, random_tool, life_tool, page_getter]
# Reuse the reminiscence from earlier
conversational_agent = initialize_agent(
agent="chat-conversational-react-description",
instruments=instruments,
llm=turbo_llm,
verbose=True,
max_iterations=3,
early_stopping_method='generate',
reminiscence=reminiscence
)
# Apply the up to date immediate
conversational_agent.agent.llm_chain.immediate.messages[0].immediate.template = fixed_prompt
conversational_agent.agent.llm_chain.immediate.messages[0]
Testing the WebPageTool
We will now check the agent by asking it to fetch content material from a selected webpage. For instance:
conversational_agent.run("Is there an article about Clubhouse on https://techcrunch.com/? as we speak")
The agent will:
- Use the WebPageTool to scrape TechCrunch’s homepage.
- Search by way of the primary 4000 characters of content material to verify if there are any mentions of “Clubhouse.”
- Present a response based mostly on whether or not it finds an article associated to Clubhouse in that scraped content material.
High Tales on CBS Information : Operating:
conversational_agent.run("What are the titles of the highest tales on www.cbsnews.com/?")
The agent will:
- Fetch and course of the textual content content material from CBS Information’ homepage.
- Extract potential headings or story titles by analyzing textual content patterns or sections that appear like titles.
- Reply with the highest tales or a abstract of what it discovered within the webpage content material.
Conclusion
We’ve got efficiently developed a flexible conversational agent utilizing LangChain that may seamlessly combine quite a lot of instruments and APIs. By harnessing the capabilities of Massive Language Fashions alongside customized functionalities—akin to random quantity technology, philosophical insights, and dynamic internet scraping—this agent demonstrates its skill to deal with a variety of person queries. The modular design permits for simple growth and adaptation, making certain that the agent can evolve as new instruments and options are added. This venture not solely showcases the ability of mixing AI with real-time knowledge but additionally units the stage for future enhancements, making the agent a useful useful resource for interactive and informative conversations.
Key Takeaways
- Constructing a conversational agent with LangChain permits for a modular method, making it straightforward to combine numerous instruments and increase performance.
- The combination of internet scraping and search instruments allows the agent to supply up-to-date data and reply to present occasions successfully.
- The flexibility to create customized instruments empowers builders to tailor the agent’s capabilities to particular use circumstances and person wants.
- Using reminiscence options ensures that the agent can preserve context throughout conversations, resulting in extra significant and coherent interactions with customers.
The media proven on this article will not be owned by Analytics Vidhya and is used on the Writer’s discretion.
Often Requested Questions
A. LangChain is a framework designed for constructing purposes that combine Massive Language Fashions (LLMs) with numerous exterior instruments and APIs, enabling builders to create clever brokers able to performing complicated duties.
A. The conversational agent can reply normal information questions, generate random numbers, present insights concerning the that means of life, and dynamically fetch and extract content material from specified webpages.
A. The agent makes use of instruments just like the DuckDuckGo Search and a customized internet scraping instrument to retrieve up-to-date data from the online, permitting it to supply correct and well timed responses to person queries about present occasions.
A. The venture primarily makes use of Python, together with libraries akin to LangChain for constructing the agent, OpenAI for accessing the language fashions, requests for making HTTP requests, and BeautifulSoup for parsing HTML content material.
A. You’ll be able to increase the agent’s capabilities by including extra instruments or APIs that serve particular features, modifying current instruments for higher efficiency, or integrating new knowledge sources to complement the responses.
The media proven on this article will not be owned by Analytics Vidhya and is used on the Writer’s discretion.