I Tried AISuite by AndrewNg, and It’s GREAT!

Andrew Ng not too long ago launched AISuite, an open-source Python package deal designed to streamline the usage of massive language fashions (LLMs) throughout a number of suppliers. This modern software simplifies the complexities of working with various LLMs by permitting seamless switching between fashions with a easy “supplier:mannequin” string. By considerably lowering integration overhead, AISuite enhances flexibility and accelerates software improvement, making it a useful useful resource for builders navigating the dynamic panorama of AI. On this article, we’ll see how efficient it’s.

I Tried AISuite by AndrewNg, and It’s GREAT!

What’s AISuite?

AISuite is an open-source challenge led by Andrew Ng, designed to make working with a number of massive language mannequin (LLM) suppliers simpler and extra environment friendly. Accessible on GitHub, it gives a easy, unified interface that permits seamless switching between LLMs utilizing HTTP endpoints or SDKs, following OpenAI’s interface. This software is right for college students, educators, and builders, providing constant and hassle-free interactions throughout varied suppliers.

Supported by a crew of open-source contributors, AISuite bridges the hole between totally different LLM frameworks. It allows customers to combine and examine fashions from suppliers like OpenAI, Anthropic, and Meta’s Llama with ease. The software simplifies duties similar to producing textual content, conducting analyses, and constructing interactive techniques. With options like streamlined API key administration, customizable shopper configurations, and an intuitive setup, AISuite helps each easy purposes and complicated LLM-based initiatives.

Implementation of AISuite

1. Set up Crucial Libraries

!pip set up openai
!pip set up aisuite[all]
  • !pip set up openai: Installs the OpenAI Python library, which is required to work together with OpenAI’s GPT fashions.
  • !pip set up aisuite[all]: Installs AISuite together with elective dependencies wanted to help a number of LLM suppliers.

2. Set API Keys for Authentication

os.environ['OPENAI_API_KEY'] = getpass('Enter your OPENAI API key: ')
os.environ['ANTHROPIC_API_KEY'] = getpass('Enter your ANTHROPIC API key: ')
  • os.environ: Units atmosphere variables to securely retailer the API keys required to entry LLM providers.
  • getpass(): Prompts the person to enter their OpenAI and Anthropic API keys securely (with out displaying the enter).
  • These keys authenticate your requests to the respective platforms.

Additionally learn: Generate Your Personal OpenAI API Key and Add Credit?

3. Initialize the AISuite Shopper

shopper = ai.Shopper()

This initializes an occasion of the AISuite shopper, permitting interplay with a number of LLMs in a standardized means.

4. Outline the Immediate (Messages)

messages = [
   {"role": "system", "content": "Talk using Pirate English."},
   {"role": "user", "content": "Tell a joke in 1 line."}
]
  • The messages listing defines a dialog enter:
    • function: “system”: Gives directions to the mannequin (e.g., “Speak utilizing Pirate English”).
    • function: “person”: Represents the person’s question (e.g., “Inform a joke in 1 line”).
  • This immediate ensures the responses comply with a pirate theme and embody a one-line joke.

5. Question the OpenAI Mannequin

response = shopper.chat.completions.create(mannequin="openai:gpt-4o", messages=messages, temperature=0.75)
print(response.decisions[0].message.content material)
  • mannequin=”openai:gpt-4o”: Specifies the OpenAI GPT-4o mannequin.
  • messages=messages: Sends the immediate outlined earlier to the mannequin.
  • temperature=0.75: Controls the randomness of the response. The next temperature leads to extra inventive outputs, whereas decrease values produce extra deterministic responses.
  • response.decisions[0].message.content material: Extracts the textual content content material of the mannequin’s response.

6. Question the Anthropic Mannequin

response = shopper.chat.completions.create(mannequin="anthropic:claude-3-5-sonnet-20241022", messages=messages, temperature=0.75)
print(response.decisions[0].message.content material)
  • mannequin=”anthropic:claude-3-5-sonnet-20241022″: Specifies the Anthropic Claude-3-5 mannequin.
  • The remaining parameters are similar to the OpenAI question. This demonstrates how AISuite allows simple switching between suppliers by altering the mannequin parameter.

7. Question the Ollama Mannequin

response = shopper.chat.completions.create(mannequin="ollama:llama3.1:8b", messages=messages, temperature=0.75)
print(response.decisions[0].message.content material)
  • mannequin=”ollama:llama3.1:8b”: Specifies the Ollama Llama3.1 mannequin.
  • Once more, the parameters and logic are constant, showcasing how AISuite gives a unified interface throughout suppliers.

Output

Why did the pirate go to high school? To enhance his "arrrrrrr-ticulation"!

Arrr, why do not pirates take a bathe earlier than they stroll the plank? As a result of
they're going to simply wash up on shore later! 🏴‍☠️

Why did the scurvy canine's parrot go to the physician? As a result of it had a fowl
mood, savvy?

Create a Chat Completion

!pip set up openai
!pip set up aisuite[all]
os.environ['OPENAI_API_KEY'] = getpass('Enter your OPENAI API key: ')
from getpass import getpass
import aisuite as ai
shopper = ai.Shopper()
supplier = "openai"
model_id = "gpt-4o"
messages = [
   {"role": "system", "content": "You are a helpful assistant."},
   {"role": "user", "content": "Give me a tabular comparison of RAG and AGENTIC RAG"},
]
response = shopper.chat.completions.create(
   mannequin=f"{supplier}:{model_id}",
   messages=messages,
)
print(response.decisions[0].message.content material)

Output

Definitely! Beneath is a tabular comparability of Retrieval-Augmented Era
(RAG) and Agentic RAG.

| Function                | RAG                                             |
Agentic RAG                                        |

|------------------------|-------------------------------------------------|-
---------------------------------------------------|

| Definition             | A framework that mixes retrieval from exterior
paperwork with technology. | An extension of RAG that includes actions
primarily based on exterior interactions and dynamic decision-making. |

| Elements             | - Retrieval System (e.g., a search engine or
doc database) <br> - Generator (e.g., a language mannequin) | - Retrieval
System <br> - Generator <br> - Agentic Layer (action-taking and interplay
controller) |

| Performance          | Retrieves related paperwork and generates
responses primarily based on prompted inputs mixed with the retrieved data.
| Provides the aptitude to take actions primarily based on interactions, similar to
interacting with APIs, controlling gadgets, or dynamically gathering extra
data. |

| Use Circumstances              | - Information-based query answering <br> -
Content material summarization <br> - Open-domain dialogue techniques | - Autonomous
brokers <br> - Interactive techniques <br> - Determination-making purposes <br> -
Programs requiring context-based actions |

| Interplay            | Restricted to the enter retrieval and output
technology cycle. | Can work together with exterior techniques or interfaces to
collect information, execute duties, and alter the atmosphere primarily based on goal
capabilities. |

| Complexity             | Usually less complicated because it combines retrieval with
technology with out taking actions past producing textual content. | Extra advanced due
to its capacity to work together with and modify the state of exterior
environments. |

| Instance of Utility | Answering advanced questions by retrieving components of
paperwork and synthesizing them into coherent solutions. | Implementing a
digital assistant able to performing duties like scheduling appointments
by accessing calendars, or a chatbot that manages customer support queries
by means of actions. |

| Flexibility            | Restricted to the accessible retrieval corpus and
technology mannequin capabilities. | Extra versatile attributable to action-oriented
interactions that may adapt to dynamic environments and circumstances. |

| Determination-Making Capacity| Restricted decision-making primarily based on static retrieval
and technology. | Enhanced decision-making by means of dynamic interplay and
adaptive habits. |

This comparability outlines the foundational variations and capabilities
between conventional RAG techniques and the extra superior, interaction-capable
Agentic RAG frameworks.

Every mannequin Makes use of a Totally different Supplier

1. Putting in and Importing Libraries

!pip set up aisuite[all]
from pprint import pprint as pp
  • Installs the aisuite library with all elective dependencies.
  • Imports a pretty-printing operate (pprint) to format output for higher readability. A customized pprint operate is outlined to permit a customized width.

2. Setting Up API Keys

import os
from getpass import getpass
os.environ['GROQ_API_KEY'] = getpass('Enter your GROQ API key: ')

Prompts the person to enter their GROQ API key, which is saved within the atmosphere variable GROQ_API_KEY.

3. Initializing the AI Shopper

import aisuite as ai
shopper = ai.Shopper()

Initializes an AI shopper utilizing the aisuite library to work together with totally different fashions.

4. Chat Completions

messages = [
    {"role": "system", "content": "You are a helpful agent, who answers with brevity."},
    {"role": "user", "content": 'Hi'},
]
response = shopper.chat.completions.create(mannequin="groq:llama-3.2-3b-preview", messages=messages)
print(response.decisions[0].message.content material)

Output

How can I help you?
  • Defines a chat with two messages:
    • A system message that units the tone or habits of the AI (concise responses).
    • A person message as enter.
  • Sends the messages to the AI mannequin groq:llama-3.2-3b-preview and prints the mannequin’s response.

5. Perform to Ship Queries

def ask(message, sys_message="You're a useful agent.",
         mannequin="groq:llama-3.2-3b-preview"):
    shopper = ai.Shopper()
    messages = [
        {"role": "system", "content": sys_message},
        {"role": "user", "content": message}
    ]
    response = shopper.chat.completions.create(mannequin=mannequin, messages=messages)
    return response.decisions[0].message.content material
ask("Hello. what's capital of Japan?")

Output

'Hi there. The capital of Japan is Tokyo.'
  • ask is a reusable operate to ship queries to the mannequin.
  • Accepts:
    • message: The person’s question.
    • sys_message: Non-obligatory system instruction.
    • mannequin: Specifies the AI mannequin.
  • Sends the enter and returns the AI’s response.

6. Utilizing A number of APIs

os.environ['OPENAI_API_KEY'] = getpass('Enter your OPENAI API key: ')
os.environ['ANTHROPIC_API_KEY'] = getpass('Enter your ANTHROPIC API key: ')
print(ask("Who's your creator?"))
print(ask('Who's your creator?', mannequin="anthropic:claude-3-5-sonnet-20240620"))
print(ask('Who's your creator?', mannequin="openai:gpt-4o"))

Output

I used to be created by Meta AI, a number one synthetic intelligence analysis
group. My data was developed from a big corpus of textual content, which
I exploit to generate human-like responses to person queries.

I used to be created by Anthropic.

I used to be developed by OpenAI, a company that focuses on synthetic
intelligence analysis and deployment.

  • Prompts the person for OpenAI and Anthropic API keys.
  • Sends a question (“Who’s your creator?”) to totally different fashions:
    • groq:llama-3.2-3b-preview
    • anthropic:claude-3-5-sonnet-20240620
    • openai:gpt-4o
  • Prints the response from every mannequin, exhibiting how totally different techniques interpret the identical question.

7. Querying A number of Fashions

fashions = [
    'llama-3.1-8b-instant',
    'llama-3.2-1b-preview',
    'llama-3.2-3b-preview',
    'llama3-70b-8192',
    'llama3-8b-8192'
]
ret = []
for x in fashions:
    ret.append(ask('Write a brief one sentence rationalization of the origins of AI?', mannequin=f'groq:{x}'))
  • An inventory of various mannequin identifiers (fashions) is outlined.
  • Loops by means of every mannequin and queries it with:
    • Write a brief one sentence rationalization of the origins of AI?
  • Shops responses within the listing ret.

8. Displaying Mannequin Responses

for idx, x in enumerate(ret):
    pprint(fashions[idx] + ': n ' + x + ' ')
  • Loops by means of the saved responses.
  • Codecs and prints the mannequin’s identify together with its response, making it simple to match outputs.

Output

('llama-3.1-8b-instant: n'

 ' The origins of Synthetic Intelligence (AI) date again to the 1956 Dartmouth '

 'Summer season Analysis Challenge on Synthetic Intelligence, the place a gaggle of '

 'pc scientists, led by John McCarthy, Marvin Minsky, Nathaniel '

 'Rochester, and Claude Shannon, coined the time period and laid the muse for '

 'the event of AI as a definite subject of examine. ')

('llama-3.2-1b-preview: n'

 ' The origins of Synthetic Intelligence (AI) date again to the mid-Twentieth '

 'century, when the primary pc applications, which mimicked human-like '

 'intelligence by means of algorithms and rule-based techniques, had been developed by '

 'famend mathematicians and pc scientists, together with Alan Turing, '

 'Marvin Minsky, and John McCarthy within the Nineteen Fifties. ')

('llama-3.2-3b-preview: n'

 ' The origins of Synthetic Intelligence (AI) date again to the Nineteen Fifties, with '

 'the Dartmouth Summer season Analysis Challenge on Synthetic Intelligence, led by '

 'pc scientists John McCarthy, Marvin Minsky, and Nathaniel Rochester, '

 'marking the start of AI as a proper subject of analysis. ')

('llama3-70b-8192: n'

 ' The origins of Synthetic Intelligence (AI) may be traced again to the Nineteen Fifties '

 'when pc scientist Alan Turing proposed the Turing Check, a technique for '

 'figuring out whether or not a machine may exhibit clever habits equal '

 'to, or indistinguishable from, that of a human. ')

('llama3-8b-8192: n'

 ' The origins of Synthetic Intelligence (AI) may be traced again to the '

 'Nineteen Fifties, when pc scientists DARPA funded the event of the primary AI '

 'applications, such because the Logical Theorist, which aimed to simulate human '

 'problem-solving talents and study from expertise. ')

Fashions present various responses to the question concerning the origins of AI, reflecting their coaching and reasoning capabilities. For example:

  • Some fashions reference the Dartmouth Summer season Analysis Challenge on AI.
  • Others point out Alan Turing or early DARPA-funded AI applications.

Key Options and Takeaways

  • Modularity: The script makes use of reusable capabilities (ask) to make querying environment friendly and customisable.
  • Multi-Mannequin Interplay: Showcases the power to work together with varied AI techniques, together with GROQ, OpenAI, and Anthropic.
  • Comparative Evaluation: Facilitates comparability of responses throughout fashions for insights into their strengths and biases.
  • Actual-Time Inputs: Helps dynamic enter for API keys, guaranteeing safe integration.

This script is a wonderful place to begin for exploring totally different AI mannequin capabilities and understanding their distinctive behaviours.

Conclusion

AISuite is an important software for anybody navigating the world of enormous language fashions. It empowers customers to harness the most effective of a number of AI suppliers whereas simplifying improvement and fostering innovation. Its open-source nature and considerate design underscore its potential as a contemporary AI software improvement cornerstone.

It accelerates improvement and enhances flexibility by enabling seamless switching between fashions like OpenAI, Anthropic, and Meta with minimal integration effort. Supreme for each easy and complicated purposes, AISuite helps modular workflows, API key administration, and real-time multi-model comparisons. Its ease of use, scalability, and talent to streamline cross-provider interactions make it a useful useful resource for builders, researchers, and educators, empowering environment friendly and modern utilisation of various LLMs in an evolving AI panorama.

In case you are searching for generative AI course on-line then discover: GenAI Pinnacle Program

Continuously Requested Questions

Q1. What’s AISuite?

Ans. AISuite is an open-source Python package deal created by Andrew Ng to streamline working with a number of massive language fashions (LLMs) from varied suppliers. It gives a unified interface for switching between fashions, simplifying integration and accelerating improvement.

Q3. Can I exploit AISuite with a number of AI suppliers without delay?

Ans. Sure, AISuite helps querying a number of fashions from totally different suppliers concurrently. You may ship the identical question to totally different fashions and examine their responses.

This fall. What’s the key characteristic of AISuite?

Ans. AISuite’s key characteristic is its modularity and talent to combine a number of LLMs right into a single workflow. It additionally simplifies API key administration and permits simple switching between fashions, facilitating fast comparisons and experimentation.

Q5. How do I set up AISuite?

Ans. To put in AISuite and crucial libraries, run:
!pip set up aisuite[all]
!pip set up openai

Hello, I’m Pankaj Singh Negi – Senior Content material Editor | Keen about storytelling and crafting compelling narratives that remodel concepts into impactful content material. I really like studying about know-how revolutionizing our way of life.