Browse Source

Merge pull request #84 from l-jacques/ljacques-w2d2

Exercise week2 day 2: offer multi llms dropdown and reusable AISystem.py with stream option
pull/87/head
Ed Donner 4 months ago committed by GitHub
parent
commit
8ae9f1866b
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
  1. 81
      week2/community-contributions/AISystem.py
  2. 202
      week2/community-contributions/day2-exercise_gradio_dropdown.ipynb

81
week2/community-contributions/AISystem.py

@ -0,0 +1,81 @@
from enum import Enum, auto
from openai import OpenAI
import anthropic
def formatPrompt(role, content):
return {"role": role, "content": content}
class AI(Enum):
OPEN_AI = "OPEN_AI"
CLAUDE = "CLAUDE"
GEMINI = "GEMINI"
OLLAMA = "OLLAMA"
class AISystem:
def __init__(self, processor, system_string="", model="", type=AI.OPEN_AI):
"""
Initialize the ChatSystem with a system string and empty messages list.
:param system_string: Optional initial system string description
"""
self.processor = processor
self.system = system_string
self.model = model
self.messages = []
self.type = type
def call(self, message):
self.messages.append(message)
toSend = self.messages
if self.type == AI.CLAUDE:
message = self.processor.messages.create(
model=self.model,
system=self.system,
messages=self.messages,
max_tokens=500
)
return message.content[0].text
else:
toSend.insert(0,self.system)
completion = self.processor.chat.completions.create(
model=self.model,
messages= toSend
)
return completion.choices[0].message.content
def stream(self, message, usingGradio=False):
self.messages.append(message)
if self.type == AI.CLAUDE:
result = self.processor.messages.stream(
model=self.model,
system=self.system,
messages=self.messages,
temperature=0.7,
max_tokens=500
)
response_chunks = ""
with result as stream:
for text in stream.text_stream:
if usingGradio:
response_chunks += text or ""
yield response_chunks
else:
yield text
else:
toSend = self.messages
toSend.insert(0,self.system)
stream = self.processor.chat.completions.create(
model=self.model,
messages= toSend,
stream=True
)
response_chunks = ""
for chunk in stream:
if usingGradio:
response_chunks += chunk.choices[0].delta.content or "" # need to yield the total cumulative results to gradio and not chunk by chunk
yield response_chunks
else:
yield chunk.choices[0].delta.content

202
week2/community-contributions/day2-exercise_gradio_dropdown.ipynb

@ -0,0 +1,202 @@
{
"cells": [
{
"cell_type": "markdown",
"id": "a473d607-073d-4963-bdc4-aba654523681",
"metadata": {},
"source": [
"## Day 2 Exercise\n",
"building upon the day1 exercise to offer a multi models via dropdown.\n",
"externalized the common methods into a AISystem.py file to be reused down the line"
]
},
{
"cell_type": "markdown",
"id": "f761729f-3bd5-4dd7-9e63-cbe6b4368a66",
"metadata": {},
"source": [
"## Load env, check for api keys and load up the connections"
]
},
{
"cell_type": "code",
"execution_count": 1,
"id": "fedb3d94-d096-43fd-8a76-9fdbc2d0d78e",
"metadata": {},
"outputs": [
{
"name": "stdout",
"output_type": "stream",
"text": [
"OpenAI API Key exists and begins sk-proj-\n",
"Anthropic API Key exists and begins sk-ant-\n",
"Google API Key exists and begins AIzaSyC-\n"
]
}
],
"source": [
"import os\n",
"from enum import Enum, auto\n",
"from dotenv import load_dotenv\n",
"from openai import OpenAI\n",
"import anthropic\n",
"from AISystem import formatPrompt, AI, AISystem\n",
"import gradio as gr # oh yeah!\n",
"\n",
"# Load environment variables in a file called .env\n",
"# Print the key prefixes to help with any debugging\n",
"\n",
"load_dotenv()\n",
"openai_api_key = os.getenv('OPENAI_API_KEY')\n",
"anthropic_api_key = os.getenv('ANTHROPIC_API_KEY')\n",
"google_api_key = os.getenv('GOOGLE_API_KEY')\n",
"\n",
"if openai_api_key:\n",
" print(f\"OpenAI API Key exists and begins {openai_api_key[:8]}\")\n",
"else:\n",
" print(\"OpenAI API Key not set\")\n",
" \n",
"if anthropic_api_key:\n",
" print(f\"Anthropic API Key exists and begins {anthropic_api_key[:7]}\")\n",
"else:\n",
" print(\"Anthropic API Key not set\")\n",
"\n",
"if google_api_key:\n",
" print(f\"Google API Key exists and begins {google_api_key[:8]}\")\n",
"else:\n",
" print(\"Google API Key not set\")\n",
"\n",
"openai = OpenAI()\n",
"\n",
"claude = anthropic.Anthropic()\n",
"\n",
"gemini_via_openai_client = OpenAI(\n",
" api_key=google_api_key, \n",
" base_url=\"https://generativelanguage.googleapis.com/v1beta/openai/\"\n",
")\n",
"ollama_via_openai = OpenAI(base_url='http://localhost:11434/v1', api_key='ollama')\n",
"openai_model = \"gpt-4o-mini\"\n",
"claude_model = \"claude-3-haiku-20240307\"\n",
"gemini_model = \"gemini-1.5-flash\"\n",
"ollama_model = \"llama3.2\""
]
},
{
"cell_type": "markdown",
"id": "17f7987b-2bdf-434a-8fce-6c367f148dde",
"metadata": {},
"source": [
"## Create the systems for each llms"
]
},
{
"cell_type": "code",
"execution_count": 2,
"id": "f92eef29-325e-418c-a444-879d83d5fbc9",
"metadata": {},
"outputs": [],
"source": [
"geminiSys = AISystem(gemini_via_openai_client,\n",
" formatPrompt(\"system\",\"You are a chatbot. you always try to make conversation and get more in depth\"), \n",
" gemini_model,\n",
" AI.GEMINI)\n",
"\n",
"openAiSys = AISystem(openai,\n",
" formatPrompt(\"system\",\"You are a chatbot. you always try to make conversation and get more in depth\"), \n",
" openai_model,\n",
" AI.OPEN_AI)\n",
"\n",
"claudeSys = AISystem(claude,\n",
" \"You are a chatbot. you always try to make conversation and get more in depth\", \n",
" claude_model,\n",
" AI.CLAUDE)\n",
"\n",
"ollamaSys = AISystem(ollama_via_openai,\n",
" formatPrompt(\"system\",\"You are a chatbot. you always try to make conversation and get more in depth\"), \n",
" ollama_model,\n",
" AI.OLLAMA)\n",
"sys_dict = { AI.GEMINI: geminiSys, AI.OPEN_AI: openAiSys, AI.CLAUDE: claudeSys, AI.OLLAMA: ollamaSys}\n",
"\n",
"def stream_model(prompt, model):\n",
" aiSystem = sys_dict.get(AI[model.upper()])\n",
" yield from aiSystem.stream(formatPrompt(\"user\",prompt), True)"
]
},
{
"cell_type": "markdown",
"id": "f8ecd283-92b2-454d-b1ae-8016d41e3026",
"metadata": {},
"source": [
"## Create the gradio interface linking with the AI enum for the dropdown"
]
},
{
"cell_type": "code",
"execution_count": 3,
"id": "9db8ed67-280a-400d-8543-4ab95863ce51",
"metadata": {},
"outputs": [
{
"name": "stdout",
"output_type": "stream",
"text": [
"* Running on local URL: http://127.0.0.1:7873\n",
"\n",
"To create a public link, set `share=True` in `launch()`.\n"
]
},
{
"data": {
"text/html": [
"<div><iframe src=\"http://127.0.0.1:7873/\" width=\"100%\" height=\"500\" allow=\"autoplay; camera; microphone; clipboard-read; clipboard-write;\" frameborder=\"0\" allowfullscreen></iframe></div>"
],
"text/plain": [
"<IPython.core.display.HTML object>"
]
},
"metadata": {},
"output_type": "display_data"
},
{
"data": {
"text/plain": []
},
"execution_count": 3,
"metadata": {},
"output_type": "execute_result"
}
],
"source": [
"\n",
"view = gr.Interface(\n",
" fn=stream_model,\n",
" inputs=[gr.Textbox(label=\"Your prompt:\", lines=6) , gr.Dropdown(choices=[ai.value for ai in AI], label=\"Select model\")],\n",
" outputs=[gr.Markdown(label=\"Response:\")],\n",
" flagging_mode=\"never\"\n",
")\n",
"view.launch()"
]
}
],
"metadata": {
"kernelspec": {
"display_name": "Python 3 (ipykernel)",
"language": "python",
"name": "python3"
},
"language_info": {
"codemirror_mode": {
"name": "ipython",
"version": 3
},
"file_extension": ".py",
"mimetype": "text/x-python",
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.11.11"
}
},
"nbformat": 4,
"nbformat_minor": 5
}
Loading…
Cancel
Save