File size: 3,568 Bytes
dc98d0d
 
 
 
b868aa9
af9e03a
dc98d0d
 
 
 
 
af9e03a
ee05990
dc98d0d
 
 
6212edf
dc98d0d
 
 
 
 
 
 
 
 
 
 
5bc9297
dc98d0d
 
 
 
 
 
 
 
 
081ec14
dc98d0d
081ec14
 
dc98d0d
4b8fbf1
 
 
dc98d0d
 
4c9e4a1
 
 
 
 
ee05990
dc98d0d
af9e03a
dc98d0d
 
 
 
 
 
748df68
dc98d0d
 
748df68
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
import streamlit as st
import requests
import json

st.title("OpenAI Chatbot Interface")
st.write("Interact with OpenAI's GPT-3 models in real-time using your OpenAI API. Choose from a selection of their best models, set the temperature and max tokens, and start a conversation. Delete the conversation at any time to start fresh.")

if "history" not in st.session_state:
    st.session_state.history = []

st.sidebar.markdown("## Configuration")
KEY = st.sidebar.text_input("Enter Your OpenAI API Key", placeholder="API Key", value="")
models = ['text-davinci-003', 'text-curie-001', 'text-babbage-001', 'text-ada-001']
model = st.sidebar.selectbox("Select a model", models, index=0)

temperature = st.sidebar.slider("Temperature", 0.0, 1.0, 0.7)
max_tokens = st.sidebar.slider("Max Tokens", 0, 4000, 1786)

if st.sidebar.button("Delete Conversation"):
    st.session_state.history = []
st.sidebar.markdown("## GPT-3")
st.sidebar.markdown("OpenAI's GPT-3 models can understand and generate natural language. They offer four main models with different levels of power suitable for different tasks. Davinci is the most capable model, and Ada is the fastest.")
st.sidebar.markdown("text-davinci-003 | 4,000 max tokens")
st.sidebar.markdown("text-curie-001 | 2,048 max tokens")
st.sidebar.markdown("text-babbage-001 | 2,048 max tokens")
st.sidebar.markdown("text-ada-001 | 2,048 max tokens")

def generate_answer(prompt):
    API_KEY = KEY
    API_URL = "https://api.openai.com/v1/completions"
    headers = {
        'Content-Type': 'application/json',
        'Authorization': 'Bearer ' + API_KEY
    }
    previous_messages = [chat['message'] for chat in st.session_state.history if not chat['is_user']]
    previous_messages_text = '\n'.join(previous_messages)
    full_prompt = previous_messages_text + '\n' + prompt if previous_messages_text else prompt
    data = {
        "model": model,
        "prompt": full_prompt,
        "temperature": temperature,
        "max_tokens": max_tokens
    }
    if not API_KEY:
        st.warning("Please input your API key")
        return
    response = requests.post(API_URL, headers=headers, data=json.dumps(data))
    result = response.json()
    if 'choices' in result:
        message_bot = result['choices'][0]['text'].strip()
        st.session_state.history.append({"message": prompt, "is_user": True})
        st.session_state.history.append({"message": message_bot, "is_user": False})
    else:
        st.error("An error occurred while processing the API response. If using a model other than text-davinci-003, then lower the Max Tokens.")

prompt = st.text_input("Prompt", placeholder="Prompt Here", value="")
if st.button("Submit"):
    generate_answer(prompt)
    with st.spinner("Waiting for the response from the bot..."):
        for chat in st.session_state.history:
            if chat['is_user']:
                st.markdown("<img src='https://i.ibb.co/zVSbGvb/585e4beacb11b227491c3399.png' width='50' height='50' style='float:right;'>", unsafe_allow_html=True)
                st.markdown(f"<div style='float:right; padding:10px; background-color: #2E2E2E; border-radius:10px; margin:10px;'>{chat['message']}</div>", unsafe_allow_html=True)
            else:
                st.markdown("<img src='https://i.ibb.co/LZFvDND/5841c0bda6515b1e0ad75a9e-1.png' width='50' height='50' style='float:left;'>", unsafe_allow_html=True)
                st.markdown(f"<div style='float:left; padding:10px; background-color: #2E2E2E; border-radius:10px; margin:10px;'>{chat['message']}</div>", unsafe_allow_html=True)