NativeGPT
- Go to https://chat.openai.com and open the developer tools by
F12
. - Find the
__Secure-next-auth.session-token
cookie inApplication
>Storage
>Cookies
>https://chat.openai.com
. - Copy the value in the
Cookie Value
field and add in .env file.
- Run app.py
Have you applied to GPT-4 API access but OpenAI is too busy to reply? Me too, that's why I created this package. It uses selenium webdriver to emulate user interaction on chat.openai.com. If the account has ChatGPT Plus
, the driver will use GPT-4, otherwise it will use the default GPT-3.5.
It supports both GPT4 browser, DALL-E 3, and plugins by selecting model via GPT4OpenAI(token=token, model='gpt-4')
, or model='gpt-4-plugins'
.
Note: This unofficial API library is not endorsed by OpenAI and violates their Terms of Service. Use it at your own risk; the creator assumes no liability for any consequences. Please adhere to platform's ToS and exercise caution with unofficial resources.
The core logic was taken from the IntelligenzaArtificiale/Free-Auto-GPT.
from gpt4_openai import GPT4OpenAI
# Token is the __Secure-next-auth.session-token from chat.openai.com
llm = GPT4OpenAI(token=my_session_token, headless=False, model='gpt-4')
# GPT3.5 will answer 8, while GPT4 should be smart enough to answer 10
response = llm('If there are 10 books in a room and I read 2, how many books are still in the room?')
print(response)
This code was used for the demo gif above.
from gpt4_openai import GPT4OpenAI
llm = GPT4OpenAI(token=my_session_token, headless=False,
model='gpt-4' # DALL-E 3 only works with gpt-4
)
img_bytes = llm.generate_image('Generate an isometric image of a cute doggo inside a house.', image_path = './img_save_path.png')
from gpt4_openai import GPT4OpenAI
# Token is the __Secure-next-auth.session-token from chat.openai.com
llm = GPT4OpenAI(token=my_session_token, headless=False, model='gpt-4-browsing')
# ChatGPT will first browse the web for the name/age of her boyfriend, then return the answer
response = llm('What is the age difference between Dua Lipa and her boyfriend?')
print(response)
GPT4OpenAI
actually extends LLM
class from langchain.llms.base
. So you can easily use this library inside langchain ecosystem. Example:
from gpt4_openai import GPT4OpenAI
from langchain import LLMChain
from langchain.prompts.chat import (ChatPromptTemplate, SystemMessagePromptTemplate, AIMessagePromptTemplate, HumanMessagePromptTemplate)
template="You are a helpful assistant that translates english to pirate."
system_message_prompt = SystemMessagePromptTemplate.from_template(template)
example_human = HumanMessagePromptTemplate.from_template("Hi")
example_ai = AIMessagePromptTemplate.from_template("Argh me mateys")
human_message_prompt = HumanMessagePromptTemplate.from_template("{text}")
chat_prompt = ChatPromptTemplate.from_messages([system_message_prompt, example_human, example_ai, human_message_prompt])
# Token is the __Secure-next-auth.session-token from chat.openai.com
llm = GPT4OpenAI(token=my_session_token)
chain = LLMChain(llm=llm, prompt=chat_prompt)
print(chain.run("My name is John and I like to eat pizza."))
Output will be:
AI: Ahoy, me name be John an' I be likin' ta feast on some pizza, arr!
Initially, I tried poe.com (private API implemented at gpt4free), but noticed that input context window is smaller than one of OpenAI ChatGPT. And the same goes for Bing's GPT4.
To install this Python package, run the following command:
pip install gpt4-openai-api
These dependencies get downloaded directly:
undetected-chromedriver
(selenium browser)markdownify
langchain