The official gpt4free repository | various collection of powerful language models
Written by @xtekky & maintained by @hlohaus
By using this repository or any code related to it, you agree to the legal notice. The author is not responsible for the usage of this repository nor endorses it, nor is the author responsible for any copies, forks, re-uploads made by other users, or anything else related to GPT4Free. This is the author's only account and repository. To prevent impersonation or irresponsible actions, please comply with the GNU GPL license this Repository uses.
[!Warning] "gpt4free" serves as a PoC (proof of concept), demonstrating the development of an API package with multi-provider requests, with features like timeouts, load balance and flow control.
pip install -U g4f
docker pull hlohaus789/g4f
g4f
now supports 100% local inference: đ§ local-docs
Is your site on this repository and you want to take it down? Send an email to [email protected] with proof it is yours and it will be removed as fast as possible. To prevent reproduction please secure your API. đ
You can always leave some feedback here: https://forms.gle/FeWV9RLEedfdkmFN6
As per the survey, here is a list of improvements to come
Openai()
class) | completed, use g4f.client.Client
Install Docker: Begin by downloading and installing Docker.
Set Up the Container: Use the following commands to pull the latest image and start the container:
docker pull hlohaus789/g4f
docker run -p 8080:8080 -p 1337:1337 -p 7900:7900 --shm-size="2g" -v ${PWD}/har_and_cookies:/app/har_and_cookies hlohaus789/g4f:latest
Access the Client:
(Optional) Provider Login: If required, you can access the container's desktop here: http://localhost:7900/?autoconnect=1&resize=scale&password=secret for provider login purposes.
To ensure the seamless operation of our application, please follow the instructions below. These steps are designed to guide you through the installation process on Windows operating systems.
g4f.webview.*.exe
..exe
file from your downloads folder to a directory of your choice on your system, and then execute it to run the app.By following these steps, you should be able to successfully install and run the application on your Windows system. If you encounter any issues during the installation process, please refer to our Issue Tracker or try to get contact over Discord for assistance.
Run the Webview UI on other Platfroms:
Run the Web UI on Your Smartphone:
pip install -U g4f[all]
How do I install only parts or do disable parts? Use partial requirements: /docs/requirements
How do I load the project using git and installing the project requirements? Read this tutorial and follow it step by step: /docs/git
How do I build and run composer image from source? Use docker-compose: /docs/docker
from g4f.client import Client
client = Client()
response = client.chat.completions.create(
model="gpt-3.5-turbo",
messages=[{"role": "user", "content": "Hello"}],
...
)
print(response.choices[0].message.content)
Hello! How can I assist you today?
from g4f.client import Client
client = Client()
response = client.images.generate(
model="gemini",
prompt="a white siamese cat",
...
)
image_url = response.data[0].url
Full Documentation for Python API
To start the web interface, type the following codes in python:
from g4f.gui import run_gui
run_gui()
or execute the following command:
python -m g4f.cli gui -port 8080 -debug
You can use the Interference API to serve other OpenAI integrations with G4F.
See docs: /docs/interference
Access with: http://localhost:1337/v1
Cookies are essential for using Meta AI and Microsoft Designer to create images. Additionally, cookies are required for the Google Gemini and WhiteRabbitNeo Provider. From Bing, ensure you have the "_U" cookie, and from Google, all cookies starting with "__Secure-1PSID" are needed.
You can pass these cookies directly to the create function or set them using the set_cookies
method before running G4F:
from g4f.cookies import set_cookies
set_cookies(".bing.com", {
"_U": "cookie value"
})
set_cookies(".google.com", {
"__Secure-1PSID": "cookie value"
})
Alternatively, you can place your .har and cookie files in the /har_and_cookies
directory. To export a cookie file, use the EditThisCookie extension available on the Chrome Web Store: EditThisCookie Extension.
You can also create .har files to capture cookies. If you need further assistance, refer to the next section.
python -m g4f.cli api --debug
Read .har file: ./har_and_cookies/you.com.har
Cookies added: 10 from .you.com
Read cookie file: ./har_and_cookies/google.json
Cookies added: 16 from .google.com
Starting server... [g4f v-0.0.0] (debug)
To utilize the OpenaiChat provider, a .har file is required from https://chat.openai.com/. Follow the steps below to create a valid .har file:
./har_and_cookies
directory if you are using Docker. Alternatively, you can store it in any preferred location within your current working directory.Note: Ensure that your .har file is stored securely, as it may contain sensitive information.
If you want to hide or change your IP address for the providers, you can set a proxy globally via an environment variable:
export G4F_PROXY="http://host:port"
set G4F_PROXY=http://host:port
Website | Provider | GPT-3.5 | GPT-4 | Stream | Status | Auth |
---|---|---|---|---|---|---|
bing.com | g4f.Provider.Bing |
â | âī¸ | âī¸ | â | |
chatgpt.ai | g4f.Provider.ChatgptAi |
â | âī¸ | âī¸ | â | |
liaobots.site | g4f.Provider.Liaobots |
âī¸ | âī¸ | âī¸ | â | |
chat.openai.com | g4f.Provider.OpenaiChat |
âī¸ | âī¸ | âī¸ | â+âī¸ | |
raycast.com | g4f.Provider.Raycast |
âī¸ | âī¸ | âī¸ | âī¸ | |
beta.theb.ai | g4f.Provider.Theb |
âī¸ | âī¸ | âī¸ | â | |
you.com | g4f.Provider.You |
âī¸ | âī¸ | âī¸ | â |
While we wait for gpt-5, here is a list of new models that are at least better than gpt-3.5-turbo. Some are better than gpt-4. Expect this list to grow.
Website | Provider | parameters | better than |
---|---|---|---|
mixtral-8x22b | g4f.Provider.DeepInfra |
176B / 44b active | gpt-3.5-turbo |
dbrx-instruct | g4f.Provider.DeepInfra |
132B / 36B active | gpt-3.5-turbo |
command-r+ | g4f.Provider.HuggingChat |
104B | gpt-4-0613 |
Website | Provider | GPT-3.5 | GPT-4 | Stream | Status | Auth |
---|---|---|---|---|---|---|
chat3.aiyunos.top | g4f.Provider.AItianhuSpace |
âī¸ | â | âī¸ | â | |
chat10.aichatos.xyz | g4f.Provider.Aichatos |
âī¸ | â | âī¸ | â | |
chatforai.store | g4f.Provider.ChatForAi |
âī¸ | â | âī¸ | â | |
chatgpt4online.org | g4f.Provider.Chatgpt4Online |
âī¸ | â | âī¸ | â | |
chatgpt-free.cc | g4f.Provider.ChatgptNext |
âī¸ | â | âī¸ | â | |
chatgptx.de | g4f.Provider.ChatgptX |
âī¸ | â | âī¸ | â | |
f1.cnote.top | g4f.Provider.Cnote |
âī¸ | â | âī¸ | â | |
duckduckgo.com | g4f.Provider.DuckDuckGo |
âī¸ | â | âī¸ | â | |
ecosia.org | g4f.Provider.Ecosia |
âī¸ | â | âī¸ | â | |
feedough.com | g4f.Provider.Feedough |
âī¸ | â | âī¸ | â | |
flowgpt.com | g4f.Provider.FlowGpt |
âī¸ | â | âī¸ | â | |
freegptsnav.aifree.site | g4f.Provider.FreeGpt |
âī¸ | â | âī¸ | â | |
gpttalk.ru | g4f.Provider.GptTalkRu |
âī¸ | â | âī¸ | â | |
koala.sh | g4f.Provider.Koala |
âī¸ | â | âī¸ | â | |
app.myshell.ai | g4f.Provider.MyShell |
âī¸ | â | âī¸ | â | |
perplexity.ai | g4f.Provider.PerplexityAi |
âī¸ | â | âī¸ | â | |
poe.com | g4f.Provider.Poe |
âī¸ | â | âī¸ | âī¸ | |
talkai.info | g4f.Provider.TalkAi |
âī¸ | â | âī¸ | â | |
chat.vercel.ai | g4f.Provider.Vercel |
âī¸ | â | âī¸ | â | |
aitianhu.com | g4f.Provider.AItianhu |
âī¸ | â | âī¸ | â | |
chatgpt.bestim.org | g4f.Provider.Bestim |
âī¸ | â | âī¸ | â | |
chatbase.co | g4f.Provider.ChatBase |
âī¸ | â | âī¸ | â | |
chatgptdemo.info | g4f.Provider.ChatgptDemo |
âī¸ | â | âī¸ | â | |
chat.chatgptdemo.ai | g4f.Provider.ChatgptDemoAi |
âī¸ | â | âī¸ | â | |
chatgptfree.ai | g4f.Provider.ChatgptFree |
âī¸ | â | â | â | |
chatgptlogin.ai | g4f.Provider.ChatgptLogin |
âī¸ | â | âī¸ | â | |
chat.3211000.xyz | g4f.Provider.Chatxyz |
âī¸ | â | âī¸ | â | |
gpt6.ai | g4f.Provider.Gpt6 |
âī¸ | â | âī¸ | â | |
gptchatly.com | g4f.Provider.GptChatly |
âī¸ | â | â | â | |
ai18.gptforlove.com | g4f.Provider.GptForLove |
âī¸ | â | âī¸ | â | |
gptgo.ai | g4f.Provider.GptGo |
âī¸ | â | âī¸ | â | |
gptgod.site | g4f.Provider.GptGod |
âī¸ | â | âī¸ | â | |
onlinegpt.org | g4f.Provider.OnlineGpt |
âī¸ | â | âī¸ | â |
Website | Provider | Stream | Status | Auth |
---|---|---|---|---|
openchat.team | g4f.Provider.Aura |
âī¸ | â | |
blackbox.ai | g4f.Provider.Blackbox |
âī¸ | â | |
cohereforai-c4ai-command-r-plus.hf.space | g4f.Provider.Cohere |
âī¸ | â | |
deepinfra.com | g4f.Provider.DeepInfra |
âī¸ | â | |
free.chatgpt.org.uk | g4f.Provider.FreeChatgpt |
âī¸ | â | |
gemini.google.com | g4f.Provider.Gemini |
âī¸ | âī¸ | |
ai.google.dev | g4f.Provider.GeminiPro |
âī¸ | âī¸ | |
gemini-chatbot-sigma.vercel.app | g4f.Provider.GeminiProChat |
âī¸ | â | |
developers.sber.ru | g4f.Provider.GigaChat |
âī¸ | âī¸ | |
console.groq.com | g4f.Provider.Groq |
âī¸ | âī¸ | |
huggingface.co | g4f.Provider.HuggingChat |
âī¸ | â | |
huggingface.co | g4f.Provider.HuggingFace |
âī¸ | â | |
llama2.ai | g4f.Provider.Llama |
âī¸ | â | |
meta.ai | g4f.Provider.MetaAI |
âī¸ | â | |
openrouter.ai | g4f.Provider.OpenRouter |
âī¸ | âī¸ | |
labs.perplexity.ai | g4f.Provider.PerplexityLabs |
âī¸ | â | |
pi.ai | g4f.Provider.Pi |
âī¸ | â | |
replicate.com | g4f.Provider.Replicate |
âī¸ | â | |
theb.ai | g4f.Provider.ThebApi |
âī¸ | âī¸ | |
whiterabbitneo.com | g4f.Provider.WhiteRabbitNeo |
âī¸ | âī¸ | |
bard.google.com | g4f.Provider.Bard |
â | âī¸ |
Model | Base Provider | Provider | Website |
---|---|---|---|
gpt-3.5-turbo | OpenAI | 8+ Providers | openai.com |
gpt-4 | OpenAI | 2+ Providers | openai.com |
gpt-4-turbo | OpenAI | g4f.Provider.Bing | openai.com |
Llama-2-7b-chat-hf | Meta | 2+ Providers | llama.meta.com |
Llama-2-13b-chat-hf | Meta | 2+ Providers | llama.meta.com |
Llama-2-70b-chat-hf | Meta | 3+ Providers | llama.meta.com |
Meta-Llama-3-8b-instruct | Meta | 1+ Providers | llama.meta.com |
Meta-Llama-3-70b-instruct | Meta | 2+ Providers | llama.meta.com |
CodeLlama-34b-Instruct-hf | Meta | g4f.Provider.HuggingChat | llama.meta.com |
CodeLlama-70b-Instruct-hf | Meta | 2+ Providers | llama.meta.com |
Mixtral-8x7B-Instruct-v0.1 | Huggingface | 4+ Providers | huggingface.co |
Mistral-7B-Instruct-v0.1 | Huggingface | 3+ Providers | huggingface.co |
Mistral-7B-Instruct-v0.2 | Huggingface | g4f.Provider.DeepInfra | huggingface.co |
zephyr-orpo-141b-A35b-v0.1 | Huggingface | 2+ Providers | huggingface.co |
dolphin-2.6-mixtral-8x7b | Huggingface | g4f.Provider.DeepInfra | huggingface.co |
gemini | g4f.Provider.Gemini | gemini.google.com | |
gemini-pro | 2+ Providers | gemini.google.com | |
claude-v2 | Anthropic | 1+ Providers | anthropic.com |
claude-3-opus | Anthropic | g4f.Provider.You | anthropic.com |
claude-3-sonnet | Anthropic | g4f.Provider.You | anthropic.com |
lzlv_70b_fp16_hf | Huggingface | g4f.Provider.DeepInfra | huggingface.co |
airoboros-70b | Huggingface | g4f.Provider.DeepInfra | huggingface.co |
openchat_3.5 | Huggingface | 2+ Providers | huggingface.co |
pi | Inflection | g4f.Provider.Pi | inflection.ai |
Label | Provider | Image Model | Vision Model | Website |
---|---|---|---|---|
Microsoft Copilot in Bing | g4f.Provider.Bing |
dall-e-3 | gpt-4-vision | bing.com |
DeepInfra | g4f.Provider.DeepInfra |
stability-ai/sdxl | llava-1.5-7b-hf | deepinfra.com |
Gemini | g4f.Provider.Gemini |
âī¸ | âī¸ | gemini.google.com |
Gemini API | g4f.Provider.GeminiPro |
â | gemini-1.5-pro | ai.google.dev |
Meta AI | g4f.Provider.MetaAI |
âī¸ | â | meta.ai |
OpenAI ChatGPT | g4f.Provider.OpenaiChat |
dall-e-3 | gpt-4-vision | chat.openai.com |
Replicate | g4f.Provider.Replicate |
stability-ai/sdxl | llava-v1.6-34b | replicate.com |
You.com | g4f.Provider.You |
dall-e-3 | âī¸ | you.com |
đ Projects | â Stars | đ Forks | đ Issues | đŦ Pull requests |
gpt4free | gpt4free-ts | |||
Free AI API's & Potential Providers List | ||||
ChatGPT-Clone | ||||
Ai agent | ||||
ChatGpt Discord Bot | ||||
chatGPT-discord-bot | ||||
Nyx-Bot (Discord) | ||||
LangChain gpt4free | ||||
ChatGpt Telegram Bot | ||||
ChatGpt Line Bot | ||||
Action Translate Readme | ||||
Langchain Document GPT | ||||
python-tgpt |
We welcome contributions from the community. Whether you're adding new providers or features, or simply fixing typos and making small improvements, your input is valued. Creating a pull request is all it takes â our co-pilot will handle the code review process. Once all changes have been addressed, we'll merge the pull request into the main branch and release the updates at a later time.
A list of all contributors is available here
Vercel.py
file contains code from vercel-llm-api by @ading2210
har_file.py
has input from xqdoo00o/ChatGPT-to-API
PerplexityLabs.py
has input from nathanrchn/perplexityai
Gemini.py
has input from dsdanielpark/Gemini-API
MetaAI.py
file contains code from meta-ai-api by @Strvm
Having input implies that the AI's code generation utilized it as one of many sources.
This program is licensed under the GNU GPL v3
xtekky/gpt4free: Copyright (C) 2023 xtekky
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <https://www.gnu.org/licenses/>.
|
This project is licensed under GNU_GPL_v3.0. |