Spaces:
Running
Running
| # from g4f.Provider import DeepInfraChat,Glider,Liaobots,Blackbox,ChatGptEs,LambdaChat,TypeGPT | |
| # DeepInfraChat.models = ["google/gemma-3-27b-it","deepseek-ai/DeepSeek-R1-Turbo","Qwen/QwQ-32B","deepseek-ai/DeepSeek-R1","deepseek-ai/DeepSeek-V3-0324","meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8","meta-llama/Llama-4-Scout-17B-16E-Instruct","microsoft/Phi-4-multimodal-instruct"] | |
| # deepinframodels=["meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8","microsoft/Phi-4-multimodal-instruct","google/gemma-3-27b-it","meta-llama/Llama-4-Scout-17B-16E-Instruct"] | |
| # REASONING_CORRESPONDANCE = {"DeepSeek-R1-Glider":Glider, "DeepSeekR1-LAMBDA":LambdaChat,"DeepSeekR1":DeepInfraChat,"deepseek-slow":TypeGPT} | |
| # REASONING_QWQ = {"qwq-32b":DeepInfraChat} | |
| # from g4f.client import Client | |
| # client = Client() | |
| # cunk="" | |
| # providers=REASONING_CORRESPONDANCE | |
| # model_name="deepseek-r1" | |
| # for provider in providers: | |
| # try: | |
| # response = client.chat.completions.create( | |
| # provider=providers[provider], | |
| # model=model_name, | |
| # messages=[{"role": "user", "content": f"Hi!"}], | |
| # stream=True | |
| # # Add any other necessary parameters | |
| # ) | |
| # for part in response: | |
| # # print(part) | |
| # cunk=cunk+(str(part.choices[0].delta.content) or "") | |
| # print(str(part.choices[0].delta.content),end="") | |
| # break | |
| # except Exception as e: | |
| # print(f"Error with {provider}: {e}") | |
| # pass | |