import asyncio from openai import AsyncOpenAI async def list_available_models(): """获取LiteLLM服务器支持的模型列表""" print("正在获取可用模型列表...") # 使用远程LiteLLM服务器 client = AsyncOpenAI( api_key="sk-0jdcGHZJpX2oUJmyEs7zVA", base_url="https://litellm.seekkey.tech" ) try: # 获取模型列表 models = await client.models.list() print("\n=== 可用模型列表 ===") for model in models.data: print(f"- {model.id}") print(f"\n总共找到 {len(models.data)} 个模型") # 尝试调用一个简单的模型 if models.data: first_model = models.data[0].id print(f"\n正在测试第一个模型: {first_model}") response = await client.chat.completions.create( model=first_model, messages=[ {"role": "user", "content": "Hello, please say hi in Chinese."} ], max_tokens=50 ) print(f"测试响应: {response.choices[0].message.content}") except Exception as e: print(f"获取模型列表失败: {e}") print(f"错误类型: {type(e).__name__}") # 尝试直接测试一些常见模型 common_models = ["gpt-4", "gpt-3.5-turbo", "gemini-pro", "claude-3-sonnet"] print("\n尝试测试常见模型...") for model in common_models: try: print(f"测试模型: {model}") response = await client.chat.completions.create( model=model, messages=[{"role": "user", "content": "Hi"}], max_tokens=10 ) print(f"✓ {model} 可用") break except Exception as model_error: print(f"✗ {model} 不可用: {str(model_error)[:100]}...") finally: await client.close() if __name__ == "__main__": asyncio.run(list_available_models())