| .. |
|
__init__.py
|
feat: server multi models support (#799)
|
2023-08-12 00:57:00 +08:00 |
|
anthropic_llm.py
|
feat: disable anthropic retry (#1067)
|
2023-08-31 16:44:46 +08:00 |
|
azure_chat_open_ai.py
|
feat: server multi models support (#799)
|
2023-08-12 00:57:00 +08:00 |
|
azure_open_ai.py
|
feat: server multi models support (#799)
|
2023-08-12 00:57:00 +08:00 |
|
chat_open_ai.py
|
feat: add LocalAI local embedding model support (#1021)
|
2023-08-29 22:22:02 +08:00 |
|
fake.py
|
feat: server multi models support (#799)
|
2023-08-12 00:57:00 +08:00 |
|
huggingface_endpoint_llm.py
|
feat: hf inference endpoint stream support (#1028)
|
2023-08-26 19:48:34 +08:00 |
|
huggingface_hub_llm.py
|
fix: hf hosted inference check (#1128)
|
2023-09-09 00:29:48 +08:00 |
|
open_ai.py
|
feat: add LocalAI local embedding model support (#1021)
|
2023-08-29 22:22:02 +08:00 |
|
openllm.py
|
fix: openllm generate cutoff (#945)
|
2023-08-22 13:43:36 +08:00 |
|
replicate_llm.py
|
feat: server multi models support (#799)
|
2023-08-12 00:57:00 +08:00 |
|
spark.py
|
feat: add spark v2 support (#885)
|
2023-08-17 15:08:57 +08:00 |
|
tongyi_llm.py
|
feat: server multi models support (#799)
|
2023-08-12 00:57:00 +08:00 |
|
wenxin.py
|
fix: wenxin error not raise when stream mode (#884)
|
2023-08-17 13:40:00 +08:00 |
|
xinference_llm.py
|
fix: xinference last token being ignored (#1013)
|
2023-08-25 18:15:05 +08:00 |