feat: add support for Gemini, Claude through Langchain (#225) (bump:patch)
This commit is contained in:
committed by
GitHub
parent
8be8a4a9d0
commit
05245f501c
@@ -175,6 +175,33 @@ if config("LOCAL_MODEL", default=""):
|
||||
"default": False,
|
||||
}
|
||||
|
||||
# additional LLM configurations
|
||||
KH_LLMS["claude"] = {
|
||||
"spec": {
|
||||
"__type__": "kotaemon.llms.chats.LCAnthropicChat",
|
||||
"model_name": "claude-3-5-sonnet-20240620",
|
||||
"api_key": "your-key",
|
||||
},
|
||||
"default": False,
|
||||
}
|
||||
KH_LLMS["gemini"] = {
|
||||
"spec": {
|
||||
"__type__": "kotaemon.llms.chats.LCGeminiChat",
|
||||
"model_name": "gemini-1.5-pro",
|
||||
"api_key": "your-key",
|
||||
},
|
||||
"default": False,
|
||||
}
|
||||
KH_LLMS["groq"] = {
|
||||
"spec": {
|
||||
"__type__": "kotaemon.llms.ChatOpenAI",
|
||||
"base_url": "https://api.groq.com/openai/v1",
|
||||
"model": "llama-3.1-8b-instant",
|
||||
"api_key": "your-key",
|
||||
},
|
||||
"default": False,
|
||||
}
|
||||
|
||||
KH_REASONINGS = [
|
||||
"ktem.reasoning.simple.FullQAPipeline",
|
||||
"ktem.reasoning.simple.FullDecomposeQAPipeline",
|
||||
|
Reference in New Issue
Block a user