@@ -18,7 +18,6 @@ from langchain.document_loaders import UnstructuredFileLoader
# pip install pinecone-client,记得换源
import pinecone
import sentence_transformers
-from configs import *
from models import *
# 写到 import torch前面,否则多显卡情况有异常
@@ -1 +1 @@
-from .model_config import *
+from .model_config import *
-from .chatglm_llm import *
+from .chatglm_llm import *
@@ -29,7 +29,7 @@ def torch_gc():
torch.cuda.ipc_collect()
-class ChatGLM(LLM):
+class ChatGLM():
max_token: int = 10000
temperature: float = 0.01
top_p = 0.9