GitHub Actions commited on
Commit
96f3e65
·
1 Parent(s): 099bee1

Sync from GitHub Actions

Browse files
Files changed (3) hide show
  1. .gitignore +1 -1
  2. Lawverse/pipeline/llm_loader.py +21 -20
  3. requirements.txt +0 -0
.gitignore CHANGED
@@ -152,7 +152,7 @@ venv.bak/
152
  .mypy_cache/
153
  .dmypy.json
154
  dmypy.json
155
-
156
  # Pyre type checker
157
  .pyre/
158
 
 
152
  .mypy_cache/
153
  .dmypy.json
154
  dmypy.json
155
+ Demo.py
156
  # Pyre type checker
157
  .pyre/
158
 
Lawverse/pipeline/llm_loader.py CHANGED
@@ -2,32 +2,33 @@ from typing import Optional, List, Any
2
  from langchain_core.language_models.llms import LLM
3
  from openai import OpenAI
4
  import os
 
5
 
6
  # env = environ.Env()
7
  # environ.Env.read_env(Path(__file__).resolve().parent.parent.parent / ".env")
8
 
9
- class LightningLLM(LLM):
10
- def __init__(self, api_key: str, model: str, **kwargs):
11
- super().__init__(**kwargs)
12
- object.__setattr__(self, "client", OpenAI(base_url="https://lightning.ai/api/v1/", api_key=api_key))
13
- object.__setattr__(self, "model", model)
14
 
15
- def _call(self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional = None, **kwargs: Any) -> str:
16
 
17
- resp = self.client.chat.completions.create(
18
- model=self.model,
19
- messages=[{"role": "user", "content": [{"type": "text", "text": prompt}]}],
20
- temperature=0.1,
21
- max_tokens=10240
22
- )
23
- return resp.choices[0].message.content
24
 
25
- @property
26
- def _identifying_params(self):
27
- return {"model": self.model}
28
 
29
- @property
30
- def _llm_type(self):
31
- return "lightning_gpt"
32
 
33
- llm = LightningLLM(api_key=os.getenv("API_KEY"), model=os.getenv("MODEL"))
 
2
  from langchain_core.language_models.llms import LLM
3
  from openai import OpenAI
4
  import os
5
+ from langchain_openai import ChatOpenAI
6
 
7
  # env = environ.Env()
8
  # environ.Env.read_env(Path(__file__).resolve().parent.parent.parent / ".env")
9
 
10
+ # class LightningLLM(LLM):
11
+ # def __init__(self, api_key: str, model: str, **kwargs):
12
+ # super().__init__(**kwargs)
13
+ # object.__setattr__(self, "client", OpenAI(base_url="https://lightning.ai/api/v1/", api_key=api_key))
14
+ # object.__setattr__(self, "model", model)
15
 
16
+ # def _call(self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional = None, **kwargs: Any) -> str:
17
 
18
+ # resp = self.client.chat.completions.create(
19
+ # model=self.model,
20
+ # messages=[{"role": "user", "content": [{"type": "text", "text": prompt}]}],
21
+ # temperature=0.1,
22
+ # max_tokens=10240
23
+ # )
24
+ # return resp.choices[0].message.content
25
 
26
+ # @property
27
+ # def _identifying_params(self):
28
+ # return {"model": self.model}
29
 
30
+ # @property
31
+ # def _llm_type(self):
32
+ # return "lightning_gpt"
33
 
34
+ llm = ChatOpenAI(base_url='https://lightning.ai/api/v1/', api_key=os.getenv("API_KEY"), model=os.getenv("MODEL"))
requirements.txt CHANGED
Binary files a/requirements.txt and b/requirements.txt differ