doberst commited on
Commit
8fd9b55
·
verified ·
1 Parent(s): 5fcf207

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -7,9 +7,9 @@ inference: false
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
- **slim-combo-sa-ner-3b** is part of the SLIM ("**S**tructured **L**anguage **I**nstruction **M**odel") model series, consisting of small, specialized decoder-based models, fine-tuned for function-calling.
11
 
12
- slim-combo-sa-ner-3b combines two of the most popular traditional classifier functions (Sentiment Analysis and Named Entity Recognition), and reimagines them as function calls on a specialized decoder-based LLM, generating output consisting of a python dictionary with keys corresponding to sentiment, and NER identifiers, such as people, organization, and place, e.g.:
13
 
14
  &nbsp;&nbsp;&nbsp;&nbsp;`{'sentiment': ['positive'], 'people': ['...'], 'organization': ['...'], 'place': [''] }`
15
 
@@ -17,7 +17,7 @@ This 'combo' model is designed to illustrate the potential power of using functi
17
 
18
  This model is fine-tuned on top of [**llmware/bling-stable-lm-3b-4e1t-v0**](https://huggingface.co/llmware/bling-stable-lm-3b-4e1t-v0), which in turn, is a fine-tune of stabilityai/stablelm-3b-4elt.
19
 
20
- Each slim model has a 'quantized tool' version, e.g., [**'slim-combo-tool-3b'**](https://huggingface.co/llmware/slim-combo-tool-3b).
21
 
22
 
23
  ## Prompt format:
@@ -31,8 +31,8 @@ Each slim model has a 'quantized tool' version, e.g., [**'slim-combo-tool-3b'**
31
  <details>
32
  <summary>Transformers Script </summary>
33
 
34
- model = AutoModelForCausalLM.from_pretrained("llmware/slim-combo-sa-ner-3b")
35
- tokenizer = AutoTokenizer.from_pretrained("llmware/slim-combo-sa-ner-3b")
36
 
37
  function = "classify"
38
  params = "topic"
@@ -74,7 +74,7 @@ Each slim model has a 'quantized tool' version, e.g., [**'slim-combo-tool-3b'**
74
  <summary>Using as Function Call in LLMWare</summary>
75
 
76
  from llmware.models import ModelCatalog
77
- slim_model = ModelCatalog().load_model("llmware/slim-combo-sa-ner-3b")
78
  response = slim_model.function_call(text,params=["sentiment", "people", "organization", "place"], function="classify")
79
 
80
  print("llmware - llm_response: ", response)
 
7
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
 
10
+ **slim-sa-ner-3b** is part of the SLIM ("**S**tructured **L**anguage **I**nstruction **M**odel") model series, consisting of small, specialized decoder-based models, fine-tuned for function-calling.
11
 
12
+ slim-sa-ner-3b combines two of the most popular traditional classifier functions (Sentiment Analysis and Named Entity Recognition), and reimagines them as function calls on a specialized decoder-based LLM, generating output consisting of a python dictionary with keys corresponding to sentiment, and NER identifiers, such as people, organization, and place, e.g.:
13
 
14
  &nbsp;&nbsp;&nbsp;&nbsp;`{'sentiment': ['positive'], 'people': ['...'], 'organization': ['...'], 'place': [''] }`
15
 
 
17
 
18
  This model is fine-tuned on top of [**llmware/bling-stable-lm-3b-4e1t-v0**](https://huggingface.co/llmware/bling-stable-lm-3b-4e1t-v0), which in turn, is a fine-tune of stabilityai/stablelm-3b-4elt.
19
 
20
+ Each slim model has a 'quantized tool' version, e.g., [**'slim-sa-ner-3b-tool'**](https://huggingface.co/llmware/slim-sa-ner-3b-tool).
21
 
22
 
23
  ## Prompt format:
 
31
  <details>
32
  <summary>Transformers Script </summary>
33
 
34
+ model = AutoModelForCausalLM.from_pretrained("llmware/slim-sa-ner-3b")
35
+ tokenizer = AutoTokenizer.from_pretrained("llmware/slim-sa-ner-3b")
36
 
37
  function = "classify"
38
  params = "topic"
 
74
  <summary>Using as Function Call in LLMWare</summary>
75
 
76
  from llmware.models import ModelCatalog
77
+ slim_model = ModelCatalog().load_model("llmware/slim-sa-ner-3b")
78
  response = slim_model.function_call(text,params=["sentiment", "people", "organization", "place"], function="classify")
79
 
80
  print("llmware - llm_response: ", response)