debug
Browse files
data/models/llama3-1-70b.py
CHANGED
|
@@ -9,8 +9,7 @@ app = Flask(__name__)
|
|
| 9 |
|
| 10 |
# Load the API key and initialize the InferenceClient
|
| 11 |
api_key = os.getenv("ai")
|
| 12 |
-
|
| 13 |
-
client = InferenceClient("meta-llama/Meta-Llama-3.1-70B-Instruct")
|
| 14 |
|
| 15 |
@app.route('/chat', methods=['POST'])
|
| 16 |
@spaces.GPU()
|
|
|
|
| 9 |
|
| 10 |
# Load the API key and initialize the InferenceClient
|
| 11 |
api_key = os.getenv("ai")
|
| 12 |
+
client = InferenceClient("meta-llama/Meta-Llama-3.1-70B-Instruct",token=api_key)
|
|
|
|
| 13 |
|
| 14 |
@app.route('/chat', methods=['POST'])
|
| 15 |
@spaces.GPU()
|