{ "name": "llama3-api", "version": "1.0.0", "description": "API d'inférence pour Llama 3", "main": "index.js", "type": "module", "scripts": { "start": "node index.js" }, "dependencies": { "@huggingface/inference": "^2.6.4", "cors": "^2.8.5", "express": "^4.19.2", "express-rate-limit": "^7.3.1" } }