| import os | |
| from huggingface_hub import InferenceClient | |
| client = InferenceClient( | |
| provider="featherless-ai", | |
| api_key=os.environ["HF_TOKEN"], | |
| ) | |
| result = client.text_generation( | |
| "Can you please let us know more details about your ", | |
| model="meta-llama/Meta-Llama-3-8B", | |
| ) |