repo
stringclasses
147 values
number
int64
1
172k
title
stringlengths
2
476
body
stringlengths
0
5k
url
stringlengths
39
70
state
stringclasses
2 values
labels
listlengths
0
9
created_at
timestamp[ns, tz=UTC]date
2017-01-18 18:50:08
2026-01-06 07:33:18
updated_at
timestamp[ns, tz=UTC]date
2017-01-18 19:20:07
2026-01-06 08:03:39
comments
int64
0
58
user
stringlengths
2
28
huggingface/transformers
33,489
passing past_key_values as a tuple is deprecated, but unclear how to resolve
### System Info Copy-and-paste the text below in your GitHub issue and FILL OUT the two last points. - `transformers` version: 4.44.2 - Platform: Linux-5.4.0-167-generic-x86_64-with-glibc2.35 - Python version: 3.10.12 - Huggingface_hub version: 0.24.7 - Safetensors version: 0.4.5 - Accelerate version: 0.34.2 ...
https://github.com/huggingface/transformers/issues/33489
closed
[ "bug" ]
2024-09-14T13:58:18Z
2025-11-29T04:50:43Z
null
RonanKMcGovern
pytorch/PiPPy
1,142
How to train a model with pippy
It seems that the examples here are all examples of inference, where are the examples of training?
https://github.com/pytorch/PiPPy/issues/1142
open
[]
2024-09-14T09:27:38Z
2024-11-20T07:18:01Z
null
sunkun1997
pytorch/data
1,317
StatefulDataloader is slower than Dataloader. Is there any best practice of StatefulDataloader?
### 📚 The doc issue Hello, Thank you for your awesome implementation of StatefulDataloader. I use the compare the speed of Dataloader and StatefulDataloader, the StatefulDataloader is much slower than Dataloader. For example, Dataloader costs 10ms per iter, but StatefulDataloader costs about 2s per iter. Is ...
https://github.com/meta-pytorch/data/issues/1317
closed
[]
2024-09-13T09:43:45Z
2024-09-13T09:50:08Z
1
by2101
pytorch/torchtitan
577
DDP (replicate) + TP?
Currently, when there are two device meshes (`tp` and `dp`), torchtitan should choose FSDP as the **only** backend for DP. Ref: https://github.com/pytorch/torchtitan/blob/d2a4904f58accc683c17c66a360026cb3c8109af/torchtitan/parallelisms/parallelize_llama.py#L97-L98 However, the `replicate` should support >1D mesh an...
https://github.com/pytorch/torchtitan/issues/577
closed
[ "question" ]
2024-09-13T08:10:05Z
2025-03-19T21:22:12Z
null
yzs981130
pytorch/xla
8,000
[RFC] `torch_xla` Backward Compatibility Proposal
Recently, we have started the process to reduce the torch_xla API footprint in favor of torch API to improve the usability. This RFC focuses on the process to deprecate any functions. ## Backward compatibility We propose to offer a 6 months (2 releases) grace period before completely removing the deprecated API. As...
https://github.com/pytorch/xla/issues/8000
open
[ "documentation", "2.5 release" ]
2024-09-12T20:58:55Z
2025-07-11T17:38:19Z
4
zpcore
huggingface/lerobot
436
Image storage format
I am quite interested in using `LeRobotDataset` for large scale training. I am interested to get more context on the options for storing images so I am aware of the implications this might have: - Did you by chance study if the mp4 video compression has any negative effects on the image quality in terms of model perfo...
https://github.com/huggingface/lerobot/issues/436
closed
[ "question", "dataset", "stale" ]
2024-09-12T16:38:21Z
2025-10-23T02:29:14Z
null
nikonikolov
huggingface/lerobot
435
Open-X datasets
Thanks for the great work! I am interested in converting more of the open-x datasets to `LeRobotDataset`. - I was wondering if there was any particular reason the entire open-x wasn't added already, e.g. some difficulties you encountered with some specific datasets? - Do you have any tips where I should be extra care...
https://github.com/huggingface/lerobot/issues/435
closed
[ "enhancement", "question", "dataset" ]
2024-09-12T16:29:40Z
2025-10-08T08:25:55Z
null
nikonikolov
huggingface/lerobot
432
some questions about real world env
### System Info ```Shell all software cfg match author's project ``` ### Information - [ ] One of the scripts in the examples/ folder of LeRobot - [X] My own task or dataset (give details below) ### Reproduction I am planning to control my own robot left-arm. I've almost figure out all the parts if lerobot-datase...
https://github.com/huggingface/lerobot/issues/432
closed
[ "question" ]
2024-09-12T09:53:23Z
2025-10-08T08:27:48Z
null
NNsauce
huggingface/chat-ui
1,463
Some bugs
## Bug description There are several issues that I have with the site, such as slow performance both on mobile and PC. When trying to select specific parts of the text, it goes back to the original message. Sometimes it occurs in errors that force me to always refresh the conversation. When I switch conversation I h...
https://github.com/huggingface/chat-ui/issues/1463
open
[ "bug" ]
2024-09-12T08:13:35Z
2024-09-12T09:03:58Z
0
Ruyeex
huggingface/transformers.js
929
what is pipeline?
https://github.com/huggingface/transformers.js/issues/929
closed
[ "question" ]
2024-09-12T05:09:05Z
2024-10-04T10:24:42Z
null
chakravarthi-vatala
pytorch/torchchat
1,134
Failures when using PyTorch local build vs. binaries
### 🐛 Describe the bug I ran into an issue with loading the tokenizer, which was root caused to me using my local PyTorch build. After building the aoti runner, I ran the following command: `cmake-out/aoti_run exportedModels/stories15M.so -z /home/angelayi/.torchchat/model-cache/stories15M/tokenizer.model -i "Once...
https://github.com/pytorch/torchchat/issues/1134
open
[ "bug", "enhancement" ]
2024-09-11T23:57:18Z
2024-09-12T01:01:24Z
0
angelayi
huggingface/diffusers
9,417
Suggestion for speeding up `index_for_timestep` by removing sequential `nonzero()` calls in samplers
**Is your feature request related to a problem? Please describe.** First off, thanks for the great codebase and providing so many resources! I just wanted to provide some insight into an improvement I made for myself, in case you'd like to include it for all samplers. I'm using the `FlowMatchEulerDiscreteScheduler` an...
https://github.com/huggingface/diffusers/issues/9417
open
[ "help wanted", "wip", "contributions-welcome", "performance" ]
2024-09-11T14:54:37Z
2025-02-08T10:26:47Z
11
ethanweber
huggingface/cosmopedia
29
What is the best way to cite the work?
This is absolutely fantastic work. Thank you very much for making it public. What is the best way to cite this dataset/project? Is there any paper I can cite or should I cite the blog-post?
https://github.com/huggingface/cosmopedia/issues/29
closed
[]
2024-09-11T14:34:54Z
2024-09-11T14:36:15Z
null
vijetadeshpande
huggingface/diffusers
9,416
[Schedulers] Add SGMUniform
Thanks to @rollingcookies, we can see in this [issue](https://github.com/huggingface/diffusers/issues/9397) that this schedulers works great with the Hyper and probably also Lighting loras/unets. It'd be fantastic if someone can contribute this scheduler to diffusers. Please let me know if someone is willing to ...
https://github.com/huggingface/diffusers/issues/9416
closed
[ "help wanted", "contributions-welcome", "advanced" ]
2024-09-11T13:59:27Z
2024-09-23T23:39:56Z
12
asomoza
huggingface/transformers
33,416
The examples in the examples directory are mostly for fine-tuning pre-trained models?how to trian from scratch
### Model description no ### Open source status - [X] The model implementation is available - [X] The model weights are available ### Provide useful links for the implementation _No response_
https://github.com/huggingface/transformers/issues/33416
open
[ "New model" ]
2024-09-11T03:32:53Z
2024-10-03T23:28:42Z
null
zc-Chao
pytorch/pytorch
135,645
[ONNX] How to export the FlashAttention kernel
### 🐛 Describe the bug 1. code ``` import sys import torch from modeling_intern_vit import FlashAttention # FlashAttention of InternVL2-2B model sys.path.append("/home/InternVL2-2B") qkv=torch.load("/home/qkv.pth") falsh=FlashAttention().eval().cuda() out=falsh(qkv.cuda()) ...
https://github.com/pytorch/pytorch/issues/135645
closed
[ "module: onnx", "triaged", "onnx-triaged" ]
2024-09-11T01:40:30Z
2024-09-27T01:46:09Z
null
scuizhibin
huggingface/diffusers
9,407
callback / cannot yield intermediate images on the fly during inference
Hi, in advance apologies if this has been asked already, or if I'm just misusing the diffusers API. Using `diffusers==0.30.2` **What API design would you like to have changed or added to the library? Why?** I will illustrate straight away the general issue with my use case: I need to call a (FLUX) diffuser...
https://github.com/huggingface/diffusers/issues/9407
closed
[]
2024-09-10T16:32:04Z
2024-09-25T12:28:20Z
8
Clement-Lelievre
huggingface/transformers.js
928
The inference speed on the mobile end is a bit slow
### Question If it is a mobile device that does not support WebGPU, how can we improve the inference speed of the model? I have tried WebWorker, but the results were not satisfactory
https://github.com/huggingface/transformers.js/issues/928
open
[ "question" ]
2024-09-10T09:14:16Z
2024-09-11T08:46:33Z
null
Gratifyyy
pytorch/tutorials
3,050
Improve example by adding missing import
The [example "Creating a Custom Dataset for your files"](https://github.com/pytorch/tutorials/blob/8a8331eb2796c05113c8a98bc03a7a164407fcbf/beginner_source/basics/data_tutorial.py#L123) is missing the import `from torch.utils.data import Dataset`. Since other imports are shown and the purpose of this example is to show...
https://github.com/pytorch/tutorials/issues/3050
closed
[]
2024-09-10T09:04:01Z
2025-04-14T18:43:31Z
0
avitase
pytorch/xla
7,987
Speeding up computation while using SPMD on large TPU pod
## ❓ Questions and Help When running on vp-128 TPU pod (even when sharding only by batch dimension) we are experiencing very low performance comparing to the same pod without SPMD. Do you have any tips how to increase the performance? some SPMD arguments? things we need to think about when using it? anything that m...
https://github.com/pytorch/xla/issues/7987
closed
[ "question", "performance" ]
2024-09-10T07:59:14Z
2025-03-31T15:57:15Z
null
dudulightricks
huggingface/transformers.js
927
Error with Using require for ES Modules in @xenova/transformers Package
### Question trying to use require to import the Pipeline class from the @xenova/transformers package, but encounter the following error: const { Pipeline } = require('@xenova/transformers'); ^ Error [ERR_REQUIRE_ESM]: require() of ES Module D:\Z-charity\dating_app_backend\node_modules@xenova\transformers\src\t...
https://github.com/huggingface/transformers.js/issues/927
closed
[ "question" ]
2024-09-10T06:02:53Z
2024-12-08T19:17:31Z
null
qamarali205
huggingface/transformers.js
925
V3 - WebGPU Whisper in Chrome Extention
### Question Can [webGPU accelerated whisper](https://huggingface.co/spaces/Xenova/whisper-webgpu) run in a chrome extension? I checked the space and found the dependency `"@xenova/transformers": "github:xenova/transformers.js#v3"` which I imported in a chrome extension. When I tried to import it, it didn't work. ...
https://github.com/huggingface/transformers.js/issues/925
open
[ "question" ]
2024-09-10T02:52:41Z
2025-01-18T16:03:26Z
null
chandeldivyam
huggingface/diffusers
9,402
[Flux ControlNet] Add img2img and inpaint pipelines
We recently added img2img and inpainting pipelines for Flux thanks to @Gothos contribution. We also have controlnet support for Flux thanks to @wangqixun. It'd be nice to have controlnet versions of these pipelines since there's been requests to have them. Basically, we need to create two new pipelines that a...
https://github.com/huggingface/diffusers/issues/9402
closed
[ "help wanted", "Good second issue", "contributions-welcome" ]
2024-09-10T02:08:32Z
2024-10-25T02:22:19Z
11
asomoza
huggingface/transformers.js
924
Steps for suppressing strings
### Question What is the syntax for suppressing strings from showing up in the output text? Should I be doing that in my code, or is there a config option for it? I'm trying to remove everything that isn't a word: ``` const suppressedStrings = [ "[BLANK_AUDIO]", "[CLEARS THROAT]", "[Coughing]", "[inaudib...
https://github.com/huggingface/transformers.js/issues/924
open
[ "question" ]
2024-09-09T21:44:16Z
2025-01-24T17:53:47Z
null
stinoga
huggingface/diffusers
9,395
[Q] Possibly unused `self.final_alpha_cumprod`
Hello team, quick question to make sure I understand the behavior of the `step` function in LCM Scheduler. https://github.com/huggingface/diffusers/blob/a7361dccdc581147620bbd74a6d295cd92daf616/src/diffusers/schedulers/scheduling_lcm.py#L534-L543 Here, it seems that the condition `prev_timestep >= 0` is always `T...
https://github.com/huggingface/diffusers/issues/9395
open
[ "stale" ]
2024-09-09T17:35:08Z
2024-11-09T15:03:23Z
7
fdtomasi
huggingface/chat-ui
1,458
Chat ui sends message prompt 404
``` MONGODB_URL='mongodb://localhost:27017' PLAYWRIGHT_ADBLOCKER='false' MODELS=`[ { "name": "Local minicpm", "tokenizer": "minicpm", "preprompt": "", "chatPromptTemplate": "<s>{{preprompt}}{{#each messages}}{{#ifUser}}<|user|>\n{{content}}<|end|>\n<|assistant|>\n{{/ifUser}}{{#ifAssistant}}{{c...
https://github.com/huggingface/chat-ui/issues/1458
open
[ "support" ]
2024-09-09T13:31:56Z
2024-09-13T09:32:24Z
2
nextdoorUncleLiu
huggingface/chat-ui
1,456
could you provide an easy way to force output as json?
current I use preprompt:'only output json. Do not output anything that is not json. Do not use markdown format. Must begin with {.' But llama is not smart enough to output json form. It always begin with Here is the JSON answer or begin with ```(markdown format) for give me unvalid json string. It seems prepr...
https://github.com/huggingface/chat-ui/issues/1456
open
[ "enhancement" ]
2024-09-09T11:34:17Z
2024-10-06T18:35:29Z
1
ghost
pytorch/torchtitan
572
How to calculate the total batchsize
Hi, it is me again~ I have a quick simple question: I am using the following training config with 4 GPUs. What is the total number of tokens per optimizer step? Is it 2 * 2048 or 2 * 2048 * 4? ``` [training] batch_size = 2 seq_len = 2048 warmup_steps = 2000 # lr scheduler warm up, normally 20% of the train ste...
https://github.com/pytorch/torchtitan/issues/572
closed
[ "question" ]
2024-09-09T09:47:50Z
2024-09-10T05:43:47Z
null
zyushun
huggingface/diffusers
9,392
[Scheduler] Add SNR shift following SD3, would the rest of the code need to be modified?
**What API design would you like to have changed or added to the library? Why?** With the increasing resolution of image or video generation, we need to introduce more noise at smaller T, such as SNR shift following SD3. I have observed that CogVideoX's schedule has already implemented [this](https://github.com/hugg...
https://github.com/huggingface/diffusers/issues/9392
open
[ "stale" ]
2024-09-09T09:19:37Z
2025-01-05T15:05:04Z
7
LinB203
huggingface/speech-to-speech
96
How to designate Melo TTS model to use my trained model?
Hi, I am using Melo as TTS. And I trained with my datasets. How to designate Melo (here at speech to speech) to use my model? Thanks!
https://github.com/huggingface/speech-to-speech/issues/96
closed
[]
2024-09-08T20:36:23Z
2024-09-10T14:42:58Z
null
insufficient-will
huggingface/huggingface_hub
2,526
How can I rename folders in given repo? I need to rename folders
### Describe the bug I am try to rename like below but it fails :/ ``` from huggingface_hub import HfApi import os # Initialize the Hugging Face API api = HfApi() # Set the repository name repo_name = "MonsterMMORPG/3D-Cartoon-Style-FLUX" # Define the folder renaming mappings folder_renames = { ...
https://github.com/huggingface/huggingface_hub/issues/2526
closed
[ "bug" ]
2024-09-07T17:23:54Z
2024-09-09T10:49:26Z
null
FurkanGozukara
pytorch/xla
7,972
Registering CUDA custom calls with the C++ FFI
## ❓ Questions and Help Curious how to build and register a CUDA custom call with XLAC - have followed https://jax.readthedocs.io/en/latest/ffi.html and read https://openxla.org/xla/custom_call and wondering what the equivalent process is for torch / whether it is currently supported.
https://github.com/pytorch/xla/issues/7972
open
[ "question" ]
2024-09-07T01:27:35Z
2025-03-31T16:08:31Z
null
skrider
huggingface/transformers
33,359
[Docs] How to build offline HTML or Docset files for other documentation viewers?
### Feature request How can I build the docs into HTML files for use with other documentation viewers like [Dash](https://www.kapeli.com/dash) , [Dash-User-Contributions](https://github.com/Kapeli/Dash-User-Contributions)? I successfully built the PyTorch docs for Dash by working directly in their `docs/` directory...
https://github.com/huggingface/transformers/issues/33359
closed
[ "Documentation", "Feature request" ]
2024-09-06T15:51:35Z
2024-09-10T23:43:57Z
null
ueoo
huggingface/transformers
33,343
How to install transformers==4.45, two or three days I can install successfully, but today cannot.
### System Info torch2.2 ### Who can help? _No response_ ### Information - [ ] The official example scripts - [ ] My own modified scripts ### Tasks - [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...) - [ ] My own task or dataset (give details below) ### Reproduction pip instal...
https://github.com/huggingface/transformers/issues/33343
closed
[ "Installation", "bug" ]
2024-09-06T08:23:00Z
2024-10-16T08:04:10Z
null
HyacinthJingjing
pytorch/torchchat
1,114
What is the future plan of this torchchat project?
### 🐛 Describe the bug Torchchat provides a solution of running LLM with PyTorch optimization on servers, desktop and mobile. May I know what is the future plan of this project? Is there any new features to finish to encourage users to use Torchchat as a solution?
https://github.com/pytorch/torchchat/issues/1114
closed
[]
2024-09-06T06:03:17Z
2024-09-09T15:40:39Z
null
yanbing-j
huggingface/optimum-nvidia
149
How to use TensorRT model converter
Referring to [src/optimum/nvidia/export/converter.py] -> class 'TensorRTModelConverter' this could 'Take a local model and create the TRTLLM checkpoint and engine' Questions: - What are applicable local model format? e.g. JAX, HuggingFace, DeepSpeed - How to use this script individually to generate TRTLLM checkpoint...
https://github.com/huggingface/optimum-nvidia/issues/149
open
[]
2024-09-05T18:55:15Z
2024-09-05T18:55:15Z
null
FortunaZhang
huggingface/datasets
7,139
Use load_dataset to load imagenet-1K But find a empty dataset
### Describe the bug ```python def get_dataset(data_path, train_folder="train", val_folder="val"): traindir = os.path.join(data_path, train_folder) valdir = os.path.join(data_path, val_folder) def transform_val_examples(examples): transform = Compose([ Resize(256), ...
https://github.com/huggingface/datasets/issues/7139
open
[]
2024-09-05T15:12:22Z
2024-10-09T04:02:41Z
2
fscdc
huggingface/datasets
7,138
Cache only changed columns?
### Feature request Cache only the actual changes to the dataset i.e. changed columns. ### Motivation I realized that caching actually saves the complete dataset again. This is especially problematic for image datasets if one wants to only change another column e.g. some metadata and then has to save 5 TB again. #...
https://github.com/huggingface/datasets/issues/7138
open
[ "enhancement" ]
2024-09-05T12:56:47Z
2024-09-20T13:27:20Z
2
Modexus
huggingface/lerobot
413
Compatible off-the-shelf robots?
Huge thanks for making all of this available! Can you recommend any (low-cost) off-the-shelf robots to work with?
https://github.com/huggingface/lerobot/issues/413
closed
[ "question" ]
2024-09-05T10:21:24Z
2025-10-08T08:27:56Z
null
danielfriis
huggingface/diffusers
9,362
IndexError: index 29 is out of bounds for dimension 0 with size 29
### Describe the bug I have three problems because of the same reason. 1) TypeError: unsupported operand type(s) for +=: 'NoneType' and 'int' # upon completion increase step index by one self._step_index += 1 <---Error [here](https://github.com/huggingface/diffusers/blob/main/src/diffusers/schedu...
https://github.com/huggingface/diffusers/issues/9362
open
[ "bug", "stale" ]
2024-09-04T11:02:49Z
2024-11-25T15:04:22Z
8
Anvarka
pytorch/pytorch
135,098
How to gracefully mask CompositeImplicitAutograd for different backends
### 🐛 Describe the bug I implemented torch.compile’s backend for my hardware via privateUserOne. I also found that torch.compile by default decomposes upsample_nearest2d into a bunch of small operators, just like _upsample_nearest does. But on my hardware, the _unsafe_index operator doesn’t perform well, so I’d like ...
https://github.com/pytorch/pytorch/issues/135098
closed
[ "oncall: pt2", "oncall: export" ]
2024-09-04T09:11:28Z
2024-11-01T06:20:49Z
null
yangxiaorun
huggingface/tokenizers
1,627
Rust: How to handle models with `precompiled_charsmap = null`
Hi guys, I'm currently working on https://github.com/supabase/edge-runtime/pull/368 that pretends to add a rust implementation of `pipeline()`. While I was coding the `translation` task I figured out that I can't load the `Tokenizer` instance for [Xenova/opus-mt-en-fr](https://huggingface.co/Xenova/opus-mt-en-fr) ...
https://github.com/huggingface/tokenizers/issues/1627
open
[ "Feature Request" ]
2024-09-04T08:33:06Z
2024-10-06T15:34:06Z
null
kallebysantos
huggingface/optimum
2,013
Is it possible convert decoder_model_merged.onnx to tensorrt via trtexec command ?
At the first I convert whisper-tiny to onnx via optimum-cli `optimum-cli export onnx --model openai/whisper-tiny --task automatic-speech-recognition-with-past whisper-tiny-onnx` I got the some config, encoder and decoder_merged model then I brought encoder and decoder_merged to convert to tensorrt via NGC versio...
https://github.com/huggingface/optimum/issues/2013
closed
[]
2024-09-03T17:52:40Z
2024-09-15T10:16:34Z
3
ccyrene
huggingface/lerobot
407
Multi-Image support for VQ-BeT
Hello, I wanted to ask if there is a possibility to have VQ-BeT running on multiple camera's for some environments that have different views, like Robomimic? If so can someone give me points on what exactly I need to change, I would be happy to submit a PR once I get it working on my side and finish the ICLR deadline! ...
https://github.com/huggingface/lerobot/issues/407
closed
[ "question", "policies" ]
2024-09-03T17:00:23Z
2025-10-08T08:27:39Z
null
bkpcoding
pytorch/vision
8,626
Better decoder docs
Our decoding docs are poor, disorganized, and don't have any example. We should improve those to clarify what is supported, how, and encourage users to rely on those.
https://github.com/pytorch/vision/issues/8626
closed
[]
2024-09-03T14:47:11Z
2024-10-01T12:19:14Z
0
NicolasHug
huggingface/optimum
2,009
[Feature request] Add kwargs or additional options for torch.onnx.export
### Feature request In `optimum.exporters.onnx.convert import export_pytorch`, there could be an option to add additional kwargs to the function which could be passed to the torch.onnx.export function. ### Motivation If such an option possible or will this ruin any of the other features, or is there a reason why the...
https://github.com/huggingface/optimum/issues/2009
open
[ "onnx" ]
2024-09-03T13:52:50Z
2024-10-08T15:27:26Z
0
martinkorelic
huggingface/speech-to-speech
74
How to integrate it with frontend
Hi, What steps should I follow to create a web app UI and integrate it? Many thanks for considering my request.
https://github.com/huggingface/speech-to-speech/issues/74
open
[]
2024-09-03T12:18:52Z
2024-09-03T13:52:08Z
null
shrinivasait
huggingface/diffusers
9,356
pipeline_stable_diffusion_xl_adapter
### Describe the bug I want to rewrite the call function of the pipeline_stable_diffusion_xl_adapter. When I want to use the function prepare_ip_adapter_image_embeds, there is an error called "AttributeError: 'NoneType' object has no attribute 'image_projection_layers'". The error tells me that the attribution self.un...
https://github.com/huggingface/diffusers/issues/9356
open
[ "bug", "stale" ]
2024-09-03T10:25:57Z
2024-10-28T15:03:18Z
6
Yuhan291
huggingface/diffusers
9,352
Text generation?
Hi thanks for this great library! There seems to be some diffusion models that generate text, instead of images. (For example, these two surveys: https://arxiv.org/abs/2303.06574, https://www.semanticscholar.org/paper/Diffusion-models-in-text-generation%3A-a-survey-Yi-Chen/41941f072db18972b610de9979e755afba35f11e). ...
https://github.com/huggingface/diffusers/issues/9352
open
[ "wip" ]
2024-09-03T06:54:38Z
2024-11-23T04:57:37Z
13
fzyzcjy
huggingface/speech-to-speech
71
How to run in ubuntu
I am trying to run it locally in my Ubuntu machine I have nvidia gpu and already setup CUDA. ``` python s2s_pipeline.py \ --recv_host 0.0.0.0 \ --send_host 0.0.0.0 \ --lm_model_name microsoft/Phi-3-mini-4k-instruct \ --init_chat_role system \ --stt_compile_mode reduce-overhead \ --tts_compile_mode defau...
https://github.com/huggingface/speech-to-speech/issues/71
closed
[]
2024-09-03T06:02:45Z
2024-10-01T07:45:20Z
null
Basal-Analytics
huggingface/optimum
2,006
Support for gemma2-2b-it(gemma 2nd version) Model Export in Optimum for OpenVINO
### Feature request please provide Support for gemma2 Model Export in Optimum for OpenVINO version:optimum(1.21.4) transformers:4.43.4 ### Motivation I encountered an issue while trying to export a gemma2 model using the optimum library for ONNX export. The error message suggests that the gemma2 model is either a...
https://github.com/huggingface/optimum/issues/2006
open
[ "onnx" ]
2024-09-03T05:54:51Z
2025-01-22T15:40:04Z
2
chakka12345677
huggingface/transformers
33,270
Static KV cache status: How to use it? Does it work for all models?
I see that there are many PRs about [StaticCache](https://github.com/huggingface/transformers/pulls?q=is%3Apr+StaticCache), but I couldn't find a clear documentation on how to use it. #### What I want * To not have Transformers allocate memory dynamically for the KV cache when using `model.generate()`, as that le...
https://github.com/huggingface/transformers/issues/33270
closed
[]
2024-09-03T02:17:54Z
2024-11-25T16:17:25Z
null
oobabooga
huggingface/transformers.js
917
Where should I get `decoder_model_merged` file from?
### Question Hey, I'm trying to use `whisper-web` demo with my finetuned model. After I managed connecting my model to the demo application, I'm getting errors related to this: https://github.com/xenova/transformers.js/blob/7f5081da29c3f77ee830269ab801344776e61bcb/src/models.js#L771 Basically, when `transforme...
https://github.com/huggingface/transformers.js/issues/917
closed
[ "question" ]
2024-09-02T07:30:57Z
2025-02-26T12:05:05Z
null
abuchnick-aiola
huggingface/diffusers
9,339
SD3 inpatinting
I found the StableDiffusion3InpaintPipeline, where can i found the weight of SD3 inpainting
https://github.com/huggingface/diffusers/issues/9339
closed
[ "stale" ]
2024-09-02T05:00:19Z
2024-10-02T15:43:24Z
5
ucasyjz
pytorch/torchtitan
566
Multi-node training without AWS EFA clusters
Thank you so much for releasing code for this great project! For multi-node training, right now I've only found commands in `multinode_trainer.slurm`, which seem to be specific to AWS EFA slurm clusters. I'm wondering if it's possible to try multi-node training without ASW setup, say with simply the IPs of 2 node...
https://github.com/pytorch/torchtitan/issues/566
closed
[ "question" ]
2024-08-31T22:41:04Z
2024-09-04T20:55:50Z
null
LeoXinhaoLee
huggingface/transformers
33,232
How to use hugginface for training: google-t5/t5-base
### Feature request How to use hugginface for training / 如何使用huggingface来训练: https://github.com/huggingface/transformers/tree/main/examples/pytorch/translation #What is the format and how do I write it? / 这个格式是怎么样的,怎么写呢? def batch_collator(data): print(data) #?????????????????????????????????????????????...
https://github.com/huggingface/transformers/issues/33232
open
[ "Usage", "Feature request" ]
2024-08-31T07:41:18Z
2024-09-09T08:45:50Z
null
gg22mm
pytorch/pytorch
134,901
How to calculate second derivative using PyTorch with GPU (cuda)
### 🚀 The feature, motivation and pitch I have a python code segment related to a deep RL algorithm where it calculates the second order optimization and second derivative with Hessian matrix and fisher information matrix. Normally I run the whole code on GPU (cuda), but since I got a computational issue to calcula...
https://github.com/pytorch/pytorch/issues/134901
open
[ "module: double backwards", "module: cudnn", "module: autograd", "module: rnn", "triaged", "module: functorch" ]
2024-08-31T04:01:40Z
2024-09-04T01:48:21Z
null
Damika-Anupama
huggingface/transformers
33,228
How to obtain batch index of validation dataset?
Hi, I wanted to know how would we fetch the batch id/index of the eval dataset in ```preprocess_logits_for_metrics()``` ? Thanks in advance!
https://github.com/huggingface/transformers/issues/33228
closed
[ "Usage" ]
2024-08-31T00:11:13Z
2024-10-13T08:04:26Z
null
SoumiDas
huggingface/transformers
33,210
The model's address is https://huggingface.co/Xenova/nllb-200-distilled-600M/tree/main/onnx。I don't know how to load encode.onnx and decoder.onnx, and successfully translate a sentence into another language. Can you help me write an inference code to achieve the translation effect through the encoder and decoder? thank...
### Feature request hello,The model's address is [https://huggingface.co/Xenova/nllb-200-distilled-600M/tree/main/onnx。I](https://huggingface.co/Xenova/nllb-200-distilled-600M/tree/main/onnx%E3%80%82I) don't know how to load encode.onnx and decoder.onnx, and successfully translate a sentence into another language. Can...
https://github.com/huggingface/transformers/issues/33210
open
[ "Feature request" ]
2024-08-30T09:33:01Z
2024-10-22T07:18:15Z
null
pengpengtao
huggingface/dataset-viewer
3,054
Image URL detection
[`is_image_url`](https://github.com/huggingface/dataset-viewer/blob/946b0788fa426007161f2077a70b5ae64b211cf8/libs/libcommon/src/libcommon/utils.py#L131-L134) relies on a filename and extension being present, however, in some cases an image URL does not contain a filename. Example [dataset](https://huggingface.co/datase...
https://github.com/huggingface/dataset-viewer/issues/3054
open
[ "question", "improvement / optimization", "P2" ]
2024-08-29T23:17:55Z
2025-07-04T09:37:23Z
null
hlky
huggingface/transformers.js
911
Next.js example breaks with v3
### Question Are there steps documented anywhere for running V3 in your app? I'm trying to test it out via these steps: 1. Pointing to the alpha in my `package.json`: `"@huggingface/transformers": "^3.0.0-alpha.10",` 2. `npm i` 3. `cd node_modules/@hugginface/transformers && npm i` 4. copy the [webpack.config.js...
https://github.com/huggingface/transformers.js/issues/911
closed
[ "question" ]
2024-08-29T20:17:03Z
2025-02-16T12:35:47Z
null
stinoga
pytorch/xla
7,925
Prepare a documentation to explain the use cases for `torch.compile`, `torch_xla.compile`, torch_xla eager mode, torchxla2
## 📚 Documentation Author a documentation to explain the use cases for `torch.compile`, `torch_xla.compile`, torch_xla eager mode, torchxla2. Users and customers look for clarity on the "the utility" of each option, pros/cons, small example to demonstrate correct use. cc @ManfeiBai @JackCaoG @will-cromar @qihqi ...
https://github.com/pytorch/xla/issues/7925
closed
[ "documentation" ]
2024-08-29T17:01:06Z
2024-09-24T18:33:39Z
2
miladm
pytorch/torchtitan
562
Pipeline Parallelism + FSDP
On `PP + FSDP` and `PP + TP + FSDP`: - Is there any documentation on how these different parallelisms compose? - What are the largest training runs these strategies have been tested on? - Are there benchmarks for how these strategies compare against other distributed training frameworks that expose similar paralleli...
https://github.com/pytorch/torchtitan/issues/562
open
[ "enhancement", "question", "module: pipelining" ]
2024-08-29T14:19:58Z
2025-10-30T06:21:51Z
null
jeromeku
huggingface/diffusers
9,317
Finetuning on dataset
dear @thedarkzeno and @patil-suraj Thank you so much for putting your work out there. I wanted to ask, how would the training be for training on a dataset and not a single instance image as mentioned in train_dreambooth_inpaint. And can I finetune models trained from https://github.com/CompVis/latent-diffusion repo...
https://github.com/huggingface/diffusers/issues/9317
closed
[ "stale" ]
2024-08-29T12:20:51Z
2024-10-23T16:10:47Z
4
ultiwinter
pytorch/pytorch
134,760
How to correctly release the memory of a tensor
i have fined the memory increase at this line. [param.copy_(input_param)](https://github.com/pytorch/pytorch/blob/d01a7a9faa5a742a3df7374b97bbc1db1205b6ed/torch/nn/modules/module.py#L2425) but the memory cant be released clean after the module use. what happen in it and how to correctly release the memory of a tenso...
https://github.com/pytorch/pytorch/issues/134760
closed
[ "module: nn", "module: memory usage", "triaged" ]
2024-08-29T11:39:12Z
2024-08-30T08:24:23Z
null
huangqiaobo
huggingface/optimum-quanto
300
How to quantize, save and load Stable Diffusion 3 model.
import torch from optimum.quanto import qint2, qint4, qint8, quantize, freeze from diffusers import StableDiffusion3Pipeline pipe = StableDiffusion3Pipeline.from_pretrained("stabilityai/stable-diffusion-3-medium-diffusers", torch_dtype=torch.bfloat16) quantize(pipe.text_encoder, weights=qint4) freeze(pip...
https://github.com/huggingface/optimum-quanto/issues/300
closed
[ "Stale" ]
2024-08-29T06:24:02Z
2024-10-06T02:06:30Z
null
jainrahul52
huggingface/optimum
2,002
Is it possible to infer the model separately through encoder.onnx and decoder.onnx
### Feature request Is it possible to infer the model separately through encoder.onnx and decoder.onnx ### Motivation Is it possible to infer the model separately through encoder.onnx and decoder.onnx ### Your contribution Is it possible to infer the model separately through encoder.onnx and decoder.onnx
https://github.com/huggingface/optimum/issues/2002
open
[ "onnx" ]
2024-08-29T03:26:20Z
2024-10-08T15:28:59Z
0
pengpengtao
pytorch/TensorRT
3,124
❓ [Question] dynamo conversion failing w/ TRTInterpreter
## ❓ Question im able to `torch.export` and generate an ExportedProgram with no issues for my model. upon compiling with `torch_tensorrt`... ```python ep = torch.export.load("...") example_inputs = ep.example_inputs[0] model = ep.module().to("cuda") compile_spec = { "ir": "torch_compile", "inputs":...
https://github.com/pytorch/TensorRT/issues/3124
open
[ "question" ]
2024-08-28T20:09:48Z
2024-09-06T19:36:58Z
null
patrick-botco
pytorch/tutorials
3,017
💡 [REQUEST] - What is purpose of `out.backward(torch.randn(1, 10))` in neural_networks_tutorial
### 🚀 Describe the improvement or the new tutorial In [neural networks tutorial for beginners](https://pytorch.org/tutorials/beginner/blitz/neural_networks_tutorial.html), we have the following: Zero the gradient buffers of all parameters and backprops with random gradients: ``` net.zero_grad() out.backward(tor...
https://github.com/pytorch/tutorials/issues/3017
open
[ "question", "intro", "core" ]
2024-08-28T14:51:46Z
2025-04-16T18:24:08Z
null
Lovkush-A
huggingface/diffusers
9,303
[Add] VEnhancer - the interpolation and upscaler for CogVideoX-5b
### Model/Pipeline/Scheduler description VEnhancer, a generative space-time enhancement framework that can improve the existing T2V results. https://github.com/Vchitect/VEnhancer ### Open source status - [X] The model implementation is available. - [X] The model weights are available (Only relevant if addition is...
https://github.com/huggingface/diffusers/issues/9303
open
[ "stale" ]
2024-08-28T14:43:32Z
2024-12-11T15:04:32Z
3
tin2tin
huggingface/text-generation-inference
2,466
Guide on how to use TensorRT-LLM Backend
### Feature request Does any documentation exist, or would it be possible to add documentation, on how to use the TensorRT-LLM backend? #2458 makes mention that the TRT-LLM backend exists, and I can see that there's a Dockerfile for TRT-LLM, but I don't see any guides on how to build/use it. ### Motivation I would l...
https://github.com/huggingface/text-generation-inference/issues/2466
open
[]
2024-08-28T13:24:26Z
2025-05-18T16:23:14Z
null
michaelthreet
huggingface/lerobot
390
[Feature Request] Add end effector pos field in lerobot dataset?
Aloha style joint space dataset will limit data set to the specific robot. Can we change joint space data or add a field of end effector to cartesian space data base on the robot URDF file? It may help robotics community build a more generalized policy.
https://github.com/huggingface/lerobot/issues/390
closed
[ "question", "dataset", "robots" ]
2024-08-28T13:19:15Z
2024-08-29T09:55:27Z
null
hilookas
huggingface/datasets
7,129
Inconsistent output in documentation example: `num_classes` not displayed in `ClassLabel` output
In the documentation for [ClassLabel](https://huggingface.co/docs/datasets/v2.21.0/en/package_reference/main_classes#datasets.ClassLabel), there is an example of usage with the following code: ```` from datasets import Features features = Features({'label': ClassLabel(num_classes=3, names=['bad', 'ok', 'good'])}) ...
https://github.com/huggingface/datasets/issues/7129
closed
[]
2024-08-28T12:27:48Z
2024-12-06T11:32:02Z
0
sergiopaniego
huggingface/diffusers
9,299
CUDAGRAPHs for Flux position embeddings
@yiyixuxu Is it possible to refactor the Flux positional embeddings so that we can fully make use of CUDAGRAPHs? ```bash skipping cudagraphs due to skipping cudagraphs due to cpu device (device_put). Found from : File "/home/sayak/diffusers/src/diffusers/models/transformers/transformer_flux.py", line 469,...
https://github.com/huggingface/diffusers/issues/9299
closed
[]
2024-08-28T11:33:16Z
2024-08-29T19:37:17Z
0
sayakpaul
pytorch/pytorch
134,668
Whether tensor parallelism supports the overlap of communication calculations for gradient computation, and how to implement it
### 🚀 The feature, motivation and pitch I want to know How to achieve the overlap of communication calculations when finding the gradient after row cutting/column cutting of the linear layer,thanks The following is https://pytorch.org/docs/2.3/distributed.tensor.parallel.html ### Alternatives _No response_ ### A...
https://github.com/pytorch/pytorch/issues/134668
open
[ "oncall: distributed", "triaged" ]
2024-08-28T11:06:58Z
2024-08-30T17:54:43Z
null
Xingzhi107
huggingface/transformers.js
906
Unsupported model type: jais
### Question ### System Info macOS, node v20.10, @xenova/transformers 2.17.2 ### Environment/Platform - [ ] Website/web-app - [ ] Browser extension - [x] Server-side (e.g., Node.js, Deno, Bun) - [ ] Desktop app (e.g., Electron) - [ ] Other (e.g., VSCode extension) ### Description ``` Error: Unsuppor...
https://github.com/huggingface/transformers.js/issues/906
closed
[ "question" ]
2024-08-28T09:46:17Z
2024-08-28T21:01:10Z
null
SherifElfadaly
huggingface/trl
1,986
how to convert dpodata to ktodata
### Feature request how to convert dpodata to ktodata ### Motivation how to convert dpodata to ktodata ### Your contribution how to convert dpodata to ktodata
https://github.com/huggingface/trl/issues/1986
closed
[]
2024-08-28T06:23:13Z
2024-08-28T09:02:35Z
null
dotsonliu
pytorch/ao
763
How to reduce autoquant compilation time
Autoquant has been popular among the diffusers crowd since its OOB performance has been the best but the main issue is compile times are quite long. There's a few strategies to mitigate this 1. Tune faster: either with better heuristics or a faster tuning core loop 2. Cache things: It's fine if tuning takes a long ti...
https://github.com/pytorch/ao/issues/763
open
[]
2024-08-27T20:49:09Z
2024-08-28T17:36:03Z
null
msaroufim
huggingface/datasets
7,128
Filter Large Dataset Entry by Entry
### Feature request I am not sure if this is a new feature, but I wanted to post this problem here, and hear if others have ways of optimizing and speeding up this process. Let's say I have a really large dataset that I cannot load into memory. At this point, I am only aware of `streaming=True` to load the dataset....
https://github.com/huggingface/datasets/issues/7128
open
[ "enhancement" ]
2024-08-27T20:31:09Z
2024-10-07T23:37:44Z
4
QiyaoWei
huggingface/huggingface_hub
2,491
How to uplaod folders into repo with most effective way - on error continue resume max speed
Hello. I have the below tasks for uploading however I am not sure if they are most effective way of doing #### This cell is used to upload single file into a repo with certain name ``` from huggingface_hub import HfApi api = HfApi() api.upload_file( path_or_fileobj=r"/home/Ubuntu/apps/stable-diffusion-w...
https://github.com/huggingface/huggingface_hub/issues/2491
closed
[ "bug" ]
2024-08-27T16:36:04Z
2024-08-28T08:24:22Z
null
FurkanGozukara
huggingface/Google-Cloud-Containers
73
Download model files from GCS (Instead of HF Hub)
When deploying an HF model to Vertex AI, I would like to download a fine-tuned model from GCS, instead of from HF Hub, like so: ``` model = aiplatform.Model.upload( display_name="my-model", serving_container_image_uri=os.getenv("CONTAINER_URI"), serving_container_environment_variables={ "AIP...
https://github.com/huggingface/Google-Cloud-Containers/issues/73
closed
[ "tei", "question" ]
2024-08-27T12:14:10Z
2024-09-16T07:07:11Z
null
rm-jeremyduplessis
pytorch/ao
750
Question RE AO MX formats
I noticed the [MX readme](https://github.com/pytorch/ao/tree/main/torchao/prototype/mx_formats) has this line: "we match bitwise to other implementations of the OCP MX spec (code not in this repo), with a couple of edge cases left to resolve." Is there a list of edge cases where AO does not match reference implementati...
https://github.com/pytorch/ao/issues/750
closed
[ "question", "mx" ]
2024-08-26T17:37:40Z
2024-08-27T17:15:01Z
null
tsengalb99
huggingface/chat-ui
1,436
MODELS=`[ variable problem when I docker run
Hello, I want to use Ollama to use Mistral model and I followed the documentation below : https://huggingface.co/docs/chat-ui/configuration/models/providers/ollama `deploy.sh` : ```sh #!/bin/bash sudo docker compose down sudo docker rm -f mongodb && sudo docker rm -f chat-ui # nginx and ollama sudo d...
https://github.com/huggingface/chat-ui/issues/1436
closed
[ "support" ]
2024-08-26T14:00:26Z
2024-08-27T11:04:39Z
5
avirgos
huggingface/diffusers
9,276
How can I manually update some of their checkpoints of UNet2/3DConditionModel objects?
### Discussed in https://github.com/huggingface/diffusers/discussions/9273 <div type='discussions-op-text'> <sup>Originally posted by **justin4ai** August 26, 2024</sup> Hello, I'm quite new to diffusers package and trying to implement fine-tuning code that uses the saved checkpoints initialized with ```UNet2/3D...
https://github.com/huggingface/diffusers/issues/9276
open
[ "stale" ]
2024-08-26T07:49:23Z
2024-09-25T15:03:01Z
1
justin4ai
huggingface/transformers
33,115
How to get the score of each token when using pipeline
pipe = pipeline( "text-generation", model=model, tokenizer=tokenizer, max_new_tokens=512, do_sample=True, temperature=0.7, top_p=0.95, top_k=40, repetition_penalty=1.1, output_scores=True ) The model I use is Qwen2-7B-Instruct. When I try to output the score of each...
https://github.com/huggingface/transformers/issues/33115
closed
[ "Usage" ]
2024-08-26T07:00:54Z
2025-03-06T08:23:58Z
null
xin0623
huggingface/diffusers
9,271
The different quality between ComfyUI and Diffusers ?
### Discussed in https://github.com/huggingface/diffusers/discussions/9265 <div type='discussions-op-text'> <sup>Originally posted by **vuongminh1907** August 25, 2024</sup> I had a problem using InstantID (https://github.com/instantX-research/InstantID), which uses Diffusers as its base. Additionally, I tried C...
https://github.com/huggingface/diffusers/issues/9271
closed
[ "stale" ]
2024-08-26T02:53:23Z
2024-10-15T18:10:42Z
3
vuongminh1907
huggingface/diffusers
9,264
Could you make an inpainting model for flux?
### Model/Pipeline/Scheduler description The [stable-diffusion-xl-1.0-inpainting-0.1](https://huggingface.co/diffusers/stable-diffusion-xl-1.0-inpainting-0.1) model helps a lot. Could you make a similar inpainting model for flux? https://huggingface.co/black-forest-labs/FLUX.1-dev ### Open source status - [ ] The...
https://github.com/huggingface/diffusers/issues/9264
closed
[]
2024-08-24T17:32:32Z
2024-08-24T17:37:59Z
2
snowbedding
huggingface/transformers
33,106
how to fine tune TrOCR on specifique langage guide.
### Model description hello , just passed through issues and other , but none of them talked on how to fine-tune TrOCR on specifique langage , like how to pick encoder and decoder , model .. etc , can you @NielsRogge , write a simple instructions/guide on this topic ? ### Open source status - [ ] The model imple...
https://github.com/huggingface/transformers/issues/33106
closed
[]
2024-08-24T14:33:02Z
2025-06-15T08:07:10Z
null
MohamedLahmeri01
pytorch/xla
7,911
Documentation: Discoverability of http://pytorch.org/xla
## 📚 Documentation: Discoverability of http://pytorch.org/xla The docs are very hard to find _despite_ being hosted on [pytorch.org](http://pytorch.org/). If I visit [pytorch.org](http://pytorch.org/) I can't find any link that goes to [pytorch.org/xla](http://pytorch.org/xla). The closest I could find is somewhere...
https://github.com/pytorch/xla/issues/7911
closed
[ "documentation" ]
2024-08-24T02:58:03Z
2024-11-04T17:38:23Z
7
tengyifei
huggingface/datasets
7,123
Make dataset viewer more flexible in displaying metadata alongside images
### Feature request To display images with their associated metadata in the dataset viewer, a `metadata.csv` file is required. In the case of a dataset with multiple subsets, this would require the CSVs to be contained in the same folder as the images since they all need to be named `metadata.csv`. The request is th...
https://github.com/huggingface/datasets/issues/7123
open
[ "enhancement" ]
2024-08-23T22:56:01Z
2024-10-17T09:13:47Z
3
egrace479
pytorch/vision
8,608
loss_box_reg increasing while training mask rcnn
### 🐛 Describe the bug I am trying to train maskRcnn model using detectron2 on my custom LVO deteset. My dataset is a single class dataset and some of the image have no annotation in it. The architecture need to learn negative examples as well for proper training as the test data contains both positive and negative...
https://github.com/pytorch/vision/issues/8608
closed
[]
2024-08-23T15:15:27Z
2024-08-27T10:13:48Z
1
ArpanGyawali
huggingface/diffusers
9,258
Kohya SS FLUX LoRA training is way faster on Linux than Windows any ideas to debug? Same settings, libraries and GPU
### Describe the bug I am using Kohya SS to train FLUX LoRA On Linux RTX 3090 gets like 5.5 second / it - batch size 1 and 1024x1024 px resolution On Windows RTX 3090 TI gets 7.7 second / it - has the most powerful CPU 13900 K This speed dispercany is huge between Windows and Linux for some reason Torch...
https://github.com/huggingface/diffusers/issues/9258
closed
[ "bug" ]
2024-08-23T11:42:53Z
2024-08-23T11:55:18Z
1
FurkanGozukara
huggingface/datasets
7,122
[interleave_dataset] sample batches from a single source at a time
### Feature request interleave_dataset and [RandomlyCyclingMultiSourcesExamplesIterable](https://github.com/huggingface/datasets/blob/3813ce846e52824b38e53895810682f0a496a2e3/src/datasets/iterable_dataset.py#L816) enable us to sample data examples from different sources. But can we also sample batches in a similar man...
https://github.com/huggingface/datasets/issues/7122
open
[ "enhancement" ]
2024-08-23T07:21:15Z
2024-08-23T07:21:15Z
0
memray
huggingface/text-generation-inference
2,452
How to get the token probability by curl request?
### Feature request curl -v -X POST http://.....srv/generate -H "Content-Type: application/json" -d '{"inputs": "xxxxx:","parameters": {"max_new_tokens": 256}}' user this curl request, get output like {"generated_text": xxxx} how to get generated text probability from llm in TGI service? ### Motivation no ...
https://github.com/huggingface/text-generation-inference/issues/2452
closed
[]
2024-08-23T03:01:17Z
2024-08-27T01:32:44Z
null
TWSFar
huggingface/speech-to-speech
37
[Feature request] How about adding an optional speech to viseme model at the end of our chain?
Hi there, Thank you so much for your work on this project. It's truly amazing, and I’m excited to see all the innovative tools that people will build based on it. I can already imagine many will integrate your speech-to-speech pipeline with avatar or robot embodiments, where lip sync will be crucial. To support ...
https://github.com/huggingface/speech-to-speech/issues/37
open
[]
2024-08-22T21:32:47Z
2024-09-09T17:16:45Z
null
fabiocat93
pytorch/TensorRT
3,115
❓ [Question] JetPack 6.0
## ❓ Question <!-- Your question --> i'd like to use torch_tensorrt w/ JetPack 6.0, but from `setup.py`, it seems like latest supported version is JetPack 5.0 https://github.com/pytorch/TensorRT/blob/main/setup.py#L147-L164 ## What you have already tried 1. added JetPack 6.0 to setup.py, setting `JETPACK_VERSIO...
https://github.com/pytorch/TensorRT/issues/3115
closed
[ "question" ]
2024-08-22T17:52:45Z
2024-08-26T17:11:33Z
null
patrick-botco
pytorch/TensorRT
3,114
❓ [Question] Revisit the argument types of normalization converters
## ❓ Question https://github.com/pytorch/TensorRT/pull/3099#issuecomment-2303600863 ## What you have already tried <!-- A clear and concise description of what you have already done. --> ## Environment > Build information about Torch-TensorRT can be found by turning on debug messages - PyTorch Versio...
https://github.com/pytorch/TensorRT/issues/3114
open
[ "question" ]
2024-08-22T16:17:01Z
2024-08-22T18:04:10Z
null
peri044
huggingface/huggingface_hub
2,480
How to use the HF Nvidia NIM API with the HF inference client?
### Describe the bug We recently introduced the [Nvidia NIM API](https://huggingface.co/blog/inference-dgx-cloud) for selected models. The recommended use is via the OAI client like this (with a specific fine-grained token for an enterprise org): ```py from openai import OpenAI client = OpenAI( base_url...
https://github.com/huggingface/huggingface_hub/issues/2480
closed
[ "bug" ]
2024-08-22T12:32:16Z
2024-08-26T12:45:55Z
null
MoritzLaurer
huggingface/transformers.js
896
How to use this model: Xenova/bge-reranker-base
### Question I see that it supports transformers.js, but I can't find the instructions for use. Please help me with using it.
https://github.com/huggingface/transformers.js/issues/896
closed
[ "question" ]
2024-08-22T07:33:42Z
2024-08-29T00:12:52Z
null
gy9527
pytorch/pytorch
134,207
How to fallback the operators those are unsupported by XLA back to cpu backend?
I'm using the xla backend, and there are some operators that are not supported by the xla backend. How can I use the backend fallback mechanism to fallback these unsupported operators to CPU backend? Thanks! cc @bdhirsh
https://github.com/pytorch/pytorch/issues/134207
closed
[ "triaged", "module: xla" ]
2024-08-22T06:46:21Z
2024-09-05T06:52:20Z
null
wwtghx