| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """ |
| | Doc utilities: Utilities related to documentation |
| | """ |
| |
|
| | import functools |
| | import re |
| | import types |
| |
|
| |
|
| | def add_start_docstrings(*docstr): |
| | def docstring_decorator(fn): |
| | fn.__doc__ = "".join(docstr) + (fn.__doc__ if fn.__doc__ is not None else "") |
| | return fn |
| |
|
| | return docstring_decorator |
| |
|
| |
|
| | def add_start_docstrings_to_model_forward(*docstr): |
| | def docstring_decorator(fn): |
| | docstring = "".join(docstr) + (fn.__doc__ if fn.__doc__ is not None else "") |
| | class_name = f"[`{fn.__qualname__.split('.')[0]}`]" |
| | intro = f" The {class_name} forward method, overrides the `__call__` special method." |
| | note = r""" |
| | |
| | <Tip> |
| | |
| | Although the recipe for forward pass needs to be defined within this function, one should call the [`Module`] |
| | instance afterwards instead of this since the former takes care of running the pre and post processing steps while |
| | the latter silently ignores them. |
| | |
| | </Tip> |
| | """ |
| |
|
| | fn.__doc__ = intro + note + docstring |
| | return fn |
| |
|
| | return docstring_decorator |
| |
|
| |
|
| | def add_end_docstrings(*docstr): |
| | def docstring_decorator(fn): |
| | fn.__doc__ = (fn.__doc__ if fn.__doc__ is not None else "") + "".join(docstr) |
| | return fn |
| |
|
| | return docstring_decorator |
| |
|
| |
|
| | PT_RETURN_INTRODUCTION = r""" |
| | Returns: |
| | [`{full_output_type}`] or `tuple(torch.FloatTensor)`: A [`{full_output_type}`] or a tuple of |
| | `torch.FloatTensor` (if `return_dict=False` is passed or when `config.return_dict=False`) comprising various |
| | elements depending on the configuration ([`{config_class}`]) and inputs. |
| | |
| | """ |
| |
|
| |
|
| | TF_RETURN_INTRODUCTION = r""" |
| | Returns: |
| | [`{full_output_type}`] or `tuple(tf.Tensor)`: A [`{full_output_type}`] or a tuple of `tf.Tensor` (if |
| | `return_dict=False` is passed or when `config.return_dict=False`) comprising various elements depending on the |
| | configuration ([`{config_class}`]) and inputs. |
| | |
| | """ |
| |
|
| |
|
| | def _get_indent(t): |
| | """Returns the indentation in the first line of t""" |
| | search = re.search(r"^(\s*)\S", t) |
| | return "" if search is None else search.groups()[0] |
| |
|
| |
|
| | def _convert_output_args_doc(output_args_doc): |
| | """Convert output_args_doc to display properly.""" |
| | |
| | indent = _get_indent(output_args_doc) |
| | blocks = [] |
| | current_block = "" |
| | for line in output_args_doc.split("\n"): |
| | |
| | if _get_indent(line) == indent: |
| | if len(current_block) > 0: |
| | blocks.append(current_block[:-1]) |
| | current_block = f"{line}\n" |
| | else: |
| | |
| | |
| | current_block += f"{line[2:]}\n" |
| | blocks.append(current_block[:-1]) |
| |
|
| | |
| | for i in range(len(blocks)): |
| | blocks[i] = re.sub(r"^(\s+)(\S+)(\s+)", r"\1- **\2**\3", blocks[i]) |
| | blocks[i] = re.sub(r":\s*\n\s*(\S)", r" -- \1", blocks[i]) |
| |
|
| | return "\n".join(blocks) |
| |
|
| |
|
| | def _prepare_output_docstrings(output_type, config_class, min_indent=None): |
| | """ |
| | Prepares the return part of the docstring using `output_type`. |
| | """ |
| | output_docstring = output_type.__doc__ |
| |
|
| | |
| | lines = output_docstring.split("\n") |
| | i = 0 |
| | while i < len(lines) and re.search(r"^\s*(Args|Parameters):\s*$", lines[i]) is None: |
| | i += 1 |
| | if i < len(lines): |
| | params_docstring = "\n".join(lines[(i + 1) :]) |
| | params_docstring = _convert_output_args_doc(params_docstring) |
| | else: |
| | raise ValueError( |
| | f"No `Args` or `Parameters` section is found in the docstring of `{output_type.__name__}`. Make sure it has" |
| | "docstring and contain either `Args` or `Parameters`." |
| | ) |
| |
|
| | |
| | full_output_type = f"{output_type.__module__}.{output_type.__name__}" |
| | intro = TF_RETURN_INTRODUCTION if output_type.__name__.startswith("TF") else PT_RETURN_INTRODUCTION |
| | intro = intro.format(full_output_type=full_output_type, config_class=config_class) |
| | result = intro + params_docstring |
| |
|
| | |
| | if min_indent is not None: |
| | lines = result.split("\n") |
| | |
| | i = 0 |
| | while len(lines[i]) == 0: |
| | i += 1 |
| | indent = len(_get_indent(lines[i])) |
| | |
| | if indent < min_indent: |
| | to_add = " " * (min_indent - indent) |
| | lines = [(f"{to_add}{line}" if len(line) > 0 else line) for line in lines] |
| | result = "\n".join(lines) |
| |
|
| | return result |
| |
|
| |
|
| | FAKE_MODEL_DISCLAIMER = """ |
| | <Tip warning={true}> |
| | |
| | This example uses a random model as the real ones are all very big. To get proper results, you should use |
| | {real_checkpoint} instead of {fake_checkpoint}. If you get out-of-memory when loading that checkpoint, you can try |
| | adding `device_map="auto"` in the `from_pretrained` call. |
| | |
| | </Tip> |
| | """ |
| |
|
| |
|
| | PT_TOKEN_CLASSIFICATION_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | >>> import torch |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = tokenizer( |
| | ... "HuggingFace is a company based in Paris and New York", add_special_tokens=False, return_tensors="pt" |
| | ... ) |
| | |
| | >>> with torch.no_grad(): |
| | ... logits = model(**inputs).logits |
| | |
| | >>> predicted_token_class_ids = logits.argmax(-1) |
| | |
| | >>> # Note that tokens are classified rather then input words which means that |
| | >>> # there might be more predicted token classes than words. |
| | >>> # Multiple token classes might account for the same word |
| | >>> predicted_tokens_classes = [model.config.id2label[t.item()] for t in predicted_token_class_ids[0]] |
| | >>> predicted_tokens_classes |
| | {expected_output} |
| | |
| | >>> labels = predicted_token_class_ids |
| | >>> loss = model(**inputs, labels=labels).loss |
| | >>> round(loss.item(), 2) |
| | {expected_loss} |
| | ``` |
| | """ |
| |
|
| | PT_QUESTION_ANSWERING_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | >>> import torch |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> question, text = "Who was Jim Henson?", "Jim Henson was a nice puppet" |
| | |
| | >>> inputs = tokenizer(question, text, return_tensors="pt") |
| | >>> with torch.no_grad(): |
| | ... outputs = model(**inputs) |
| | |
| | >>> answer_start_index = outputs.start_logits.argmax() |
| | >>> answer_end_index = outputs.end_logits.argmax() |
| | |
| | >>> predict_answer_tokens = inputs.input_ids[0, answer_start_index : answer_end_index + 1] |
| | >>> tokenizer.decode(predict_answer_tokens, skip_special_tokens=True) |
| | {expected_output} |
| | |
| | >>> # target is "nice puppet" |
| | >>> target_start_index = torch.tensor([{qa_target_start_index}]) |
| | >>> target_end_index = torch.tensor([{qa_target_end_index}]) |
| | |
| | >>> outputs = model(**inputs, start_positions=target_start_index, end_positions=target_end_index) |
| | >>> loss = outputs.loss |
| | >>> round(loss.item(), 2) |
| | {expected_loss} |
| | ``` |
| | """ |
| |
|
| | PT_SEQUENCE_CLASSIFICATION_SAMPLE = r""" |
| | Example of single-label classification: |
| | |
| | ```python |
| | >>> import torch |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = tokenizer("Hello, my dog is cute", return_tensors="pt") |
| | |
| | >>> with torch.no_grad(): |
| | ... logits = model(**inputs).logits |
| | |
| | >>> predicted_class_id = logits.argmax().item() |
| | >>> model.config.id2label[predicted_class_id] |
| | {expected_output} |
| | |
| | >>> # To train a model on `num_labels` classes, you can pass `num_labels=num_labels` to `.from_pretrained(...)` |
| | >>> num_labels = len(model.config.id2label) |
| | >>> model = {model_class}.from_pretrained("{checkpoint}", num_labels=num_labels) |
| | |
| | >>> labels = torch.tensor([1]) |
| | >>> loss = model(**inputs, labels=labels).loss |
| | >>> round(loss.item(), 2) |
| | {expected_loss} |
| | ``` |
| | |
| | Example of multi-label classification: |
| | |
| | ```python |
| | >>> import torch |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}", problem_type="multi_label_classification") |
| | |
| | >>> inputs = tokenizer("Hello, my dog is cute", return_tensors="pt") |
| | |
| | >>> with torch.no_grad(): |
| | ... logits = model(**inputs).logits |
| | |
| | >>> predicted_class_ids = torch.arange(0, logits.shape[-1])[torch.sigmoid(logits).squeeze(dim=0) > 0.5] |
| | |
| | >>> # To train a model on `num_labels` classes, you can pass `num_labels=num_labels` to `.from_pretrained(...)` |
| | >>> num_labels = len(model.config.id2label) |
| | >>> model = {model_class}.from_pretrained( |
| | ... "{checkpoint}", num_labels=num_labels, problem_type="multi_label_classification" |
| | ... ) |
| | |
| | >>> labels = torch.sum( |
| | ... torch.nn.functional.one_hot(predicted_class_ids[None, :].clone(), num_classes=num_labels), dim=1 |
| | ... ).to(torch.float) |
| | >>> loss = model(**inputs, labels=labels).loss |
| | ``` |
| | """ |
| |
|
| | PT_MASKED_LM_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | >>> import torch |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = tokenizer("The capital of France is {mask}.", return_tensors="pt") |
| | |
| | >>> with torch.no_grad(): |
| | ... logits = model(**inputs).logits |
| | |
| | >>> # retrieve index of {mask} |
| | >>> mask_token_index = (inputs.input_ids == tokenizer.mask_token_id)[0].nonzero(as_tuple=True)[0] |
| | |
| | >>> predicted_token_id = logits[0, mask_token_index].argmax(axis=-1) |
| | >>> tokenizer.decode(predicted_token_id) |
| | {expected_output} |
| | |
| | >>> labels = tokenizer("The capital of France is Paris.", return_tensors="pt")["input_ids"] |
| | >>> # mask labels of non-{mask} tokens |
| | >>> labels = torch.where(inputs.input_ids == tokenizer.mask_token_id, labels, -100) |
| | |
| | >>> outputs = model(**inputs, labels=labels) |
| | >>> round(outputs.loss.item(), 2) |
| | {expected_loss} |
| | ``` |
| | """ |
| |
|
| | PT_BASE_MODEL_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | >>> import torch |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = tokenizer("Hello, my dog is cute", return_tensors="pt") |
| | >>> outputs = model(**inputs) |
| | |
| | >>> last_hidden_states = outputs.last_hidden_state |
| | ``` |
| | """ |
| |
|
| | PT_MULTIPLE_CHOICE_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | >>> import torch |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> prompt = "In Italy, pizza served in formal settings, such as at a restaurant, is presented unsliced." |
| | >>> choice0 = "It is eaten with a fork and a knife." |
| | >>> choice1 = "It is eaten while held in the hand." |
| | >>> labels = torch.tensor(0).unsqueeze(0) # choice0 is correct (according to Wikipedia ;)), batch size 1 |
| | |
| | >>> encoding = tokenizer([prompt, prompt], [choice0, choice1], return_tensors="pt", padding=True) |
| | >>> outputs = model(**{{k: v.unsqueeze(0) for k, v in encoding.items()}}, labels=labels) # batch size is 1 |
| | |
| | >>> # the linear classifier still needs to be trained |
| | >>> loss = outputs.loss |
| | >>> logits = outputs.logits |
| | ``` |
| | """ |
| |
|
| | PT_CAUSAL_LM_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> import torch |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = tokenizer("Hello, my dog is cute", return_tensors="pt") |
| | >>> outputs = model(**inputs, labels=inputs["input_ids"]) |
| | >>> loss = outputs.loss |
| | >>> logits = outputs.logits |
| | ``` |
| | """ |
| |
|
| | PT_SPEECH_BASE_MODEL_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoProcessor, {model_class} |
| | >>> import torch |
| | >>> from datasets import load_dataset |
| | |
| | >>> dataset = load_dataset("hf-internal-testing/librispeech_asr_demo", "clean", split="validation") |
| | >>> dataset = dataset.sort("id") |
| | >>> sampling_rate = dataset.features["audio"].sampling_rate |
| | |
| | >>> processor = AutoProcessor.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> # audio file is decoded on the fly |
| | >>> inputs = processor(dataset[0]["audio"]["array"], sampling_rate=sampling_rate, return_tensors="pt") |
| | >>> with torch.no_grad(): |
| | ... outputs = model(**inputs) |
| | |
| | >>> last_hidden_states = outputs.last_hidden_state |
| | >>> list(last_hidden_states.shape) |
| | {expected_output} |
| | ``` |
| | """ |
| |
|
| | PT_SPEECH_CTC_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoProcessor, {model_class} |
| | >>> from datasets import load_dataset |
| | >>> import torch |
| | |
| | >>> dataset = load_dataset("hf-internal-testing/librispeech_asr_demo", "clean", split="validation") |
| | >>> dataset = dataset.sort("id") |
| | >>> sampling_rate = dataset.features["audio"].sampling_rate |
| | |
| | >>> processor = AutoProcessor.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> # audio file is decoded on the fly |
| | >>> inputs = processor(dataset[0]["audio"]["array"], sampling_rate=sampling_rate, return_tensors="pt") |
| | >>> with torch.no_grad(): |
| | ... logits = model(**inputs).logits |
| | >>> predicted_ids = torch.argmax(logits, dim=-1) |
| | |
| | >>> # transcribe speech |
| | >>> transcription = processor.batch_decode(predicted_ids) |
| | >>> transcription[0] |
| | {expected_output} |
| | |
| | >>> inputs["labels"] = processor(text=dataset[0]["text"], return_tensors="pt").input_ids |
| | |
| | >>> # compute loss |
| | >>> loss = model(**inputs).loss |
| | >>> round(loss.item(), 2) |
| | {expected_loss} |
| | ``` |
| | """ |
| |
|
| | PT_SPEECH_SEQ_CLASS_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoFeatureExtractor, {model_class} |
| | >>> from datasets import load_dataset |
| | >>> import torch |
| | |
| | >>> dataset = load_dataset("hf-internal-testing/librispeech_asr_demo", "clean", split="validation") |
| | >>> dataset = dataset.sort("id") |
| | >>> sampling_rate = dataset.features["audio"].sampling_rate |
| | |
| | >>> feature_extractor = AutoFeatureExtractor.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> # audio file is decoded on the fly |
| | >>> inputs = feature_extractor(dataset[0]["audio"]["array"], sampling_rate=sampling_rate, return_tensors="pt") |
| | |
| | >>> with torch.no_grad(): |
| | ... logits = model(**inputs).logits |
| | |
| | >>> predicted_class_ids = torch.argmax(logits, dim=-1).item() |
| | >>> predicted_label = model.config.id2label[predicted_class_ids] |
| | >>> predicted_label |
| | {expected_output} |
| | |
| | >>> # compute loss - target_label is e.g. "down" |
| | >>> target_label = model.config.id2label[0] |
| | >>> inputs["labels"] = torch.tensor([model.config.label2id[target_label]]) |
| | >>> loss = model(**inputs).loss |
| | >>> round(loss.item(), 2) |
| | {expected_loss} |
| | ``` |
| | """ |
| |
|
| |
|
| | PT_SPEECH_FRAME_CLASS_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoFeatureExtractor, {model_class} |
| | >>> from datasets import load_dataset |
| | >>> import torch |
| | |
| | >>> dataset = load_dataset("hf-internal-testing/librispeech_asr_demo", "clean", split="validation") |
| | >>> dataset = dataset.sort("id") |
| | >>> sampling_rate = dataset.features["audio"].sampling_rate |
| | |
| | >>> feature_extractor = AutoFeatureExtractor.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> # audio file is decoded on the fly |
| | >>> inputs = feature_extractor(dataset[0]["audio"]["array"], return_tensors="pt", sampling_rate=sampling_rate) |
| | >>> with torch.no_grad(): |
| | ... logits = model(**inputs).logits |
| | |
| | >>> probabilities = torch.sigmoid(logits[0]) |
| | >>> # labels is a one-hot array of shape (num_frames, num_speakers) |
| | >>> labels = (probabilities > 0.5).long() |
| | >>> labels[0].tolist() |
| | {expected_output} |
| | ``` |
| | """ |
| |
|
| |
|
| | PT_SPEECH_XVECTOR_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoFeatureExtractor, {model_class} |
| | >>> from datasets import load_dataset |
| | >>> import torch |
| | |
| | >>> dataset = load_dataset("hf-internal-testing/librispeech_asr_demo", "clean", split="validation") |
| | >>> dataset = dataset.sort("id") |
| | >>> sampling_rate = dataset.features["audio"].sampling_rate |
| | |
| | >>> feature_extractor = AutoFeatureExtractor.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> # audio file is decoded on the fly |
| | >>> inputs = feature_extractor( |
| | ... [d["array"] for d in dataset[:2]["audio"]], sampling_rate=sampling_rate, return_tensors="pt", padding=True |
| | ... ) |
| | >>> with torch.no_grad(): |
| | ... embeddings = model(**inputs).embeddings |
| | |
| | >>> embeddings = torch.nn.functional.normalize(embeddings, dim=-1).cpu() |
| | |
| | >>> # the resulting embeddings can be used for cosine similarity-based retrieval |
| | >>> cosine_sim = torch.nn.CosineSimilarity(dim=-1) |
| | >>> similarity = cosine_sim(embeddings[0], embeddings[1]) |
| | >>> threshold = 0.7 # the optimal threshold is dataset-dependent |
| | >>> if similarity < threshold: |
| | ... print("Speakers are not the same!") |
| | >>> round(similarity.item(), 2) |
| | {expected_output} |
| | ``` |
| | """ |
| |
|
| | PT_VISION_BASE_MODEL_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoImageProcessor, {model_class} |
| | >>> import torch |
| | >>> from datasets import load_dataset |
| | |
| | >>> dataset = load_dataset("huggingface/cats-image") |
| | >>> image = dataset["test"]["image"][0] |
| | |
| | >>> image_processor = AutoImageProcessor.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = image_processor(image, return_tensors="pt") |
| | |
| | >>> with torch.no_grad(): |
| | ... outputs = model(**inputs) |
| | |
| | >>> last_hidden_states = outputs.last_hidden_state |
| | >>> list(last_hidden_states.shape) |
| | {expected_output} |
| | ``` |
| | """ |
| |
|
| | PT_VISION_SEQ_CLASS_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoImageProcessor, {model_class} |
| | >>> import torch |
| | >>> from datasets import load_dataset |
| | |
| | >>> dataset = load_dataset("huggingface/cats-image") |
| | >>> image = dataset["test"]["image"][0] |
| | |
| | >>> image_processor = AutoImageProcessor.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = image_processor(image, return_tensors="pt") |
| | |
| | >>> with torch.no_grad(): |
| | ... logits = model(**inputs).logits |
| | |
| | >>> # model predicts one of the 1000 ImageNet classes |
| | >>> predicted_label = logits.argmax(-1).item() |
| | >>> print(model.config.id2label[predicted_label]) |
| | {expected_output} |
| | ``` |
| | """ |
| |
|
| |
|
| | PT_SAMPLE_DOCSTRINGS = { |
| | "SequenceClassification": PT_SEQUENCE_CLASSIFICATION_SAMPLE, |
| | "QuestionAnswering": PT_QUESTION_ANSWERING_SAMPLE, |
| | "TokenClassification": PT_TOKEN_CLASSIFICATION_SAMPLE, |
| | "MultipleChoice": PT_MULTIPLE_CHOICE_SAMPLE, |
| | "MaskedLM": PT_MASKED_LM_SAMPLE, |
| | "LMHead": PT_CAUSAL_LM_SAMPLE, |
| | "BaseModel": PT_BASE_MODEL_SAMPLE, |
| | "SpeechBaseModel": PT_SPEECH_BASE_MODEL_SAMPLE, |
| | "CTC": PT_SPEECH_CTC_SAMPLE, |
| | "AudioClassification": PT_SPEECH_SEQ_CLASS_SAMPLE, |
| | "AudioFrameClassification": PT_SPEECH_FRAME_CLASS_SAMPLE, |
| | "AudioXVector": PT_SPEECH_XVECTOR_SAMPLE, |
| | "VisionBaseModel": PT_VISION_BASE_MODEL_SAMPLE, |
| | "ImageClassification": PT_VISION_SEQ_CLASS_SAMPLE, |
| | } |
| |
|
| |
|
| | TF_TOKEN_CLASSIFICATION_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | >>> import tensorflow as tf |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = tokenizer( |
| | ... "HuggingFace is a company based in Paris and New York", add_special_tokens=False, return_tensors="tf" |
| | ... ) |
| | |
| | >>> logits = model(**inputs).logits |
| | >>> predicted_token_class_ids = tf.math.argmax(logits, axis=-1) |
| | |
| | >>> # Note that tokens are classified rather then input words which means that |
| | >>> # there might be more predicted token classes than words. |
| | >>> # Multiple token classes might account for the same word |
| | >>> predicted_tokens_classes = [model.config.id2label[t] for t in predicted_token_class_ids[0].numpy().tolist()] |
| | >>> predicted_tokens_classes |
| | {expected_output} |
| | ``` |
| | |
| | ```python |
| | >>> labels = predicted_token_class_ids |
| | >>> loss = tf.math.reduce_mean(model(**inputs, labels=labels).loss) |
| | >>> round(float(loss), 2) |
| | {expected_loss} |
| | ``` |
| | """ |
| |
|
| | TF_QUESTION_ANSWERING_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | >>> import tensorflow as tf |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> question, text = "Who was Jim Henson?", "Jim Henson was a nice puppet" |
| | |
| | >>> inputs = tokenizer(question, text, return_tensors="tf") |
| | >>> outputs = model(**inputs) |
| | |
| | >>> answer_start_index = int(tf.math.argmax(outputs.start_logits, axis=-1)[0]) |
| | >>> answer_end_index = int(tf.math.argmax(outputs.end_logits, axis=-1)[0]) |
| | |
| | >>> predict_answer_tokens = inputs.input_ids[0, answer_start_index : answer_end_index + 1] |
| | >>> tokenizer.decode(predict_answer_tokens) |
| | {expected_output} |
| | ``` |
| | |
| | ```python |
| | >>> # target is "nice puppet" |
| | >>> target_start_index = tf.constant([{qa_target_start_index}]) |
| | >>> target_end_index = tf.constant([{qa_target_end_index}]) |
| | |
| | >>> outputs = model(**inputs, start_positions=target_start_index, end_positions=target_end_index) |
| | >>> loss = tf.math.reduce_mean(outputs.loss) |
| | >>> round(float(loss), 2) |
| | {expected_loss} |
| | ``` |
| | """ |
| |
|
| | TF_SEQUENCE_CLASSIFICATION_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | >>> import tensorflow as tf |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = tokenizer("Hello, my dog is cute", return_tensors="tf") |
| | |
| | >>> logits = model(**inputs).logits |
| | |
| | >>> predicted_class_id = int(tf.math.argmax(logits, axis=-1)[0]) |
| | >>> model.config.id2label[predicted_class_id] |
| | {expected_output} |
| | ``` |
| | |
| | ```python |
| | >>> # To train a model on `num_labels` classes, you can pass `num_labels=num_labels` to `.from_pretrained(...)` |
| | >>> num_labels = len(model.config.id2label) |
| | >>> model = {model_class}.from_pretrained("{checkpoint}", num_labels=num_labels) |
| | |
| | >>> labels = tf.constant(1) |
| | >>> loss = model(**inputs, labels=labels).loss |
| | >>> round(float(loss), 2) |
| | {expected_loss} |
| | ``` |
| | """ |
| |
|
| | TF_MASKED_LM_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | >>> import tensorflow as tf |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = tokenizer("The capital of France is {mask}.", return_tensors="tf") |
| | >>> logits = model(**inputs).logits |
| | |
| | >>> # retrieve index of {mask} |
| | >>> mask_token_index = tf.where((inputs.input_ids == tokenizer.mask_token_id)[0]) |
| | >>> selected_logits = tf.gather_nd(logits[0], indices=mask_token_index) |
| | |
| | >>> predicted_token_id = tf.math.argmax(selected_logits, axis=-1) |
| | >>> tokenizer.decode(predicted_token_id) |
| | {expected_output} |
| | ``` |
| | |
| | ```python |
| | >>> labels = tokenizer("The capital of France is Paris.", return_tensors="tf")["input_ids"] |
| | >>> # mask labels of non-{mask} tokens |
| | >>> labels = tf.where(inputs.input_ids == tokenizer.mask_token_id, labels, -100) |
| | |
| | >>> outputs = model(**inputs, labels=labels) |
| | >>> round(float(outputs.loss), 2) |
| | {expected_loss} |
| | ``` |
| | """ |
| |
|
| | TF_BASE_MODEL_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | >>> import tensorflow as tf |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = tokenizer("Hello, my dog is cute", return_tensors="tf") |
| | >>> outputs = model(inputs) |
| | |
| | >>> last_hidden_states = outputs.last_hidden_state |
| | ``` |
| | """ |
| |
|
| | TF_MULTIPLE_CHOICE_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | >>> import tensorflow as tf |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> prompt = "In Italy, pizza served in formal settings, such as at a restaurant, is presented unsliced." |
| | >>> choice0 = "It is eaten with a fork and a knife." |
| | >>> choice1 = "It is eaten while held in the hand." |
| | |
| | >>> encoding = tokenizer([prompt, prompt], [choice0, choice1], return_tensors="tf", padding=True) |
| | >>> inputs = {{k: tf.expand_dims(v, 0) for k, v in encoding.items()}} |
| | >>> outputs = model(inputs) # batch size is 1 |
| | |
| | >>> # the linear classifier still needs to be trained |
| | >>> logits = outputs.logits |
| | ``` |
| | """ |
| |
|
| | TF_CAUSAL_LM_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | >>> import tensorflow as tf |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = tokenizer("Hello, my dog is cute", return_tensors="tf") |
| | >>> outputs = model(inputs) |
| | >>> logits = outputs.logits |
| | ``` |
| | """ |
| |
|
| | TF_SPEECH_BASE_MODEL_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoProcessor, {model_class} |
| | >>> from datasets import load_dataset |
| | |
| | >>> dataset = load_dataset("hf-internal-testing/librispeech_asr_demo", "clean", split="validation") |
| | >>> dataset = dataset.sort("id") |
| | >>> sampling_rate = dataset.features["audio"].sampling_rate |
| | |
| | >>> processor = AutoProcessor.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> # audio file is decoded on the fly |
| | >>> inputs = processor(dataset[0]["audio"]["array"], sampling_rate=sampling_rate, return_tensors="tf") |
| | >>> outputs = model(**inputs) |
| | |
| | >>> last_hidden_states = outputs.last_hidden_state |
| | >>> list(last_hidden_states.shape) |
| | {expected_output} |
| | ``` |
| | """ |
| |
|
| | TF_SPEECH_CTC_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoProcessor, {model_class} |
| | >>> from datasets import load_dataset |
| | >>> import tensorflow as tf |
| | |
| | >>> dataset = load_dataset("hf-internal-testing/librispeech_asr_demo", "clean", split="validation") |
| | >>> dataset = dataset.sort("id") |
| | >>> sampling_rate = dataset.features["audio"].sampling_rate |
| | |
| | >>> processor = AutoProcessor.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> # audio file is decoded on the fly |
| | >>> inputs = processor(dataset[0]["audio"]["array"], sampling_rate=sampling_rate, return_tensors="tf") |
| | >>> logits = model(**inputs).logits |
| | >>> predicted_ids = tf.math.argmax(logits, axis=-1) |
| | |
| | >>> # transcribe speech |
| | >>> transcription = processor.batch_decode(predicted_ids) |
| | >>> transcription[0] |
| | {expected_output} |
| | ``` |
| | |
| | ```python |
| | >>> inputs["labels"] = processor(text=dataset[0]["text"], return_tensors="tf").input_ids |
| | |
| | >>> # compute loss |
| | >>> loss = model(**inputs).loss |
| | >>> round(float(loss), 2) |
| | {expected_loss} |
| | ``` |
| | """ |
| |
|
| | TF_VISION_BASE_MODEL_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoImageProcessor, {model_class} |
| | >>> from datasets import load_dataset |
| | |
| | >>> dataset = load_dataset("huggingface/cats-image") |
| | >>> image = dataset["test"]["image"][0] |
| | |
| | >>> image_processor = AutoImageProcessor.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = image_processor(image, return_tensors="tf") |
| | >>> outputs = model(**inputs) |
| | |
| | >>> last_hidden_states = outputs.last_hidden_state |
| | >>> list(last_hidden_states.shape) |
| | {expected_output} |
| | ``` |
| | """ |
| |
|
| | TF_VISION_SEQ_CLASS_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoImageProcessor, {model_class} |
| | >>> import tensorflow as tf |
| | >>> from datasets import load_dataset |
| | |
| | >>> dataset = load_dataset("huggingface/cats-image") |
| | >>> image = dataset["test"]["image"][0] |
| | |
| | >>> image_processor = AutoImageProcessor.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = image_processor(image, return_tensors="tf") |
| | >>> logits = model(**inputs).logits |
| | |
| | >>> # model predicts one of the 1000 ImageNet classes |
| | >>> predicted_label = int(tf.math.argmax(logits, axis=-1)) |
| | >>> print(model.config.id2label[predicted_label]) |
| | {expected_output} |
| | ``` |
| | """ |
| |
|
| | TF_SAMPLE_DOCSTRINGS = { |
| | "SequenceClassification": TF_SEQUENCE_CLASSIFICATION_SAMPLE, |
| | "QuestionAnswering": TF_QUESTION_ANSWERING_SAMPLE, |
| | "TokenClassification": TF_TOKEN_CLASSIFICATION_SAMPLE, |
| | "MultipleChoice": TF_MULTIPLE_CHOICE_SAMPLE, |
| | "MaskedLM": TF_MASKED_LM_SAMPLE, |
| | "LMHead": TF_CAUSAL_LM_SAMPLE, |
| | "BaseModel": TF_BASE_MODEL_SAMPLE, |
| | "SpeechBaseModel": TF_SPEECH_BASE_MODEL_SAMPLE, |
| | "CTC": TF_SPEECH_CTC_SAMPLE, |
| | "VisionBaseModel": TF_VISION_BASE_MODEL_SAMPLE, |
| | "ImageClassification": TF_VISION_SEQ_CLASS_SAMPLE, |
| | } |
| |
|
| |
|
| | FLAX_TOKEN_CLASSIFICATION_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = tokenizer("Hello, my dog is cute", return_tensors="jax") |
| | |
| | >>> outputs = model(**inputs) |
| | >>> logits = outputs.logits |
| | ``` |
| | """ |
| |
|
| | FLAX_QUESTION_ANSWERING_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> question, text = "Who was Jim Henson?", "Jim Henson was a nice puppet" |
| | >>> inputs = tokenizer(question, text, return_tensors="jax") |
| | |
| | >>> outputs = model(**inputs) |
| | >>> start_scores = outputs.start_logits |
| | >>> end_scores = outputs.end_logits |
| | ``` |
| | """ |
| |
|
| | FLAX_SEQUENCE_CLASSIFICATION_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = tokenizer("Hello, my dog is cute", return_tensors="jax") |
| | |
| | >>> outputs = model(**inputs) |
| | >>> logits = outputs.logits |
| | ``` |
| | """ |
| |
|
| | FLAX_MASKED_LM_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = tokenizer("The capital of France is {mask}.", return_tensors="jax") |
| | |
| | >>> outputs = model(**inputs) |
| | >>> logits = outputs.logits |
| | ``` |
| | """ |
| |
|
| | FLAX_BASE_MODEL_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = tokenizer("Hello, my dog is cute", return_tensors="jax") |
| | >>> outputs = model(**inputs) |
| | |
| | >>> last_hidden_states = outputs.last_hidden_state |
| | ``` |
| | """ |
| |
|
| | FLAX_MULTIPLE_CHOICE_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> prompt = "In Italy, pizza served in formal settings, such as at a restaurant, is presented unsliced." |
| | >>> choice0 = "It is eaten with a fork and a knife." |
| | >>> choice1 = "It is eaten while held in the hand." |
| | |
| | >>> encoding = tokenizer([prompt, prompt], [choice0, choice1], return_tensors="jax", padding=True) |
| | >>> outputs = model(**{{k: v[None, :] for k, v in encoding.items()}}) |
| | |
| | >>> logits = outputs.logits |
| | ``` |
| | """ |
| |
|
| | FLAX_CAUSAL_LM_SAMPLE = r""" |
| | Example: |
| | |
| | ```python |
| | >>> from transformers import AutoTokenizer, {model_class} |
| | |
| | >>> tokenizer = AutoTokenizer.from_pretrained("{checkpoint}") |
| | >>> model = {model_class}.from_pretrained("{checkpoint}") |
| | |
| | >>> inputs = tokenizer("Hello, my dog is cute", return_tensors="np") |
| | >>> outputs = model(**inputs) |
| | |
| | >>> # retrieve logts for next token |
| | >>> next_token_logits = outputs.logits[:, -1] |
| | ``` |
| | """ |
| |
|
| | FLAX_SAMPLE_DOCSTRINGS = { |
| | "SequenceClassification": FLAX_SEQUENCE_CLASSIFICATION_SAMPLE, |
| | "QuestionAnswering": FLAX_QUESTION_ANSWERING_SAMPLE, |
| | "TokenClassification": FLAX_TOKEN_CLASSIFICATION_SAMPLE, |
| | "MultipleChoice": FLAX_MULTIPLE_CHOICE_SAMPLE, |
| | "MaskedLM": FLAX_MASKED_LM_SAMPLE, |
| | "BaseModel": FLAX_BASE_MODEL_SAMPLE, |
| | "LMHead": FLAX_CAUSAL_LM_SAMPLE, |
| | } |
| |
|
| |
|
| | def filter_outputs_from_example(docstring, **kwargs): |
| | """ |
| | Removes the lines testing an output with the doctest syntax in a code sample when it's set to `None`. |
| | """ |
| | for key, value in kwargs.items(): |
| | if value is not None: |
| | continue |
| |
|
| | doc_key = "{" + key + "}" |
| | docstring = re.sub(rf"\n([^\n]+)\n\s+{doc_key}\n", "\n", docstring) |
| |
|
| | return docstring |
| |
|
| |
|
| | def add_code_sample_docstrings( |
| | *docstr, |
| | processor_class=None, |
| | checkpoint=None, |
| | output_type=None, |
| | config_class=None, |
| | mask="[MASK]", |
| | qa_target_start_index=14, |
| | qa_target_end_index=15, |
| | model_cls=None, |
| | modality=None, |
| | expected_output=None, |
| | expected_loss=None, |
| | real_checkpoint=None, |
| | ): |
| | def docstring_decorator(fn): |
| | |
| | model_class = fn.__qualname__.split(".")[0] if model_cls is None else model_cls |
| |
|
| | if model_class[:2] == "TF": |
| | sample_docstrings = TF_SAMPLE_DOCSTRINGS |
| | elif model_class[:4] == "Flax": |
| | sample_docstrings = FLAX_SAMPLE_DOCSTRINGS |
| | else: |
| | sample_docstrings = PT_SAMPLE_DOCSTRINGS |
| |
|
| | |
| | |
| | |
| | doc_kwargs = { |
| | "model_class": model_class, |
| | "processor_class": processor_class, |
| | "checkpoint": checkpoint, |
| | "mask": mask, |
| | "qa_target_start_index": qa_target_start_index, |
| | "qa_target_end_index": qa_target_end_index, |
| | "expected_output": expected_output, |
| | "expected_loss": expected_loss, |
| | "real_checkpoint": real_checkpoint, |
| | "fake_checkpoint": checkpoint, |
| | "true": "{true}", |
| | } |
| |
|
| | if ("SequenceClassification" in model_class or "AudioClassification" in model_class) and modality == "audio": |
| | code_sample = sample_docstrings["AudioClassification"] |
| | elif "SequenceClassification" in model_class: |
| | code_sample = sample_docstrings["SequenceClassification"] |
| | elif "QuestionAnswering" in model_class: |
| | code_sample = sample_docstrings["QuestionAnswering"] |
| | elif "TokenClassification" in model_class: |
| | code_sample = sample_docstrings["TokenClassification"] |
| | elif "MultipleChoice" in model_class: |
| | code_sample = sample_docstrings["MultipleChoice"] |
| | elif "MaskedLM" in model_class or model_class in ["FlaubertWithLMHeadModel", "XLMWithLMHeadModel"]: |
| | code_sample = sample_docstrings["MaskedLM"] |
| | elif "LMHead" in model_class or "CausalLM" in model_class: |
| | code_sample = sample_docstrings["LMHead"] |
| | elif "CTC" in model_class: |
| | code_sample = sample_docstrings["CTC"] |
| | elif "AudioFrameClassification" in model_class: |
| | code_sample = sample_docstrings["AudioFrameClassification"] |
| | elif "XVector" in model_class and modality == "audio": |
| | code_sample = sample_docstrings["AudioXVector"] |
| | elif "Model" in model_class and modality == "audio": |
| | code_sample = sample_docstrings["SpeechBaseModel"] |
| | elif "Model" in model_class and modality == "vision": |
| | code_sample = sample_docstrings["VisionBaseModel"] |
| | elif "Model" in model_class or "Encoder" in model_class: |
| | code_sample = sample_docstrings["BaseModel"] |
| | elif "ImageClassification" in model_class: |
| | code_sample = sample_docstrings["ImageClassification"] |
| | else: |
| | raise ValueError(f"Docstring can't be built for model {model_class}") |
| |
|
| | code_sample = filter_outputs_from_example( |
| | code_sample, expected_output=expected_output, expected_loss=expected_loss |
| | ) |
| | if real_checkpoint is not None: |
| | code_sample = FAKE_MODEL_DISCLAIMER + code_sample |
| | func_doc = (fn.__doc__ or "") + "".join(docstr) |
| | output_doc = "" if output_type is None else _prepare_output_docstrings(output_type, config_class) |
| | built_doc = code_sample.format(**doc_kwargs) |
| | fn.__doc__ = func_doc + output_doc + built_doc |
| | return fn |
| |
|
| | return docstring_decorator |
| |
|
| |
|
| | def replace_return_docstrings(output_type=None, config_class=None): |
| | def docstring_decorator(fn): |
| | func_doc = fn.__doc__ |
| | lines = func_doc.split("\n") |
| | i = 0 |
| | while i < len(lines) and re.search(r"^\s*Returns?:\s*$", lines[i]) is None: |
| | i += 1 |
| | if i < len(lines): |
| | indent = len(_get_indent(lines[i])) |
| | lines[i] = _prepare_output_docstrings(output_type, config_class, min_indent=indent) |
| | func_doc = "\n".join(lines) |
| | else: |
| | raise ValueError( |
| | f"The function {fn} should have an empty 'Return:' or 'Returns:' in its docstring as placeholder, " |
| | f"current docstring is:\n{func_doc}" |
| | ) |
| | fn.__doc__ = func_doc |
| | return fn |
| |
|
| | return docstring_decorator |
| |
|
| |
|
| | def copy_func(f): |
| | """Returns a copy of a function f.""" |
| | |
| | g = types.FunctionType(f.__code__, f.__globals__, name=f.__name__, argdefs=f.__defaults__, closure=f.__closure__) |
| | g = functools.update_wrapper(g, f) |
| | g.__kwdefaults__ = f.__kwdefaults__ |
| | return g |
| |
|