| | --- |
| | language: |
| | - en |
| | license: cc-by-nc-nd-4.0 |
| | library_name: transformers |
| | pipeline_tag: text-classification |
| | widget: |
| | - text: Mr. Jones, an architect is going to surprise his family by building them a |
| | new house. |
| | example_title: Pow |
| | - text: They want the research to go well and be productive. |
| | example_title: Ach |
| | - text: The man is trying to see a friend on board, but the officer will not let him |
| | go as the whistle for all ashore who are not going has already blown. |
| | example_title: Aff |
| | - text: The recollection of skating on the Charles, and the time she had pushed me |
| | through the ice, brought a laugh to the conversation; but it quickly faded in |
| | the murky waters of the river that could no longer freeze over. |
| | example_title: Pow + Aff |
| | - text: They are also well-known research scientists and are quite talented in this |
| | field. |
| | example_title: Pow + Ach |
| | - text: After a nice evening with his family, he will be back at work tomorrow, doing |
| | the best job he can on his drafting. |
| | example_title: Ach + Aff |
| | - text: She is surprised that she is able to make these calls and pleasantly surprised |
| | that her friends respond to her request. |
| | example_title: Pow + Aff |
| | --- |
| | This is a version of a classifier for implicit motives based on ModernBert. The classifier identifies the |
| | presence of implicit motive imagery in sentences, namely the three felt needs for Power, Achievement, |
| | and Affiliation. |
| |
|
| | This model is being made available to other researchers via download. The |
| | current license allows for free use without modification for non-commercial purposes. If you would |
| | like to use this model commercially, get in touch with us for access to our most recent model. |
| |
|
| |
|
| | ## Inference guide |
| |
|
| | This model can be directly downloaded and used with the following code. |
| |
|
| | ```python |
| | from transformers import AutoModelForSequenceClassification, AutoTokenizer, pipeline |
| | |
| | mbert = "encodingai/mBERT-im-multilabel" |
| | |
| | tokenizer = AutoTokenizer.from_pretrained(mbert, use_fast=True) |
| | model = AutoModelForSequenceClassification.from_pretrained(mbert, |
| | problem_type="multi_label_classification", |
| | ) |
| | |
| | # load model using the pipeline, returning the top 3 classifications |
| | classifier = pipeline("text-classification", model=model, device=0, tokenizer=tokenizer, top_k=3) |
| | |
| | sample = ["""The recollection of skating on the Charles, and the time she had |
| | pushed me through the ice, brought a laugh to the conversation; but |
| | it quickly faded in the murky waters of the river that could no |
| | longer freeze over."""] |
| | |
| | # predict on a sentence |
| | pred = classifier(sample) |
| | print(pred) |
| | # The labels are arranged according to likelihood of classification |
| | repdict = {"LABEL_0": "Pow", "LABEL_1": "Ach", "LABEL_2": "Aff"} |
| | # so we replace them in the output |
| | for y in pred: |
| | scores = {repdict[x['label']]: x['score'] for x in y} |
| | print(scores) |
| | ``` |
| |
|
| | ## References |
| |
|
| | McClelland, D. C. (1965). Toward a theory of motive acquisition. American Psychologist, 20,321-333. |
| |
|
| | Pang, J. S., & Ring, H. (2020). Automated Coding of Implicit Motives: A Machine-Learning Approach. |
| | Motivation and Emotion, 44(4), 549-566. DOI: 10.1007/s11031-020-09832-8. |
| |
|
| | Winter, D.G. (1994). Manual for scoring motive imagery in running text. Unpublished Instrument. Ann Arbor: University of Michigan. |