| {\rtf1\ansi\ansicpg1252\cocoartf2636 |
| \cocoatextscaling0\cocoaplatform0{\fonttbl\f0\froman\fcharset0 Times-Roman;\f1\fnil\fcharset0 HelveticaNeue;\f2\fnil\fcharset0 HelveticaNeue-Bold; |
| } |
| {\colortbl;\red255\green255\blue255;\red226\green224\blue220;\red15\green84\blue167;\red0\green0\blue0; |
| } |
| {\*\expandedcolortbl;;\cssrgb\c90980\c90196\c89020;\cssrgb\c3922\c41569\c71373\c23922;\cssrgb\c0\c0\c0; |
| } |
| \paperw11900\paperh16840\margl1440\margr1440\vieww18380\viewh8400\viewkind0 |
| \deftab720 |
| \pard\pardeftab720\partightenfactor0 |
|
|
| \f0\fs24 \cf2 \expnd0\expndtw0\kerning0 |
| \outl0\strokewidth0 \strokec2 ***** Running training *****\ |
| Num examples = 12460\ |
| Num Epochs = 2\ |
| Instantaneous batch size per device = 10\ |
| Total train batch size (w. parallel, distributed & accumulation) = 80\ |
| Gradient Accumulation steps = 4\ |
| Total optimization steps = 310\ |
| Number of trainable parameters = 161844480\ |
| \pard\pardeftab720\partightenfactor0 |
|
|
| \f1\fs28 \cf2 \cb3 \strokec2 \ |
| \'a0[161/310 7:07:44 < 6:40:50, 0.01 it/s, Epoch 1.03/2]\ |
| \ |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrt\brdrnil \trbrdrl\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
|
|
| \f2\b\fs24 \cf0 \cb1 \strokec4 Step\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 Training Loss\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 Validation Loss\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 Rouge2 Precision\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 Rouge2 Recall\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 Rouge2 Fmeasure\cell \row |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrl\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
|
|
| \f1\b0 \cf0 10\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.635400\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.757870\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.055400\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.185500\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.082700\cell \row |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrl\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 20\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.504800\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.658047\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.062500\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.192100\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.091700\cell \row |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrl\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 30\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.419400\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.546135\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.061300\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.207100\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.091900\cell \row |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrl\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 40\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.353900\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.522316\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.060600\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.206800\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.090900\cell \row |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrl\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 50\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.332600\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.483913\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.059400\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.211300\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.090200\cell \row |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrl\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 60\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.278900\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.551230\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.063900\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.212300\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.095400\cell \row |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrl\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 70\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.286100\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.461083\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.065100\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.223400\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.098000\cell \row |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrl\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 80\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.297700\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.452156\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.067700\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.227200\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.101300\cell \row |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrl\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 90\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.240300\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.419080\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.069500\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.237600\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.104400\cell \row |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrl\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 100\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.223500\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.425878\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.065800\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.220100\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.098800\cell \row |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrl\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 110\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.294400\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.397157\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.066800\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.230300\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.100800\cell \row |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrl\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 120\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.218400\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.404413\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.068800\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.243400\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.104500\cell \row |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrl\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 130\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.257900\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.397923\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.067800\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.231700\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.102000\cell \row |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrl\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 140\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.187500\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.431973\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.069400\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.233300\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.104000\cell \row |
|
|
| \itap1\trowd \taflags5 \trgaph108\trleft-108 \trcbpat3 \tamarb240 \trbrdrl\brdrnil \trbrdrt\brdrnil \trbrdrr\brdrnil |
| \clvertalc \clshdrawnil \clwWidth559\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx1440 |
| \clvertalc \clshdrawnil \clwWidth1599\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx2880 |
| \clvertalc \clshdrawnil \clwWidth1821\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx4320 |
| \clvertalc \clshdrawnil \clwWidth2098\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx5760 |
| \clvertalc \clshdrawnil \clwWidth1708\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx7200 |
| \clvertalc \clshdrawnil \clwWidth2162\clftsWidth3 \clbrdrt\brdrnil \clbrdrl\brdrnil \clbrdrb\brdrnil \clbrdrr\brdrnil \clpadt120 \clpadl120 \clpadb120 \clpadr120 \gaph\cellx8640 |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 150\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.211600\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 1.417688\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.069000\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.233400\cell |
| \pard\intbl\itap1\pardeftab720\qr\partightenfactor0 |
| \cf0 0.103400\cell \lastrow\row |
| \pard\pardeftab720\partightenfactor0 |
|
|
| \f0 \cf2 \strokec2 \ |
| \ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/transformers/generation/utils.py:1186: UserWarning: You have modified the pretrained model configuration to control generation. This is a deprecated strategy to control generation and will be removed soon, in a future version. Please use a generation configuration file (see {\field{\*\fldinst{HYPERLINK "https://huggingface.co/docs/transformers/main_classes/text_generation"}}{\fldrslt https://huggingface.co/docs/transformers/main_classes/text_generation}})\ |
| warnings.warn(\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Saving model checkpoint to ./checkpoint-10\ |
| Configuration saved in ./checkpoint-10/config.json\ |
| Configuration saved in ./checkpoint-10/generation_config.json\ |
| Model weights saved in ./checkpoint-10/pytorch_model.bin\ |
| tokenizer config file saved in ./checkpoint-10/tokenizer_config.json\ |
| Special tokens file saved in ./checkpoint-10/special_tokens_map.json\ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\ |
| warnings.warn('Was asked to gather along dimension 0, but all '\ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Saving model checkpoint to ./checkpoint-20\ |
| Configuration saved in ./checkpoint-20/config.json\ |
| Configuration saved in ./checkpoint-20/generation_config.json\ |
| Model weights saved in ./checkpoint-20/pytorch_model.bin\ |
| tokenizer config file saved in ./checkpoint-20/tokenizer_config.json\ |
| Special tokens file saved in ./checkpoint-20/special_tokens_map.json\ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\ |
| warnings.warn('Was asked to gather along dimension 0, but all '\ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Saving model checkpoint to ./checkpoint-30\ |
| Configuration saved in ./checkpoint-30/config.json\ |
| Configuration saved in ./checkpoint-30/generation_config.json\ |
| Model weights saved in ./checkpoint-30/pytorch_model.bin\ |
| tokenizer config file saved in ./checkpoint-30/tokenizer_config.json\ |
| Special tokens file saved in ./checkpoint-30/special_tokens_map.json\ |
| Deleting older checkpoint [checkpoint-10] due to args.save_total_limit\ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\ |
| warnings.warn('Was asked to gather along dimension 0, but all '\ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Saving model checkpoint to ./checkpoint-40\ |
| Configuration saved in ./checkpoint-40/config.json\ |
| Configuration saved in ./checkpoint-40/generation_config.json\ |
| Model weights saved in ./checkpoint-40/pytorch_model.bin\ |
| tokenizer config file saved in ./checkpoint-40/tokenizer_config.json\ |
| Special tokens file saved in ./checkpoint-40/special_tokens_map.json\ |
| Deleting older checkpoint [checkpoint-20] due to args.save_total_limit\ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\ |
| warnings.warn('Was asked to gather along dimension 0, but all '\ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Saving model checkpoint to ./checkpoint-50\ |
| Configuration saved in ./checkpoint-50/config.json\ |
| Configuration saved in ./checkpoint-50/generation_config.json\ |
| Model weights saved in ./checkpoint-50/pytorch_model.bin\ |
| tokenizer config file saved in ./checkpoint-50/tokenizer_config.json\ |
| Special tokens file saved in ./checkpoint-50/special_tokens_map.json\ |
| Deleting older checkpoint [checkpoint-30] due to args.save_total_limit\ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\ |
| warnings.warn('Was asked to gather along dimension 0, but all '\ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Saving model checkpoint to ./checkpoint-60\ |
| Configuration saved in ./checkpoint-60/config.json\ |
| Configuration saved in ./checkpoint-60/generation_config.json\ |
| Model weights saved in ./checkpoint-60/pytorch_model.bin\ |
| tokenizer config file saved in ./checkpoint-60/tokenizer_config.json\ |
| Special tokens file saved in ./checkpoint-60/special_tokens_map.json\ |
| Deleting older checkpoint [checkpoint-40] due to args.save_total_limit\ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\ |
| warnings.warn('Was asked to gather along dimension 0, but all '\ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Saving model checkpoint to ./checkpoint-70\ |
| Configuration saved in ./checkpoint-70/config.json\ |
| Configuration saved in ./checkpoint-70/generation_config.json\ |
| Model weights saved in ./checkpoint-70/pytorch_model.bin\ |
| tokenizer config file saved in ./checkpoint-70/tokenizer_config.json\ |
| Special tokens file saved in ./checkpoint-70/special_tokens_map.json\ |
| Deleting older checkpoint [checkpoint-50] due to args.save_total_limit\ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\ |
| warnings.warn('Was asked to gather along dimension 0, but all '\ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Saving model checkpoint to ./checkpoint-80\ |
| Configuration saved in ./checkpoint-80/config.json\ |
| Configuration saved in ./checkpoint-80/generation_config.json\ |
| Model weights saved in ./checkpoint-80/pytorch_model.bin\ |
| tokenizer config file saved in ./checkpoint-80/tokenizer_config.json\ |
| Special tokens file saved in ./checkpoint-80/special_tokens_map.json\ |
| Deleting older checkpoint [checkpoint-60] due to args.save_total_limit\ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\ |
| warnings.warn('Was asked to gather along dimension 0, but all '\ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Saving model checkpoint to ./checkpoint-90\ |
| Configuration saved in ./checkpoint-90/config.json\ |
| Configuration saved in ./checkpoint-90/generation_config.json\ |
| Model weights saved in ./checkpoint-90/pytorch_model.bin\ |
| tokenizer config file saved in ./checkpoint-90/tokenizer_config.json\ |
| Special tokens file saved in ./checkpoint-90/special_tokens_map.json\ |
| Deleting older checkpoint [checkpoint-70] due to args.save_total_limit\ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\ |
| warnings.warn('Was asked to gather along dimension 0, but all '\ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Saving model checkpoint to ./checkpoint-100\ |
| Configuration saved in ./checkpoint-100/config.json\ |
| Configuration saved in ./checkpoint-100/generation_config.json\ |
| Model weights saved in ./checkpoint-100/pytorch_model.bin\ |
| tokenizer config file saved in ./checkpoint-100/tokenizer_config.json\ |
| Special tokens file saved in ./checkpoint-100/special_tokens_map.json\ |
| Deleting older checkpoint [checkpoint-80] due to args.save_total_limit\ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\ |
| warnings.warn('Was asked to gather along dimension 0, but all '\ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Saving model checkpoint to ./checkpoint-110\ |
| Configuration saved in ./checkpoint-110/config.json\ |
| Configuration saved in ./checkpoint-110/generation_config.json\ |
| Model weights saved in ./checkpoint-110/pytorch_model.bin\ |
| tokenizer config file saved in ./checkpoint-110/tokenizer_config.json\ |
| Special tokens file saved in ./checkpoint-110/special_tokens_map.json\ |
| Deleting older checkpoint [checkpoint-90] due to args.save_total_limit\ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\ |
| warnings.warn('Was asked to gather along dimension 0, but all '\ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Saving model checkpoint to ./checkpoint-120\ |
| Configuration saved in ./checkpoint-120/config.json\ |
| Configuration saved in ./checkpoint-120/generation_config.json\ |
| Model weights saved in ./checkpoint-120/pytorch_model.bin\ |
| tokenizer config file saved in ./checkpoint-120/tokenizer_config.json\ |
| Special tokens file saved in ./checkpoint-120/special_tokens_map.json\ |
| Deleting older checkpoint [checkpoint-100] due to args.save_total_limit\ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\ |
| warnings.warn('Was asked to gather along dimension 0, but all '\ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Saving model checkpoint to ./checkpoint-130\ |
| Configuration saved in ./checkpoint-130/config.json\ |
| Configuration saved in ./checkpoint-130/generation_config.json\ |
| Model weights saved in ./checkpoint-130/pytorch_model.bin\ |
| tokenizer config file saved in ./checkpoint-130/tokenizer_config.json\ |
| Special tokens file saved in ./checkpoint-130/special_tokens_map.json\ |
| Deleting older checkpoint [checkpoint-110] due to args.save_total_limit\ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\ |
| warnings.warn('Was asked to gather along dimension 0, but all '\ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Saving model checkpoint to ./checkpoint-140\ |
| Configuration saved in ./checkpoint-140/config.json\ |
| Configuration saved in ./checkpoint-140/generation_config.json\ |
| Model weights saved in ./checkpoint-140/pytorch_model.bin\ |
| tokenizer config file saved in ./checkpoint-140/tokenizer_config.json\ |
| Special tokens file saved in ./checkpoint-140/special_tokens_map.json\ |
| Deleting older checkpoint [checkpoint-120] due to args.save_total_limit\ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\ |
| warnings.warn('Was asked to gather along dimension 0, but all '\ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Saving model checkpoint to ./checkpoint-150\ |
| Configuration saved in ./checkpoint-150/config.json\ |
| Configuration saved in ./checkpoint-150/generation_config.json\ |
| Model weights saved in ./checkpoint-150/pytorch_model.bin\ |
| tokenizer config file saved in ./checkpoint-150/tokenizer_config.json\ |
| Special tokens file saved in ./checkpoint-150/special_tokens_map.json\ |
| Deleting older checkpoint [checkpoint-130] due to args.save_total_limit\ |
| /home/pageocr/bart-decoder-san/venv/testing/venv/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.\ |
| warnings.warn('Was asked to gather along dimension 0, but all '\ |
| ***** Running Evaluation *****\ |
| Num examples = 500\ |
| Batch size = 10\ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| \ |
| Generate config GenerationConfig \{\ |
| "bos_token_id": 0,\ |
| "decoder_start_token_id": 2,\ |
| "early_stopping": true,\ |
| "eos_token_id": 2,\ |
| "length_penalty": 2.0,\ |
| "max_length": 512,\ |
| "min_length": 100,\ |
| "no_repeat_ngram_size": 3,\ |
| "num_beams": 2,\ |
| "pad_token_id": 1,\ |
| "transformers_version": "4.26.1",\ |
| "use_cache": false\ |
| \}\ |
| } |