{ "type": "sequential", "layers": "Sequential(\n (0): Linear(in_features=1024, out_features=256, bias=True)\n (1): ReLU()\n (2): Dropout(p=0.1, inplace=False)\n (3): Linear(in_features=256, out_features=1, bias=True)\n)", "num_labels": 1, "hidden_size": 1024 }