RHQE_aya-expanse-8b_layer_-1 / head_configs.json
ArchSid's picture
Upload folder using huggingface_hub
ff91619 verified
raw
history blame contribute delete
398 Bytes
{"mean_regression": {"name": "mean_regression", "in_size": 4096, "num_outputs": 1, "layer_hook": -1, "hidden_size": 0, "num_layers": 1, "output_activation": "linear", "target": "mean_regression", "is_causal_lm": false, "pred_for_sequence": true, "is_regression": true, "output_bias": false, "loss_fct": "mse", "trainable": true, "loss_weight": 0.002, "ignore_pads": true, "block_gradients": false}}