Edit model card

decision_transformer_1

This model is a fine-tuned version of on the city_learn dataset.

Model description

state_mean = np.array([6.52472527e+00, 4.00000000e+00, 1.25000000e+01, 1.68241415e+01, 1.68242216e+01, 1.68249313e+01, 1.68268315e+01, 7.29934753e+01, 7.29969093e+01, 7.29977106e+01, 7.29979396e+01, 2.08098329e+02, 2.08098329e+02, 2.07998283e+02, 2.08040522e+02, 2.01204785e+02, 2.01204785e+02, 2.00978709e+02, 2.01073375e+02, 1.56447270e-01, 1.06496225e+00, 6.98845768e-01, 2.90539899e-01, 4.02466726e-01, 2.73094091e-01, 2.73094091e-01, 2.73094091e-01, 2.73094091e-01]) state_std = np.array([3.45249551e+00, 2.00000100e+00, 6.92218755e+00, 3.55839049e+00, 3.55843321e+00, 3.55972060e+00, 3.56299330e+00, 1.64936264e+01, 1.64957718e+01, 1.64978640e+01, 1.65000009e+01, 2.92600647e+02, 2.92600647e+02, 2.92543689e+02, 2.92592247e+02, 2.96262436e+02, 2.96262436e+02, 2.96151575e+02, 2.96175911e+02, 3.53418023e-02, 8.88195655e-01, 1.01691038e+00, 3.23315111e-01, 9.21189104e-01, 1.17759695e-01, 1.17759695e-01, 1.17759695e-01, 1.17759695e-01])

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0001
  • train_batch_size: 64
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_ratio: 0.1
  • num_epochs: 500

Training results

Framework versions

  • Transformers 4.26.1
  • Pytorch 1.13.1+cu116
  • Datasets 2.10.0
  • Tokenizers 0.13.2
Downloads last month
2
Inference API
Unable to determine this model’s pipeline type. Check the docs .