textGenerationHyperParameters
The hyperparameters used to configure and optimize the learning process of the base model. You can set any combination of the following hyperparameters for all base models. For more information on each supported hyperparameter, see Optimize the learning process of your text generation models with hyperparameters.
"epochCount"
: The number of times the model goes through the entire training dataset. Its value should be a string containing an integer value within the range of "1" to "10"."batchSize"
: The number of data samples used in each iteration of training. Its value should be a string containing an integer value within the range of "1" to "64"."learningRate"
: The step size at which a model's parameters are updated during training. Its value should be a string containing a floating-point value within the range of "0" to "1"."learningRateWarmupSteps"
: The number of training steps during which the learning rate gradually increases before reaching its target or maximum value. Its value should be a string containing an integer value within the range of "0" to "250".
Here is an example where all four hyperparameters are configured.
{ "epochCount":"5", "learningRate":"0.5", "batchSize": "32", "learningRateWarmupSteps": "10" }