You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thank you for your outstanding work. I have some questions about the training process of the Bolt model:
When training the Bolt model based on the chronos_datasets with a learning rate of 1e-5 and a batch size of 32, the loss always fluctuates and does not show a downward trend. Are there any techniques to address this?
During the training process, if there are instances where the context is completely identical but the targets are slightly different, loss explosions occur in the instance normalization step because the scale becomes too small (e.g., less than 1e-8).
While retaining the original structure of the Bolt model, if I add some new layers (e.g., an input_embedding_layer), after training, the model completely fails to capture any periodic patterns. Have others encountered a similar situation? Is there any trick to avoid it?
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
-
Thank you for your outstanding work. I have some questions about the training process of the Bolt model:
chronos_datasets
with a learning rate of 1e-5 and a batch size of 32, the loss always fluctuates and does not show a downward trend. Are there any techniques to address this?input_embedding_layer
), after training, the model completely fails to capture any periodic patterns. Have others encountered a similar situation? Is there any trick to avoid it?Beta Was this translation helpful? Give feedback.
All reactions