0925 use custom x_transformers for easy develop
This commit is contained in:
@ -1,8 +1,8 @@
|
||||
defaults:
|
||||
# - nn_params: nb8_embSum_NMT
|
||||
# - nn_params: remi8
|
||||
- nn_params: nb8_embSum_diff_t2m_150M_finetunning
|
||||
# - nn_params: nb8_embSum_diff_t2m_150M_pretraining
|
||||
# - nn_params: nb8_embSum_diff_t2m_150M_finetunning
|
||||
- nn_params: nb8_embSum_diff_t2m_150M_pretrainingv2
|
||||
# - nn_params: nb8_embSum_subPararell
|
||||
# - nn_params: nb8_embSum_diff_t2m_150M
|
||||
|
||||
|
||||
@ -0,0 +1,19 @@
|
||||
encoding_scheme: nb
|
||||
num_features: 8
|
||||
vocab_name: MusicTokenVocabNB
|
||||
model_name: AmadeusModel
|
||||
input_embedder_name: SummationEmbedder
|
||||
main_decoder_name: XtransformerNewPretrainingDecoder
|
||||
sub_decoder_name: DiffusionDecoder
|
||||
model_dropout: 0
|
||||
input_embedder:
|
||||
num_layer: 1
|
||||
num_head: 8
|
||||
main_decoder:
|
||||
dim_model: 768
|
||||
num_layer: 20
|
||||
num_head: 12
|
||||
sub_decoder:
|
||||
decout_window_size: 1 # 1 means no previous decoding output added
|
||||
num_layer: 1
|
||||
feature_enricher_use: False
|
||||
Reference in New Issue
Block a user