codefuse-admin commited on
Commit
f32f8fe
·
verified ·
1 Parent(s): dcbc05d

Upload 4 files

Browse files
Files changed (4) hide show
  1. config.json +49 -0
  2. special_tokens_map.json +14 -0
  3. tokenizer.json +0 -0
  4. tokenizer_config.json +28 -0
config.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_dropout": 0.0,
3
+ "architectures": [
4
+ "RodimusForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "attn_cfg": {
8
+ "ffn_expand_ratio": 1.3333333333333333,
9
+ "num_heads": 24,
10
+ "num_heads_k": 1,
11
+ "num_heads_v": 24,
12
+ "rotary_emb_dim": -1,
13
+ "rotary_emb_scale_base": null,
14
+ "window_size": 4096
15
+ },
16
+ "bos_token_id": 126080,
17
+ "d_model": 3072,
18
+ "dropout": 0.0,
19
+ "eos_token_id": 126081,
20
+ "initializer_range": 0.02,
21
+ "max_position_embeddings": 4096,
22
+ "mixer_cfg": {
23
+ "expand_ratio": 2.0,
24
+ "input_gate_low_rank": "auto",
25
+ "mem_size": 64,
26
+ "mode": "fused_chunk",
27
+ "normalize_epsilon": 1e-05
28
+ },
29
+ "model_type": "rodimus",
30
+ "n_layer": 32,
31
+ "norm_epsilon": 1e-05,
32
+ "rescale_prenorm_residual": true,
33
+ "residual_in_fp32": true,
34
+ "tie_word_embeddings": false,
35
+ "torch_dtype": "float32",
36
+ "transformers_version": "4.42.0",
37
+ "use_cache": true,
38
+ "use_fast_path": true,
39
+ "use_fused_cross_entropy": true,
40
+ "use_norm_embedding": false,
41
+ "use_scale_embedding": false,
42
+ "vocab_size": 126464,
43
+ "block_type": "rodimus_plus",
44
+ "auto_map": {
45
+ "AutoConfig": "configuration_rodimus.RodimusConfig",
46
+ "AutoModel": "modeling_rodimus.RodimusForCausalLM",
47
+ "AutoModelForCausalLM": "modeling_rodimus.RodimusForCausalLM"
48
+ }
49
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|fim_begin|>",
4
+ "<|fim_hole|>",
5
+ "<|fim_end|>",
6
+ "<|fim_pad|>",
7
+ "<|repo_name|>",
8
+ "<|file_sep|>"
9
+ ],
10
+ "bos_token": "<|startoftext|>",
11
+ "cls_token": "[CLS]",
12
+ "eos_token": "<|endoftext|>",
13
+ "gmask_token": "[gMASK]"
14
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_eos_token": false,
4
+ "additional_special_tokens": [
5
+ "<|fim_begin|>",
6
+ "<|fim_hole|>",
7
+ "<|fim_end|>",
8
+ "<|fim_pad|>",
9
+ "<|repo_name|>",
10
+ "<|file_sep|>"
11
+ ],
12
+ "bos_token": "<|startoftext|>",
13
+ "clean_up_tokenization_spaces": false,
14
+ "cls_token": "[CLS]",
15
+ "eos_token": "<|endoftext|>",
16
+ "pad_token": "<|endoftext|>",
17
+ "gmask_token": "[gMASK]",
18
+ "merges_file": null,
19
+ "model_max_length": 1000000000000000019884624838656,
20
+ "tokenizer_class": "RodimusTokenizer",
21
+ "vocab_file": null,
22
+ "auto_map": {
23
+ "AutoTokenizer": [
24
+ null,
25
+ "tokenization_rodimus_fast.RodimusTokenizer"
26
+ ]
27
+ }
28
+ }