LG-AI-EXAONE commited on
Commit
ee43914
·
1 Parent(s): c3ab95d

Update HuggingFace transformers support

Browse files
Files changed (3) hide show
  1. README.md +2 -8
  2. config.json +33 -1
  3. generation_config.json +1 -1
README.md CHANGED
@@ -18,6 +18,7 @@ library_name: transformers
18
  <img src="assets/EXAONE_Symbol+BI_3d.png", width="300", style="margin: 40 auto;">
19
  🎉 License Updated! We are pleased to announce our more flexible licensing terms 🤗
20
  <br>✈️ Try on <a href="https://friendli.ai/suite/~/serverless-endpoints/LGAI-EXAONE/EXAONE-4.0-32B/overview">FriendliAI</a>
 
21
  <br>
22
 
23
  # EXAONE-4.0-1.2B
@@ -48,14 +49,7 @@ For more details, please refer to our [technical report](https://arxiv.org/abs/2
48
 
49
  ## Quickstart
50
 
51
- You should install the transformers library forked from the original, available in our [PR](https://github.com/huggingface/transformers/pull/39129).
52
- Once this PR is merged and released, we will update this section.
53
-
54
- You can install the latest version of transformers with support for EXAONE 4.0 by following the command:
55
-
56
- ```bash
57
- pip install git+https://github.com/lgai-exaone/transformers@add-exaone4
58
- ```
59
 
60
  ### Non-reasoning mode
61
 
 
18
  <img src="assets/EXAONE_Symbol+BI_3d.png", width="300", style="margin: 40 auto;">
19
  🎉 License Updated! We are pleased to announce our more flexible licensing terms 🤗
20
  <br>✈️ Try on <a href="https://friendli.ai/suite/~/serverless-endpoints/LGAI-EXAONE/EXAONE-4.0-32B/overview">FriendliAI</a>
21
+ <br><br><i>📢 EXAONE 4.0 is officially supported by HuggingFace transformers! Please check out the guide <a href="#quickstart">below</a></i>
22
  <br>
23
 
24
  # EXAONE-4.0-1.2B
 
49
 
50
  ## Quickstart
51
 
52
+ You should install the transformers library with version >= `4.54.0`.
 
 
 
 
 
 
 
53
 
54
  ### Non-reasoning mode
55
 
config.json CHANGED
@@ -10,6 +10,38 @@
10
  "hidden_size": 2048,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 4096,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  "max_position_embeddings": 65536,
14
  "model_type": "exaone4",
15
  "num_attention_heads": 32,
@@ -29,7 +61,7 @@
29
  "sliding_window_pattern": null,
30
  "tie_word_embeddings": true,
31
  "torch_dtype": "bfloat16",
32
- "transformers_version": "4.54.0.dev0",
33
  "use_cache": false,
34
  "vocab_size": 102400
35
  }
 
10
  "hidden_size": 2048,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 4096,
13
+ "layer_types": [
14
+ "full_attention",
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention"
44
+ ],
45
  "max_position_embeddings": 65536,
46
  "model_type": "exaone4",
47
  "num_attention_heads": 32,
 
61
  "sliding_window_pattern": null,
62
  "tie_word_embeddings": true,
63
  "torch_dtype": "bfloat16",
64
+ "transformers_version": "4.54.0",
65
  "use_cache": false,
66
  "vocab_size": 102400
67
  }
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "bos_token_id": 1,
4
  "eos_token_id": 361,
5
  "pad_token_id": 0,
6
- "transformers_version": "4.54.0.dev0"
7
  }
 
3
  "bos_token_id": 1,
4
  "eos_token_id": 361,
5
  "pad_token_id": 0,
6
+ "transformers_version": "4.54.0"
7
  }