AliceThirty commited on
Commit
19538ae
1 Parent(s): 595991f

Support gated models

Browse files
Files changed (1) hide show
  1. index.html +30 -6
index.html CHANGED
@@ -76,19 +76,24 @@
76
  "Q8_0": 8.5,
77
  }
78
 
79
- async function modelConfig(hf_model) {
 
 
 
 
 
80
  let config = await fetch(
81
- `https://huggingface.co/${hf_model}/raw/main/config.json`
82
  ).then(r => r.json())
83
  let model_size = 0
84
  try {
85
- model_size = (await fetch(`https://huggingface.co/${hf_model}/resolve/main/model.safetensors.index.json`).then(r => r.json()))["metadata"]["total_size"] / 2
86
  if (isNaN(model_size)) {
87
  throw new Erorr("no size in safetensors metadata")
88
  }
89
  } catch (e) {
90
  try {
91
- model_size = (await fetch(`https://huggingface.co/${hf_model}/resolve/main/pytorch_model.bin.index.json`).then(r => r.json()))["metadata"]["total_size"] / 2
92
  if (isNaN(model_size)) {
93
  throw new Erorr("no size in pytorch metadata")
94
  }
@@ -157,7 +162,7 @@
157
 
158
  async function calculateSizes(format) {
159
  try {
160
- const model_config = await modelConfig(document.getElementById("modelsearch").value)
161
  const context = parseInt(document.getElementById("contextsize").value)
162
  let bsz = 512
163
  let cache_bit = 16
@@ -204,10 +209,29 @@
204
  LLM Model, Can I run it?
205
  </h1>
206
  <p>
207
- *This does not support gated or private repos
208
  </p>
 
209
  <div class="flex flex-col gap-10">
210
  <div class="w-auto flex flex-col gap-4">
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
211
  <!-- GPU Selector -->
212
  <div
213
  class="relative"
 
76
  "Q8_0": 8.5,
77
  }
78
 
79
+ async function modelConfig(hf_model, hf_token) {
80
+ auth = hf_token == "" ? {} : {
81
+ headers: {
82
+ 'Authorization': `Bearer ${hf_token}`
83
+ }
84
+ }
85
  let config = await fetch(
86
+ `https://huggingface.co/${hf_model}/raw/main/config.json`, auth
87
  ).then(r => r.json())
88
  let model_size = 0
89
  try {
90
+ model_size = (await fetch(`https://huggingface.co/${hf_model}/resolve/main/model.safetensors.index.json`, auth).then(r => r.json()))["metadata"]["total_size"] / 2
91
  if (isNaN(model_size)) {
92
  throw new Erorr("no size in safetensors metadata")
93
  }
94
  } catch (e) {
95
  try {
96
+ model_size = (await fetch(`https://huggingface.co/${hf_model}/resolve/main/pytorch_model.bin.index.json`, auth).then(r => r.json()))["metadata"]["total_size"] / 2
97
  if (isNaN(model_size)) {
98
  throw new Erorr("no size in pytorch metadata")
99
  }
 
162
 
163
  async function calculateSizes(format) {
164
  try {
165
+ const model_config = await modelConfig(document.getElementById("modelsearch").value, document.getElementById("hf_token").value)
166
  const context = parseInt(document.getElementById("contextsize").value)
167
  let bsz = 512
168
  let cache_bit = 16
 
209
  LLM Model, Can I run it?
210
  </h1>
211
  <p>
212
+ To support gated or private repos, you need to <a href="https://huggingface.co/settings/tokens" style="color: #4444ff"><b>create an authentification token</b></a>, to check the box <span style="color: #6e1818"><b>"Read access to contents of all public gated repos you can access"</b></span> and then enter the token in the field below.
213
  </p>
214
+
215
  <div class="flex flex-col gap-10">
216
  <div class="w-auto flex flex-col gap-4">
217
+ <!-- Huggingface Authentification Token -->
218
+ <div
219
+ class="relative"
220
+ x-data="{
221
+ results: null,
222
+ query: null
223
+ }"
224
+ >
225
+ <label
226
+ for="gpusearch"
227
+ class="absolute -top-2 left-2 inline-block bg-white px-1 text-xs font-medium text-gray-900"
228
+ >Huggingface Token (optional)</label
229
+ >
230
+ <input
231
+ class="block w-full rounded-md border-0 p-3 text-gray-900 shadow-sm ring-1 ring-inset ring-gray-300 placeholder:text-gray-400 focus:ring-2 focus:ring-inset focus:ring-indigo-600 sm:text-sm sm:leading-6"
232
+ id="hf_token"
233
+ />
234
+ </div>
235
  <!-- GPU Selector -->
236
  <div
237
  class="relative"