prithivMLmods commited on
Commit
bc99b31
·
verified ·
1 Parent(s): 0fd1195

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,17 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Gliese-OCR-7B-Post2.0-final.mmproj-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Gliese-OCR-7B-Post2.0-final.mmproj-f16.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Gliese-OCR-7B-Post2.0-final.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Gliese-OCR-7B-Post2.0-final.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Gliese-OCR-7B-Post2.0-final.f16.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Gliese-OCR-7B-Post2.0-final.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Gliese-OCR-7B-Post2.0-final.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Gliese-OCR-7B-Post2.0-final.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
44
+ Gliese-OCR-7B-Post2.0-final.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Gliese-OCR-7B-Post2.0-final.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Gliese-OCR-7B-Post2.0-final.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Gliese-OCR-7B-Post2.0-final.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
48
+ Gliese-OCR-7B-Post2.0-final.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
49
+ Gliese-OCR-7B-Post2.0-final.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
Gliese-OCR-7B-Post2.0-final.IQ4_XS.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21cc56531094367648779804f8a1fbc567a3420fd867606e70a81c29b9fa57fc
3
+ size 4250298144
Gliese-OCR-7B-Post2.0-final.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fa7de541a6733a5875edfc430cf35f610394bcde13a16750673db88044cc35b
3
+ size 3015939872
Gliese-OCR-7B-Post2.0-final.Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:812422572a418ba4d09f9799998482856a27e7728f62a74b45b556424a7a7156
3
+ size 4088459040
Gliese-OCR-7B-Post2.0-final.Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:536b7a0ff6eb9c7751edd3da70ce9d4987d72150f614669ae1e0e975e90a0f21
3
+ size 3808390944
Gliese-OCR-7B-Post2.0-final.Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:538adb37c0a74c081dddd66d1aea8930a2460edc48e0152e1bffe1a5110787ed
3
+ size 3492368160
Gliese-OCR-7B-Post2.0-final.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e4c01e6ff507fde19026688c3b000171581b1eedfa26b46ea434d2e4adbcbc9
3
+ size 4683073312
Gliese-OCR-7B-Post2.0-final.Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16931eeb0395ea48d1a2141bb34ed061e077158ec8f360e241ce16ccc354abb7
3
+ size 4457768736
Gliese-OCR-7B-Post2.0-final.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbdcb3675764fcfea73a83b88a83ae72882feaf0f0d363a08b7498c5d2366ccf
3
+ size 5444831008
Gliese-OCR-7B-Post2.0-final.Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:993603d8a5b5dc25c78e5b61a3101cfded0e6bd3bb1cf00b75abbce846359580
3
+ size 5315176224
Gliese-OCR-7B-Post2.0-final.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f124306cb18c370745aa3c633b804ad37dfbf667b38d0cc6cbbc3038eff08ba0
3
+ size 6254198560
Gliese-OCR-7B-Post2.0-final.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0c15ada2f3b3c1e3cea9167f3c32f14db699a78b0698705a13acdc74d97513a
3
+ size 8098524960
Gliese-OCR-7B-Post2.0-final.f16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d17b1d857a8ef69685f7a879569ca0d7c355f903cdd4c9e2237c8c2e21e141b
3
+ size 15237852960
Gliese-OCR-7B-Post2.0-final.mmproj-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8c476e359e7c6471908f4a38abb43511ee0d1e86b8a07052d281e3ffccdddf6
3
+ size 856131072
Gliese-OCR-7B-Post2.0-final.mmproj-f16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a01b2fb90bde7c7ee956047bce7674b9438dd6dfb14bb08256e264595b998da6
3
+ size 1354163712
README.md CHANGED
@@ -1,11 +1,96 @@
1
  ---
2
- license: apache-2.0
 
 
 
 
 
3
  language:
4
  - en
5
- base_model:
6
- - prithivMLmods/Gliese-OCR-7B-Post2.0-final
7
- pipeline_tag: image-text-to-text
8
  library_name: transformers
 
 
 
 
9
  tags:
 
10
  - text-generation-inference
11
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model: prithivMLmods/Gliese-OCR-7B-Post2.0-final
3
+ datasets:
4
+ - prithivMLmods/OpenDoc-Pdf-Preview
5
+ - prithivMLmods/Opendoc1-Analysis-Recognition
6
+ - allenai/olmOCR-mix-0225
7
+ - prithivMLmods/Openpdf-Analysis-Recognition
8
  language:
9
  - en
 
 
 
10
  library_name: transformers
11
+ license: apache-2.0
12
+ mradermacher:
13
+ readme_rev: 1
14
+ quantized_by: mradermacher
15
  tags:
16
+ - trl
17
  - text-generation-inference
18
+ - Document
19
+ - VLM
20
+ - KIE
21
+ - OCR
22
+ - VL
23
+ - Camel
24
+ - Openpdf
25
+ - Extraction
26
+ - Linking
27
+ - Markdown
28
+ - .Md
29
+ - Document Digitization
30
+ - Intelligent Document Processing (IDP)
31
+ - Intelligent Word Recognition (IWR)
32
+ ---
33
+ ## About
34
+
35
+ <!-- ### quantize_version: 2 -->
36
+ <!-- ### output_tensor_quantised: 1 -->
37
+ <!-- ### convert_type: hf -->
38
+ <!-- ### vocab_type: -->
39
+ <!-- ### tags: -->
40
+ <!-- ### quants: x-f16 Q4_K_S Q2_K Q8_0 Q6_K Q3_K_M Q3_K_S Q3_K_L Q4_K_M Q5_K_S Q5_K_M IQ4_XS -->
41
+ <!-- ### quants_skip: -->
42
+ <!-- ### skip_mmproj: -->
43
+ static quants of https://huggingface.co/prithivMLmods/Gliese-OCR-7B-Post2.0-final
44
+
45
+ <!-- provided-files -->
46
+
47
+ ***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Gliese-OCR-7B-Post2.0-final-GGUF).***
48
+
49
+ weighted/imatrix quants are available at https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-i1-GGUF
50
+ ## Usage
51
+
52
+ If you are unsure how to use GGUF files, refer to one of [TheBloke's
53
+ READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
54
+ more details, including on how to concatenate multi-part files.
55
+
56
+ ## Provided Quants
57
+
58
+ (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
59
+
60
+ | Link | Type | Size/GB | Notes |
61
+ |:-----|:-----|--------:|:------|
62
+ | [GGUF](https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-GGUF/resolve/main/Gliese-OCR-7B-Post2.0-final.mmproj-Q8_0.gguf) | mmproj-Q8_0 | 1.0 | multi-modal supplement |
63
+ | [GGUF](https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-GGUF/resolve/main/Gliese-OCR-7B-Post2.0-final.mmproj-f16.gguf) | mmproj-f16 | 1.5 | multi-modal supplement |
64
+ | [GGUF](https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-GGUF/resolve/main/Gliese-OCR-7B-Post2.0-final.Q2_K.gguf) | Q2_K | 3.1 | |
65
+ | [GGUF](https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-GGUF/resolve/main/Gliese-OCR-7B-Post2.0-final.Q3_K_S.gguf) | Q3_K_S | 3.6 | |
66
+ | [GGUF](https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-GGUF/resolve/main/Gliese-OCR-7B-Post2.0-final.Q3_K_M.gguf) | Q3_K_M | 3.9 | lower quality |
67
+ | [GGUF](https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-GGUF/resolve/main/Gliese-OCR-7B-Post2.0-final.Q3_K_L.gguf) | Q3_K_L | 4.2 | |
68
+ | [GGUF](https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-GGUF/resolve/main/Gliese-OCR-7B-Post2.0-final.IQ4_XS.gguf) | IQ4_XS | 4.4 | |
69
+ | [GGUF](https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-GGUF/resolve/main/Gliese-OCR-7B-Post2.0-final.Q4_K_S.gguf) | Q4_K_S | 4.6 | fast, recommended |
70
+ | [GGUF](https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-GGUF/resolve/main/Gliese-OCR-7B-Post2.0-final.Q4_K_M.gguf) | Q4_K_M | 4.8 | fast, recommended |
71
+ | [GGUF](https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-GGUF/resolve/main/Gliese-OCR-7B-Post2.0-final.Q5_K_S.gguf) | Q5_K_S | 5.4 | |
72
+ | [GGUF](https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-GGUF/resolve/main/Gliese-OCR-7B-Post2.0-final.Q5_K_M.gguf) | Q5_K_M | 5.5 | |
73
+ | [GGUF](https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-GGUF/resolve/main/Gliese-OCR-7B-Post2.0-final.Q6_K.gguf) | Q6_K | 6.4 | very good quality |
74
+ | [GGUF](https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-GGUF/resolve/main/Gliese-OCR-7B-Post2.0-final.Q8_0.gguf) | Q8_0 | 8.2 | fast, best quality |
75
+ | [GGUF](https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-GGUF/resolve/main/Gliese-OCR-7B-Post2.0-final.f16.gguf) | f16 | 15.3 | 16 bpw, overkill |
76
+
77
+ Here is a handy graph by ikawrakow comparing some lower-quality quant
78
+ types (lower is better):
79
+
80
+ ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png)
81
+
82
+ And here are Artefact2's thoughts on the matter:
83
+ https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
84
+
85
+ ## FAQ / Model Request
86
+
87
+ See https://huggingface.co/mradermacher/model_requests for some answers to
88
+ questions you might have and/or if you want some other model quantized.
89
+
90
+ ## Thanks
91
+
92
+ I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
93
+ me use its servers and providing upgrades to my workstation to enable
94
+ this work in my free time.
95
+
96
+ <!-- end -->