nielsr HF staff commited on
Commit
9b08c37
1 Parent(s): 292af50

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +47 -0
README.md ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ tags:
4
+ - vision
5
+ ---
6
+
7
+ # LiLT-RoBERTa (base-sized model)
8
+
9
+ Language-Independent Layout Transformer - RoBERTa model by stitching a pre-trained RoBERTa (English) and a pre-trained Language-Independent Layout Transformer (LiLT) together. It was introduced in the paper [LiLT: A Simple yet Effective Language-Independent Layout Transformer for Structured Document Understanding](https://arxiv.org/abs/2202.13669) by Wang et al. and first released in [this repository](https://github.com/jpwang/lilt).
10
+
11
+ Disclaimer: The team releasing LiLT did not write a model card for this model so this model card has been written by the Hugging Face team.
12
+
13
+ ## Model description
14
+
15
+ The Language-Independent Layout Transformer (LiLT) allows to combine any pre-trained RoBERTa encoder from the hub (hence, in any language) with a lightweight Layout Transformer to have a LayoutLM-like model for any language.
16
+
17
+ <img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/model_doc/lilt_architecture.jpg" alt="drawing" width="600"/>
18
+
19
+ ## Intended uses & limitations
20
+
21
+ The model is meant to be fine-tuned on tasks like document image classification, document parsing and document QA. See the [model hub](https://huggingface.co/models?search=lilt) to look for fine-tuned versions on a task that interests you.
22
+
23
+ ### How to use
24
+
25
+ For code examples, we refer to the [documentation](https://huggingface.co/transformers/main/model_doc/lilt.html).
26
+
27
+ ### BibTeX entry and citation info
28
+
29
+ ```bibtex
30
+ @misc{https://doi.org/10.48550/arxiv.2202.13669,
31
+ doi = {10.48550/ARXIV.2202.13669},
32
+
33
+ url = {https://arxiv.org/abs/2202.13669},
34
+
35
+ author = {Wang, Jiapeng and Jin, Lianwen and Ding, Kai},
36
+
37
+ keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
38
+
39
+ title = {LiLT: A Simple yet Effective Language-Independent Layout Transformer for Structured Document Understanding},
40
+
41
+ publisher = {arXiv},
42
+
43
+ year = {2022},
44
+
45
+ copyright = {arXiv.org perpetual, non-exclusive license}
46
+ }
47
+ ```