Boriscii commited on
Commit
82fb7e2
1 Parent(s): 256ca36

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -1
README.md CHANGED
@@ -5,4 +5,31 @@ datasets:
5
  language:
6
  - en
7
  pipeline_tag: sentence-similarity
8
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  language:
6
  - en
7
  pipeline_tag: sentence-similarity
8
+ ---
9
+
10
+ # LayoutLM-Byne
11
+ ## The new SOTA in page retrieval from visually-rich documents.
12
+
13
+ [![Logo](https://armalytix.s3.eu-west-2.amazonaws.com/TRUST+THE+COUNSEL+(1).png "Logo")](https://bynedocs.com "Logo")
14
+
15
+ We're glad to introduce one of the first document page embedding models, LayoutLM-Byne.
16
+
17
+ With the rise of multimodal LLMs, there is a growing adoption of applying models directly to a document without pre-processing it first, as was done before with RAG. This approach is significantly more robust than text-only RAG on a large subset of documents, especially visually rich ones.
18
+
19
+ On the other hand, there is a significant lack of research focused on extracting a relevant page from a PDF or a DOCX document. Most practitioners would parse the page into text and apply regular text embeddings to the text, losing much positional context in the process.
20
+
21
+ LayoutLM [1] is an excellent solution for the problems because, at its core, it is a regular BERT-alike model, but it is uniquely capable of embedding positional information about the text alongside the text itself.
22
+
23
+ We have fine-tuned the model on the DocVQA [2] dataset, far surpassing the current SOTA (all-mpnet-base-v2) [3]:
24
+
25
+ | Model | HR@3 | HR@5 | HR@10 |
26
+ |-------|------|------|-------|
27
+ | all-mpnet-base-v2 (Baseline) | 0.2505 | 0.2941 | 0.3624 |
28
+ | LayoutLM (Our Model) | 0.3159 | 0.3909 | 0.5019 |
29
+ | Relative Improvement | +26.1% | +32.9% | +38.5% |
30
+
31
+ ### Usage
32
+ Please refer to the Colab workbook or the blog post to learn more!
33
+
34
+ ### Get in touch
35
+ Reach out to [[email protected]](mailto:[email protected]) if you'd like help with deploying the model in commerical setting.