Item request has been placed! ×
Item request cannot be made. ×
loading  Processing Request

Language Models for Document Understanding ; Modèles de Langages pour les Compréhension de Documents

Item request has been placed! ×
Item request cannot be made. ×
loading   Processing Request
  • معلومة اضافية
    • Contributors:
      Laboratoire d'InfoRmatique en Image et Systèmes d'information (LIRIS); Université Lumière - Lyon 2 (UL2)-École Centrale de Lyon (ECL); Université de Lyon-Université de Lyon-Université Claude Bernard Lyon 1 (UCBL); Université de Lyon-Institut National des Sciences Appliquées de Lyon (INSA Lyon); Université de Lyon-Institut National des Sciences Appliquées (INSA)-Institut National des Sciences Appliquées (INSA)-Centre National de la Recherche Scientifique (CNRS); Esker SA; Extraction de Caractéristiques et Identification (imagine); Université de Lyon-Institut National des Sciences Appliquées (INSA)-Institut National des Sciences Appliquées (INSA)-Centre National de la Recherche Scientifique (CNRS)-Université Lumière - Lyon 2 (UL2)-École Centrale de Lyon (ECL); CIFRE with Esker SA; INSA LYON; Christophe Garcia; Stefan Duffner
    • بيانات النشر:
      HAL CCSD
    • الموضوع:
      2023
    • Collection:
      HAL Lyon 1 (University Claude Bernard Lyon 1)
    • نبذة مختصرة :
      Every day, an uncountable amount of documents are received and processed by com-panies worldwide. In an effort to reduce the cost of processing each document, thelargest companies have resorted to document automation technologies. In an idealworld, a document can be automatically processed without any human intervention:its content is read, and information is extracted and forwarded to the relevant ser-vice. The state-of-the-art techniques have quickly evolved in the last decades, fromrule-based algorithms to statistical models. This thesis focuses on machine learningmodels for document information extraction.Recent advances in model architecture for natural language processing haveshown the importance of the attention mechanism. Transformers have revolution-ized the field by generalizing the use of attention and by pushing self-supervisedpre-training to the next level. In the first part, we confirm that transformers withappropriate pre-training were able to perform document understanding tasks withhigh performance. We show that, when used as a token classifier for informationextraction, transformers are able to exceptionally efficiently learn the task comparedto recurrent networks. Transformers only need a small proportion of the trainingdata to reach close to maximum performance. This highlights the importance ofself-supervised pre-training for future fine-tuning.In the following part, we design specialized pre-training tasks, to better preparethe model for specific data distributions such as business documents. By acknowl-edging the specificities of business documents such as their table structure and theirover-representation of numeric figures, we can target specific skills useful for themodel in its future tasks. We show that those new tasks improve the model’s down-stream performances, even with small models. Using this pre-training approach,we are able to reach the performances of significantly bigger models without anyadditional cost during finetuning or inference.Finally, in the last part, we address ...
    • Relation:
      tel-04459378; https://hal.science/tel-04459378; https://hal.science/tel-04459378/document; https://hal.science/tel-04459378/file/final_manuscript.pdf
    • Rights:
      info:eu-repo/semantics/OpenAccess
    • الرقم المعرف:
      edsbas.87C16D00