MEANOTEK Building Gapping Resolution System Overnight

Slide Note
Embed
Share

Explore the journey of Denis Tarasov, Tatyana Matveeva, and Nailia Galliulina in developing a system for gapping resolution in computational linguistics. The goal is to test a rapid NLP model prototyping system for a novel task, driven by the motivation to efficiently build NLP models for various problems. Utilizing character-level embeddings and LSTM language models, they address challenges such as maintainability and understanding in model improvement.


Uploaded on Nov 16, 2024 | 0 Views


Download Presentation

Please find below an Image/Link to download the presentation.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author. Download presentation by click this link. If you encounter any issues during the download, it is possible that the publisher has removed the file from their server.

E N D

Presentation Transcript


  1. MEANOTEK Building gapping resolution system overnight: Lessons Learned Denis Tarasov, Tatyana Matveeva, Nailia Galliulina Denis Tarasov, Tatyana Matveeva, Nailia Galliulina Dialogue 2019 international conference on computational linguistics Email for correspondence: dtarasov@meanotek.io

  2. THE GOAL Test of NLP rapid model prototyping system on novel type of the task

  3. MOTIVATION The need to quickly and reliably build NLP models in large quantities for different types of problems The need for techology to be extensible and improvable

  4. FIRST REQUIREMENT The need to quickly and reliably build NLP models in large quantities for different types of problems

  5. SECOND REQUIREMENT The usual way to quickly obtain competive result is to find out current SOTA model, get its code from github, adapt it, if necessary or just train on new data

  6. SECOND REQUIREMENT PROBLEM #1: This leads to unmaintainable software code when combined into complex pipelines

  7. SECOND REQUIREMENT PROBLEM NUMBER 2: We cannot improve things that we do not really understand We don t really understand things that we can t duplicate ourselves Copying someone s else research puts us in position of forever catching up party

  8. METHODS Character level context sensetive embeddeings based on language model Model parameters: 3192*2048*2048 LSTM language model trained on 2.2 GB of text (cleaned common crawl+books dataset) with the goal of predicting next character. Long BPTT length 350 characters

  9. SIMPLIFICATIONS Task is considered to be sequence labeling task Position of V is start of R2 Gapping is present if R2 is present

  10. MODEL OVERVIEW Softmax LSTM 256 LSTM 256 The cat sits on mat LSTM 2048 Pre-trained Part (fixed) LSTM 2048 LSTM 3192 Character embeddings, size 50

  11. NeuThink Library Model definition using expression trees syntax Automatic generation of inference and training code Automatic guessing of suitable hyperparameters

  12. RESULTS

  13. DISCUSSION Need to extend system desing with new format converstion tools, to assist conversion from/to various data format types, since this seems to be main failure mode now Interesting that character-level models can form representations that are useful for representing long-distance relations Overall, results are sensible, given the time constraint

  14. NOTES ON COMPETITIONS ORGANIZATION Automatic scoring during competition would be nice to have Standartization of formats and eval scripts Clear and consistent policy on after-deadline submissions

  15. THANK YOU FOR YOUR ATTENTION

  16. APPENDIX 1. How NeuThink differential programming model works

Related


More Related Content