all AI news
[P] Open type Named Entity Recognition with Transformer Encoder
Feb. 11, 2024, 9:37 a.m. | /u/Substantial-Push-179
Machine Learning www.reddit.com
I'd like to share our project on open-type Named Entity Recognition (NER). Our model uses a transformer encoder (BERT-like), making the computation overhead very minimal compared to use of LLMs. I've developed a demo that runs on CPU on Google Colab.
Colab Demo: [https://colab.research.google.com/drive/1mhalKWzmfSTqMnR0wQBZvt9-ktTsATHB?usp=sharing](https://colab.research.google.com/drive/1mhalKWzmfSTqMnR0wQBZvt9-ktTsATHB?usp=sharing)
Code: [https://github.com/urchade/GLiNER](https://github.com/urchade/GLiNER)
Paper: [https://arxiv.org/abs/2311.08526](https://arxiv.org/abs/2311.08526)
bert colab computation cpu demo encoder google llms machinelearning making ner project recognition transformer transformer encoder type
More from www.reddit.com / Machine Learning
[P] Table Extraction , Text Extraction
1 day, 1 hour ago |
www.reddit.com
[D] Kolmogorov-Arnold Network is just an MLP
1 day, 3 hours ago |
www.reddit.com
[D] Why Gemma has such crazy big MLP hidden dim size?
1 day, 4 hours ago |
www.reddit.com
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Consultant - Artificial Intelligence & Data (Google Cloud Data Engineer) - MY / TH
@ Deloitte | Kuala Lumpur, MY