all AI news
How to add additional module to BERT architecture, then load the original weight and use it
May 18, 2022, 1:40 a.m. | /u/scp-8989
Natural Language Processing www.reddit.com
1. Add an additional module to BERT architecture (huggingface’s transformers)
2. Load the BERT’s weight to the BERT model with new architecture
3. Then use BERT directly or continue train BERT
I’m very confused how to do it. Since we usually use `from_pretrained` directly to load the model (both weight and architecture) from huggingface.
In more detail, I’m working on both [prajjwal1/bert-tiny](https://huggingface.co/prajjwal1/bert-tiny) and [bert-base-uncased](https://huggingface.co/bert-base-uncased).
More from www.reddit.com / Natural Language Processing
What Do You Love About NLP?
22 hours ago |
www.reddit.com
How to Install and Deploy LLaMA 3 Into Production
1 day, 15 hours ago |
www.reddit.com
The Languages AI Is Leaving Behind
4 days, 14 hours ago |
www.reddit.com
Feeling so inferior in the NLP job market.
5 days, 11 hours ago |
www.reddit.com
NLP: building a sentiment model
5 days, 12 hours ago |
www.reddit.com
ReFT: Representation Finetuning for Language Models
1 week, 6 days ago |
www.reddit.com
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Data Engineer
@ Parker | New York City
Sr. Data Analyst | Home Solutions
@ Three Ships | Raleigh or Charlotte, NC