all AI news
[R][D] Code to benchmark and train Vision Transformers vs CLIP vs ResNet50 for real time facial emotion recognition.
Dec. 14, 2023, 4:13 p.m. | /u/FallMindless3563
Machine Learning www.reddit.com
[https://blog.oxen.ai/practical-ml-dive-how-to-customize-a-vision-transformer-on-your-own-data/](https://blog.oxen.ai/practical-ml-dive-how-to-customize-a-vision-transformer-on-your-own-data/)
\~ TLDR \~ ViT works the best in this small experiment, with minimal code. The experiment was classifying 7 different facial emotions such as "happy", "sad", "angry", etc...
|Model|Accuracy|
|:-|:-|
|ViT|69%|
|ResNet50|64%|
|Zero-Shot CLIP|53%|
Was honestly most …
benchmark benchmarking clip code dataset emotion fine-tuning hey machinelearning papers practical recognition resnet50 train transformers vision vision transformers vit
More from www.reddit.com / Machine Learning
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Data Engineer (m/f/d)
@ Project A Ventures | Berlin, Germany
Principle Research Scientist
@ Analog Devices | US, MA, Boston