all AI news
SATformer: Transformer-Based UNSAT Core Learning
March 13, 2024, 4:43 a.m. | Zhengyuan Shi (The Chinese University of Hong Kong), Min Li (The Chinese University of Hong Kong), Yi Liu (The Chinese University of Hong Kong), Sadaf
cs.LG updates on arXiv.org arxiv.org
Abstract: This paper introduces SATformer, a novel Transformer-based approach for the Boolean Satisfiability (SAT) problem. Rather than solving the problem directly, SATformer approaches the problem from the opposite direction by focusing on unsatisfiability. Specifically, it models clause interactions to identify any unsatisfiable sub-problems. Using a graph neural network, we convert clauses into clause embeddings and employ a hierarchical Transformer-based model to understand clause correlation. SATformer is trained through a multi-task learning approach, using the single-bit satisfiability …
abstract arxiv core cs.ai cs.lg cs.lo graph graph neural network identify interactions network neural network novel paper transformer type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Senior Machine Learning Engineer
@ Samsara | Canada - Remote