March 21, 2024, 12:25 p.m. | /u/PipeTrance

Data Science www.reddit.com

We (a small startup) have recently seen considerable success fine-tuning LLMs (primarily OpenAI models) to generate data explorations and reports based on user requests. We provide relevant details of data schema as input and expect the LLM to generate a response written in our custom domain-specific language, which we then convert into a UI exploration.

We've shared more details in a blog post: https://www.supersimple.io/blog/gpt-4-fine-tuning-early-access

I'm curious if anyone has explored similar approaches in other domains or perhaps used entirely different …

data datascience domain expect fine-tuning generate gpt gpt-4 language llm llms openai openai models reports schema small startup success

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Intern Large Language Models Planning (f/m/x)

@ BMW Group | Munich, DE

Data Engineer Analytics

@ Meta | Menlo Park, CA | Remote, US