I played around with GPT-3 to build this demo. Select a public BigQuery dataset and describe your query in natural English, then edit the generated SQL as needed and execute it.
This demo sent us on a warpath today. We have a fairly clean SQL schema for which we need to craft a lot of queries that handle things like business logic, reporting and configuration.
If we could get even 50% success rate on a reasonable starting point for the generated SQL each time, that would be the biggest value-add our organization has ever seen.
I think our use case is compelling because we have to implement the same SQL targets for every customer. The only variations are typically customer-specific parameters/codes/etc.
We also have a huge corpus of examples to pull from for training data.
We are thinking about initially implementing some higher order views/functions in our SQL dialect to make things easier on ourselves with the GPT model. Complex joins across many tables seems to be something that would still elude these techniques. Most of our joins are of a very particular shape, so we can abstract the super nasty stuff away.
Worst case scenario, this concludes like my cynical mind assumes it will, but I am open to being surprised this time. We aren't going to put everything behind this, more of a "if it works..." kind of 1-2 week experiment.
I briefly worked on a startup to commercialize this tech, but we decided it wasn't accurate enough to be useful. It was very cool when it actually worked. If you can only produce what you want half the time on simple queries, that doesn't seem very useful to me though.
Have you considered using templates? can you elaborate more on why they can't be used (I guess there is no way to cleanly separate parameters but i may be wrong)
At Veezoo (http://www.veezoo.com) we have been tackling this problem for over 5 years now.
Under the hood we are using our own models. With GPT-3 we're a bit worried about the lack of fine-grained control needed for productive use-cases and obv. also lock in.
Will try out against the same dataset and see how it compares!
Nice product, will give it a spin. For the banking market. Do you have an API exposed as well to feed questions and get (json) data? Our clients would definitely need it integrated fully.
Nice! Yeah I have no doubt that a specialized model could beat this general one, although I find the output from the general one to be uncanny at times. Would love to hear your expert opinion on how they compare!
and I'm wondering if:
1. Is there a paid-for version of your app/website where I can plug in a diff dataset?
2. Have you considered sharing the source code for others to recreate and plug in diff datasets?
3. Or, :sweat_smile:, perhaps this is as simple as adding new datasets to the drop-down menu? say, finance data / Bitcoin transactions data?
For other public BigQuery datasets I’d certainly consider just adding them to the demo for free—which specific Bitcoin/finance datasets did you have in mind?
For private datasets, we’re looking at adding that functionality to the core Tabby service (that’s the SaaS this relates to). Please email for info!
We did a similar analysis a while back, works surprisingly well! I OpenAI increases the amount of "training data" you can send in, I think it could get really good at generating SQL.
Yeah many rough edges indeed. The generated SQL is the plain output from GPT-3; I have not done anything to customize the model or validate syntax outside it, so the roughness is expected. No idea if folks will find value in this despite that, hence the demo.
/*
Describe the data in the all_geoclustered table.
*/
SELECT
COUNT(*) AS total_rows,
COUNT(*) FILTER (WHERE date IS NOT NULL) AS total_rows_with_date,
COUNT(*) FILTER (WHERE wban IS NOT NULL) AS total_rows_with_wban,
COUNT(*) FILTER (WHERE stn IS NOT NULL) AS total_rows_with_stn,
COUNT(*) FILTER (WHERE year IS NOT NULL) AS total_rows_with_year,
COUNT(*) FILTER (WHERE mo IS NOT NULL) AS total_rows_with_mo,
COUNT(*) FILTER (WHERE da IS NOT NULL) AS total_rows_with_da,
COUNT(*) FILTER (WHERE temp IS NOT NULL) AS total_rows_with_temp,
COUNT(*) FILTER (WHERE count_temp IS NOT NULL) AS total_rows_with_count_temp,
COUNT(*) FILTER (WHERE dewp IS NOT NULL) AS total_rows_with_dewp,
COUNT(*) FILTER (WHERE count_dewp IS NOT NULL) AS total_rows_with_count_dewp,
COUNT(*) FILTER (WHERE slp IS NOT NULL) AS total_rows_with_slp,
COUNT(*) FILTER (WHERE count_slp IS NOT NULL) AS total_rows_with_count_slp,
COUNT(*) FILTER (WHERE stp IS NOT NULL) AS total_rows_with_stp,
COUNT(*) FILTER (WHERE count_stp IS NOT NULL) AS total_rows_with_count_stp,
COUNT(*) FILTER (WHERE visib IS NOT NULL) AS total_rows_with_visib,
COUNT(*) FILTER (WHERE count_visib IS NOT NULL) AS total_rows_with_count_visib,
COUNT(*) FILTER (WHERE wdsp IS NOT NULL) AS total_rows_with_wdsp,
COUNT(*) FILTER (WHERE count_wdsp IS NOT NULL) AS total_rows_with_count_wdsp,
COUNT(*) FILTER (WHERE mxpsd IS NOT NULL
/*
How often do hail and thunder happen together?
*/
SELECT
COUNT(*) AS count,
SUM(IF(hail=1 AND thunder=1, 1, 0)) AS sum_hail_thunder
FROM
`fh-bigquery.weather_gsod.all`
WHERE
year = 2019
AND mo = 11
AND da = 14
AND stn = '725030'
SELECT
date,
wban,
stn,
year,
mo,
da,
temp,
count_temp,
dewp,
count_dewp,
slp,
count_slp,
stp,
count_stp,
visib,
count_visib,
wdsp,
count_wdsp,
mxpsd,
gust,
max,
flag_max,
min,
flag_min,
prcp,
flag_prcp,
sndp,
fog,
rain_drizzle,
snow_ice_pellets,
hail,
thunder,
tornado_funnel_cloud,
usaf,
name,
country,
state,
call,
lat,
lon,
elev,
begin,
end,
point_gis,
fake_date
FROM
`fh-bigquery.weather_gsod.all_geoclustered`
WHERE
lat IS NOT NULL
AND lon IS NOT NULL
AND lat != 0
AND lon != 0
AND lat != 1
AND lon != 1
AND lat != -1
AND lon != -1
AND lat != 2
AND lon != 2
AND lat != -2
AND lon != -2
AND lat != 3
AND lon != 3
AND lat != -3
AND lon != -3
AND lat != 4
AND lon != 4
AND lat != -4
AND lon != -4
AND lat != 5
AND lon != 5
AND lat != -5
AND lon != -5
AND lat != 6
AND lon != 6
AND lat != -6
AND lon != -6
AND lat != 7
AND lon != 7
AND lat != -7
AND lon != -7
AND lat != 8
AND lon != 8
AND lat != -8
AND lon
Nice idea, but the default natural language query generated a query that didn't run. Perhaps turn down the temperature or tune other parameters. This could be better if you fine-tune it such that anyone could bring their own dataset, with GPT3 just making it available recently. My favorite recent app for no code sql query is trevor.io that has allowed me to build a dashboard so easily. I only knew basic sql query and didn't have to tamper with the actual SQL code; it was all interactive and the results is always in the correct format. Comparing to the GPT3 approach, I would say I prefer trevor over using natural language for the precision. The appeal of using GPT3 exists only if it gets high enough accuracy, imo.
The weather dataset has a bit of an unorthodox schema IMO, which gives the model more trouble than usual. That’s kind of the point of this demo, though: to what extent is generated SQL like this useful—despite its flaws—in the context of real-world datasets? Jury is still out :)
For google-trends it translated "Is there anything cat-related that people are excited about?" to something very reasonable that worked (answer: popcat apparently). Nice!
As an interesting test case, check out the very strange and seemingly recursive query generated for “Get the top 10 authors of caching libraries, ranked by commit volume”
Thanks for noting this! The functionality is definitely not perfect, and the opaque nature of the underlying model does not give much opportunity for tweaking. I suspect slightly altering your input might help drive better output.
Are you looking to set it up for another public dataset or your own private one? Either way, shoot me an email (on the demo page and also in my HN profile) and I’ll see what I can do.
/*
What is the range of elevations where tornadoes happen?
*/
SELECT
elev,
COUNT(*) AS count
FROM
`fh-bigquery.weather_gsod.all`
WHERE
tornado_funnel_cloud = 'T'
GROUP BY
elev
ORDER BY
count DESC
bob1029|4 years ago
If we could get even 50% success rate on a reasonable starting point for the generated SQL each time, that would be the biggest value-add our organization has ever seen.
I think our use case is compelling because we have to implement the same SQL targets for every customer. The only variations are typically customer-specific parameters/codes/etc.
We also have a huge corpus of examples to pull from for training data.
We are thinking about initially implementing some higher order views/functions in our SQL dialect to make things easier on ourselves with the GPT model. Complex joins across many tables seems to be something that would still elude these techniques. Most of our joins are of a very particular shape, so we can abstract the super nasty stuff away.
Worst case scenario, this concludes like my cynical mind assumes it will, but I am open to being surprised this time. We aren't going to put everything behind this, more of a "if it works..." kind of 1-2 week experiment.
atom_arranger|4 years ago
Possibly relevant: https://yale-lily.github.io/spider
I briefly worked on a startup to commercialize this tech, but we decided it wasn't accurate enough to be useful. It was very cool when it actually worked. If you can only produce what you want half the time on simple queries, that doesn't seem very useful to me though.
napoleond|4 years ago
newusertoday|4 years ago
tillvz|4 years ago
At Veezoo (http://www.veezoo.com) we have been tackling this problem for over 5 years now.
Under the hood we are using our own models. With GPT-3 we're a bit worried about the lack of fine-grained control needed for productive use-cases and obv. also lock in.
Will try out against the same dataset and see how it compares!
tluyben2|4 years ago
napoleond|4 years ago
cafed00d|4 years ago
How can I extend this to use other datasets? There seem to be quite a few interesting publicly available datasets out there: https://console.cloud.google.com/marketplace/browse?filter=s...
and I'm wondering if: 1. Is there a paid-for version of your app/website where I can plug in a diff dataset? 2. Have you considered sharing the source code for others to recreate and plug in diff datasets? 3. Or, :sweat_smile:, perhaps this is as simple as adding new datasets to the drop-down menu? say, finance data / Bitcoin transactions data?
Regardless, really cool app!
napoleond|4 years ago
For private datasets, we’re looking at adding that functionality to the core Tabby service (that’s the SaaS this relates to). Please email for info!
mritchie712|4 years ago
https://blog.seekwell.io/gpt3
napoleond|4 years ago
earleybird|4 years ago
ie, "ERROR: Syntax error: Expected end of input but got keyword END at [49:3]"
napoleond|4 years ago
napoleond|4 years ago
perl4ever|4 years ago
lswainemoore|4 years ago
SELECT date, wban, stn, year, mo, da, temp, count_temp, dewp, count_dewp, slp, count_slp, stp, count_stp, visib, count_visib, wdsp, count_wdsp, mxpsd, gust, max, flag_max, min, flag_min, prcp, flag_prcp, sndp, fog, rain_drizzle, snow_ice_pellets, hail, thunder, tornado_funnel_cloud, usaf, name, country, state, call, lat, lon, elev, begin, end, point_gis, fake_date FROM `fh-bigquery.weather_gsod.all_geoclustered` WHERE lat IS NOT NULL AND lon IS NOT NULL AND lat != 0 AND lon != 0 AND lat != 1 AND lon != 1 AND lat != -1 AND lon != -1 AND lat != 2 AND lon != 2 AND lat != -2 AND lon != -2 AND lat != 3 AND lon != 3 AND lat != -3 AND lon != -3 AND lat != 4 AND lon != 4 AND lat != -4 AND lon != -4 AND lat != 5 AND lon != 5 AND lat != -5 AND lon != -5 AND lat != 6 AND lon != 6 AND lat != -6 AND lon != -6 AND lat != 7 AND lon != 7 AND lat != -7 AND lon != -7 AND lat != 8 AND lon != 8 AND lat != -8 AND lon
perl4ever|4 years ago
easylearnai|4 years ago
i_like_apis|4 years ago
I thought it was funny in that in the weather dataset that "NULL" comes in first for the win for some questions:
> what is the all time rainiest city? > what are the top 5 most dry states?
The query conversion is impressive!
napoleond|4 years ago
mjirv|4 years ago
Would be happy to chat and compare notes if you'd like!
napoleond|4 years ago
paulfitz|4 years ago
napoleond|4 years ago
talos2110|4 years ago
As an interesting test case, check out the very strange and seemingly recursive query generated for “Get the top 10 authors of caching libraries, ranked by commit volume”
napoleond|4 years ago
331c8c71|4 years ago
napoleond|4 years ago
perl4ever|4 years ago
gertjandewilde|4 years ago
matthewmorgan|4 years ago