PRQL
Pipelined Relational Query Language, pronounced "Prequel".
PRQL is a modern language for transforming data — a simpler and more powerful SQL. Like SQL, it's readable, explicit and declarative. Unlike SQL, it forms a logical pipeline of transformations, and supports abstractions such as variables and functions. It can be used with any database that uses SQL, since it transpiles to SQL.
An example
Here's a fairly simple SQL query:
SELECT TOP 20
title,
country,
AVG(salary) AS average_salary,
SUM(salary) AS sum_salary,
AVG(salary + payroll_tax) AS average_gross_salary,
SUM(salary + payroll_tax) AS sum_gross_salary,
AVG(salary + payroll_tax + healthcare_cost) AS average_gross_cost,
SUM(salary + payroll_tax + healthcare_cost) AS sum_gross_cost,
COUNT(*) as count
FROM employees
WHERE salary + payroll_tax + healthcare_cost > 0 AND country = 'USA'
GROUP BY title, country
ORDER BY sum_gross_cost
HAVING count > 200
Even this simple query demonstrates some of the problems with SQL's lack of abstractions:
- Unnecessary repetition — the calculations for each measure are repeated, despite deriving from a previous measure. The repetition in the
WHERE
clause obfuscates the meaning of the expression. - Functions have multiple operators —
HAVING
&WHERE
are fundamentally similar operations applied at different stages of the pipeline but SQL's lack of pipeline-based precedence requires it to have two different operators. - Operators have multiple functions — the
SELECT
operator both creates new aggregations, and selects which columns to include. - Awkward syntax — when developing the query, commenting out the final line of the
SELECT
list causes a syntax error because of how commas and handled, and we need to repeat the columns in theGROUP BY
clause in theSELECT
list.
Here's the same query with PRQL:
from employees
filter country = "USA" # Each line transforms the previous result.
gross_salary = salary + payroll_tax # This _adds_ a column / variable.
gross_cost = gross_salary + healthcare_cost # Variable can use other variables.
filter gross_cost > 0
aggregate split:[title, country] [ # Split are the columns to group by.
average salary, # These are the calcs to run on the groups.
sum salary,
average gross_salary,
sum gross_salary,
average gross_cost,
sum gross_cost,
count,
]
sort sum_gross_cost # Uses the auto-generated column name.
filter count > 200
top 20
As well as using variables to reduce unnecessary repetition, the query is also more readable — it flows from top to bottom, each line representing a transformation of the previous line's result. For example, TOP 20
modifies the final result in both queries — but only PRQL represents it as the final transformation. And context is localized — the aggregate
function contains both the calculations and the columns to group by.
An example using Functions
Here's another SQL query, which calculates returns from prices on days with valid prices.
SELECT
date,
-- Can't use a `WHERE` clause, as it would affect the row that the `LAG` function referenced.
IF(is_valid_price, price_adjusted / LAG(price_adjusted, 1) OVER
(PARTITION BY sec_id ORDER BY date) - 1 + dividend_return, NULL) AS return_total,
IF(is_valid_price, price_adjusted_usd / LAG(price_adjusted_usd, 1) OVER
(PARTITION BY sec_id ORDER BY date) - 1 + dividend_return, NULL) AS return_usd,
IF(is_valid_price, price_adjusted / LAG(price_adjusted, 1) OVER
(PARTITION BY sec_id ORDER BY date) - 1 + dividend_return, NULL)
- interest_rate / 252 AS return_excess,
IF(is_valid_price, price_adjusted_usd / LAG(price_adjusted_usd, 1) OVER
(PARTITION BY sec_id ORDER BY date) - 1 + dividend_return, NULL)
- interest_rate / 252 AS return_usd_excess
FROM prices
This might seem like a convoluted example, but it's taken from a real query. Indeed, it's also simpler and smaller than the full logic — note that it starts from
price_adjusted
, whose logic had to be split into a previous query to avoid the SQL becoming even less readable.
Here's the same query with PRQL:
prql version:0.0.1 db:snowflake # Version number & database name.
func lag_day x = (
window x
split sec_id
sort date
lag 1
)
func ret x = x / (x | lag_day) - 1 + dividend_return
func excess x = (x - interest_rate) / 252
func if_valid x = is_valid_price ? x : null
from prices
return_total = prices_adj | ret | if_valid # `|` can be used rather than newlines.
return_usd = prices_usd | ret | if_valid
return_excess = return_total | excess
return_usd_excess = return_usd | excess
select [
date,
sec_id,
return_total,
return_usd,
return_excess,
return_usd_excess,
]
Because we define the functions once rather than copying & pasting the code, we get all the benefits of encapsulation and extensibility — we can have reliable & tested functions, whose purpose is explicit, which we can share across queries and colleagues.
Principles
PRQL is intended to be a modern, simple, declarative language for transforming data, with abstractions such as variables & functions. It's intended to replace SQL, but doesn't have ambitions as a general-purpose programming language. While it's at a pre-alpha stage, it has some immutable principles:
- Pipelined — PRQL is a linear pipeline of transformations — each line of the query is a transformation of the previous line's result. This makes it easy to read, and simple to write.
- Simple — PRQL serves both sophisticated engineers and analysts without coding experience. By providing simple, clean abstractions, the language can be both powerful and easy to use.
- Compatible — PRQL transpiles to SQL, so it can be used with any database that uses SQL. Where possible PRQL can unify syntax across databases. PRQL should allow for a gradual onramp — it should be practical to mix SQL into a PRQL query where PRQL doesn't yet have an implementation.
- Analytical — PRQL's focus is analytical queries; we de-emphasize other SQL features such as inserting data or transactions.
- Extensible — PRQL can be extended through its abstractions, and can evolve without breaking backward-compatibility, because its queries can specify their PRQL version.
TODOs
- Write a basic parser
- Currently writing it using
nom
.
- Currently writing it using
- Write a basic complier
- This should be fairly easy since it's just generating SQL.
- Demonstrate some more complicated examples — e.g. most of the examples in https://github.com/dbt-labs/dbt-utils could all be covered much better by this.
Notes
Joins
-
Joins are implemented as
{join_type} {table} {[conditions]}
. For example:from employees left_join positions [id=employee_id]
...is equivalent to...
SELECT * FROM employees LEFT JOIN positions ON id = employee_id
-
Possibly we could shorten
[id=id]
toid
, and use SQL'sUSING
, but it may be ambiguous with usingid
as a boolean column.
Functions
-
Functions can take two disjoint types of arguments:
- Positional arguments. Callers must pass these.
- Named arguments, which can optionally have a default value.
-
So a function like:
func lag col sort_col split_col=id = ( window col split split_col sort sort_col lag 1 )
...is called
lag
, takes three argumentscol
,sort_col
&split_col
, of which the first two much be supplied, the third can optionally be supplied withsplit_col:sec_id
.
Assignments
- To create a column, we use
{column_name} = {calculation}
in a pipeline. Technically this is "upserts" the column — it'll either create or overwrite a column, depending on whether it already exists. - I'd be open to alternative syntax, given that this syntax is generally a new statement in most programming languages.
- But I can't think of any syntax that's more familiar than this.
- Possibly
let {column_name} {calculation}
would be more consistent with the other keywords?
Lists
-
Currently lists require brackets; there's no implicit list like:
from employees select salary # fails, would require `select [salary]`
-
For some functions where we're only expecting a single arg, like
select
, we could accept a single arg not as a list?
Pipelines
-
A line-break generally creates a pipelined transformation. For example:
from tbl select [ col1, col2, ] filter col1 = col2
...is equivalent to:
from tbl | select [col1, col2] | filter col1 = col2
-
A line-break doesn't created a pipeline in a few cases:
- Within a list (e.g. the
select
example above). - When the following line is a new statement, by starting with a keyword such as
func
.
- Within a list (e.g. the
Thinking about
-
The previous result is passed as the final argument of a function; i.e.
aggregate
would be like; whereX
is taken from the line above:aggregate split=[] calcs X
-
CTE syntax — something like
table =
? -
Raw syntax — I think we should have backticks represent raw SQL; i.e.
UPPER
could be defined as:func upper col = `UPPER(`col`)` # or with f-string-like syntax func upper col = `UPPER({col})` # or with " rather than ` func upper col = "UPPER({col})"
-
Arrays — PRQL is in part inspired by DataPipes.jl, which demonstrates how effective point-free pipelines can be (Chain.jl is similar). One benefit of this is how well it deals with arbitrarily nested pipelines — which are difficult to read in SQL and even in
jq
. Could we do something similar for nested data in PRQL?-
Here's a snippet from
DataPipes.jl
— and we could avoid the macros /do
/end
):@p begin text strip split(__, "\n") map() do __ collect map() do __ __ == chars[begin] ? 1 : 0 end end hcat(__...)' end
-
-
Partials — how functional do we want to make the lang? e.g. should we have partial functions? e.g. [now based on an old version of
window
] potentially we don't need thecol
inlag
here?func lag col = window col split:sec_id sort:date lag:1
-
Boolean logic — how should we represent boolean logic like
or
? With someor
function that takes*args
(which we don't currently have a design for)? Or implement dyadic operators; eitheror
or||
? (Same fornot
) -
from
— do we needfrom
? A previous version of this proposal didn't require this — just start with the table name. But some initial feedback was that removingfrom
made it less clear. -
Readme syntax — we can't get syntax highlighting in GitHub's markdown — is there a solution to this aside from submitting a parser to GitHub / screenshots / creating a website?
- Currently we use
elm
as it coincidentally provides the best syntax highlight (open to suggestions for others!).
- Currently we use
-
In advance of a full parser & compiler, could we use something like Codex to generate the transformations, and let us explore the space? We can provide our owen examples, by using fine-tuning. Changing examples is easier than changing compilers!
Inspired by
- dplyr is a beautiful language for manipulating data, in R. It's the most similar language to PRQL.
- Julia's DataPipes.jl & Chain.jl, which demonstrate how effective point-free pipelines can be, and how line-breaks can work as pipes.
- Ocaml's elegant and simple syntax.