1
Fork 0
simple-sql-parser/Language/SQL/SimpleSQL/Lex.hs

1035 lines
33 KiB
Haskell
Raw Permalink Normal View History

{-
The parser uses a separate lexer for two reasons:
1. sql syntax is very awkward to parse, the separate lexer makes it
easier to handle this in most places (in some places it makes it
harder or impossible, the fix is to switch to something better than
parsec)
2. using a separate lexer gives a huge speed boost because it reduces
backtracking. (We could get this by making the parsing code a lot more
complex also.)
3. we can test the lexer relatively exhaustively, then even when we
don't do nearly as comprehensive testing on the syntax level, we still
have a relatively high assurance of the low level of bugs. This is
much more difficult to get parity with when testing the syntax parser
directly without the separately testing lexing stage.
2024-01-09 18:47:04 +01:00
TODO:
optimisations:
check for left factor opportunities
check for places where it parses a few substrings from the source,
then puts them back together with a concatenate of some flavour
-> this is better if can find a way to parse the entire string
from the source and lift it in one go into the lexical token
before this is done, a smaller optimisation is when any code matches
a constant string in the lexer, use that constant string instead
of the string from the parser, it might make a small difference in
a few places
maybe every token should carry the exact source as well as any fields
it's been broken into - so pretty printing is trivial
2024-01-09 18:47:04 +01:00
make the tokenswill print more dialect accurate. Maybe add symbol
chars and identifier chars to the dialect definition and use them from
here
start adding negative / different parse dialect tests
add token tables and tests for oracle, sql server
review existing tables
look for refactoring opportunities, especially the token
generation tables in the tests
do some user documentation on lexing, and lexing/dialects
start thinking about a more separated design for the dialect handling
lexing tests are starting to take a really long time, so split the
tests so it is much easier to run all the tests except the lexing
tests which only need to be run when working on the lexer (which
should be relatively uncommon), or doing a commit or finishing off a
series of commits,
start writing the error message tests:
generate/write a large number of syntax errors
create a table with the source and the error message
try to compare some different versions of code to compare the
quality of the error messages by hand
get this checked in so improvements and regressions in the error
message quality can be tracked a little more easily (although it will
still be manual)
try again to add annotation to the ast
-}
-- | Lexer for SQL.
{-# LANGUAGE OverloadedStrings #-}
{-# LANGUAGE RecordWildCards #-}
{-# LANGUAGE TypeFamilies #-}
module Language.SQL.SimpleSQL.Lex
(Token(..)
,WithPos(..)
,lexSQL
,lexSQLWithPositions
,prettyToken
,prettyTokens
,ParseError
,prettyError
,tokenListWillPrintAndLex
,ansi2011
,SQLStream(..)
) where
import Language.SQL.SimpleSQL.Dialect
(Dialect(..)
,ansi2011
)
import Text.Megaparsec
(Parsec
,runParser'
,PosState(..)
,TraversableStream(..)
,VisualStream(..)
,ParseErrorBundle(..)
,errorBundlePretty
,SourcePos(..)
,getSourcePos
,getOffset
,pstateSourcePos
,statePosState
,mkPos
,hidden
,setErrorOffset
,choice
,satisfy
,takeWhileP
,takeWhile1P
,eof
,many
,try
,option
,(<|>)
,notFollowedBy
,lookAhead
,match
,optional
,label
,chunk
,region
,anySingle
)
import qualified Text.Megaparsec as M
import Text.Megaparsec.Char
(string
,char
)
import Text.Megaparsec.State (initialState)
import qualified Data.List as DL
import qualified Data.List.NonEmpty as NE
import Data.Proxy (Proxy(..))
import Data.Void (Void)
import Data.Char
(isAlphaNum
,isAlpha
,isSpace
,isDigit
)
import Control.Monad (void)
import Data.Text (Text)
import qualified Data.Text as T
2024-01-11 16:34:07 +01:00
import Data.Maybe (fromMaybe)
--import Text.Megaparsec.Debug (dbg)
2024-01-09 18:47:04 +01:00
------------------------------------------------------------------------------
-- syntax
-- | Represents a lexed token
data Token
-- | A symbol (in ansi dialect) is one of the following
--
-- * multi char symbols <> \<= \>= != ||
-- * single char symbols: * + - < > ^ / % ~ & | ? ( ) [ ] , ; ( )
--
= Symbol Text
-- | This is an identifier or keyword. The first field is
-- the quotes used, or nothing if no quotes were used. The quotes
-- can be " or u& or something dialect specific like []
| Identifier (Maybe (Text,Text)) Text
-- | This is a prefixed variable symbol, such as :var, @var or #var
-- (only :var is used in ansi dialect)
| PrefixedVariable Char Text
-- | This is a positional arg identifier e.g. $1
| PositionalArg Int
-- | This is a string literal. The first two fields are the --
-- start and end quotes, which are usually both ', but can be
-- the character set (one of nNbBxX, or u&, U&), or a dialect
-- specific string quoting (such as $$ in postgres)
| SqlString Text Text Text
-- | A number literal (integral or otherwise), stored in original format
-- unchanged
| SqlNumber Text
-- | Whitespace, one or more of space, tab or newline.
| Whitespace Text
-- | A commented line using --, contains every character starting with the
-- \'--\' and including the terminating newline character if there is one
-- - this will be missing if the last line in the source is a line comment
-- with no trailing newline
| LineComment Text
-- | A block comment, \/* stuff *\/, includes the comment delimiters
| BlockComment Text
-- | Used for generating better error messages when using the
-- output of the lexer in a parser
| InvalidToken Text
deriving (Eq,Show,Ord)
2024-01-09 18:47:04 +01:00
------------------------------------------------------------------------------
2024-01-09 18:47:04 +01:00
-- main api functions
-- | Lex some SQL to a list of tokens. The invalid token setting
-- changes the behaviour so that if there's a parse error at the start
-- of parsing an invalid token, it adds a final InvalidToken with the
-- character to the result then stop parsing. This can then be used to
-- produce a parse error with more context in the parser. Parse errors
-- within tokens still produce Left errors.
lexSQLWithPositions
:: Dialect
-- ^ dialect of SQL to use
-> Bool
-- ^ produce InvalidToken
-> Text
-- ^ filename to use in error messages
-> Maybe (Int,Int)
-- ^ line number and column number of the first character
-- in the source to use in error messages
-> Text
-- ^ the SQL source to lex
-> Either ParseError [WithPos Token]
lexSQLWithPositions dialect pit fn p src = myParse fn p (tokens dialect pit) src
-- | Lex some SQL to a list of tokens.
lexSQL
:: Dialect
-- ^ dialect of SQL to use
-> Bool
-- ^ produce InvalidToken, see lexSQLWithPositions
-> Text
-- ^ filename to use in error messages
-> Maybe (Int,Int)
-- ^ line number and column number of the first character
-- in the source to use in error messages
-> Text
-- ^ the SQL source to lex
-> Either ParseError [Token]
lexSQL dialect pit fn p src =
map tokenVal <$> lexSQLWithPositions dialect pit fn p src
2024-01-11 16:34:07 +01:00
myParse :: Text -> Maybe (Int,Int) -> Parser a -> Text -> Either ParseError a
myParse name sp' p s =
2024-01-11 16:34:07 +01:00
let sp = fromMaybe (1,1) sp'
ps = SourcePos (T.unpack name) (mkPos $ fst sp) (mkPos $ snd sp)
2024-01-11 16:34:07 +01:00
is = initialState (T.unpack name) s
sps = (statePosState is) {pstateSourcePos = ps}
is' = is {statePosState = sps}
in snd $ runParser' p is'
prettyError :: ParseError -> Text
prettyError = T.pack . errorBundlePretty
------------------------------------------------------------------------------
-- parsing boilerplate
type ParseError = ParseErrorBundle Text Void
type Parser = Parsec Void Text
-- | Positional information added to tokens to preserve source positions
-- for the parser
data WithPos a = WithPos
{ startPos :: SourcePos
, endPos :: SourcePos
, tokenLength :: Int
, tokenVal :: a
} deriving (Eq, Ord, Show)
2024-01-09 18:47:04 +01:00
------------------------------------------------------------------------------
-- pretty print
2024-01-09 18:47:04 +01:00
-- | Pretty printing, if you lex a bunch of tokens, then pretty
-- print them, should should get back exactly the same string
prettyToken :: Dialect -> Token -> Text
2024-01-09 18:47:04 +01:00
prettyToken _ (Symbol s) = s
prettyToken _ (Identifier Nothing t) = t
prettyToken _ (Identifier (Just (q1,q2)) t) = q1 <> t <> q2
prettyToken _ (PrefixedVariable c p) = T.cons c p
prettyToken _ (PositionalArg p) = T.cons '$' $ T.pack $ show p
prettyToken _ (SqlString s e t) = s <> t <> e
2024-01-09 18:47:04 +01:00
prettyToken _ (SqlNumber r) = r
prettyToken _ (Whitespace t) = t
prettyToken _ (LineComment l) = l
prettyToken _ (BlockComment c) = c
prettyToken _ (InvalidToken t) = t
2024-01-09 18:47:04 +01:00
prettyTokens :: Dialect -> [Token] -> Text
prettyTokens d ts = T.concat $ map (prettyToken d) ts
2024-01-09 18:47:04 +01:00
------------------------------------------------------------------------------
-- token parsers
-- | parser for a sql token
sqlToken :: Dialect -> Parser (WithPos Token)
sqlToken d =
withPos $ hidden $ choice $
[sqlString d
,identifier d
,lineComment d
,blockComment d
,sqlNumber d
,positionalArg d
,dontParseEndBlockComment d
,prefixedVariable d
,symbol d
,sqlWhitespace d]
--fakeSourcePos :: SourcePos
--fakeSourcePos = SourcePos "" (mkPos 1) (mkPos 1)
--------------------------------------
-- position and error helpers
withPos :: Parser a -> Parser (WithPos a)
withPos p = do
sp <- getSourcePos
off <- getOffset
a <- p
off1 <- getOffset
ep <- getSourcePos
pure $ WithPos sp ep (off1 - off) a
{-
TODO: extend this idea, to recover to parsing regular tokens after an
invalid one. This can then support resumption after error in the parser.
This would also need something similar being done for parse errors
within lexical tokens.
-}
invalidToken :: Dialect -> Parser (WithPos Token)
invalidToken _ =
withPos $ (hidden eof *> fail "") <|> (InvalidToken . T.singleton <$> anySingle)
tokens :: Dialect -> Bool -> Parser [WithPos Token]
tokens d pit = do
x <- many (sqlToken d)
if pit
then choice [x <$ hidden eof
,(\y -> x ++ [y]) <$> hidden (invalidToken d)]
else x <$ hidden eof
2024-01-09 18:47:04 +01:00
--------------------------------------
{-
Parse a SQL string. Examples:
'basic string'
'string with '' a quote'
n'international text'
b'binary string'
x'hexidecimal string'
-}
sqlString :: Dialect -> Parser Token
sqlString d =
(if (diDollarString d)
then (dollarString <|>)
else id) csString <|> normalString
2024-01-09 18:47:04 +01:00
where
dollarString = do
-- use try because of ambiguity with symbols and with
-- positional arg
delim <- fstMatch (try (char '$' *> hoptional_ identifierString <* char '$'))
let moreDollarString =
label (T.unpack delim) $ takeWhileP_ Nothing (/='$') *> checkDollar
checkDollar = label (T.unpack delim) $
choice
[lookAhead (chunk_ delim) *> pure () -- would be nice not to parse it twice?
-- but makes the whole match trick much less neat
,char_ '$' *> moreDollarString]
str <- fstMatch moreDollarString
chunk_ delim
pure $ SqlString delim delim str
lq = label "'" $ char_ '\''
normalString = SqlString "'" "'" <$> (lq *> normalStringSuffix False)
normalStringSuffix allowBackslash = label "'" $ do
let regularChar = if allowBackslash
then (\x -> x /= '\'' && x /='\\')
else (\x -> x /= '\'')
nonQuoteStringChar = takeWhileP_ Nothing regularChar
nonRegularContinue =
(hchunk_ "''" <|> hchunk_ "\\'" <|> hchar_ '\\')
moreChars = nonQuoteStringChar
*> (option () (nonRegularContinue *> moreChars))
fstMatch moreChars <* lq
2024-01-09 18:47:04 +01:00
-- try is used to to avoid conflicts with
-- identifiers which can start with n,b,x,u
-- once we read the quote type and the starting '
-- then we commit to a string
-- it's possible that this will reject some valid syntax
-- but only pathalogical stuff, and I think the improved
-- error messages and user predictability make it a good
-- pragmatic choice
csString
| diEString d =
choice [SqlString <$> try (string "e'" <|> string "E'")
<*> pure "'" <*> normalStringSuffix True
2024-01-09 18:47:04 +01:00
,csString']
| otherwise = csString'
csString' = SqlString
<$> try cs
<*> pure "'"
<*> normalStringSuffix False
2024-01-11 16:34:07 +01:00
csPrefixes = map (`T.cons` "'") "nNbBxX" ++ ["u&'", "U&'"]
cs :: Parser Text
cs = choice $ map string csPrefixes
2024-01-09 18:47:04 +01:00
--------------------------------------
{-
Parses identifiers:
simple_identifier_23
u&"unicode quoted identifier"
"quoted identifier"
"quoted identifier "" with double quote char"
`mysql quoted identifier`
-}
identifier :: Dialect -> Parser Token
identifier d =
choice $
[quotedIden
,unicodeQuotedIden
,regularIden]
++ [mySqlQuotedIden | diBackquotedIden d]
++ [sqlServerQuotedIden | diSquareBracketQuotedIden d]
where
regularIden = Identifier Nothing <$> identifierString
quotedIden = Identifier (Just ("\"","\"")) <$> qiden
failEmptyIden c = failOnThis (char_ c) "empty identifier"
mySqlQuotedIden =
Identifier (Just ("`","`")) <$>
(char_ '`' *>
(failEmptyIden '`'
<|> (takeWhile1P Nothing (/='`') <* char_ '`')))
sqlServerQuotedIden =
Identifier (Just ("[","]")) <$>
(char_ '[' *>
(failEmptyIden ']'
<|> (takeWhileP Nothing (`notElemChar` "[]")
<* choice [char_ ']'
-- should probably do this error message as
-- a proper unexpected message
,failOnThis (char_ '[') "unexpected ["])))
-- try is used here to avoid a conflict with identifiers
-- and quoted strings which also start with a 'u'
unicodeQuotedIden = Identifier
2024-01-11 16:34:07 +01:00
<$> (f <$> try (oneOf "uU" <* string "&"))
<*> qiden
where f x = Just (T.cons x "&\"", "\"")
qiden =
char_ '"' *> (failEmptyIden '"' <|> fstMatch moreQIden <* char_ '"')
moreQIden =
label "\""
(takeWhileP_ Nothing (/='"')
*> hoptional_ (chunk "\"\"" *> moreQIden))
identifierString :: Parser Text
identifierString = label "identifier" $ do
c <- satisfy isFirstLetter
choice
[T.cons c <$> takeWhileP Nothing isIdentifierChar
,pure $ T.singleton c]
where
isFirstLetter c = c == '_' || isAlpha c
isIdentifierChar :: Char -> Bool
isIdentifierChar c = c == '_' || isAlphaNum c
2024-01-09 18:47:04 +01:00
--------------------------------------
2024-01-09 18:47:04 +01:00
lineComment :: Dialect -> Parser Token
lineComment _ = LineComment <$> fstMatch (do
hidden (string_ "--")
takeWhileP_ Nothing (/='\n')
-- can you optionally read the \n to terminate the takewhilep without reparsing it?
hoptional_ $ char_ '\n')
2024-01-09 18:47:04 +01:00
--------------------------------------
-- TODO: the parser before the switch to megaparsec parsed nested block comments
-- I don't know any dialects that use this, but I think it's useful, if needed,
-- add it back in under a dialect flag?
blockComment :: Dialect -> Parser Token
blockComment _ = BlockComment <$> fstMatch bc
where
bc = chunk_ "/*" *> moreBlockChars
regularBlockCommentChars = label "*/" $
takeWhileP_ Nothing (\x -> x /= '*' && x /= '/')
continueBlockComment = label "*/" (char_ '*' <|> char_ '/') *> moreBlockChars
endComment = label "*/" $ chunk_ "*/"
moreBlockChars = label "*/" $
regularBlockCommentChars
*> (endComment
<|> (label "*/" bc *> moreBlockChars) -- nest
<|> continueBlockComment)
2024-01-09 18:47:04 +01:00
{-
This is to improve user experience: provide an error if we see */
outside a comment. This could potentially break postgres ops with */
in them (it is not sensible to use operators that contain this as a
substring). In other cases, the user should write * / instead (I can't
think of any cases when this would be valid syntax).
-}
2024-01-09 18:47:04 +01:00
dontParseEndBlockComment :: Dialect -> Parser Token
dontParseEndBlockComment _ =
failOnThis (chunk_ "*/") "comment end without comment start"
--------------------------------------
{-
numbers
digits
digits.[digits][e[+-]digits]
[digits].digits[e[+-]digits]
digitse[+-]digits
where digits is one or more decimal digits (0 through 9). At least one
digit must be before or after the decimal point, if one is used. At
least one digit must follow the exponent marker (e), if one is
present. There cannot be any spaces or other characters embedded in
the constant. Note that any leading plus or minus sign is not actually
considered part of the constant; it is an operator applied to the
constant.
algorithm:
either
parse 1 or more digits
then an optional dot which isn't two dots
then optional digits
or: parse a dot which isn't two dots
then digits
followed by an optional exponent
-}
sqlNumber :: Dialect -> Parser Token
sqlNumber d =
SqlNumber <$> fstMatch
((numStartingWithDigits <|> numStartingWithDot)
*> hoptional_ expo *> trailingCheck)
where
numStartingWithDigits = digits_ *> hoptional_ (safeDot *> hoptional_ digits_)
-- use try, so we don't commit to a number when there's a . with no following digit
numStartingWithDot = try (safeDot *> digits_)
expo = (char_ 'e' <|> char_ 'E') *> optional_ (char_ '-' <|> char_ '+') *> digits_
digits_ = label "digits" $ takeWhile1P_ Nothing isDigit
-- if there's a '..' next to the number, and it's a dialect that has .. as a
-- lexical token, parse what we have so far and leave the dots in the chamber
-- otherwise, give an error
safeDot =
if diPostgresSymbols d
then try (char_ '.' <* notFollowedBy (char_ '.'))
else char_ '.' <* notFollowedBy (char_ '.')
-- additional check to give an error if the number is immediately
-- followed by e, E or . with an exception for .. if this symbol is supported
trailingCheck =
if diPostgresSymbols d
then -- special case to allow e.g. 1..2
void (lookAhead $ hidden $ chunk_ "..")
<|> void (notFollowedBy (oneOf "eE."))
else notFollowedBy (oneOf "eE.")
digits :: Parser Text
digits = label "digits" $ takeWhile1P Nothing isDigit
2024-01-09 18:47:04 +01:00
--------------------------------------
2024-01-09 18:47:04 +01:00
positionalArg :: Dialect -> Parser Token
positionalArg d =
-- use try to avoid ambiguities with other syntax which starts with dollar
choice [PositionalArg <$>
try (char_ '$' *> (read . T.unpack <$> digits)) | diPositionalArg d]
2024-01-09 18:47:04 +01:00
--------------------------------------
-- todo: I think the try here should read a prefix char, then a single valid
-- identifier char, then commit
prefixedVariable :: Dialect -> Parser Token
prefixedVariable d = try $ choice $
[PrefixedVariable <$> char ':' <*> identifierString]
++ [PrefixedVariable <$> char '@' <*> identifierString | diAtIdentifier d]
++ [PrefixedVariable <$> char '#' <*> identifierString | diHashIdentifier d]
2024-01-09 18:47:04 +01:00
--------------------------------------
{-
Symbols
A symbol is an operator, or one of the misc symbols which include:
. .. := : :: ( ) ? ; , { } (for odbc)
The postgresql operator syntax allows a huge range of operators
compared with ansi and other dialects
-}
symbol :: Dialect -> Parser Token
symbol d = Symbol <$> choice (concat
[dots
,if diPostgresSymbols d
then postgresExtraSymbols
else []
,miscSymbol
,if diOdbc d then odbcSymbol else []
,if diPostgresSymbols d
then generalizedPostgresqlOperator
else basicAnsiOps
])
where
dots = [takeWhile1P Nothing (=='.')]
odbcSymbol = [string "{", string "}"]
postgresExtraSymbols =
[try (string ":=")
-- parse :: and : and avoid allowing ::: or more
,try (string "::" <* notFollowedBy (char ':'))
,try (string ":" <* notFollowedBy (char ':'))]
miscSymbol = map (string . T.singleton) $
case () of
_ | diSqlServerSymbols d -> ",;():?"
| diPostgresSymbols d -> "[],;()"
| otherwise -> "[],;():?"
{-
try is used because most of the first characters of the two character
symbols can also be part of a single character symbol
-}
basicAnsiOps = map (try . string) [">=","<=","!=","<>"]
++ map (string . T.singleton) "+-^*/%~&<>="
++ pipes
pipes = -- what about using many1 (char '|'), then it will
-- fail in the parser? Not sure exactly how
-- standalone the lexer should be
[char '|' *>
choice ["||" <$ char '|' <* notFollowedBy (char '|')
,pure "|"]]
{-
postgresql generalized operators
this includes the custom operators that postgres supports,
plus all the standard operators which could be custom operators
according to their grammar
rules
An operator name is a sequence of up to NAMEDATALEN-1 (63 by default) characters from the following list:
+ - * / < > = ~ ! @ # % ^ & | ` ?
There are a few restrictions on operator names, however:
-- and /* cannot appear anywhere in an operator name, since they will be taken as the start of a comment.
A multiple-character operator name cannot end in + or -, unless the name also contains at least one of these characters:
~ ! @ # % ^ & | ` ?
which allows the last character of a multi character symbol to be + or
-
-}
generalizedPostgresqlOperator :: [Parser Text]
generalizedPostgresqlOperator = [singlePlusMinus,opMoreChars]
where
allOpSymbols = "+-*/<>=~!@#%^&|`?"
-- these are the symbols when if part of a multi character
-- operator permit the operator to end with a + or - symbol
exceptionOpSymbols = "~!@#%^&|`?"
-- special case for parsing a single + or - symbol
singlePlusMinus = try $ do
c <- oneOf "+-"
notFollowedBy $ oneOf allOpSymbols
pure $ T.singleton c
-- this is used when we are parsing a potentially multi symbol
-- operator and we have alread seen one of the 'exception chars'
-- and so we can end with a + or -
moreOpCharsException = do
c <- oneOf (filter (`notElemChar` "-/*") allOpSymbols)
-- make sure we don't parse a comment starting token
-- as part of an operator
<|> try (char '/' <* notFollowedBy (char '*'))
<|> try (char '-' <* notFollowedBy (char '-'))
-- and make sure we don't parse a block comment end
-- as part of another symbol
<|> try (char '*' <* notFollowedBy (char '/'))
T.cons c <$> option "" moreOpCharsException
opMoreChars = choice
[-- parse an exception char, now we can finish with a + -
T.cons
<$> oneOf exceptionOpSymbols
<*> option "" moreOpCharsException
,T.cons
<$> (-- parse +, make sure it isn't the last symbol
try (char '+' <* lookAhead (oneOf allOpSymbols))
<|> -- parse -, make sure it isn't the last symbol
-- or the start of a -- comment
try (char '-'
<* notFollowedBy (char '-')
<* lookAhead (oneOf allOpSymbols))
<|> -- parse / check it isn't the start of a /* comment
try (char '/' <* notFollowedBy (char '*'))
<|> -- make sure we don't parse */ as part of a symbol
try (char '*' <* notFollowedBy (char '/'))
<|> -- any other ansi operator symbol
oneOf "<>=")
<*> option "" opMoreChars
]
2024-01-09 18:47:04 +01:00
--------------------------------------
sqlWhitespace :: Dialect -> Parser Token
sqlWhitespace _ = Whitespace <$> takeWhile1P Nothing isSpace
2024-01-09 18:47:04 +01:00
----------------------------------------------------------------------------
2024-01-09 18:47:04 +01:00
-- parser helpers
char_ :: Char -> Parser ()
char_ = void . char
hchar_ :: Char -> Parser ()
hchar_ = void . hidden . char
string_ :: Text -> Parser ()
string_ = void . string
oneOf :: [Char] -> Parser Char
oneOf = M.oneOf
notElemChar :: Char -> [Char] -> Bool
notElemChar a b = a `notElem` (b :: [Char])
fstMatch :: Parser () -> Parser Text
fstMatch x = fst <$> match x
hoptional_ :: Parser a -> Parser ()
hoptional_ = void . hoptional
hoptional :: Parser a -> Parser (Maybe a)
hoptional = hidden . optional
optional_ :: Parser a -> Parser ()
optional_ = void . optional
--hoption :: a -> Parser a -> Parser a
--hoption a p = hidden $ option a p
takeWhileP_ :: Maybe String -> (Char -> Bool) -> Parser ()
takeWhileP_ m p = void $ takeWhileP m p
takeWhile1P_ :: Maybe String -> (Char -> Bool) -> Parser ()
takeWhile1P_ m p = void $ takeWhile1P m p
chunk_ :: Text -> Parser ()
chunk_ = void . chunk
hchunk_ :: Text -> Parser ()
hchunk_ = void . hidden . chunk
failOnThis :: Parser () -> Text -> Parser a
failOnThis p msg = do
o <- getOffset
hidden p
region (setErrorOffset o) $ fail $ T.unpack msg
2024-01-09 18:47:04 +01:00
----------------------------------------------------------------------------
{-
This utility function will accurately report if the two tokens are
pretty printed, if they should lex back to the same two tokens. This
function is used in testing (and can be used in other places), and
must not be implemented by actually trying to print both tokens and
then lex them back from a single string (because then we would have
the risk of thinking two tokens cannot be together when there is bug
in the lexer, which the testing is supposed to find).
maybe do some quick checking to make sure this function only gives
true negatives: check pairs which return false actually fail to lex or
give different symbols in return: could use quickcheck for this
a good sanity test for this function is to change it to always return
true, then check that the automated tests return the same number of
successes. I don't think it succeeds this test at the moment
-}
-- | Utility function to tell you if a list of tokens
-- will pretty print then lex back to the same set of tokens.
-- Used internally, might be useful for generating SQL via lexical tokens.
tokenListWillPrintAndLex :: Dialect -> [Token] -> Bool
tokenListWillPrintAndLex _ [] = True
tokenListWillPrintAndLex _ [_] = True
tokenListWillPrintAndLex d (a:b:xs) =
tokensWillPrintAndLex d a b && tokenListWillPrintAndLex d (b:xs)
tokensWillPrintAndLex :: Dialect -> Token -> Token -> Bool
tokensWillPrintAndLex d a b
{-
a : followed by an identifier character will look like a host param
followed by = or : makes a different symbol
-}
| Symbol ":" <- a
, checkFirstBChar (\x -> isIdentifierChar x || x `T.elem` ":=") = False
{-
two symbols next to eachother will fail if the symbols can combine and
(possibly just the prefix) look like a different symbol
-}
| diPostgresSymbols d
, Symbol a' <- a
, Symbol b' <- b
2024-01-11 16:34:07 +01:00
, b' `notElem` ["+", "-"] || any (`T.elem` a') ("~!@#%^&|`?" :: [Char]) = False
{-
check two adjacent symbols in non postgres where the combination
possibilities are much more limited. This is ansi behaviour, it might
be different when the other dialects are done properly
-}
| Symbol a' <- a
, Symbol b' <- b
, (a',b') `elem` [("<",">")
,("<","=")
,(">","=")
,("!","=")
,("|","|")
,("||","|")
,("|","||")
,("||","||")
,("<",">=")
] = False
-- two whitespaces will be combined
| Whitespace {} <- a
, Whitespace {} <- b = False
-- line comment without a newline at the end will eat the next token
| LineComment {} <- a
, checkLastAChar (/='\n') = False
{-
check the last character of the first token and the first character of
the second token forming a comment start or end symbol
-}
| let f '-' '-' = True
f '/' '*' = True
f '*' '/' = True
f _ _ = False
in checkBorderChars f = False
{-
a symbol will absorb a following .
TODO: not 100% on this always being bad
-}
| Symbol {} <- a
, checkFirstBChar (=='.') = False
-- cannot follow a symbol ending in : with another token starting with :
| let f ':' ':' = True
f _ _ = False
in checkBorderChars f = False
-- unquoted identifier followed by an identifier letter
| Identifier Nothing _ <- a
, checkFirstBChar isIdentifierChar = False
-- a quoted identifier using ", followed by a " will fail
| Identifier (Just (_,"\"")) _ <- a
, checkFirstBChar (=='"') = False
-- prefixed variable followed by an identifier char will be absorbed
| PrefixedVariable {} <- a
, checkFirstBChar isIdentifierChar = False
-- a positional arg will absorb a following digit
| PositionalArg {} <- a
, checkFirstBChar isDigit = False
-- a string ending with ' followed by a token starting with ' will be absorbed
| SqlString _ "'" _ <- a
, checkFirstBChar (=='\'') = False
-- a number followed by a . will fail or be absorbed
| SqlNumber {} <- a
, checkFirstBChar (=='.') = False
-- a number followed by an e or E will fail or be absorbed
| SqlNumber {} <- a
, checkFirstBChar (\x -> x =='e' || x == 'E') = False
-- two numbers next to eachother will fail or be absorbed
| SqlNumber {} <- a
, SqlNumber {} <- b = False
| otherwise = True
where
prettya = prettyToken d a
prettyb = prettyToken d b
-- helper function to run a predicate on the
-- last character of the first token and the first
-- character of the second token
checkBorderChars f =
case (T.unsnoc prettya, T.uncons prettyb) of
(Just (_,la), Just (fb,_)) -> f la fb
_ -> False
checkFirstBChar f = case T.uncons prettyb of
Just (b',_) -> f b'
_ -> False
checkLastAChar f = case T.unsnoc prettya of
Just (_,la) -> f la
_ -> False
------------------------------------------------------------------------------
-- megaparsec stream boilerplate
-- | Wrapper to allow using the lexer as input to a megaparsec parser.
data SQLStream = SQLStream
{ sqlStreamInput :: String
, unSQLStream :: [WithPos Token]
}
instance M.Stream SQLStream where
type Token SQLStream = WithPos Token
type Tokens SQLStream = [WithPos Token]
tokenToChunk Proxy x = [x]
tokensToChunk Proxy xs = xs
chunkToTokens Proxy = id
chunkLength Proxy = length
chunkEmpty Proxy = null
take1_ (SQLStream _ []) = Nothing
take1_ (SQLStream str (t:ts)) = Just
( t
, SQLStream (drop (tokensLength pxy (t NE.:|[])) str) ts
)
takeN_ n (SQLStream str s)
| n <= 0 = Just ([], SQLStream str s)
| null s = Nothing
| otherwise =
let (x, s') = splitAt n s
in case NE.nonEmpty x of
Nothing -> Just (x, SQLStream str s')
Just nex -> Just (x, SQLStream (drop (tokensLength pxy nex) str) s')
takeWhile_ f (SQLStream str s) =
let (x, s') = DL.span f s
in case NE.nonEmpty x of
Nothing -> (x, SQLStream str s')
Just nex -> (x, SQLStream (drop (tokensLength pxy nex) str) s')
instance VisualStream SQLStream where
showTokens Proxy = DL.intercalate " "
. NE.toList
. fmap (showMyToken . tokenVal)
tokensLength Proxy xs = sum (tokenLength <$> xs)
instance TraversableStream SQLStream where
2024-02-02 18:20:20 +01:00
-- I have no idea what all this is doing
reachOffset o _x@(M.PosState {..}) =
( Just $ actualLine
, PosState
{ pstateInput = SQLStream
{ sqlStreamInput = postStr
, unSQLStream = post
}
, pstateOffset = max pstateOffset o
, pstateSourcePos = newSourcePos
, pstateTabWidth = pstateTabWidth
, pstateLinePrefix = prefix
}
)
where
2024-02-02 18:20:20 +01:00
maybeitsthefullsource = sqlStreamInput pstateInput
targetLineNo = M.unPos $ sourceLine newSourcePos
actualLine = case drop (targetLineNo - 1) $ lines maybeitsthefullsource of
(x:_) -> x
[] -> "<empty line>"
prefix =
if sameLine
then pstateLinePrefix ++ preLine
else preLine
sameLine = sourceLine newSourcePos == sourceLine pstateSourcePos
newSourcePos =
case post of
[] -> case unSQLStream pstateInput of
[] -> pstateSourcePos
xs -> endPos (last xs)
(x:_) -> startPos x
(pre, post) = splitAt (o - pstateOffset) (unSQLStream pstateInput)
(preStr, postStr) = splitAt tokensConsumed (sqlStreamInput pstateInput)
preLine = reverse . takeWhile (/= '\n') . reverse $ preStr
tokensConsumed =
case NE.nonEmpty pre of
Nothing -> 0
Just nePre -> tokensLength pxy nePre
pxy :: Proxy SQLStream
pxy = Proxy
showMyToken :: Token -> String
-- todo: how to do this properly?
showMyToken = T.unpack . prettyToken ansi2011