%% %% %CopyrightBegin% %% %% Copyright Ericsson AB 1998-2017. All Rights Reserved. %% %% Licensed under the Apache License, Version 2.0 (the "License"); %% you may not use this file except in compliance with the License. %% You may obtain a copy of the License at %% %% http://www.apache.org/licenses/LICENSE-2.0 %% %% Unless required by applicable law or agreed to in writing, software %% distributed under the License is distributed on an "AS IS" BASIS, %% WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. %% See the License for the specific language governing permissions and %% limitations under the License. %% %% %CopyrightEnd% -module(erl_scan_SUITE). -export([all/0, suite/0,groups/0,init_per_suite/1, end_per_suite/1, init_per_testcase/2, end_per_testcase/2, init_per_group/2,end_per_group/2]). -export([error_1/1, error_2/1, iso88591/1, otp_7810/1, otp_10302/1, otp_10990/1, otp_10992/1, otp_11807/1]). -import(lists, [nth/2,flatten/1]). -import(io_lib, [print/1]). %% %% Define to run outside of test server %% %%-define(STANDALONE,1). -ifdef(STANDALONE). -compile(export_all). -define(line, put(line, ?LINE), ). -define(config(A,B),config(A,B)). -define(t, test_server). %% config(priv_dir, _) -> %% "."; %% config(data_dir, _) -> %% ".". -else. -include_lib("common_test/include/ct.hrl"). -endif. init_per_testcase(_Case, Config) -> Config. end_per_testcase(_Case, _Config) -> ok. suite() -> [{ct_hooks,[ts_install_cth]}, {timetrap,{minutes,20}}]. all() -> [{group, error}, iso88591, otp_7810, otp_10302, otp_10990, otp_10992, otp_11807]. groups() -> [{error, [], [error_1, error_2]}]. init_per_suite(Config) -> Config. end_per_suite(_Config) -> ok. init_per_group(_GroupName, Config) -> Config. end_per_group(_GroupName, Config) -> Config. %% (OTP-2347) error_1(Config) when is_list(Config) -> {error, _, _} = erl_scan:string("'a"), ok. %% Checks that format_error works on the error cases. error_2(Config) when is_list(Config) -> lists:foreach(fun check/1, error_cases()), ok. error_cases() -> ["'a", "\"a", "'\\", "\"\\", "$", "$\\", "2.3e", "2.3e-", "91#9" ]. assert_type(N, integer) when is_integer(N) -> ok; assert_type(N, atom) when is_atom(N) -> ok. check(String) -> Error = erl_scan:string(String), check_error(Error, erl_scan). %%% (This should be useful for all format_error functions.) check_error({error, Info, EndLine}, Module0) -> {ErrorLine, Module, Desc} = Info, true = (Module == Module0), assert_type(EndLine, integer), assert_type(ErrorLine, integer), true = (ErrorLine =< EndLine), String = lists:flatten(Module0:format_error(Desc)), true = io_lib:printable_list(String). %% Tests the support for ISO-8859-1 i.e Latin-1. iso88591(Config) when is_list(Config) -> ok = case catch begin %% Some atom and variable names V1s = [$Á,$á,$é,$ë], V2s = [$N,$ä,$r], A1s = [$h,$ä,$r], A2s = [$ö,$r,$e], %% Test parsing atom and variable characters. {ok,Ts1,_} = erl_scan_string(V1s ++ " " ++ V2s ++ "\327" ++ A1s ++ " " ++ A2s), V1s = atom_to_list(element(3, nth(1, Ts1))), V2s = atom_to_list(element(3, nth(2, Ts1))), A1s = atom_to_list(element(3, nth(4, Ts1))), A2s = atom_to_list(element(3, nth(5, Ts1))), %% Test printing atoms A1s = flatten(print(element(3, nth(4, Ts1)))), A2s = flatten(print(element(3, nth(5, Ts1)))), %% Test parsing and printing strings. S1 = V1s ++ "\327" ++ A1s ++ "\250" ++ A2s, S1s = "\"" ++ S1 ++ "\"", {ok,Ts2,_} = erl_scan_string(S1s), S1 = element(3, nth(1, Ts2)), S1s = flatten(print(element(3, nth(1, Ts2)))), ok %It all worked end of {'EXIT',R} -> %Something went wrong! {error,R}; ok -> ok %Aok end. %% OTP-7810. White spaces, comments, and more... otp_7810(Config) when is_list(Config) -> ok = reserved_words(), ok = atoms(), ok = punctuations(), ok = comments(), ok = errors(), ok = integers(), ok = base_integers(), ok = floats(), ok = dots(), ok = chars(), ok = variables(), ok = eof(), ok = illegal(), ok = crashes(), ok = options(), ok = token_info(), ok = column_errors(), ok = white_spaces(), ok = unicode(), ok = more_chars(), ok = more_options(), ok = anno_info(), ok. reserved_words() -> L = ['after', 'begin', 'case', 'try', 'cond', 'catch', 'andalso', 'orelse', 'end', 'fun', 'if', 'let', 'of', 'receive', 'when', 'bnot', 'not', 'div', 'rem', 'band', 'and', 'bor', 'bxor', 'bsl', 'bsr', 'or', 'xor'], [begin {RW, true} = {RW, erl_scan:reserved_word(RW)}, S = atom_to_list(RW), Ts = [{RW,{1,1}}], test_string(S, Ts) end || RW <- L], ok. atoms() -> test_string("a b", [{atom,{1,1},a},{atom,{2,18},b}]), test_string("'a b'", [{atom,{1,1},'a b'}]), test_string("a", [{atom,{1,1},a}]), test_string("a@2", [{atom,{1,1},a@2}]), test_string([39,65,200,39], [{atom,{1,1},'AÈ'}]), test_string("ärlig östen", [{atom,{1,1},ärlig},{atom,{1,7},östen}]), {ok,[{atom,_,'$a'}],{1,6}} = erl_scan_string("'$\\a'", {1,1}), test("'$\\a'"), ok. punctuations() -> L = ["<<", "<-", "<=", "<", ">>", ">=", ">", "->", "--", "-", "++", "+", "=:=", "=/=", "=<", "=>", "==", "=", "/=", "/", "||", "|", ":=", "::", ":"], %% One token at a time: [begin W = list_to_atom(S), Ts = [{W,{1,1}}], test_string(S, Ts) end || S <- L], Three = ["/=:=", "<=:=", "==:=", ">=:="], % three tokens... No = Three ++ L, SL0 = [{S1++S2,{-length(S1),S1,S2}} || S1 <- L, S2 <- L, not lists:member(S1++S2, No)], SL = family_list(SL0), %% Two tokens. When there are several answers, the one with %% the longest first token is chosen: %% [the special case "=<<" is among the tested ones] [begin W1 = list_to_atom(S1), W2 = list_to_atom(S2), Ts = [{W1,{1,1}},{W2,{1,-L2+1}}], test_string(S, Ts) end || {S,[{L2,S1,S2}|_]} <- SL], PTs1 = [{'!',{1,1}},{'(',{1,2}},{')',{1,3}},{',',{1,4}},{';',{1,5}}, {'=',{1,6}},{'[',{1,7}},{']',{1,8}},{'{',{1,9}},{'|',{1,10}}, {'}',{1,11}}], test_string("!(),;=[]{|}", PTs1), PTs2 = [{'#',{1,1}},{'&',{1,2}},{'*',{1,3}},{'+',{1,4}},{'/',{1,5}}, {':',{1,6}},{'<',{1,7}},{'>',{1,8}},{'?',{1,9}},{'@',{1,10}}, {'\\',{1,11}},{'^',{1,12}},{'`',{1,13}},{'~',{1,14}}], test_string("#&*+/:<>?@\\^`~", PTs2), test_string(".. ", [{'..',{1,1}}]), test_string("1 .. 2", [{integer,{1,1},1},{'..',{1,3}},{integer,{1,6},2}]), test_string("...", [{'...',{1,1}}]), ok. comments() -> test("a %%\n b"), {ok,[],1} = erl_scan_string("%"), test("a %%\n b"), {ok,[{atom,{1,1},a},{atom,{2,2},b}],{2,3}} = erl_scan_string("a %%\n b", {1,1}), {ok,[{atom,{1,1},a},{comment,{1,3},"%%"},{atom,{2,2},b}],{2,3}} = erl_scan_string("a %%\n b",{1,1}, [return_comments]), {ok,[{atom,{1,1},a}, {white_space,{1,2}," "}, {white_space,{1,5},"\n "}, {atom,{2,2},b}], {2,3}} = erl_scan_string("a %%\n b",{1,1},[return_white_spaces]), {ok,[{atom,{1,1},a}, {white_space,{1,2}," "}, {comment,{1,3},"%%"}, {white_space,{1,5},"\n "}, {atom,{2,2},b}], {2,3}} = erl_scan_string("a %%\n b",{1,1},[return]), ok. errors() -> {error,{1,erl_scan,{string,$',"qa"}},1} = erl_scan:string("'qa"), %' {error,{{1,1},erl_scan,{string,$',"qa"}},{1,4}} = %' erl_scan:string("'qa", {1,1}, []), %' {error,{1,erl_scan,{string,$","str"}},1} = %" erl_scan:string("\"str"), %" {error,{{1,1},erl_scan,{string,$","str"}},{1,5}} = %" erl_scan:string("\"str", {1,1}, []), %" {error,{1,erl_scan,char},1} = erl_scan:string("$"), {error,{{1,1},erl_scan,char},{1,2}} = erl_scan:string("$", {1,1}, []), test_string([34,65,200,34], [{string,{1,1},"AÈ"}]), test_string("\\", [{'\\',{1,1}}]), {'EXIT',_} = (catch {foo, erl_scan:string('$\\a', {1,1})}), % type error {'EXIT',_} = (catch {foo, erl_scan:tokens([], '$\\a', {1,1})}), % type error "{a,tuple}" = erl_scan:format_error({a,tuple}), ok. integers() -> [begin I = list_to_integer(S), Ts = [{integer,{1,1},I}], test_string(S, Ts) end || S <- [[N] || N <- lists:seq($0, $9)] ++ ["2323","000"] ], ok. base_integers() -> [begin B = list_to_integer(BS), I = erlang:list_to_integer(S, B), Ts = [{integer,{1,1},I}], test_string(BS++"#"++S, Ts) end || {BS,S} <- [{"2","11"}, {"5","23234"}, {"12","05a"}, {"16","abcdef"}, {"16","ABCDEF"}] ], {error,{1,erl_scan,{base,1}},1} = erl_scan:string("1#000"), {error,{{1,1},erl_scan,{base,1}},{1,2}} = erl_scan:string("1#000", {1,1}, []), test_string("12#bc", [{integer,{1,1},11},{atom,{1,5},c}]), [begin Str = BS ++ "#" ++ S, {error,{1,erl_scan,{illegal,integer}},1} = erl_scan:string(Str) end || {BS,S} <- [{"3","3"},{"15","f"}, {"12","c"}] ], {ok,[{integer,1,239},{'@',1}],1} = erl_scan_string("16#ef@"), {ok,[{integer,{1,1},239},{'@',{1,6}}],{1,7}} = erl_scan_string("16#ef@", {1,1}, []), {ok,[{integer,{1,1},14},{atom,{1,5},g@}],{1,7}} = erl_scan_string("16#eg@", {1,1}, []), ok. floats() -> [begin F = list_to_float(FS), Ts = [{float,{1,1},F}], test_string(FS, Ts) end || FS <- ["1.0","001.17","3.31200","1.0e0","1.0E17", "34.21E-18", "17.0E+14"]], test_string("1.e2", [{integer,{1,1},1},{'.',{1,2}},{atom,{1,3},e2}]), {error,{1,erl_scan,{illegal,float}},1} = erl_scan:string("1.0e400"), {error,{{1,1},erl_scan,{illegal,float}},{1,8}} = erl_scan:string("1.0e400", {1,1}, []), [begin {error,{1,erl_scan,{illegal,float}},1} = erl_scan:string(S), {error,{{1,1},erl_scan,{illegal,float}},{1,_}} = erl_scan:string(S, {1,1}, []) end || S <- ["1.14Ea"]], ok. dots() -> Dot = [{".", {ok,[{dot,1}],1}, {ok,[{dot,{1,1}}],{1,2}}}, {". ", {ok,[{dot,1}],1}, {ok,[{dot,{1,1}}],{1,3}}}, {".\n", {ok,[{dot,1}],2}, {ok,[{dot,{1,1}}],{2,1}}}, {".%", {ok,[{dot,1}],1}, {ok,[{dot,{1,1}}],{1,3}}}, {".\210",{ok,[{dot,1}],1}, {ok,[{dot,{1,1}}],{1,3}}}, {".% öh",{ok,[{dot,1}],1}, {ok,[{dot,{1,1}}],{1,6}}}, {".%\n", {ok,[{dot,1}],2}, {ok,[{dot,{1,1}}],{2,1}}}, {".$", {error,{1,erl_scan,char},1}, {error,{{1,2},erl_scan,char},{1,3}}}, {".$\\", {error,{1,erl_scan,char},1}, {error,{{1,2},erl_scan,char},{1,4}}}, {".a", {ok,[{'.',1},{atom,1,a}],1}, {ok,[{'.',{1,1}},{atom,{1,2},a}],{1,3}}} ], [begin R = erl_scan_string(S), R2 = erl_scan_string(S, {1,1}, []) end || {S, R, R2} <- Dot], {ok,[{dot,_}=T1],{1,2}} = erl_scan:string(".", {1,1}, text), [1, 1, "."] = token_info(T1), {ok,[{dot,_}=T2],{1,3}} = erl_scan:string(".%", {1,1}, text), [1, 1, "."] = token_info(T2), {ok,[{dot,_}=T3],{1,6}} = erl_scan:string(".% öh", {1,1}, text), [1, 1, "."] = token_info(T3), {error,{{1,2},erl_scan,char},{1,3}} = erl_scan:string(".$", {1,1}), {error,{{1,2},erl_scan,char},{1,4}} = erl_scan:string(".$\\", {1,1}), test_string(". ", [{dot,{1,1}}]), test_string(". ", [{dot,{1,1}}]), test_string(".\n", [{dot,{1,1}}]), test_string(".\n\n", [{dot,{1,1}}]), test_string(".\n\r", [{dot,{1,1}}]), test_string(".\n\n\n", [{dot,{1,1}}]), test_string(".\210", [{dot,{1,1}}]), test_string(".%\n", [{dot,{1,1}}]), test_string(".a", [{'.',{1,1}},{atom,{1,2},a}]), test_string("%. \n. ", [{dot,{2,1}}]), {more,C} = erl_scan:tokens([], "%. ",{1,1}, return), {done,{ok,[{comment,{1,1},"%. "}, {white_space,{1,4},"\n"}, {dot,{2,1}}], {2,3}}, ""} = erl_scan_tokens(C, "\n. ", {1,1}, return), % any loc, any options [test_string(S, R) || {S, R} <- [{".$\n", [{'.',{1,1}},{char,{1,2},$\n}]}, {"$\\\n", [{char,{1,1},$\n}]}, {"'\\\n'", [{atom,{1,1},'\n'}]}, {"$\n", [{char,{1,1},$\n}]}] ], ok. chars() -> [begin L = lists:flatten(io_lib:format("$\\~.8b", [C])), Ts = [{char,{1,1},C}], test_string(L, Ts) end || C <- lists:seq(0, 255)], %% Leading zeroes... [begin L = lists:flatten(io_lib:format("$\\~3.8.0b", [C])), Ts = [{char,{1,1},C}], test_string(L, Ts) end || C <- lists:seq(0, 255)], %% $\^\n now increments the line... [begin L = "$\\^" ++ [C], Ts = [{char,{1,1},C band 2#11111}], test_string(L, Ts) end || C <- lists:seq(0, 255)], [begin L = "$\\" ++ [C], Ts = [{char,{1,1},V}], test_string(L, Ts) end || {C,V} <- [{$n,$\n}, {$r,$\r}, {$t,$\t}, {$v,$\v}, {$b,$\b}, {$f,$\f}, {$e,$\e}, {$s,$\s}, {$d,$\d}]], EC = [$\n,$\r,$\t,$\v,$\b,$\f,$\e,$\s,$\d], Ds = lists:seq($0, $9), X = [$^,$n,$r,$t,$v,$b,$f,$e,$s,$d], New = [${,$x], No = EC ++ Ds ++ X ++ New, [begin L = "$\\" ++ [C], Ts = [{char,{1,1},C}], test_string(L, Ts) end || C <- lists:seq(0, 255) -- No], [begin L = "'$\\" ++ [C] ++ "'", Ts = [{atom,{1,1},list_to_atom("$"++[C])}], test_string(L, Ts) end || C <- lists:seq(0, 255) -- No], test_string("\"\\013a\\\n\"", [{string,{1,1},"\va\n"}]), test_string("'\n'", [{atom,{1,1},'\n'}]), test_string("\"\n\a\"", [{string,{1,1},"\na"}]), %% No escape [begin L = "$" ++ [C], Ts = [{char,{1,1},C}], test_string(L, Ts) end || C <- lists:seq(0, 255) -- (No ++ [$\\])], test_string("$\n", [{char,{1,1},$\n}]), {error,{{1,1},erl_scan,char},{1,4}} = erl_scan:string("$\\^",{1,1}), test_string("$\\\n", [{char,{1,1},$\n}]), %% Robert's scanner returns line 1: test_string("$\\\n", [{char,{1,1},$\n}]), test_string("$\n\n", [{char,{1,1},$\n}]), test("$\n\n"), ok. variables() -> test_string(" \237_Aouåeiyäö", [{var,{1,7},'_Aouåeiyäö'}]), test_string("A_b_c@", [{var,{1,1},'A_b_c@'}]), test_string("V@2", [{var,{1,1},'V@2'}]), test_string("ABDÀ", [{var,{1,1},'ABDÀ'}]), test_string("Ärlig Östen", [{var,{1,1},'Ärlig'},{var,{1,7},'Östen'}]), ok. eof() -> {done,{eof,1},eof} = erl_scan:tokens([], eof, 1), {more, C1} = erl_scan:tokens([]," \n", 1), {done,{eof,2},eof} = erl_scan:tokens(C1, eof, 1), {more, C2} = erl_scan:tokens([], "abra", 1), %% An error before R13A. %% {done,Err={error,{1,erl_scan,scan},1},eof} = {done,{ok,[{atom,1,abra}],1},eof} = erl_scan_tokens(C2, eof, 1), %% With column. {more, C3} = erl_scan:tokens([]," \n",{1,1}), {done,{eof,{2,1}},eof} = erl_scan:tokens(C3, eof, 1), {more, C4} = erl_scan:tokens([], "abra", {1,1}), %% An error before R13A. %% {done,{error,{{1,1},erl_scan,scan},{1,5}},eof} = {done,{ok,[{atom,_,abra}],{1,5}},eof} = erl_scan_tokens(C4, eof, 1), %% Robert's scanner returns "" as LeftoverChars; %% the R12B scanner returns eof as LeftoverChars: (eof is correct) {more, C5} = erl_scan:tokens([], "a", 1), %% An error before R13A. %% {done,{error,{1,erl_scan,scan},1},eof} = {done,{ok,[{atom,1,a}],1},eof} = erl_scan_tokens(C5,eof,1), %% With column. {more, C6} = erl_scan:tokens([], "a", {1,1}), %% An error before R13A. %% {done,{error,{1,erl_scan,scan},1},eof} = {done,{ok,[{atom,{1,1},a}],{1,2}},eof} = erl_scan_tokens(C6,eof,1), %% A dot followed by eof is special: {more, C} = erl_scan:tokens([], "a.", 1), {done,{ok,[{atom,1,a},{dot,1}],1},eof} = erl_scan_tokens(C,eof,1), {ok,[{atom,1,foo},{dot,1}],1} = erl_scan_string("foo."), %% With column. {more, CCol} = erl_scan:tokens([], "a.", {1,1}), {done,{ok,[{atom,{1,1},a},{dot,{1,2}}],{1,3}},eof} = erl_scan_tokens(CCol,eof,1), {ok,[{atom,{1,1},foo},{dot,{1,4}}],{1,5}} = erl_scan_string("foo.", {1,1}, []), ok. illegal() -> Atom = lists:duplicate(1000, $a), {error,{1,erl_scan,{illegal,atom}},1} = erl_scan:string(Atom), {done,{error,{1,erl_scan,{illegal,atom}},1},". "} = erl_scan:tokens([], Atom++". ", 1), QAtom = "'" ++ Atom ++ "'", {error,{1,erl_scan,{illegal,atom}},1} = erl_scan:string(QAtom), {done,{error,{1,erl_scan,{illegal,atom}},1},". "} = erl_scan:tokens([], QAtom++". ", 1), Var = lists:duplicate(1000, $A), {error,{1,erl_scan,{illegal,var}},1} = erl_scan:string(Var), {done,{error,{1,erl_scan,{illegal,var}},1},". "} = erl_scan:tokens([], Var++". ", 1), Float = "1" ++ lists:duplicate(400, $0) ++ ".0", {error,{1,erl_scan,{illegal,float}},1} = erl_scan:string(Float), {done,{error,{1,erl_scan,{illegal,float}},1},". "} = erl_scan:tokens([], Float++". ", 1), String = "\"43\\x{aaaaaa}34\"", {error,{1,erl_scan,{illegal,character}},1} = erl_scan:string(String), {done,{error,{1,erl_scan,{illegal,character}},1},"34\". "} = %% Would be nice if `34\"' were skipped... %% Maybe, but then the LeftOverChars would not be the characters %% immediately following the end location of the error. erl_scan:tokens([], String++". ", 1), {error,{{1,1},erl_scan,{illegal,atom}},{1,1001}} = erl_scan:string(Atom, {1,1}), {done,{error,{{1,5},erl_scan,{illegal,atom}},{1,1005}},". "} = erl_scan:tokens([], "foo "++Atom++". ", {1,1}), {error,{{1,1},erl_scan,{illegal,atom}},{1,1003}} = erl_scan:string(QAtom, {1,1}), {done,{error,{{1,5},erl_scan,{illegal,atom}},{1,1007}},". "} = erl_scan:tokens([], "foo "++QAtom++". ", {1,1}), {error,{{1,1},erl_scan,{illegal,var}},{1,1001}} = erl_scan:string(Var, {1,1}), {done,{error,{{1,5},erl_scan,{illegal,var}},{1,1005}},". "} = erl_scan:tokens([], "foo "++Var++". ", {1,1}), {error,{{1,1},erl_scan,{illegal,float}},{1,404}} = erl_scan:string(Float, {1,1}), {done,{error,{{1,5},erl_scan,{illegal,float}},{1,408}},". "} = erl_scan:tokens([], "foo "++Float++". ", {1,1}), {error,{{1,4},erl_scan,{illegal,character}},{1,14}} = erl_scan:string(String, {1,1}), {done,{error,{{1,4},erl_scan,{illegal,character}},{1,14}},"34\". "} = erl_scan:tokens([], String++". ", {1,1}), ok. crashes() -> {'EXIT',_} = (catch {foo, erl_scan:string([-1])}), % type error {'EXIT',_} = (catch {foo, erl_scan:string("$"++[-1])}), {'EXIT',_} = (catch {foo, erl_scan:string("$\\"++[-1])}), {'EXIT',_} = (catch {foo, erl_scan:string("$\\^"++[-1])}), {'EXIT',_} = (catch {foo, erl_scan:string([$",-1,$"],{1,1})}), {'EXIT',_} = (catch {foo, erl_scan:string("\"\\v"++[-1,$"])}), %$" {'EXIT',_} = (catch {foo, erl_scan:string([$",-1,$"])}), {'EXIT',_} = (catch {foo, erl_scan:string("% foo"++[-1])}), {'EXIT',_} = (catch {foo, erl_scan:string("% foo"++[-1],{1,1})}), {'EXIT',_} = (catch {foo, erl_scan:string([a])}), % type error {'EXIT',_} = (catch {foo, erl_scan:string("$"++[a])}), {'EXIT',_} = (catch {foo, erl_scan:string("$\\"++[a])}), {'EXIT',_} = (catch {foo, erl_scan:string("$\\^"++[a])}), {'EXIT',_} = (catch {foo, erl_scan:string([$",a,$"],{1,1})}), {'EXIT',_} = (catch {foo, erl_scan:string("\"\\v"++[a,$"])}), %$" {'EXIT',_} = (catch {foo, erl_scan:string([$",a,$"])}), {'EXIT',_} = (catch {foo, erl_scan:string("% foo"++[a])}), {'EXIT',_} = (catch {foo, erl_scan:string("% foo"++[a],{1,1})}), {'EXIT',_} = (catch {foo, erl_scan:string([3.0])}), % type error ok. options() -> %% line and column are not options, but tested here {ok,[{atom,1,foo},{white_space,1," "},{comment,1,"% bar"}], 1} = erl_scan_string("foo % bar", 1, return), {ok,[{atom,1,foo},{white_space,1," "}],1} = erl_scan_string("foo % bar", 1, return_white_spaces), {ok,[{atom,1,foo},{comment,1,"% bar"}],1} = erl_scan_string("foo % bar", 1, return_comments), {ok,[{atom,17,foo}],17} = erl_scan_string("foo % bar", 17), {'EXIT',{function_clause,_}} = (catch {foo, erl_scan:string("foo % bar", {a,1}, [])}), % type error {ok,[{atom,_,foo}],{17,18}} = erl_scan_string("foo % bar", {17,9}, []), {'EXIT',{function_clause,_}} = (catch {foo, erl_scan:string("foo % bar", {1,0}, [])}), % type error {ok,[{foo,1}],1} = erl_scan_string("foo % bar",1, [{reserved_word_fun, fun(W) -> W =:= foo end}]), {'EXIT',{badarg,_}} = (catch {foo, erl_scan:string("foo % bar",1, % type error [{reserved_word_fun, fun(W,_) -> W =:= foo end}])}), ok. more_options() -> {ok,[{atom,_,foo}=T1],{19,20}} = erl_scan:string("foo", {19,17},[]), {19,17} = erl_scan:location(T1), {done,{ok,[{atom,_,foo}=T2,{dot,_}],{19,22}},[]} = erl_scan:tokens([], "foo. ", {19,17}, [bad_opt]), % type error {19,17} = erl_scan:location(T2), {ok,[{atom,_,foo}=T3],{19,20}} = erl_scan:string("foo", {19,17},[text]), {19,17} = erl_scan:location(T3), "foo" = erl_scan:text(T3), {ok,[{atom,_,foo}=T4],1} = erl_scan:string("foo", 1, [text]), 1 = erl_scan:line(T4), 1 = erl_scan:location(T4), "foo" = erl_scan:text(T4), ok. token_info() -> {ok,[T1],_} = erl_scan:string("foo", {1,18}, [text]), {'EXIT',{badarg,_}} = (catch {foo, erl_scan:category(foo)}), % type error {'EXIT',{badarg,_}} = (catch {foo, erl_scan:symbol(foo)}), % type error atom = erl_scan:category(T1), foo = erl_scan:symbol(T1), {ok,[T2],_} = erl_scan:string("foo", 1, []), 1 = erl_scan:line(T2), undefined = erl_scan:column(T2), undefined = erl_scan:text(T2), 1 = erl_scan:location(T2), {ok,[T3],_} = erl_scan:string("=", 1, []), '=' = erl_scan:category(T3), '=' = erl_scan:symbol(T3), ok. anno_info() -> {'EXIT',_} = (catch {foo,erl_scan:line(foo)}), % type error {ok,[{atom,_,foo}=T0],_} = erl_scan:string("foo", 19, [text]), 19 = erl_scan:location(T0), 19 = erl_scan:end_location(T0), {ok,[{atom,_,foo}=T3],_} = erl_scan:string("foo", {1,3}, [text]), 1 = erl_scan:line(T3), 3 = erl_scan:column(T3), {1,3} = erl_scan:location(T3), {1,6} = erl_scan:end_location(T3), "foo" = erl_scan:text(T3), {ok,[{atom,_,foo}=T4],_} = erl_scan:string("foo", 2, [text]), 2 = erl_scan:line(T4), undefined = erl_scan:column(T4), 2 = erl_scan:location(T4), "foo" = erl_scan:text(T4), {ok,[{atom,_,foo}=T5],_} = erl_scan:string("foo", {1,3}, []), 1 = erl_scan:line(T5), 3 = erl_scan:column(T5), {1,3} = erl_scan:location(T5), undefined = erl_scan:text(T5), ok. column_errors() -> {error,{{1,1},erl_scan,{string,$',""}},{1,3}} = % $' erl_scan:string("'\\",{1,1}), {error,{{1,1},erl_scan,{string,$",""}},{1,3}} = % $" erl_scan:string("\"\\",{1,1}), {error,{{1,1},erl_scan,{string,$',""}},{1,2}} = % $' erl_scan:string("'",{1,1}), {error,{{1,1},erl_scan,{string,$",""}},{1,2}} = % $" erl_scan:string("\"",{1,1}), {error,{{1,1},erl_scan,char},{1,2}} = erl_scan:string("$",{1,1}), {error,{{1,2},erl_scan,{string,$',"1234567890123456"}},{1,20}} = %' erl_scan:string(" '12345678901234567", {1,1}), {error,{{1,2},erl_scan,{string,$',"123456789012345 "}}, {1,20}} = %' erl_scan:string(" '123456789012345\\s", {1,1}), {error,{{1,2},erl_scan,{string,$","1234567890123456"}},{1,20}} = %" erl_scan:string(" \"12345678901234567", {1,1}), {error,{{1,2},erl_scan,{string,$","123456789012345 "}}, {1,20}} = %" erl_scan:string(" \"123456789012345\\s", {1,1}), {error,{{1,2},erl_scan,{string,$',"1234567890123456"}},{2,1}} = %' erl_scan:string(" '12345678901234567\n", {1,1}), ok. white_spaces() -> {ok,[{white_space,_,"\r"}, {white_space,_," "}, {atom,_,a}, {white_space,_,"\n"}], _} = erl_scan_string("\r a\n", {1,1}, return), test("\r a\n"), L = "{\"a\nb\", \"a\\nb\",\nabc\r,def}.\n\n", {ok,[{'{',_}, {string,_,"a\nb"}, {',',_}, {white_space,_," "}, {string,_,"a\nb"}, {',',_}, {white_space,_,"\n"}, {atom,_,abc}, {white_space,_,"\r"}, {',',_}, {atom,_,def}, {'}',_}, {dot,_}, {white_space,_,"\n"}], _} = erl_scan_string(L, {1,1}, return), test(L), test("\"\n\"\n"), test("\n\r\n"), test("\n\r"), test("\r\n"), test("\n\f"), [test(lists:duplicate(N, $\t)) || N <- lists:seq(1, 20)], [test([$\n|lists:duplicate(N, $\t)]) || N <- lists:seq(1, 20)], [test(lists:duplicate(N, $\s)) || N <- lists:seq(1, 20)], [test([$\n|lists:duplicate(N, $\s)]) || N <- lists:seq(1, 20)], test("\v\f\n\v "), test("\n\e\n\b\f\n\da\n"), ok. unicode() -> {ok,[{char,1,83},{integer,1,45}],1} = erl_scan_string("$\\12345"), % not unicode {error,{1,erl_scan,{illegal,character}},1} = erl_scan:string([1089]), {error,{{1,1},erl_scan,{illegal,character}},{1,2}} = erl_scan:string([1089], {1,1}), {error,{{1,3},erl_scan,{illegal,character}},{1,4}} = erl_scan:string("'a" ++ [999999999] ++ "c'", {1,1}), test("\"a"++[1089]++"b\""), {ok,[{char,1,1}],1} = erl_scan_string([$$,$\\,$^,1089], 1), {error,{1,erl_scan,Error},1} = erl_scan:string("\"qa\x{aaa}", 1), "unterminated string starting with \"qa"++[2730]++"\"" = erl_scan:format_error(Error), {error,{{1,1},erl_scan,_},{1,11}} = erl_scan:string("\"qa\\x{aaa}",{1,1}), {error,{{1,1},erl_scan,_},{1,11}} = erl_scan:string("'qa\\x{aaa}",{1,1}), {ok,[{char,1,1089}],1} = erl_scan_string([$$,1089], 1), {ok,[{char,1,1089}],1} = erl_scan_string([$$,$\\,1089], 1), Qs = "$\\x{aaa}", {ok,[{char,1,$\x{aaa}}],1} = erl_scan_string(Qs, 1), {ok,[Q2],{1,9}} = erl_scan:string("$\\x{aaa}", {1,1}, [text]), [{category,char},{column,1},{line,1},{symbol,16#aaa},{text,Qs}] = token_info_long(Q2), U1 = "\"\\x{aaa}\"", {ok,[{string,_,[2730]}=T1],{1,10}} = erl_scan:string(U1, {1,1}, [text]), {1,1} = erl_scan:location(T1), "\"\\x{aaa}\"" = erl_scan:text(T1), {ok,[{string,1,[2730]}],1} = erl_scan_string(U1, 1), U2 = "\"\\x41\\x{fff}\\x42\"", {ok,[{string,1,[$\x41,$\x{fff},$\x42]}],1} = erl_scan_string(U2, 1), U3 = "\"a\n\\x{fff}\n\"", {ok,[{string,1,[$a,$\n,$\x{fff},$\n]}],3} = erl_scan_string(U3, 1), U4 = "\"\\^\n\\x{aaa}\\^\n\"", {ok,[{string,1,[$\n,$\x{aaa},$\n]}],3} = erl_scan_string(U4, 1), %% Keep these tests: test(Qs), test(U1), test(U2), test(U3), test(U4), Str1 = "\"ab" ++ [1089] ++ "cd\"", {ok,[{string,1,[$a,$b,1089,$c,$d]}],1} = erl_scan_string(Str1, 1), {ok,[{string,{1,1},[$a,$b,1089,$c,$d]}],{1,8}} = erl_scan_string(Str1, {1,1}), test(Str1), Comment = "%% "++[1089], {ok,[{comment,1,[$%,$%,$\s,1089]}],1} = erl_scan_string(Comment, 1, [return]), {ok,[{comment,{1,1},[$%,$%,$\s,1089]}],{1,5}} = erl_scan_string(Comment, {1,1}, [return]), ok. more_chars() -> %% Due to unicode, the syntax has been incompatibly augmented: %% $\x{...}, $\xHH %% All kinds of tests... {ok,[{char,_,123}],{1,4}} = erl_scan_string("$\\{",{1,1}), {more, C1} = erl_scan:tokens([], "$\\{", {1,1}), {done,{ok,[{char,_,123}],{1,4}},eof} = erl_scan_tokens(C1, eof, 1), {ok,[{char,1,123},{atom,1,a},{'}',1}],1} = erl_scan_string("$\\{a}"), {error,{{1,1},erl_scan,char},{1,4}} = erl_scan:string("$\\x", {1,1}), {error,{{1,1},erl_scan,char},{1,5}} = erl_scan:string("$\\x{",{1,1}), {more, C3} = erl_scan:tokens([], "$\\x", {1,1}), {done,{error,{{1,1},erl_scan,char},{1,4}},eof} = erl_scan:tokens(C3, eof, 1), {error,{{1,1},erl_scan,char},{1,5}} = erl_scan:string("$\\x{",{1,1}), {more, C2} = erl_scan:tokens([], "$\\x{", {1,1}), {done,{error,{{1,1},erl_scan,char},{1,5}},eof} = erl_scan:tokens(C2, eof, 1), {error,{1,erl_scan,{illegal,character}},1} = erl_scan:string("$\\x{g}"), {error,{{1,1},erl_scan,{illegal,character}},{1,5}} = erl_scan:string("$\\x{g}", {1,1}), {error,{{1,1},erl_scan,{illegal,character}},{1,6}} = erl_scan:string("$\\x{}",{1,1}), test("\"\\{0}\""), test("\"\\x{0}\""), test("\'\\{0}\'"), test("\'\\x{0}\'"), {error,{{2,3},erl_scan,{illegal,character}},{2,6}} = erl_scan:string("\"ab \n $\\x{g}\"",{1,1}), {error,{{2,3},erl_scan,{illegal,character}},{2,6}} = erl_scan:string("\'ab \n $\\x{g}\'",{1,1}), test("$\\{34}"), test("$\\x{34}"), test("$\\{377}"), test("$\\x{FF}"), test("$\\{400}"), test("$\\x{100}"), test("$\\x{10FFFF}"), test("$\\x{10ffff}"), test("\"$\n \\{1}\""), {error,{1,erl_scan,{illegal,character}},1} = erl_scan:string("$\\x{110000}"), {error,{{1,1},erl_scan,{illegal,character}},{1,12}} = erl_scan:string("$\\x{110000}", {1,1}), {error,{{1,1},erl_scan,{illegal,character}},{1,4}} = erl_scan:string("$\\xfg", {1,1}), test("$\\xffg"), {error,{{1,1},erl_scan,{illegal,character}},{1,4}} = erl_scan:string("$\\xg", {1,1}), ok. %% OTP-10302. Unicode characters scanner/parser. otp_10302(Config) when is_list(Config) -> %% From unicode(): {ok,[{atom,1,'aсb'}],1} = erl_scan_string("'a"++[1089]++"b'", 1), {ok,[{atom,{1,1},'qaપ'}],{1,12}} = erl_scan_string("'qa\\x{aaa}'",{1,1}), {ok,[{char,1,1089}],1} = erl_scan_string([$$,1089], 1), {ok,[{char,1,1089}],1} = erl_scan_string([$$,$\\,1089],1), Qs = "$\\x{aaa}", {ok,[{char,1,2730}],1} = erl_scan_string(Qs, 1), {ok,[Q2],{1,9}} = erl_scan:string(Qs,{1,1},[text]), [{category,char},{column,1},{line,1},{symbol,16#aaa},{text,Qs}] = token_info_long(Q2), U1 = "\"\\x{aaa}\"", {ok,[T1],{1,10}} = erl_scan:string(U1, {1,1}, [text]), [{category,string},{column,1},{line,1},{symbol,[16#aaa]},{text,U1}] = token_info_long(T1), U2 = "\"\\x41\\x{fff}\\x42\"", {ok,[{string,1,[65,4095,66]}],1} = erl_scan_string(U2, 1), U3 = "\"a\n\\x{fff}\n\"", {ok,[{string,1,[97,10,4095,10]}],3} = erl_scan_string(U3, 1), U4 = "\"\\^\n\\x{aaa}\\^\n\"", {ok,[{string,1,[10,2730,10]}],3} = erl_scan_string(U4, 1,[]), Str1 = "\"ab" ++ [1089] ++ "cd\"", {ok,[{string,1,[97,98,1089,99,100]}],1} = erl_scan_string(Str1,1), {ok,[{string,{1,1},[97,98,1089,99,100]}],{1,8}} = erl_scan_string(Str1, {1,1}), OK1 = 16#D800-1, OK2 = 16#DFFF+1, OK3 = 16#FFFE-1, OK4 = 16#FFFF+1, OKL = [OK1,OK2,OK3,OK4], Illegal1 = 16#D800, Illegal2 = 16#DFFF, Illegal3 = 16#FFFE, Illegal4 = 16#FFFF, IllegalL = [Illegal1,Illegal2,Illegal3,Illegal4], [{ok,[{comment,1,[$%,$%,$\s,OK]}],1} = erl_scan_string("%% "++[OK], 1, [return]) || OK <- OKL], {ok,[{comment,_,[$%,$%,$\s,OK1]}],{1,5}} = erl_scan_string("%% "++[OK1], {1,1}, [return]), [{error,{1,erl_scan,{illegal,character}},1} = erl_scan:string("%% "++[Illegal], 1, [return]) || Illegal <- IllegalL], {error,{{1,1},erl_scan,{illegal,character}},{1,5}} = erl_scan:string("%% "++[Illegal1], {1,1}, [return]), [{ok,[],1} = erl_scan_string("%% "++[OK], 1, []) || OK <- OKL], {ok,[],{1,5}} = erl_scan_string("%% "++[OK1], {1,1}, []), [{error,{1,erl_scan,{illegal,character}},1} = erl_scan:string("%% "++[Illegal], 1, []) || Illegal <- IllegalL], {error,{{1,1},erl_scan,{illegal,character}},{1,5}} = erl_scan:string("%% "++[Illegal1], {1,1}, []), [{ok,[{string,{1,1},[OK]}],{1,4}} = erl_scan_string("\""++[OK]++"\"",{1,1}) || OK <- OKL], [{error,{{1,2},erl_scan,{illegal,character}},{1,3}} = erl_scan:string("\""++[OK]++"\"",{1,1}) || OK <- IllegalL], [{error,{{1,1},erl_scan,{illegal,character}},{1,2}} = erl_scan:string([Illegal],{1,1}) || Illegal <- IllegalL], {ok,[{char,{1,1},OK1}],{1,3}} = erl_scan_string([$$,OK1],{1,1}), {error,{{1,1},erl_scan,{illegal,character}},{1,2}} = erl_scan:string([$$,Illegal1],{1,1}), {ok,[{char,{1,1},OK1}],{1,4}} = erl_scan_string([$$,$\\,OK1],{1,1}), {error,{{1,1},erl_scan,{illegal,character}},{1,4}} = erl_scan:string([$$,$\\,Illegal1],{1,1}), {ok,[{string,{1,1},[55295]}],{1,5}} = erl_scan_string("\"\\"++[OK1]++"\"",{1,1}), {error,{{1,2},erl_scan,{illegal,character}},{1,4}} = erl_scan:string("\"\\"++[Illegal1]++"\"",{1,1}), {ok,[{char,{1,1},OK1}],{1,10}} = erl_scan_string("$\\x{D7FF}",{1,1}), {error,{{1,1},erl_scan,{illegal,character}},{1,10}} = erl_scan:string("$\\x{D800}",{1,1}), %% Not erl_scan, but erl_parse. {integer,0,1} = erl_parse_abstract(1), Float = 3.14, {float,0,Float} = erl_parse_abstract(Float), {nil,0} = erl_parse_abstract([]), {bin,0, [{bin_element,0,{integer,0,1},default,default}, {bin_element,0,{integer,0,2},default,default}]} = erl_parse_abstract(<<1,2>>), {cons,0,{tuple,0,[{atom,0,a}]},{atom,0,b}} = erl_parse_abstract([{a} | b]), {string,0,"str"} = erl_parse_abstract("str"), {cons,0, {integer,0,$a}, {cons,0,{integer,0,55296},{string,0,"c"}}} = erl_parse_abstract("a"++[55296]++"c"), Line = 17, {integer,Line,1} = erl_parse_abstract(1, Line), Float = 3.14, {float,Line,Float} = erl_parse_abstract(Float, Line), {nil,Line} = erl_parse_abstract([], Line), {bin,Line, [{bin_element,Line,{integer,Line,1},default,default}, {bin_element,Line,{integer,Line,2},default,default}]} = erl_parse_abstract(<<1,2>>, Line), {cons,Line,{tuple,Line,[{atom,Line,a}]},{atom,Line,b}} = erl_parse_abstract([{a} | b], Line), {string,Line,"str"} = erl_parse_abstract("str", Line), {cons,Line, {integer,Line,$a}, {cons,Line,{integer,Line,55296},{string,Line,"c"}}} = erl_parse_abstract("a"++[55296]++"c", Line), Opts1 = [{line,17}], {integer,Line,1} = erl_parse_abstract(1, Opts1), Float = 3.14, {float,Line,Float} = erl_parse_abstract(Float, Opts1), {nil,Line} = erl_parse_abstract([], Opts1), {bin,Line, [{bin_element,Line,{integer,Line,1},default,default}, {bin_element,Line,{integer,Line,2},default,default}]} = erl_parse_abstract(<<1,2>>, Opts1), {cons,Line,{tuple,Line,[{atom,Line,a}]},{atom,Line,b}} = erl_parse_abstract([{a} | b], Opts1), {string,Line,"str"} = erl_parse_abstract("str", Opts1), {cons,Line, {integer,Line,$a}, {cons,Line,{integer,Line,55296},{string,Line,"c"}}} = erl_parse_abstract("a"++[55296]++"c", Opts1), [begin {integer,Line,1} = erl_parse_abstract(1, Opts2), Float = 3.14, {float,Line,Float} = erl_parse_abstract(Float, Opts2), {nil,Line} = erl_parse_abstract([], Opts2), {bin,Line, [{bin_element,Line,{integer,Line,1},default,default}, {bin_element,Line,{integer,Line,2},default,default}]} = erl_parse_abstract(<<1,2>>, Opts2), {cons,Line,{tuple,Line,[{atom,Line,a}]},{atom,Line,b}} = erl_parse_abstract([{a} | b], Opts2), {string,Line,"str"} = erl_parse_abstract("str", Opts2), {string,Line,[97,1024,99]} = erl_parse_abstract("a"++[1024]++"c", Opts2) end || Opts2 <- [[{encoding,unicode},{line,Line}], [{encoding,utf8},{line,Line}]]], {cons,0, {integer,0,97}, {cons,0,{integer,0,1024},{string,0,"c"}}} = erl_parse_abstract("a"++[1024]++"c", [{encoding,latin1}]), ok. %% OTP-10990. Floating point number in input string. otp_10990(Config) when is_list(Config) -> {'EXIT',_} = (catch {foo, erl_scan:string([$",42.0,$"],1)}), ok. %% OTP-10992. List of floats to abstract format. otp_10992(Config) when is_list(Config) -> {cons,0,{float,0,42.0},{nil,0}} = erl_parse_abstract([42.0], [{encoding,unicode}]), {cons,0,{float,0,42.0},{nil,0}} = erl_parse_abstract([42.0], [{encoding,utf8}]), {cons,0,{integer,0,65},{cons,0,{float,0,42.0},{nil,0}}} = erl_parse_abstract([$A,42.0], [{encoding,unicode}]), {cons,0,{integer,0,65},{cons,0,{float,0,42.0},{nil,0}}} = erl_parse_abstract([$A,42.0], [{encoding,utf8}]), ok. %% OTP-11807. Generalize erl_parse:abstract/2. otp_11807(Config) when is_list(Config) -> {cons,0,{integer,0,97},{cons,0,{integer,0,98},{nil,0}}} = erl_parse_abstract("ab", [{encoding,none}]), {cons,0,{integer,0,-1},{nil,0}} = erl_parse_abstract([-1], [{encoding,latin1}]), ASCII = fun(I) -> I >= 0 andalso I < 128 end, {string,0,"xyz"} = erl_parse_abstract("xyz", [{encoding,ASCII}]), {cons,0,{integer,0,228},{nil,0}} = erl_parse_abstract([228], [{encoding,ASCII}]), {cons,0,{integer,0,97},{atom,0,a}} = erl_parse_abstract("a"++a, [{encoding,latin1}]), {'EXIT', {{badarg,bad},_}} = % minor backward incompatibility (catch erl_parse:abstract("string", [{encoding,bad}])), ok. test_string(String, ExpectedWithCol) -> {ok, ExpectedWithCol, _EndWithCol} = erl_scan_string(String, {1, 1}, []), Expected = [ begin {L,_C} = element(2, T), setelement(2, T, L) end || T <- ExpectedWithCol ], {ok, Expected, _End} = erl_scan_string(String), test(String). erl_scan_string(String) -> erl_scan_string(String, 1, []). erl_scan_string(String, StartLocation) -> erl_scan_string(String, StartLocation, []). erl_scan_string(String, StartLocation, Options) -> case erl_scan:string(String, StartLocation, Options) of {ok, Tokens, EndLocation} -> {ok, unopaque_tokens(Tokens), EndLocation}; Else -> Else end. erl_scan_tokens(C, S, L) -> erl_scan_tokens(C, S, L, []). erl_scan_tokens(C, S, L, O) -> case erl_scan:tokens(C, S, L, O) of {done, {ok, Ts, End}, R} -> {done, {ok, unopaque_tokens(Ts), End}, R}; Else -> Else end. unopaque_tokens([]) -> []; unopaque_tokens([Token|Tokens]) -> Attrs = element(2, Token), Term = erl_anno:to_term(Attrs), T = setelement(2, Token, Term), [T | unopaque_tokens(Tokens)]. erl_parse_abstract(Term) -> erl_parse_abstract(Term, []). erl_parse_abstract(Term, Options) -> Abstr = erl_parse:abstract(Term, Options), unopaque_abstract(Abstr). unopaque_abstract(Abstr) -> erl_parse:anno_to_term(Abstr). %% test_string(String, Expected, StartLocation, Options) -> %% {ok, Expected, _End} = erl_scan:string(String, StartLocation, Options), %% test(String). %% There are no checks of the tags... test(String) -> %% io:format("Testing `~ts'~n", [String]), [{Tokens, End}, {Wtokens, Wend}, {Ctokens, Cend}, {CWtokens, CWend}, {CWtokens2, _}] = [scan_string_with_column(String, X) || X <- [[], [return_white_spaces], [return_comments], [return], [return]]], % for white space compaction test {end1,End,Wend} = {end1,Wend,End}, {end2,Wend,Cend} = {end2,Cend,Wend}, {end3,Cend,CWend} = {end3,CWend,Cend}, %% Test that the tokens that are common to two token lists are identical. {none,Tokens} = {none, filter_tokens(CWtokens, [white_space,comment])}, {comments,Ctokens} = {comments,filter_tokens(CWtokens, [white_space])}, {white_spaces,Wtokens} = {white_spaces,filter_tokens(CWtokens, [comment])}, %% Use token attributes to extract parts from the original string, %% and check that the parts are identical to the token strings. {Line,Column} = test_decorated_tokens(String, CWtokens), {deco,{Line,Column},End} = {deco,End,{Line,Column}}, %% Almost the same again: concat texts to get the original: Text = get_text(CWtokens), {text,Text,String} = {text,String,Text}, %% Test that white spaces occupy less heap than the worst case. ok = test_white_space_compaction(CWtokens, CWtokens2), %% Test that white newlines are always first in text: WhiteTokens = select_tokens(CWtokens, [white_space]), ok = newlines_first(WhiteTokens), %% Line attribute only: [Simple,Wsimple,Csimple,WCsimple] = Simples = [element(2, erl_scan:string(String, 1, Opts)) || Opts <- [[], [return_white_spaces], [return_comments], [return]]], {consistent,true} = {consistent,consistent_attributes(Simples)}, {simple_wc,WCsimple} = {simple_wc,simplify(CWtokens)}, {simple,Simple} = {simple,filter_tokens(WCsimple, [white_space,comment])}, {simple_c,Csimple} = {simple_c,filter_tokens(WCsimple, [white_space])}, {simple_w,Wsimple} = {simple_w,filter_tokens(WCsimple, [comment])}, %% Line attribute only, with text: [SimpleTxt,WsimpleTxt,CsimpleTxt,WCsimpleTxt] = SimplesTxt = [element(2, erl_scan:string(String, 1, [text|Opts])) || Opts <- [[], [return_white_spaces], [return_comments], [return]]], TextTxt = get_text(WCsimpleTxt), {text_txt,TextTxt,String} = {text_txt,String,TextTxt}, {consistent_txt,true} = {consistent_txt,consistent_attributes(SimplesTxt)}, {simple_txt,SimpleTxt} = {simple_txt,filter_tokens(WCsimpleTxt, [white_space,comment])}, {simple_c_txt,CsimpleTxt} = {simple_c_txt,filter_tokens(WCsimpleTxt, [white_space])}, {simple_w_txt,WsimpleTxt} = {simple_w_txt,filter_tokens(WCsimpleTxt, [comment])}, ok. test_white_space_compaction(Tokens, Tokens2) when Tokens =:= Tokens2 -> [WS, WS2] = [select_tokens(Ts, [white_space]) || Ts <- [Tokens, Tokens2]], test_wsc(WS, WS2). test_wsc([], []) -> ok; test_wsc([Token|Tokens], [Token2|Tokens2]) -> [Text, Text2] = [Text || Text <- [erl_scan:text(T) || T <- [Token, Token2]]], Sz = erts_debug:size(Text), Sz2 = erts_debug:size({Text, Text2}), IsCompacted = Sz2 < 2*Sz+erts_debug:size({a,a}), ToBeCompacted = is_compacted(Text), if IsCompacted =:= ToBeCompacted -> test_wsc(Tokens, Tokens2); true -> {compaction_error, Token} end. is_compacted("\r") -> true; is_compacted("\n\r") -> true; is_compacted("\n\f") -> true; is_compacted([$\n|String]) -> all_spaces(String) orelse all_tabs(String); is_compacted(String) -> all_spaces(String) orelse all_tabs(String). all_spaces(L) -> all_same(L, $\s). all_tabs(L) -> all_same(L, $\t). all_same(L, Char) -> lists:all(fun(C) -> C =:= Char end, L). newlines_first([]) -> ok; newlines_first([Token|Tokens]) -> Text = erl_scan:text(Token), Nnls = length([C || C <- Text, C =:= $\n]), OK = case Text of [$\n|_] -> Nnls =:= 1; _ -> Nnls =:= 0 end, if OK -> newlines_first(Tokens); true -> OK end. filter_tokens(Tokens, Tags) -> lists:filter(fun(T) -> not lists:member(element(1, T), Tags) end, Tokens). select_tokens(Tokens, Tags) -> lists:filter(fun(T) -> lists:member(element(1, T), Tags) end, Tokens). simplify([Token|Tokens]) -> Line = erl_scan:line(Token), [setelement(2, Token, erl_anno:new(Line)) | simplify(Tokens)]; simplify([]) -> []. get_text(Tokens) -> lists:flatten( [T || Token <- Tokens, (T = erl_scan:text(Token)) =/= []]). test_decorated_tokens(String, Tokens) -> ToksAttrs = token_attrs(Tokens), test_strings(ToksAttrs, String, 1, 1). token_attrs(Tokens) -> [{L,C,length(T),T} || Token <- Tokens, ([C,L,T] = token_info(Token)) =/= []]. token_info(T) -> Column = erl_scan:column(T), Line = erl_scan:line(T), Text = erl_scan:text(T), [Column, Line, Text]. token_info_long(T) -> Column = erl_scan:column(T), Line = erl_scan:line(T), Text = erl_scan:text(T), Category = erl_scan:category(T), Symbol = erl_scan:symbol(T), [{category,Category},{column,Column},{line,Line}, {symbol,Symbol},{text,Text}]. test_strings([], _S, Line, Column) -> {Line,Column}; test_strings([{L,C,Len,T}=Attr|Attrs], String0, Line0, Column0) -> {String1, Column1} = skip_newlines(String0, L, Line0, Column0), String = skip_chars(String1, C-Column1), {Str,Rest} = lists:split(Len, String), if Str =:= T -> {Line,Column} = string_newlines(T, L, C), test_strings(Attrs, Rest, Line, Column); true -> {token_error, Attr, Str} end. skip_newlines(String, Line, Line, Column) -> {String, Column}; skip_newlines([$\n|String], L, Line, _Column) -> skip_newlines(String, L, Line+1, 1); skip_newlines([_|String], L, Line, Column) -> skip_newlines(String, L, Line, Column+1). skip_chars(String, 0) -> String; skip_chars([_|String], N) -> skip_chars(String, N-1). string_newlines([$\n|String], Line, _Column) -> string_newlines(String, Line+1, 1); string_newlines([], Line, Column) -> {Line, Column}; string_newlines([_|String], Line, Column) -> string_newlines(String, Line, Column+1). scan_string_with_column(String, Options0) -> Options = [text | Options0], StartLoc = {1, 1}, {ok, Ts1, End1} = erl_scan:string(String, StartLoc, Options), TString = String ++ ". ", {ok,Ts2,End2} = scan_tokens(TString, Options, [], StartLoc), {ok, Ts3, End3} = scan_tokens_1({more, []}, TString, Options, [], StartLoc), {end_2,End2,End3} = {end_2,End3,End2}, {EndLine1,EndColumn1} = End1, End2 = {EndLine1,EndColumn1+2}, {ts_1,Ts2,Ts3} = {ts_1,Ts3,Ts2}, Ts2 = Ts1 ++ [lists:last(Ts2)], %% Attributes are keylists, but have no text. {ok, Ts7, End7} = erl_scan:string(String, {1,1}, Options), {ok, Ts8, End8} = scan_tokens(TString, Options, [], {1,1}), {end1, End1} = {end1, End7}, {end2, End2} = {end2, End8}, Ts8 = Ts7 ++ [lists:last(Ts8)], {cons,true} = {cons,consistent_attributes([Ts1,Ts2,Ts3,Ts7,Ts8])}, {Ts1, End1}. scan_tokens(String, Options, Rs, Location) -> case erl_scan:tokens([], String, Location, Options) of {done, {ok,Ts,End}, ""} -> {ok, lists:append(lists:reverse([Ts|Rs])), End}; {done, {ok,Ts,End}, Rest} -> scan_tokens(Rest, Options, [Ts|Rs], End) end. scan_tokens_1({done, {ok,Ts,End}, ""}, "", _Options, Rs, _Location) -> {ok,lists:append(lists:reverse([Ts|Rs])),End}; scan_tokens_1({done, {ok,Ts,End}, Rest}, Cs, Options, Rs, _Location) -> scan_tokens_1({more,[]}, Rest++Cs, Options, [Ts|Rs], End); scan_tokens_1({more, Cont}, [C | Cs], Options, Rs, Loc) -> R = erl_scan:tokens(Cont, [C], Loc, Options), scan_tokens_1(R, Cs, Options, Rs, Loc). consistent_attributes([]) -> true; consistent_attributes([Ts | TsL]) -> L = [T || T <- Ts, is_integer(element(2, T))], case L of [] -> TagsL = [[Tag || {Tag,_} <- defined(token_info_long(T))] || T <- Ts], case lists:usort(TagsL) of [_] -> consistent_attributes(TsL); [] when Ts =:= [] -> consistent_attributes(TsL); _ -> Ts end; Ts -> consistent_attributes(TsL); _ -> Ts end. defined(L) -> [{T,V} || {T,V} <- L, V =/= undefined]. family_list(L) -> sofs:to_external(family(L)). family(L) -> sofs:relation_to_family(sofs:relation(L)).