Regenerate code with ragel 6.10

This commit is contained in:
Radek Simko 2019-01-25 16:46:54 +00:00
parent a9ca194bcd
commit cb33095462
2 changed files with 116 additions and 202 deletions

View File

@ -1,10 +1,10 @@
// line 1 "scan_string_lit.rl" //line scan_string_lit.rl:1
package hclsyntax package hclsyntax
// This file is generated from scan_string_lit.rl. DO NOT EDIT. // This file is generated from scan_string_lit.rl. DO NOT EDIT.
// line 9 "scan_string_lit.go" //line scan_string_lit.go:9
var _hclstrtok_actions []byte = []byte{ var _hclstrtok_actions []byte = []byte{
0, 1, 0, 1, 1, 2, 1, 0, 0, 1, 0, 1, 1, 2, 1, 0,
} }
@ -114,12 +114,12 @@ const hclstrtok_error int = 0
const hclstrtok_en_quoted int = 10 const hclstrtok_en_quoted int = 10
const hclstrtok_en_unquoted int = 4 const hclstrtok_en_unquoted int = 4
// line 10 "scan_string_lit.rl" //line scan_string_lit.rl:10
func scanStringLit(data []byte, quoted bool) [][]byte { func scanStringLit(data []byte, quoted bool) [][]byte {
var ret [][]byte var ret [][]byte
// line 61 "scan_string_lit.rl" //line scan_string_lit.rl:61
// Ragel state // Ragel state
p := 0 // "Pointer" into data p := 0 // "Pointer" into data
@ -144,11 +144,11 @@ func scanStringLit(data []byte, quoted bool) [][]byte {
ret = append(ret, data[ts:te]) ret = append(ret, data[ts:te])
}*/ }*/
// line 154 "scan_string_lit.go" //line scan_string_lit.go:154
{ {
} }
// line 158 "scan_string_lit.go" //line scan_string_lit.go:158
{ {
var _klen int var _klen int
var _trans int var _trans int
@ -229,7 +229,7 @@ func scanStringLit(data []byte, quoted bool) [][]byte {
_acts++ _acts++
switch _hclstrtok_actions[_acts-1] { switch _hclstrtok_actions[_acts-1] {
case 0: case 0:
// line 40 "scan_string_lit.rl" //line scan_string_lit.rl:40
// If te is behind p then we've skipped over some literal // If te is behind p then we've skipped over some literal
// characters which we must now return. // characters which we must now return.
@ -239,12 +239,12 @@ func scanStringLit(data []byte, quoted bool) [][]byte {
ts = p ts = p
case 1: case 1:
// line 48 "scan_string_lit.rl" //line scan_string_lit.rl:48
te = p te = p
ret = append(ret, data[ts:te]) ret = append(ret, data[ts:te])
// line 255 "scan_string_lit.go" //line scan_string_lit.go:253
} }
} }
@ -267,12 +267,12 @@ func scanStringLit(data []byte, quoted bool) [][]byte {
__acts++ __acts++
switch _hclstrtok_actions[__acts-1] { switch _hclstrtok_actions[__acts-1] {
case 1: case 1:
// line 48 "scan_string_lit.rl" //line scan_string_lit.rl:48
te = p te = p
ret = append(ret, data[ts:te]) ret = append(ret, data[ts:te])
// line 281 "scan_string_lit.go" //line scan_string_lit.go:278
} }
} }
} }
@ -282,7 +282,7 @@ func scanStringLit(data []byte, quoted bool) [][]byte {
} }
} }
// line 89 "scan_string_lit.rl" //line scan_string_lit.rl:89
if te < p { if te < p {
// Collect any leftover literal characters at the end of the input // Collect any leftover literal characters at the end of the input

View File

@ -1,4 +1,4 @@
// line 1 "scan_tokens.rl" //line scan_tokens.rl:1
package hclsyntax package hclsyntax
@ -10,7 +10,7 @@ import (
// This file is generated from scan_tokens.rl. DO NOT EDIT. // This file is generated from scan_tokens.rl. DO NOT EDIT.
// line 15 "scan_tokens.go" //line scan_tokens.go:15
var _hcltok_actions []byte = []byte{ var _hcltok_actions []byte = []byte{
0, 1, 0, 1, 1, 1, 2, 1, 3, 0, 1, 0, 1, 1, 1, 2, 1, 3,
1, 4, 1, 6, 1, 7, 1, 8, 1, 4, 1, 6, 1, 7, 1, 8,
@ -264,7 +264,7 @@ var _hcltok_trans_keys []byte = []byte{
228, 233, 234, 237, 239, 240, 243, 48, 228, 233, 234, 237, 239, 240, 243, 48,
57, 65, 90, 97, 122, 196, 218, 229, 57, 65, 90, 97, 122, 196, 218, 229,
236, 10, 170, 181, 183, 186, 128, 150, 236, 10, 170, 181, 183, 186, 128, 150,
152, 182, 184, 255, 192, 255, 0, 127, 152, 182, 184, 255, 192, 255, 128, 255,
173, 130, 133, 146, 159, 165, 171, 175, 173, 130, 133, 146, 159, 165, 171, 175,
255, 181, 190, 184, 185, 192, 255, 140, 255, 181, 190, 184, 185, 192, 255, 140,
134, 138, 142, 161, 163, 255, 182, 130, 134, 138, 142, 161, 163, 255, 182, 130,
@ -573,7 +573,7 @@ var _hcltok_trans_keys []byte = []byte{
159, 150, 153, 131, 140, 255, 160, 163, 159, 150, 153, 131, 140, 255, 160, 163,
164, 165, 184, 185, 186, 161, 162, 133, 164, 165, 184, 185, 186, 161, 162, 133,
255, 170, 181, 183, 186, 128, 150, 152, 255, 170, 181, 183, 186, 128, 150, 152,
182, 184, 255, 192, 255, 128, 255, 173, 182, 184, 255, 192, 255, 0, 127, 173,
130, 133, 146, 159, 165, 171, 175, 255, 130, 133, 146, 159, 165, 171, 175, 255,
181, 190, 184, 185, 192, 255, 140, 134, 181, 190, 184, 185, 192, 255, 140, 134,
138, 142, 161, 163, 255, 182, 130, 136, 138, 142, 161, 163, 255, 182, 130, 136,
@ -1268,7 +1268,7 @@ var _hcltok_trans_keys []byte = []byte{
128, 140, 141, 143, 144, 153, 154, 189, 128, 140, 141, 143, 144, 153, 154, 189,
190, 191, 128, 136, 137, 191, 173, 128, 190, 191, 128, 136, 137, 191, 173, 128,
168, 169, 177, 178, 180, 181, 182, 183, 168, 169, 177, 178, 180, 181, 182, 183,
191, 0, 127, 192, 255, 150, 151, 158, 191, 128, 255, 192, 255, 150, 151, 158,
159, 152, 154, 156, 158, 134, 135, 142, 159, 152, 154, 156, 158, 134, 135, 142,
143, 190, 191, 192, 255, 181, 189, 191, 143, 190, 191, 192, 255, 181, 189, 191,
128, 190, 133, 181, 128, 129, 130, 140, 128, 190, 133, 181, 128, 129, 130, 140,
@ -2202,8 +2202,8 @@ var _hcltok_indicies []int16 = []int16{
61, 62, 63, 64, 39, 41, 65, 43, 61, 62, 63, 64, 39, 41, 65, 43,
66, 67, 68, 13, 13, 13, 16, 40, 66, 67, 68, 13, 13, 13, 16, 40,
3, 46, 3, 13, 13, 13, 13, 3, 3, 46, 3, 13, 13, 13, 13, 3,
13, 13, 13, 3, 13, 3, 13, 3, 13, 13, 13, 3, 13, 3, 13, 13,
13, 3, 3, 3, 3, 3, 13, 3, 3, 3, 3, 3, 3, 3, 13, 3,
3, 3, 3, 13, 13, 13, 13, 13, 3, 3, 3, 13, 13, 13, 13, 13,
3, 3, 13, 3, 3, 13, 3, 13, 3, 3, 13, 3, 3, 13, 3, 13,
3, 3, 13, 3, 3, 3, 13, 13, 3, 3, 13, 3, 3, 3, 13, 13,
@ -2455,7 +2455,7 @@ var _hcltok_indicies []int16 = []int16{
3, 3, 13, 18, 419, 18, 267, 302, 3, 3, 13, 18, 419, 18, 267, 302,
303, 304, 16, 3, 3, 13, 421, 421, 303, 304, 16, 3, 3, 13, 421, 421,
421, 421, 420, 421, 421, 421, 420, 421, 421, 421, 420, 421, 421, 421, 420, 421,
420, 421, 421, 420, 420, 420, 420, 420, 420, 421, 420, 421, 420, 420, 420, 420,
420, 421, 420, 420, 420, 420, 421, 421, 420, 421, 420, 420, 420, 420, 421, 421,
421, 421, 421, 420, 420, 421, 420, 420, 421, 421, 421, 420, 420, 421, 420, 420,
421, 420, 421, 420, 420, 421, 420, 420, 421, 420, 421, 420, 420, 421, 420, 420,
@ -2993,7 +2993,7 @@ var _hcltok_indicies []int16 = []int16{
794, 1045, 794, 1045, 794, 1044, 1045, 1044, 794, 1045, 794, 1045, 794, 1044, 1045, 1044,
794, 794, 1045, 1044, 1045, 794, 1045, 794, 794, 794, 1045, 1044, 1045, 794, 1045, 794,
1045, 1044, 794, 1045, 1044, 1045, 1045, 794, 1045, 1044, 794, 1045, 1044, 1045, 1045, 794,
1045, 794, 1045, 1044, 1044, 794, 1044, 794, 1045, 794, 1045, 1044, 794, 1044, 1044, 794,
1045, 1045, 794, 1045, 1045, 1045, 1045, 1045, 1045, 1045, 794, 1045, 1045, 1045, 1045, 1045,
1045, 1045, 1044, 794, 1045, 1045, 1045, 794, 1045, 1045, 1044, 794, 1045, 1045, 1045, 794,
1044, 1045, 1045, 1045, 794, 1045, 794, 1045, 1044, 1045, 1045, 1045, 794, 1045, 794, 1045,
@ -4301,7 +4301,7 @@ const hcltok_en_bareTemplate int = 1549
const hcltok_en_identOnly int = 1556 const hcltok_en_identOnly int = 1556
const hcltok_en_main int = 1464 const hcltok_en_main int = 1464
// line 16 "scan_tokens.rl" //line scan_tokens.rl:16
func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []Token { func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []Token {
stripData := stripUTF8BOM(data) stripData := stripUTF8BOM(data)
@ -4315,7 +4315,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
StartByte: start.Byte, StartByte: start.Byte,
} }
// line 299 "scan_tokens.rl" //line scan_tokens.rl:299
// Ragel state // Ragel state
p := 0 // "Pointer" into data p := 0 // "Pointer" into data
@ -4343,7 +4343,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
var retBraces []int // stack of brace levels that cause us to use fret var retBraces []int // stack of brace levels that cause us to use fret
var heredocs []heredocInProgress // stack of heredocs we're currently processing var heredocs []heredocInProgress // stack of heredocs we're currently processing
// line 334 "scan_tokens.rl" //line scan_tokens.rl:334
// Make Go compiler happy // Make Go compiler happy
_ = ts _ = ts
@ -4363,7 +4363,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
f.emitToken(TokenType(b[0]), ts, te) f.emitToken(TokenType(b[0]), ts, te)
} }
// line 4375 "scan_tokens.go" //line scan_tokens.go:4375
{ {
top = 0 top = 0
ts = 0 ts = 0
@ -4371,7 +4371,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
act = 0 act = 0
} }
// line 4383 "scan_tokens.go" //line scan_tokens.go:4383
{ {
var _klen int var _klen int
var _trans int var _trans int
@ -4392,11 +4392,10 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
_acts++ _acts++
switch _hcltok_actions[_acts-1] { switch _hcltok_actions[_acts-1] {
case 6: case 6:
// line 1 "NONE" //line NONE:1
ts = p ts = p
// line 4407 "scan_tokens.go" //line scan_tokens.go:4406
} }
} }
@ -4468,33 +4467,27 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
_acts++ _acts++
switch _hcltok_actions[_acts-1] { switch _hcltok_actions[_acts-1] {
case 0: case 0:
// line 223 "scan_tokens.rl" //line scan_tokens.rl:223
p-- p--
case 1: case 1:
// line 224 "scan_tokens.rl" //line scan_tokens.rl:224
p-- p--
case 2: case 2:
// line 229 "scan_tokens.rl" //line scan_tokens.rl:229
p-- p--
case 3: case 3:
// line 230 "scan_tokens.rl" //line scan_tokens.rl:230
p-- p--
case 7: case 7:
// line 1 "NONE" //line NONE:1
te = p + 1 te = p + 1
case 8: case 8:
// line 160 "scan_tokens.rl" //line scan_tokens.rl:160
te = p + 1 te = p + 1
{ {
token(TokenTemplateInterp) token(TokenTemplateInterp)
@ -4512,8 +4505,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 9: case 9:
// line 170 "scan_tokens.rl" //line scan_tokens.rl:170
te = p + 1 te = p + 1
{ {
token(TokenTemplateControl) token(TokenTemplateControl)
@ -4531,8 +4523,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 10: case 10:
// line 84 "scan_tokens.rl" //line scan_tokens.rl:84
te = p + 1 te = p + 1
{ {
token(TokenCQuote) token(TokenCQuote)
@ -4545,22 +4536,19 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
case 11: case 11:
// line 244 "scan_tokens.rl" //line scan_tokens.rl:244
te = p + 1 te = p + 1
{ {
token(TokenInvalid) token(TokenInvalid)
} }
case 12: case 12:
// line 245 "scan_tokens.rl" //line scan_tokens.rl:245
te = p + 1 te = p + 1
{ {
token(TokenBadUTF8) token(TokenBadUTF8)
} }
case 13: case 13:
// line 160 "scan_tokens.rl" //line scan_tokens.rl:160
te = p te = p
p-- p--
{ {
@ -4579,8 +4567,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 14: case 14:
// line 170 "scan_tokens.rl" //line scan_tokens.rl:170
te = p te = p
p-- p--
{ {
@ -4599,54 +4586,46 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 15: case 15:
// line 243 "scan_tokens.rl" //line scan_tokens.rl:243
te = p te = p
p-- p--
{ {
token(TokenQuotedLit) token(TokenQuotedLit)
} }
case 16: case 16:
// line 244 "scan_tokens.rl" //line scan_tokens.rl:244
te = p te = p
p-- p--
{ {
token(TokenInvalid) token(TokenInvalid)
} }
case 17: case 17:
// line 245 "scan_tokens.rl" //line scan_tokens.rl:245
te = p te = p
p-- p--
{ {
token(TokenBadUTF8) token(TokenBadUTF8)
} }
case 18: case 18:
// line 243 "scan_tokens.rl" //line scan_tokens.rl:243
p = (te) - 1 p = (te) - 1
{ {
token(TokenQuotedLit) token(TokenQuotedLit)
} }
case 19: case 19:
// line 245 "scan_tokens.rl" //line scan_tokens.rl:245
p = (te) - 1 p = (te) - 1
{ {
token(TokenBadUTF8) token(TokenBadUTF8)
} }
case 20: case 20:
// line 148 "scan_tokens.rl" //line scan_tokens.rl:148
act = 10 act = 10
case 21: case 21:
// line 253 "scan_tokens.rl" //line scan_tokens.rl:253
act = 11 act = 11
case 22: case 22:
// line 160 "scan_tokens.rl" //line scan_tokens.rl:160
te = p + 1 te = p + 1
{ {
token(TokenTemplateInterp) token(TokenTemplateInterp)
@ -4664,8 +4643,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 23: case 23:
// line 170 "scan_tokens.rl" //line scan_tokens.rl:170
te = p + 1 te = p + 1
{ {
token(TokenTemplateControl) token(TokenTemplateControl)
@ -4683,8 +4661,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 24: case 24:
// line 111 "scan_tokens.rl" //line scan_tokens.rl:111
te = p + 1 te = p + 1
{ {
// This action is called specificially when a heredoc literal // This action is called specificially when a heredoc literal
@ -4729,15 +4706,13 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
token(TokenStringLit) token(TokenStringLit)
} }
case 25: case 25:
// line 253 "scan_tokens.rl" //line scan_tokens.rl:253
te = p + 1 te = p + 1
{ {
token(TokenBadUTF8) token(TokenBadUTF8)
} }
case 26: case 26:
// line 160 "scan_tokens.rl" //line scan_tokens.rl:160
te = p te = p
p-- p--
{ {
@ -4756,8 +4731,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 27: case 27:
// line 170 "scan_tokens.rl" //line scan_tokens.rl:170
te = p te = p
p-- p--
{ {
@ -4776,8 +4750,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 28: case 28:
// line 148 "scan_tokens.rl" //line scan_tokens.rl:148
te = p te = p
p-- p--
{ {
@ -4788,16 +4761,14 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
token(TokenStringLit) token(TokenStringLit)
} }
case 29: case 29:
// line 253 "scan_tokens.rl" //line scan_tokens.rl:253
te = p te = p
p-- p--
{ {
token(TokenBadUTF8) token(TokenBadUTF8)
} }
case 30: case 30:
// line 148 "scan_tokens.rl" //line scan_tokens.rl:148
p = (te) - 1 p = (te) - 1
{ {
// This action is called when a heredoc literal _doesn't_ end // This action is called when a heredoc literal _doesn't_ end
@ -4807,8 +4778,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
token(TokenStringLit) token(TokenStringLit)
} }
case 31: case 31:
// line 1 "NONE" //line NONE:1
switch act { switch act {
case 0: case 0:
{ {
@ -4833,16 +4803,13 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
case 32: case 32:
// line 156 "scan_tokens.rl" //line scan_tokens.rl:156
act = 14 act = 14
case 33: case 33:
// line 260 "scan_tokens.rl" //line scan_tokens.rl:260
act = 15 act = 15
case 34: case 34:
// line 160 "scan_tokens.rl" //line scan_tokens.rl:160
te = p + 1 te = p + 1
{ {
token(TokenTemplateInterp) token(TokenTemplateInterp)
@ -4860,8 +4827,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 35: case 35:
// line 170 "scan_tokens.rl" //line scan_tokens.rl:170
te = p + 1 te = p + 1
{ {
token(TokenTemplateControl) token(TokenTemplateControl)
@ -4879,22 +4845,19 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 36: case 36:
// line 156 "scan_tokens.rl" //line scan_tokens.rl:156
te = p + 1 te = p + 1
{ {
token(TokenStringLit) token(TokenStringLit)
} }
case 37: case 37:
// line 260 "scan_tokens.rl" //line scan_tokens.rl:260
te = p + 1 te = p + 1
{ {
token(TokenBadUTF8) token(TokenBadUTF8)
} }
case 38: case 38:
// line 160 "scan_tokens.rl" //line scan_tokens.rl:160
te = p te = p
p-- p--
{ {
@ -4913,8 +4876,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 39: case 39:
// line 170 "scan_tokens.rl" //line scan_tokens.rl:170
te = p te = p
p-- p--
{ {
@ -4933,31 +4895,27 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 40: case 40:
// line 156 "scan_tokens.rl" //line scan_tokens.rl:156
te = p te = p
p-- p--
{ {
token(TokenStringLit) token(TokenStringLit)
} }
case 41: case 41:
// line 260 "scan_tokens.rl" //line scan_tokens.rl:260
te = p te = p
p-- p--
{ {
token(TokenBadUTF8) token(TokenBadUTF8)
} }
case 42: case 42:
// line 156 "scan_tokens.rl" //line scan_tokens.rl:156
p = (te) - 1 p = (te) - 1
{ {
token(TokenStringLit) token(TokenStringLit)
} }
case 43: case 43:
// line 1 "NONE" //line NONE:1
switch act { switch act {
case 0: case 0:
{ {
@ -4978,60 +4936,51 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
case 44: case 44:
// line 264 "scan_tokens.rl" //line scan_tokens.rl:264
act = 16 act = 16
case 45: case 45:
// line 265 "scan_tokens.rl" //line scan_tokens.rl:265
act = 17 act = 17
case 46: case 46:
// line 265 "scan_tokens.rl" //line scan_tokens.rl:265
te = p + 1 te = p + 1
{ {
token(TokenBadUTF8) token(TokenBadUTF8)
} }
case 47: case 47:
// line 266 "scan_tokens.rl" //line scan_tokens.rl:266
te = p + 1 te = p + 1
{ {
token(TokenInvalid) token(TokenInvalid)
} }
case 48: case 48:
// line 264 "scan_tokens.rl" //line scan_tokens.rl:264
te = p te = p
p-- p--
{ {
token(TokenIdent) token(TokenIdent)
} }
case 49: case 49:
// line 265 "scan_tokens.rl" //line scan_tokens.rl:265
te = p te = p
p-- p--
{ {
token(TokenBadUTF8) token(TokenBadUTF8)
} }
case 50: case 50:
// line 264 "scan_tokens.rl" //line scan_tokens.rl:264
p = (te) - 1 p = (te) - 1
{ {
token(TokenIdent) token(TokenIdent)
} }
case 51: case 51:
// line 265 "scan_tokens.rl" //line scan_tokens.rl:265
p = (te) - 1 p = (te) - 1
{ {
token(TokenBadUTF8) token(TokenBadUTF8)
} }
case 52: case 52:
// line 1 "NONE" //line NONE:1
switch act { switch act {
case 16: case 16:
{ {
@ -5046,109 +4995,92 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
case 53: case 53:
// line 272 "scan_tokens.rl" //line scan_tokens.rl:272
act = 21 act = 21
case 54: case 54:
// line 285 "scan_tokens.rl" //line scan_tokens.rl:285
act = 32 act = 32
case 55: case 55:
// line 295 "scan_tokens.rl" //line scan_tokens.rl:295
act = 38 act = 38
case 56: case 56:
// line 296 "scan_tokens.rl" //line scan_tokens.rl:296
act = 39 act = 39
case 57: case 57:
// line 274 "scan_tokens.rl" //line scan_tokens.rl:274
te = p + 1 te = p + 1
{ {
token(TokenComment) token(TokenComment)
} }
case 58: case 58:
// line 275 "scan_tokens.rl" //line scan_tokens.rl:275
te = p + 1 te = p + 1
{ {
token(TokenNewline) token(TokenNewline)
} }
case 59: case 59:
// line 277 "scan_tokens.rl" //line scan_tokens.rl:277
te = p + 1 te = p + 1
{ {
token(TokenEqualOp) token(TokenEqualOp)
} }
case 60: case 60:
// line 278 "scan_tokens.rl" //line scan_tokens.rl:278
te = p + 1 te = p + 1
{ {
token(TokenNotEqual) token(TokenNotEqual)
} }
case 61: case 61:
// line 279 "scan_tokens.rl" //line scan_tokens.rl:279
te = p + 1 te = p + 1
{ {
token(TokenGreaterThanEq) token(TokenGreaterThanEq)
} }
case 62: case 62:
// line 280 "scan_tokens.rl" //line scan_tokens.rl:280
te = p + 1 te = p + 1
{ {
token(TokenLessThanEq) token(TokenLessThanEq)
} }
case 63: case 63:
// line 281 "scan_tokens.rl" //line scan_tokens.rl:281
te = p + 1 te = p + 1
{ {
token(TokenAnd) token(TokenAnd)
} }
case 64: case 64:
// line 282 "scan_tokens.rl" //line scan_tokens.rl:282
te = p + 1 te = p + 1
{ {
token(TokenOr) token(TokenOr)
} }
case 65: case 65:
// line 283 "scan_tokens.rl" //line scan_tokens.rl:283
te = p + 1 te = p + 1
{ {
token(TokenEllipsis) token(TokenEllipsis)
} }
case 66: case 66:
// line 284 "scan_tokens.rl" //line scan_tokens.rl:284
te = p + 1 te = p + 1
{ {
token(TokenFatArrow) token(TokenFatArrow)
} }
case 67: case 67:
// line 285 "scan_tokens.rl" //line scan_tokens.rl:285
te = p + 1 te = p + 1
{ {
selfToken() selfToken()
} }
case 68: case 68:
// line 180 "scan_tokens.rl" //line scan_tokens.rl:180
te = p + 1 te = p + 1
{ {
token(TokenOBrace) token(TokenOBrace)
braces++ braces++
} }
case 69: case 69:
// line 185 "scan_tokens.rl" //line scan_tokens.rl:185
te = p + 1 te = p + 1
{ {
if len(retBraces) > 0 && retBraces[len(retBraces)-1] == braces { if len(retBraces) > 0 && retBraces[len(retBraces)-1] == braces {
@ -5168,8 +5100,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 70: case 70:
// line 197 "scan_tokens.rl" //line scan_tokens.rl:197
te = p + 1 te = p + 1
{ {
// Only consume from the retBraces stack and return if we are at // Only consume from the retBraces stack and return if we are at
@ -5198,8 +5129,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 71: case 71:
// line 79 "scan_tokens.rl" //line scan_tokens.rl:79
te = p + 1 te = p + 1
{ {
token(TokenOQuote) token(TokenOQuote)
@ -5212,8 +5142,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 72: case 72:
// line 89 "scan_tokens.rl" //line scan_tokens.rl:89
te = p + 1 te = p + 1
{ {
token(TokenOHeredoc) token(TokenOHeredoc)
@ -5243,96 +5172,83 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
case 73: case 73:
// line 295 "scan_tokens.rl" //line scan_tokens.rl:295
te = p + 1 te = p + 1
{ {
token(TokenBadUTF8) token(TokenBadUTF8)
} }
case 74: case 74:
// line 296 "scan_tokens.rl" //line scan_tokens.rl:296
te = p + 1 te = p + 1
{ {
token(TokenInvalid) token(TokenInvalid)
} }
case 75: case 75:
// line 270 "scan_tokens.rl" //line scan_tokens.rl:270
te = p te = p
p-- p--
case 76: case 76:
// line 271 "scan_tokens.rl" //line scan_tokens.rl:271
te = p te = p
p-- p--
{ {
token(TokenNumberLit) token(TokenNumberLit)
} }
case 77: case 77:
// line 272 "scan_tokens.rl" //line scan_tokens.rl:272
te = p te = p
p-- p--
{ {
token(TokenIdent) token(TokenIdent)
} }
case 78: case 78:
// line 285 "scan_tokens.rl" //line scan_tokens.rl:285
te = p te = p
p-- p--
{ {
selfToken() selfToken()
} }
case 79: case 79:
// line 295 "scan_tokens.rl" //line scan_tokens.rl:295
te = p te = p
p-- p--
{ {
token(TokenBadUTF8) token(TokenBadUTF8)
} }
case 80: case 80:
// line 296 "scan_tokens.rl" //line scan_tokens.rl:296
te = p te = p
p-- p--
{ {
token(TokenInvalid) token(TokenInvalid)
} }
case 81: case 81:
// line 271 "scan_tokens.rl" //line scan_tokens.rl:271
p = (te) - 1 p = (te) - 1
{ {
token(TokenNumberLit) token(TokenNumberLit)
} }
case 82: case 82:
// line 272 "scan_tokens.rl" //line scan_tokens.rl:272
p = (te) - 1 p = (te) - 1
{ {
token(TokenIdent) token(TokenIdent)
} }
case 83: case 83:
// line 285 "scan_tokens.rl" //line scan_tokens.rl:285
p = (te) - 1 p = (te) - 1
{ {
selfToken() selfToken()
} }
case 84: case 84:
// line 295 "scan_tokens.rl" //line scan_tokens.rl:295
p = (te) - 1 p = (te) - 1
{ {
token(TokenBadUTF8) token(TokenBadUTF8)
} }
case 85: case 85:
// line 1 "NONE" //line NONE:1
switch act { switch act {
case 21: case 21:
{ {
@ -5356,7 +5272,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
// line 5222 "scan_tokens.go" //line scan_tokens.go:5138
} }
} }
@ -5368,16 +5284,14 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
_acts++ _acts++
switch _hcltok_actions[_acts-1] { switch _hcltok_actions[_acts-1] {
case 4: case 4:
// line 1 "NONE" //line NONE:1
ts = 0 ts = 0
case 5: case 5:
// line 1 "NONE" //line NONE:1
act = 0 act = 0
// line 5242 "scan_tokens.go" //line scan_tokens.go:5156
} }
} }
@ -5403,7 +5317,7 @@ func scanTokens(data []byte, filename string, start hcl.Pos, mode scanMode) []To
} }
} }
// line 357 "scan_tokens.rl" //line scan_tokens.rl:357
// If we fall out here without being in a final state then we've // If we fall out here without being in a final state then we've
// encountered something that the scanner can't match, which we'll // encountered something that the scanner can't match, which we'll