mirror of
https://github.com/prometheus/prometheus.git
synced 2024-12-30 07:59:40 -08:00
9a63ff0c32
This is to pull in bug fixes. However, it also requires code changes, see next commit.
982 lines
23 KiB
Go
982 lines
23 KiB
Go
package dns
|
|
|
|
import (
|
|
"io"
|
|
"log"
|
|
"os"
|
|
"strconv"
|
|
"strings"
|
|
)
|
|
|
|
type debugging bool
|
|
|
|
const debug debugging = false
|
|
|
|
func (d debugging) Printf(format string, args ...interface{}) {
|
|
if d {
|
|
log.Printf(format, args...)
|
|
}
|
|
}
|
|
|
|
const maxTok = 2048 // Largest token we can return.
|
|
const maxUint16 = 1<<16 - 1
|
|
|
|
// Tokinize a RFC 1035 zone file. The tokenizer will normalize it:
|
|
// * Add ownernames if they are left blank;
|
|
// * Suppress sequences of spaces;
|
|
// * Make each RR fit on one line (_NEWLINE is send as last)
|
|
// * Handle comments: ;
|
|
// * Handle braces - anywhere.
|
|
const (
|
|
// Zonefile
|
|
zEOF = iota
|
|
zString
|
|
zBlank
|
|
zQuote
|
|
zNewline
|
|
zRrtpe
|
|
zOwner
|
|
zClass
|
|
zDirOrigin // $ORIGIN
|
|
zDirTtl // $TTL
|
|
zDirInclude // $INCLUDE
|
|
zDirGenerate // $GENERATE
|
|
|
|
// Privatekey file
|
|
zValue
|
|
zKey
|
|
|
|
zExpectOwnerDir // Ownername
|
|
zExpectOwnerBl // Whitespace after the ownername
|
|
zExpectAny // Expect rrtype, ttl or class
|
|
zExpectAnyNoClass // Expect rrtype or ttl
|
|
zExpectAnyNoClassBl // The whitespace after _EXPECT_ANY_NOCLASS
|
|
zExpectAnyNoTtl // Expect rrtype or class
|
|
zExpectAnyNoTtlBl // Whitespace after _EXPECT_ANY_NOTTL
|
|
zExpectRrtype // Expect rrtype
|
|
zExpectRrtypeBl // Whitespace BEFORE rrtype
|
|
zExpectRdata // The first element of the rdata
|
|
zExpectDirTtlBl // Space after directive $TTL
|
|
zExpectDirTtl // Directive $TTL
|
|
zExpectDirOriginBl // Space after directive $ORIGIN
|
|
zExpectDirOrigin // Directive $ORIGIN
|
|
zExpectDirIncludeBl // Space after directive $INCLUDE
|
|
zExpectDirInclude // Directive $INCLUDE
|
|
zExpectDirGenerate // Directive $GENERATE
|
|
zExpectDirGenerateBl // Space after directive $GENERATE
|
|
)
|
|
|
|
// ParseError is a parsing error. It contains the parse error and the location in the io.Reader
|
|
// where the error occurred.
|
|
type ParseError struct {
|
|
file string
|
|
err string
|
|
lex lex
|
|
}
|
|
|
|
func (e *ParseError) Error() (s string) {
|
|
if e.file != "" {
|
|
s = e.file + ": "
|
|
}
|
|
s += "dns: " + e.err + ": " + strconv.QuoteToASCII(e.lex.token) + " at line: " +
|
|
strconv.Itoa(e.lex.line) + ":" + strconv.Itoa(e.lex.column)
|
|
return
|
|
}
|
|
|
|
type lex struct {
|
|
token string // text of the token
|
|
tokenUpper string // uppercase text of the token
|
|
length int // length of the token
|
|
err bool // when true, token text has lexer error
|
|
value uint8 // value: zString, _BLANK, etc.
|
|
line int // line in the file
|
|
column int // column in the file
|
|
torc uint16 // type or class as parsed in the lexer, we only need to look this up in the grammar
|
|
comment string // any comment text seen
|
|
}
|
|
|
|
// Token holds the token that are returned when a zone file is parsed.
|
|
type Token struct {
|
|
// The scanned resource record when error is not nil.
|
|
RR
|
|
// When an error occurred, this has the error specifics.
|
|
Error *ParseError
|
|
// A potential comment positioned after the RR and on the same line.
|
|
Comment string
|
|
}
|
|
|
|
// NewRR reads the RR contained in the string s. Only the first RR is
|
|
// returned. If s contains no RR, return nil with no error. The class
|
|
// defaults to IN and TTL defaults to 3600. The full zone file syntax
|
|
// like $TTL, $ORIGIN, etc. is supported. All fields of the returned
|
|
// RR are set, except RR.Header().Rdlength which is set to 0.
|
|
func NewRR(s string) (RR, error) {
|
|
if len(s) > 0 && s[len(s)-1] != '\n' { // We need a closing newline
|
|
return ReadRR(strings.NewReader(s+"\n"), "")
|
|
}
|
|
return ReadRR(strings.NewReader(s), "")
|
|
}
|
|
|
|
// ReadRR reads the RR contained in q.
|
|
// See NewRR for more documentation.
|
|
func ReadRR(q io.Reader, filename string) (RR, error) {
|
|
r := <-parseZoneHelper(q, ".", filename, 1)
|
|
if r == nil {
|
|
return nil, nil
|
|
}
|
|
|
|
if r.Error != nil {
|
|
return nil, r.Error
|
|
}
|
|
return r.RR, nil
|
|
}
|
|
|
|
// ParseZone reads a RFC 1035 style zonefile from r. It returns *Tokens on the
|
|
// returned channel, which consist out the parsed RR, a potential comment or an error.
|
|
// If there is an error the RR is nil. The string file is only used
|
|
// in error reporting. The string origin is used as the initial origin, as
|
|
// if the file would start with: $ORIGIN origin .
|
|
// The directives $INCLUDE, $ORIGIN, $TTL and $GENERATE are supported.
|
|
// The channel t is closed by ParseZone when the end of r is reached.
|
|
//
|
|
// Basic usage pattern when reading from a string (z) containing the
|
|
// zone data:
|
|
//
|
|
// for x := range dns.ParseZone(strings.NewReader(z), "", "") {
|
|
// if x.Error != nil {
|
|
// // log.Println(x.Error)
|
|
// } else {
|
|
// // Do something with x.RR
|
|
// }
|
|
// }
|
|
//
|
|
// Comments specified after an RR (and on the same line!) are returned too:
|
|
//
|
|
// foo. IN A 10.0.0.1 ; this is a comment
|
|
//
|
|
// The text "; this is comment" is returned in Token.Comment. Comments inside the
|
|
// RR are discarded. Comments on a line by themselves are discarded too.
|
|
func ParseZone(r io.Reader, origin, file string) chan *Token {
|
|
return parseZoneHelper(r, origin, file, 10000)
|
|
}
|
|
|
|
func parseZoneHelper(r io.Reader, origin, file string, chansize int) chan *Token {
|
|
t := make(chan *Token, chansize)
|
|
go parseZone(r, origin, file, t, 0)
|
|
return t
|
|
}
|
|
|
|
func parseZone(r io.Reader, origin, f string, t chan *Token, include int) {
|
|
defer func() {
|
|
if include == 0 {
|
|
close(t)
|
|
}
|
|
}()
|
|
s := scanInit(r)
|
|
c := make(chan lex)
|
|
// Start the lexer
|
|
go zlexer(s, c)
|
|
// 6 possible beginnings of a line, _ is a space
|
|
// 0. zRRTYPE -> all omitted until the rrtype
|
|
// 1. zOwner _ zRrtype -> class/ttl omitted
|
|
// 2. zOwner _ zString _ zRrtype -> class omitted
|
|
// 3. zOwner _ zString _ zClass _ zRrtype -> ttl/class
|
|
// 4. zOwner _ zClass _ zRrtype -> ttl omitted
|
|
// 5. zOwner _ zClass _ zString _ zRrtype -> class/ttl (reversed)
|
|
// After detecting these, we know the zRrtype so we can jump to functions
|
|
// handling the rdata for each of these types.
|
|
|
|
if origin == "" {
|
|
origin = "."
|
|
}
|
|
origin = Fqdn(origin)
|
|
if _, ok := IsDomainName(origin); !ok {
|
|
t <- &Token{Error: &ParseError{f, "bad initial origin name", lex{}}}
|
|
return
|
|
}
|
|
|
|
st := zExpectOwnerDir // initial state
|
|
var h RR_Header
|
|
var defttl uint32 = defaultTtl
|
|
var prevName string
|
|
for l := range c {
|
|
// Lexer spotted an error already
|
|
if l.err == true {
|
|
t <- &Token{Error: &ParseError{f, l.token, l}}
|
|
return
|
|
|
|
}
|
|
switch st {
|
|
case zExpectOwnerDir:
|
|
// We can also expect a directive, like $TTL or $ORIGIN
|
|
h.Ttl = defttl
|
|
h.Class = ClassINET
|
|
switch l.value {
|
|
case zNewline:
|
|
st = zExpectOwnerDir
|
|
case zOwner:
|
|
h.Name = l.token
|
|
if l.token[0] == '@' {
|
|
h.Name = origin
|
|
prevName = h.Name
|
|
st = zExpectOwnerBl
|
|
break
|
|
}
|
|
if h.Name[l.length-1] != '.' {
|
|
h.Name = appendOrigin(h.Name, origin)
|
|
}
|
|
_, ok := IsDomainName(l.token)
|
|
if !ok {
|
|
t <- &Token{Error: &ParseError{f, "bad owner name", l}}
|
|
return
|
|
}
|
|
prevName = h.Name
|
|
st = zExpectOwnerBl
|
|
case zDirTtl:
|
|
st = zExpectDirTtlBl
|
|
case zDirOrigin:
|
|
st = zExpectDirOriginBl
|
|
case zDirInclude:
|
|
st = zExpectDirIncludeBl
|
|
case zDirGenerate:
|
|
st = zExpectDirGenerateBl
|
|
case zRrtpe:
|
|
h.Name = prevName
|
|
h.Rrtype = l.torc
|
|
st = zExpectRdata
|
|
case zClass:
|
|
h.Name = prevName
|
|
h.Class = l.torc
|
|
st = zExpectAnyNoClassBl
|
|
case zBlank:
|
|
// Discard, can happen when there is nothing on the
|
|
// line except the RR type
|
|
case zString:
|
|
ttl, ok := stringToTtl(l.token)
|
|
if !ok {
|
|
t <- &Token{Error: &ParseError{f, "not a TTL", l}}
|
|
return
|
|
}
|
|
h.Ttl = ttl
|
|
// Don't about the defttl, we should take the $TTL value
|
|
// defttl = ttl
|
|
st = zExpectAnyNoTtlBl
|
|
|
|
default:
|
|
t <- &Token{Error: &ParseError{f, "syntax error at beginning", l}}
|
|
return
|
|
}
|
|
case zExpectDirIncludeBl:
|
|
if l.value != zBlank {
|
|
t <- &Token{Error: &ParseError{f, "no blank after $INCLUDE-directive", l}}
|
|
return
|
|
}
|
|
st = zExpectDirInclude
|
|
case zExpectDirInclude:
|
|
if l.value != zString {
|
|
t <- &Token{Error: &ParseError{f, "expecting $INCLUDE value, not this...", l}}
|
|
return
|
|
}
|
|
neworigin := origin // There may be optionally a new origin set after the filename, if not use current one
|
|
l := <-c
|
|
switch l.value {
|
|
case zBlank:
|
|
l := <-c
|
|
if l.value == zString {
|
|
if _, ok := IsDomainName(l.token); !ok || l.length == 0 || l.err {
|
|
t <- &Token{Error: &ParseError{f, "bad origin name", l}}
|
|
return
|
|
}
|
|
// a new origin is specified.
|
|
if l.token[l.length-1] != '.' {
|
|
if origin != "." { // Prevent .. endings
|
|
neworigin = l.token + "." + origin
|
|
} else {
|
|
neworigin = l.token + origin
|
|
}
|
|
} else {
|
|
neworigin = l.token
|
|
}
|
|
}
|
|
case zNewline, zEOF:
|
|
// Ok
|
|
default:
|
|
t <- &Token{Error: &ParseError{f, "garbage after $INCLUDE", l}}
|
|
return
|
|
}
|
|
// Start with the new file
|
|
r1, e1 := os.Open(l.token)
|
|
if e1 != nil {
|
|
t <- &Token{Error: &ParseError{f, "failed to open `" + l.token + "'", l}}
|
|
return
|
|
}
|
|
if include+1 > 7 {
|
|
t <- &Token{Error: &ParseError{f, "too deeply nested $INCLUDE", l}}
|
|
return
|
|
}
|
|
parseZone(r1, l.token, neworigin, t, include+1)
|
|
st = zExpectOwnerDir
|
|
case zExpectDirTtlBl:
|
|
if l.value != zBlank {
|
|
t <- &Token{Error: &ParseError{f, "no blank after $TTL-directive", l}}
|
|
return
|
|
}
|
|
st = zExpectDirTtl
|
|
case zExpectDirTtl:
|
|
if l.value != zString {
|
|
t <- &Token{Error: &ParseError{f, "expecting $TTL value, not this...", l}}
|
|
return
|
|
}
|
|
if e, _ := slurpRemainder(c, f); e != nil {
|
|
t <- &Token{Error: e}
|
|
return
|
|
}
|
|
ttl, ok := stringToTtl(l.token)
|
|
if !ok {
|
|
t <- &Token{Error: &ParseError{f, "expecting $TTL value, not this...", l}}
|
|
return
|
|
}
|
|
defttl = ttl
|
|
st = zExpectOwnerDir
|
|
case zExpectDirOriginBl:
|
|
if l.value != zBlank {
|
|
t <- &Token{Error: &ParseError{f, "no blank after $ORIGIN-directive", l}}
|
|
return
|
|
}
|
|
st = zExpectDirOrigin
|
|
case zExpectDirOrigin:
|
|
if l.value != zString {
|
|
t <- &Token{Error: &ParseError{f, "expecting $ORIGIN value, not this...", l}}
|
|
return
|
|
}
|
|
if e, _ := slurpRemainder(c, f); e != nil {
|
|
t <- &Token{Error: e}
|
|
}
|
|
if _, ok := IsDomainName(l.token); !ok {
|
|
t <- &Token{Error: &ParseError{f, "bad origin name", l}}
|
|
return
|
|
}
|
|
if l.token[l.length-1] != '.' {
|
|
if origin != "." { // Prevent .. endings
|
|
origin = l.token + "." + origin
|
|
} else {
|
|
origin = l.token + origin
|
|
}
|
|
} else {
|
|
origin = l.token
|
|
}
|
|
st = zExpectOwnerDir
|
|
case zExpectDirGenerateBl:
|
|
if l.value != zBlank {
|
|
t <- &Token{Error: &ParseError{f, "no blank after $GENERATE-directive", l}}
|
|
return
|
|
}
|
|
st = zExpectDirGenerate
|
|
case zExpectDirGenerate:
|
|
if l.value != zString {
|
|
t <- &Token{Error: &ParseError{f, "expecting $GENERATE value, not this...", l}}
|
|
return
|
|
}
|
|
if errMsg := generate(l, c, t, origin); errMsg != "" {
|
|
t <- &Token{Error: &ParseError{f, errMsg, l}}
|
|
return
|
|
}
|
|
st = zExpectOwnerDir
|
|
case zExpectOwnerBl:
|
|
if l.value != zBlank {
|
|
t <- &Token{Error: &ParseError{f, "no blank after owner", l}}
|
|
return
|
|
}
|
|
st = zExpectAny
|
|
case zExpectAny:
|
|
switch l.value {
|
|
case zRrtpe:
|
|
h.Rrtype = l.torc
|
|
st = zExpectRdata
|
|
case zClass:
|
|
h.Class = l.torc
|
|
st = zExpectAnyNoClassBl
|
|
case zString:
|
|
ttl, ok := stringToTtl(l.token)
|
|
if !ok {
|
|
t <- &Token{Error: &ParseError{f, "not a TTL", l}}
|
|
return
|
|
}
|
|
h.Ttl = ttl
|
|
// defttl = ttl // don't set the defttl here
|
|
st = zExpectAnyNoTtlBl
|
|
default:
|
|
t <- &Token{Error: &ParseError{f, "expecting RR type, TTL or class, not this...", l}}
|
|
return
|
|
}
|
|
case zExpectAnyNoClassBl:
|
|
if l.value != zBlank {
|
|
t <- &Token{Error: &ParseError{f, "no blank before class", l}}
|
|
return
|
|
}
|
|
st = zExpectAnyNoClass
|
|
case zExpectAnyNoTtlBl:
|
|
if l.value != zBlank {
|
|
t <- &Token{Error: &ParseError{f, "no blank before TTL", l}}
|
|
return
|
|
}
|
|
st = zExpectAnyNoTtl
|
|
case zExpectAnyNoTtl:
|
|
switch l.value {
|
|
case zClass:
|
|
h.Class = l.torc
|
|
st = zExpectRrtypeBl
|
|
case zRrtpe:
|
|
h.Rrtype = l.torc
|
|
st = zExpectRdata
|
|
default:
|
|
t <- &Token{Error: &ParseError{f, "expecting RR type or class, not this...", l}}
|
|
return
|
|
}
|
|
case zExpectAnyNoClass:
|
|
switch l.value {
|
|
case zString:
|
|
ttl, ok := stringToTtl(l.token)
|
|
if !ok {
|
|
t <- &Token{Error: &ParseError{f, "not a TTL", l}}
|
|
return
|
|
}
|
|
h.Ttl = ttl
|
|
// defttl = ttl // don't set the def ttl anymore
|
|
st = zExpectRrtypeBl
|
|
case zRrtpe:
|
|
h.Rrtype = l.torc
|
|
st = zExpectRdata
|
|
default:
|
|
t <- &Token{Error: &ParseError{f, "expecting RR type or TTL, not this...", l}}
|
|
return
|
|
}
|
|
case zExpectRrtypeBl:
|
|
if l.value != zBlank {
|
|
t <- &Token{Error: &ParseError{f, "no blank before RR type", l}}
|
|
return
|
|
}
|
|
st = zExpectRrtype
|
|
case zExpectRrtype:
|
|
if l.value != zRrtpe {
|
|
t <- &Token{Error: &ParseError{f, "unknown RR type", l}}
|
|
return
|
|
}
|
|
h.Rrtype = l.torc
|
|
st = zExpectRdata
|
|
case zExpectRdata:
|
|
r, e, c1 := setRR(h, c, origin, f)
|
|
if e != nil {
|
|
// If e.lex is nil than we have encounter a unknown RR type
|
|
// in that case we substitute our current lex token
|
|
if e.lex.token == "" && e.lex.value == 0 {
|
|
e.lex = l // Uh, dirty
|
|
}
|
|
t <- &Token{Error: e}
|
|
return
|
|
}
|
|
t <- &Token{RR: r, Comment: c1}
|
|
st = zExpectOwnerDir
|
|
}
|
|
}
|
|
// If we get here, we and the h.Rrtype is still zero, we haven't parsed anything, this
|
|
// is not an error, because an empty zone file is still a zone file.
|
|
}
|
|
|
|
// zlexer scans the sourcefile and returns tokens on the channel c.
|
|
func zlexer(s *scan, c chan lex) {
|
|
var l lex
|
|
str := make([]byte, maxTok) // Should be enough for any token
|
|
stri := 0 // Offset in str (0 means empty)
|
|
com := make([]byte, maxTok) // Hold comment text
|
|
comi := 0
|
|
quote := false
|
|
escape := false
|
|
space := false
|
|
commt := false
|
|
rrtype := false
|
|
owner := true
|
|
brace := 0
|
|
x, err := s.tokenText()
|
|
defer close(c)
|
|
for err == nil {
|
|
l.column = s.position.Column
|
|
l.line = s.position.Line
|
|
if stri >= maxTok {
|
|
l.token = "token length insufficient for parsing"
|
|
l.err = true
|
|
debug.Printf("[%+v]", l.token)
|
|
c <- l
|
|
return
|
|
}
|
|
if comi >= maxTok {
|
|
l.token = "comment length insufficient for parsing"
|
|
l.err = true
|
|
debug.Printf("[%+v]", l.token)
|
|
c <- l
|
|
return
|
|
}
|
|
|
|
switch x {
|
|
case ' ', '\t':
|
|
if escape {
|
|
escape = false
|
|
str[stri] = x
|
|
stri++
|
|
break
|
|
}
|
|
if quote {
|
|
// Inside quotes this is legal
|
|
str[stri] = x
|
|
stri++
|
|
break
|
|
}
|
|
if commt {
|
|
com[comi] = x
|
|
comi++
|
|
break
|
|
}
|
|
if stri == 0 {
|
|
// Space directly in the beginning, handled in the grammar
|
|
} else if owner {
|
|
// If we have a string and its the first, make it an owner
|
|
l.value = zOwner
|
|
l.token = string(str[:stri])
|
|
l.tokenUpper = strings.ToUpper(l.token)
|
|
l.length = stri
|
|
// escape $... start with a \ not a $, so this will work
|
|
switch l.tokenUpper {
|
|
case "$TTL":
|
|
l.value = zDirTtl
|
|
case "$ORIGIN":
|
|
l.value = zDirOrigin
|
|
case "$INCLUDE":
|
|
l.value = zDirInclude
|
|
case "$GENERATE":
|
|
l.value = zDirGenerate
|
|
}
|
|
debug.Printf("[7 %+v]", l.token)
|
|
c <- l
|
|
} else {
|
|
l.value = zString
|
|
l.token = string(str[:stri])
|
|
l.tokenUpper = strings.ToUpper(l.token)
|
|
l.length = stri
|
|
if !rrtype {
|
|
if t, ok := StringToType[l.tokenUpper]; ok {
|
|
l.value = zRrtpe
|
|
l.torc = t
|
|
rrtype = true
|
|
} else {
|
|
if strings.HasPrefix(l.tokenUpper, "TYPE") {
|
|
t, ok := typeToInt(l.token)
|
|
if !ok {
|
|
l.token = "unknown RR type"
|
|
l.err = true
|
|
c <- l
|
|
return
|
|
}
|
|
l.value = zRrtpe
|
|
l.torc = t
|
|
}
|
|
}
|
|
if t, ok := StringToClass[l.tokenUpper]; ok {
|
|
l.value = zClass
|
|
l.torc = t
|
|
} else {
|
|
if strings.HasPrefix(l.tokenUpper, "CLASS") {
|
|
t, ok := classToInt(l.token)
|
|
if !ok {
|
|
l.token = "unknown class"
|
|
l.err = true
|
|
c <- l
|
|
return
|
|
}
|
|
l.value = zClass
|
|
l.torc = t
|
|
}
|
|
}
|
|
}
|
|
debug.Printf("[6 %+v]", l.token)
|
|
c <- l
|
|
}
|
|
stri = 0
|
|
// I reverse space stuff here
|
|
if !space && !commt {
|
|
l.value = zBlank
|
|
l.token = " "
|
|
l.length = 1
|
|
debug.Printf("[5 %+v]", l.token)
|
|
c <- l
|
|
}
|
|
owner = false
|
|
space = true
|
|
case ';':
|
|
if escape {
|
|
escape = false
|
|
str[stri] = x
|
|
stri++
|
|
break
|
|
}
|
|
if quote {
|
|
// Inside quotes this is legal
|
|
str[stri] = x
|
|
stri++
|
|
break
|
|
}
|
|
if stri > 0 {
|
|
l.value = zString
|
|
l.token = string(str[:stri])
|
|
l.tokenUpper = strings.ToUpper(l.token)
|
|
l.length = stri
|
|
debug.Printf("[4 %+v]", l.token)
|
|
c <- l
|
|
stri = 0
|
|
}
|
|
commt = true
|
|
com[comi] = ';'
|
|
comi++
|
|
case '\r':
|
|
escape = false
|
|
if quote {
|
|
str[stri] = x
|
|
stri++
|
|
break
|
|
}
|
|
// discard if outside of quotes
|
|
case '\n':
|
|
escape = false
|
|
// Escaped newline
|
|
if quote {
|
|
str[stri] = x
|
|
stri++
|
|
break
|
|
}
|
|
// inside quotes this is legal
|
|
if commt {
|
|
// Reset a comment
|
|
commt = false
|
|
rrtype = false
|
|
stri = 0
|
|
// If not in a brace this ends the comment AND the RR
|
|
if brace == 0 {
|
|
owner = true
|
|
owner = true
|
|
l.value = zNewline
|
|
l.token = "\n"
|
|
l.tokenUpper = l.token
|
|
l.length = 1
|
|
l.comment = string(com[:comi])
|
|
debug.Printf("[3 %+v %+v]", l.token, l.comment)
|
|
c <- l
|
|
l.comment = ""
|
|
comi = 0
|
|
break
|
|
}
|
|
com[comi] = ' ' // convert newline to space
|
|
comi++
|
|
break
|
|
}
|
|
|
|
if brace == 0 {
|
|
// If there is previous text, we should output it here
|
|
if stri != 0 {
|
|
l.value = zString
|
|
l.token = string(str[:stri])
|
|
l.tokenUpper = strings.ToUpper(l.token)
|
|
|
|
l.length = stri
|
|
if !rrtype {
|
|
if t, ok := StringToType[l.tokenUpper]; ok {
|
|
l.value = zRrtpe
|
|
l.torc = t
|
|
rrtype = true
|
|
}
|
|
}
|
|
debug.Printf("[2 %+v]", l.token)
|
|
c <- l
|
|
}
|
|
l.value = zNewline
|
|
l.token = "\n"
|
|
l.tokenUpper = l.token
|
|
l.length = 1
|
|
debug.Printf("[1 %+v]", l.token)
|
|
c <- l
|
|
stri = 0
|
|
commt = false
|
|
rrtype = false
|
|
owner = true
|
|
comi = 0
|
|
}
|
|
case '\\':
|
|
// comments do not get escaped chars, everything is copied
|
|
if commt {
|
|
com[comi] = x
|
|
comi++
|
|
break
|
|
}
|
|
// something already escaped must be in string
|
|
if escape {
|
|
str[stri] = x
|
|
stri++
|
|
escape = false
|
|
break
|
|
}
|
|
// something escaped outside of string gets added to string
|
|
str[stri] = x
|
|
stri++
|
|
escape = true
|
|
case '"':
|
|
if commt {
|
|
com[comi] = x
|
|
comi++
|
|
break
|
|
}
|
|
if escape {
|
|
str[stri] = x
|
|
stri++
|
|
escape = false
|
|
break
|
|
}
|
|
space = false
|
|
// send previous gathered text and the quote
|
|
if stri != 0 {
|
|
l.value = zString
|
|
l.token = string(str[:stri])
|
|
l.tokenUpper = strings.ToUpper(l.token)
|
|
l.length = stri
|
|
|
|
debug.Printf("[%+v]", l.token)
|
|
c <- l
|
|
stri = 0
|
|
}
|
|
|
|
// send quote itself as separate token
|
|
l.value = zQuote
|
|
l.token = "\""
|
|
l.tokenUpper = l.token
|
|
l.length = 1
|
|
c <- l
|
|
quote = !quote
|
|
case '(', ')':
|
|
if commt {
|
|
com[comi] = x
|
|
comi++
|
|
break
|
|
}
|
|
if escape {
|
|
str[stri] = x
|
|
stri++
|
|
escape = false
|
|
break
|
|
}
|
|
if quote {
|
|
str[stri] = x
|
|
stri++
|
|
break
|
|
}
|
|
switch x {
|
|
case ')':
|
|
brace--
|
|
if brace < 0 {
|
|
l.token = "extra closing brace"
|
|
l.tokenUpper = l.token
|
|
l.err = true
|
|
debug.Printf("[%+v]", l.token)
|
|
c <- l
|
|
return
|
|
}
|
|
case '(':
|
|
brace++
|
|
}
|
|
default:
|
|
escape = false
|
|
if commt {
|
|
com[comi] = x
|
|
comi++
|
|
break
|
|
}
|
|
str[stri] = x
|
|
stri++
|
|
space = false
|
|
}
|
|
x, err = s.tokenText()
|
|
}
|
|
if stri > 0 {
|
|
// Send remainder
|
|
l.token = string(str[:stri])
|
|
l.tokenUpper = strings.ToUpper(l.token)
|
|
l.length = stri
|
|
l.value = zString
|
|
debug.Printf("[%+v]", l.token)
|
|
c <- l
|
|
}
|
|
}
|
|
|
|
// Extract the class number from CLASSxx
|
|
func classToInt(token string) (uint16, bool) {
|
|
offset := 5
|
|
if len(token) < offset+1 {
|
|
return 0, false
|
|
}
|
|
class, ok := strconv.Atoi(token[offset:])
|
|
if ok != nil || class > maxUint16 {
|
|
return 0, false
|
|
}
|
|
return uint16(class), true
|
|
}
|
|
|
|
// Extract the rr number from TYPExxx
|
|
func typeToInt(token string) (uint16, bool) {
|
|
offset := 4
|
|
if len(token) < offset+1 {
|
|
return 0, false
|
|
}
|
|
typ, ok := strconv.Atoi(token[offset:])
|
|
if ok != nil || typ > maxUint16 {
|
|
return 0, false
|
|
}
|
|
return uint16(typ), true
|
|
}
|
|
|
|
// Parse things like 2w, 2m, etc, Return the time in seconds.
|
|
func stringToTtl(token string) (uint32, bool) {
|
|
s := uint32(0)
|
|
i := uint32(0)
|
|
for _, c := range token {
|
|
switch c {
|
|
case 's', 'S':
|
|
s += i
|
|
i = 0
|
|
case 'm', 'M':
|
|
s += i * 60
|
|
i = 0
|
|
case 'h', 'H':
|
|
s += i * 60 * 60
|
|
i = 0
|
|
case 'd', 'D':
|
|
s += i * 60 * 60 * 24
|
|
i = 0
|
|
case 'w', 'W':
|
|
s += i * 60 * 60 * 24 * 7
|
|
i = 0
|
|
case '0', '1', '2', '3', '4', '5', '6', '7', '8', '9':
|
|
i *= 10
|
|
i += uint32(c) - '0'
|
|
default:
|
|
return 0, false
|
|
}
|
|
}
|
|
return s + i, true
|
|
}
|
|
|
|
// Parse LOC records' <digits>[.<digits>][mM] into a
|
|
// mantissa exponent format. Token should contain the entire
|
|
// string (i.e. no spaces allowed)
|
|
func stringToCm(token string) (e, m uint8, ok bool) {
|
|
if token[len(token)-1] == 'M' || token[len(token)-1] == 'm' {
|
|
token = token[0 : len(token)-1]
|
|
}
|
|
s := strings.SplitN(token, ".", 2)
|
|
var meters, cmeters, val int
|
|
var err error
|
|
switch len(s) {
|
|
case 2:
|
|
if cmeters, err = strconv.Atoi(s[1]); err != nil {
|
|
return
|
|
}
|
|
fallthrough
|
|
case 1:
|
|
if meters, err = strconv.Atoi(s[0]); err != nil {
|
|
return
|
|
}
|
|
case 0:
|
|
// huh?
|
|
return 0, 0, false
|
|
}
|
|
ok = true
|
|
if meters > 0 {
|
|
e = 2
|
|
val = meters
|
|
} else {
|
|
e = 0
|
|
val = cmeters
|
|
}
|
|
for val > 10 {
|
|
e++
|
|
val /= 10
|
|
}
|
|
if e > 9 {
|
|
ok = false
|
|
}
|
|
m = uint8(val)
|
|
return
|
|
}
|
|
|
|
func appendOrigin(name, origin string) string {
|
|
if origin == "." {
|
|
return name + origin
|
|
}
|
|
return name + "." + origin
|
|
}
|
|
|
|
// LOC record helper function
|
|
func locCheckNorth(token string, latitude uint32) (uint32, bool) {
|
|
switch token {
|
|
case "n", "N":
|
|
return LOC_EQUATOR + latitude, true
|
|
case "s", "S":
|
|
return LOC_EQUATOR - latitude, true
|
|
}
|
|
return latitude, false
|
|
}
|
|
|
|
// LOC record helper function
|
|
func locCheckEast(token string, longitude uint32) (uint32, bool) {
|
|
switch token {
|
|
case "e", "E":
|
|
return LOC_EQUATOR + longitude, true
|
|
case "w", "W":
|
|
return LOC_EQUATOR - longitude, true
|
|
}
|
|
return longitude, false
|
|
}
|
|
|
|
// "Eat" the rest of the "line". Return potential comments
|
|
func slurpRemainder(c chan lex, f string) (*ParseError, string) {
|
|
l := <-c
|
|
com := ""
|
|
switch l.value {
|
|
case zBlank:
|
|
l = <-c
|
|
com = l.comment
|
|
if l.value != zNewline && l.value != zEOF {
|
|
return &ParseError{f, "garbage after rdata", l}, ""
|
|
}
|
|
case zNewline:
|
|
com = l.comment
|
|
case zEOF:
|
|
default:
|
|
return &ParseError{f, "garbage after rdata", l}, ""
|
|
}
|
|
return nil, com
|
|
}
|
|
|
|
// Parse a 64 bit-like ipv6 address: "0014:4fff:ff20:ee64"
|
|
// Used for NID and L64 record.
|
|
func stringToNodeID(l lex) (uint64, *ParseError) {
|
|
if len(l.token) < 19 {
|
|
return 0, &ParseError{l.token, "bad NID/L64 NodeID/Locator64", l}
|
|
}
|
|
// There must be three colons at fixes postitions, if not its a parse error
|
|
if l.token[4] != ':' && l.token[9] != ':' && l.token[14] != ':' {
|
|
return 0, &ParseError{l.token, "bad NID/L64 NodeID/Locator64", l}
|
|
}
|
|
s := l.token[0:4] + l.token[5:9] + l.token[10:14] + l.token[15:19]
|
|
u, err := strconv.ParseUint(s, 16, 64)
|
|
if err != nil {
|
|
return 0, &ParseError{l.token, "bad NID/L64 NodeID/Locator64", l}
|
|
}
|
|
return u, nil
|
|
}
|