1762 lines
		
	
	
		
			44 KiB
		
	
	
	
		
			JavaScript
		
	
	
	
	
	
			
		
		
	
	
			1762 lines
		
	
	
		
			44 KiB
		
	
	
	
		
			JavaScript
		
	
	
	
	
	
| 'use strict'
 | |
| 
 | |
| const Buffer = require('buffer').Buffer
 | |
| const types = require('./types')
 | |
| const rcodes = require('./rcodes')
 | |
| const opcodes = require('./opcodes')
 | |
| const classes = require('./classes')
 | |
| const optioncodes = require('./optioncodes')
 | |
| const ip = require('@leichtgewicht/ip-codec')
 | |
| 
 | |
| const QUERY_FLAG = 0
 | |
| const RESPONSE_FLAG = 1 << 15
 | |
| const FLUSH_MASK = 1 << 15
 | |
| const NOT_FLUSH_MASK = ~FLUSH_MASK
 | |
| const QU_MASK = 1 << 15
 | |
| const NOT_QU_MASK = ~QU_MASK
 | |
| 
 | |
| const name = exports.name = {}
 | |
| 
 | |
| name.encode = function (str, buf, offset, { mail = false } = {}) {
 | |
|   if (!buf) buf = Buffer.alloc(name.encodingLength(str))
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   // strip leading and trailing .
 | |
|   const n = str.replace(/^\.|\.$/gm, '')
 | |
|   if (n.length) {
 | |
|     let list = []
 | |
|     if (mail) {
 | |
|       let localPart = ''
 | |
|       n.split('.').forEach(label => {
 | |
|         if (label.endsWith('\\')) {
 | |
|           localPart += (localPart.length ? '.' : '') + label.slice(0, -1)
 | |
|         } else {
 | |
|           if (list.length === 0 && localPart.length) {
 | |
|             list.push(localPart + '.' + label)
 | |
|           } else {
 | |
|             list.push(label)
 | |
|           }
 | |
|         }
 | |
|       })
 | |
|     } else {
 | |
|       list = n.split('.')
 | |
|     }
 | |
| 
 | |
|     for (let i = 0; i < list.length; i++) {
 | |
|       const len = buf.write(list[i], offset + 1)
 | |
|       buf[offset] = len
 | |
|       offset += len + 1
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   buf[offset++] = 0
 | |
| 
 | |
|   name.encode.bytes = offset - oldOffset
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| name.encode.bytes = 0
 | |
| 
 | |
| name.decode = function (buf, offset, { mail = false } = {}) {
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const list = []
 | |
|   let oldOffset = offset
 | |
|   let totalLength = 0
 | |
|   let consumedBytes = 0
 | |
|   let jumped = false
 | |
| 
 | |
|   while (true) {
 | |
|     if (offset >= buf.length) {
 | |
|       throw new Error('Cannot decode name (buffer overflow)')
 | |
|     }
 | |
|     const len = buf[offset++]
 | |
|     consumedBytes += jumped ? 0 : 1
 | |
| 
 | |
|     if (len === 0) {
 | |
|       break
 | |
|     } else if ((len & 0xc0) === 0) {
 | |
|       if (offset + len > buf.length) {
 | |
|         throw new Error('Cannot decode name (buffer overflow)')
 | |
|       }
 | |
|       totalLength += len + 1
 | |
|       if (totalLength > 254) {
 | |
|         throw new Error('Cannot decode name (name too long)')
 | |
|       }
 | |
|       let label = buf.toString('utf-8', offset, offset + len)
 | |
|       if (mail) {
 | |
|         label = label.replace(/\./g, '\\.')
 | |
|       }
 | |
|       list.push(label)
 | |
|       offset += len
 | |
|       consumedBytes += jumped ? 0 : len
 | |
|     } else if ((len & 0xc0) === 0xc0) {
 | |
|       if (offset + 1 > buf.length) {
 | |
|         throw new Error('Cannot decode name (buffer overflow)')
 | |
|       }
 | |
|       const jumpOffset = buf.readUInt16BE(offset - 1) - 0xc000
 | |
|       if (jumpOffset >= oldOffset) {
 | |
|         // Allow only pointers to prior data. RFC 1035, section 4.1.4 states:
 | |
|         // "[...] an entire domain name or a list of labels at the end of a domain name
 | |
|         // is replaced with a pointer to a prior occurance (sic) of the same name."
 | |
|         throw new Error('Cannot decode name (bad pointer)')
 | |
|       }
 | |
|       offset = jumpOffset
 | |
|       oldOffset = jumpOffset
 | |
|       consumedBytes += jumped ? 0 : 1
 | |
|       jumped = true
 | |
|     } else {
 | |
|       throw new Error('Cannot decode name (bad label)')
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   name.decode.bytes = consumedBytes
 | |
|   return list.length === 0 ? '.' : list.join('.')
 | |
| }
 | |
| 
 | |
| name.decode.bytes = 0
 | |
| 
 | |
| name.encodingLength = function (n) {
 | |
|   if (n === '.' || n === '..') return 1
 | |
|   return Buffer.byteLength(n.replace(/^\.|\.$/gm, '')) + 2
 | |
| }
 | |
| 
 | |
| const string = {}
 | |
| 
 | |
| string.encode = function (s, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(string.encodingLength(s))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const len = buf.write(s, offset + 1)
 | |
|   buf[offset] = len
 | |
|   string.encode.bytes = len + 1
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| string.encode.bytes = 0
 | |
| 
 | |
| string.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const len = buf[offset]
 | |
|   const s = buf.toString('utf-8', offset + 1, offset + 1 + len)
 | |
|   string.decode.bytes = len + 1
 | |
|   return s
 | |
| }
 | |
| 
 | |
| string.decode.bytes = 0
 | |
| 
 | |
| string.encodingLength = function (s) {
 | |
|   return Buffer.byteLength(s) + 1
 | |
| }
 | |
| 
 | |
| const header = {}
 | |
| 
 | |
| header.encode = function (h, buf, offset) {
 | |
|   if (!buf) buf = header.encodingLength(h)
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const flags = (h.flags || 0) & 32767
 | |
|   const type = h.type === 'response' ? RESPONSE_FLAG : QUERY_FLAG
 | |
| 
 | |
|   buf.writeUInt16BE(h.id || 0, offset)
 | |
|   buf.writeUInt16BE(flags | type, offset + 2)
 | |
|   buf.writeUInt16BE(h.questions.length, offset + 4)
 | |
|   buf.writeUInt16BE(h.answers.length, offset + 6)
 | |
|   buf.writeUInt16BE(h.authorities.length, offset + 8)
 | |
|   buf.writeUInt16BE(h.additionals.length, offset + 10)
 | |
| 
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| header.encode.bytes = 12
 | |
| 
 | |
| header.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
|   if (buf.length < 12) throw new Error('Header must be 12 bytes')
 | |
|   const flags = buf.readUInt16BE(offset + 2)
 | |
| 
 | |
|   return {
 | |
|     id: buf.readUInt16BE(offset),
 | |
|     type: flags & RESPONSE_FLAG ? 'response' : 'query',
 | |
|     flags: flags & 32767,
 | |
|     flag_qr: ((flags >> 15) & 0x1) === 1,
 | |
|     opcode: opcodes.toString((flags >> 11) & 0xf),
 | |
|     flag_aa: ((flags >> 10) & 0x1) === 1,
 | |
|     flag_tc: ((flags >> 9) & 0x1) === 1,
 | |
|     flag_rd: ((flags >> 8) & 0x1) === 1,
 | |
|     flag_ra: ((flags >> 7) & 0x1) === 1,
 | |
|     flag_z: ((flags >> 6) & 0x1) === 1,
 | |
|     flag_ad: ((flags >> 5) & 0x1) === 1,
 | |
|     flag_cd: ((flags >> 4) & 0x1) === 1,
 | |
|     rcode: rcodes.toString(flags & 0xf),
 | |
|     questions: new Array(buf.readUInt16BE(offset + 4)),
 | |
|     answers: new Array(buf.readUInt16BE(offset + 6)),
 | |
|     authorities: new Array(buf.readUInt16BE(offset + 8)),
 | |
|     additionals: new Array(buf.readUInt16BE(offset + 10))
 | |
|   }
 | |
| }
 | |
| 
 | |
| header.decode.bytes = 12
 | |
| 
 | |
| header.encodingLength = function () {
 | |
|   return 12
 | |
| }
 | |
| 
 | |
| const runknown = exports.unknown = {}
 | |
| 
 | |
| runknown.encode = function (data, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(runknown.encodingLength(data))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   buf.writeUInt16BE(data.length, offset)
 | |
|   data.copy(buf, offset + 2)
 | |
| 
 | |
|   runknown.encode.bytes = data.length + 2
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| runknown.encode.bytes = 0
 | |
| 
 | |
| runknown.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const len = buf.readUInt16BE(offset)
 | |
|   const data = buf.slice(offset + 2, offset + 2 + len)
 | |
|   runknown.decode.bytes = len + 2
 | |
|   return data
 | |
| }
 | |
| 
 | |
| runknown.decode.bytes = 0
 | |
| 
 | |
| runknown.encodingLength = function (data) {
 | |
|   return data.length + 2
 | |
| }
 | |
| 
 | |
| const rns = exports.ns = {}
 | |
| 
 | |
| rns.encode = function (data, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rns.encodingLength(data))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   name.encode(data, buf, offset + 2)
 | |
|   buf.writeUInt16BE(name.encode.bytes, offset)
 | |
|   rns.encode.bytes = name.encode.bytes + 2
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rns.encode.bytes = 0
 | |
| 
 | |
| rns.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const len = buf.readUInt16BE(offset)
 | |
|   const dd = name.decode(buf, offset + 2)
 | |
| 
 | |
|   rns.decode.bytes = len + 2
 | |
|   return dd
 | |
| }
 | |
| 
 | |
| rns.decode.bytes = 0
 | |
| 
 | |
| rns.encodingLength = function (data) {
 | |
|   return name.encodingLength(data) + 2
 | |
| }
 | |
| 
 | |
| const rsoa = exports.soa = {}
 | |
| 
 | |
| rsoa.encode = function (data, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rsoa.encodingLength(data))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const oldOffset = offset
 | |
|   offset += 2
 | |
|   name.encode(data.mname, buf, offset)
 | |
|   offset += name.encode.bytes
 | |
|   name.encode(data.rname, buf, offset, { mail: true })
 | |
|   offset += name.encode.bytes
 | |
|   buf.writeUInt32BE(data.serial || 0, offset)
 | |
|   offset += 4
 | |
|   buf.writeUInt32BE(data.refresh || 0, offset)
 | |
|   offset += 4
 | |
|   buf.writeUInt32BE(data.retry || 0, offset)
 | |
|   offset += 4
 | |
|   buf.writeUInt32BE(data.expire || 0, offset)
 | |
|   offset += 4
 | |
|   buf.writeUInt32BE(data.minimum || 0, offset)
 | |
|   offset += 4
 | |
| 
 | |
|   buf.writeUInt16BE(offset - oldOffset - 2, oldOffset)
 | |
|   rsoa.encode.bytes = offset - oldOffset
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rsoa.encode.bytes = 0
 | |
| 
 | |
| rsoa.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   const data = {}
 | |
|   offset += 2
 | |
|   data.mname = name.decode(buf, offset)
 | |
|   offset += name.decode.bytes
 | |
|   data.rname = name.decode(buf, offset, { mail: true })
 | |
|   offset += name.decode.bytes
 | |
|   data.serial = buf.readUInt32BE(offset)
 | |
|   offset += 4
 | |
|   data.refresh = buf.readUInt32BE(offset)
 | |
|   offset += 4
 | |
|   data.retry = buf.readUInt32BE(offset)
 | |
|   offset += 4
 | |
|   data.expire = buf.readUInt32BE(offset)
 | |
|   offset += 4
 | |
|   data.minimum = buf.readUInt32BE(offset)
 | |
|   offset += 4
 | |
| 
 | |
|   rsoa.decode.bytes = offset - oldOffset
 | |
|   return data
 | |
| }
 | |
| 
 | |
| rsoa.decode.bytes = 0
 | |
| 
 | |
| rsoa.encodingLength = function (data) {
 | |
|   return 22 + name.encodingLength(data.mname) + name.encodingLength(data.rname)
 | |
| }
 | |
| 
 | |
| const rtxt = exports.txt = {}
 | |
| 
 | |
| rtxt.encode = function (data, buf, offset) {
 | |
|   if (!Array.isArray(data)) data = [data]
 | |
|   for (let i = 0; i < data.length; i++) {
 | |
|     if (typeof data[i] === 'string') {
 | |
|       data[i] = Buffer.from(data[i])
 | |
|     }
 | |
|     if (!Buffer.isBuffer(data[i])) {
 | |
|       throw new Error('Must be a Buffer')
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   if (!buf) buf = Buffer.alloc(rtxt.encodingLength(data))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const oldOffset = offset
 | |
|   offset += 2
 | |
| 
 | |
|   data.forEach(function (d) {
 | |
|     buf[offset++] = d.length
 | |
|     d.copy(buf, offset, 0, d.length)
 | |
|     offset += d.length
 | |
|   })
 | |
| 
 | |
|   buf.writeUInt16BE(offset - oldOffset - 2, oldOffset)
 | |
|   rtxt.encode.bytes = offset - oldOffset
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rtxt.encode.bytes = 0
 | |
| 
 | |
| rtxt.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
|   let remaining = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
| 
 | |
|   let data = []
 | |
|   while (remaining > 0) {
 | |
|     const len = buf[offset++]
 | |
|     --remaining
 | |
|     if (remaining < len) {
 | |
|       throw new Error('Buffer overflow')
 | |
|     }
 | |
|     data.push(buf.slice(offset, offset + len))
 | |
|     offset += len
 | |
|     remaining -= len
 | |
|   }
 | |
| 
 | |
|   rtxt.decode.bytes = offset - oldOffset
 | |
|   return data
 | |
| }
 | |
| 
 | |
| rtxt.decode.bytes = 0
 | |
| 
 | |
| rtxt.encodingLength = function (data) {
 | |
|   if (!Array.isArray(data)) data = [data]
 | |
|   let length = 2
 | |
|   data.forEach(function (buf) {
 | |
|     if (typeof buf === 'string') {
 | |
|       length += Buffer.byteLength(buf) + 1
 | |
|     } else {
 | |
|       length += buf.length + 1
 | |
|     }
 | |
|   })
 | |
|   return length
 | |
| }
 | |
| 
 | |
| const rnull = exports.null = {}
 | |
| 
 | |
| rnull.encode = function (data, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rnull.encodingLength(data))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   if (typeof data === 'string') data = Buffer.from(data)
 | |
|   if (!data) data = Buffer.alloc(0)
 | |
| 
 | |
|   const oldOffset = offset
 | |
|   offset += 2
 | |
| 
 | |
|   const len = data.length
 | |
|   data.copy(buf, offset, 0, len)
 | |
|   offset += len
 | |
| 
 | |
|   buf.writeUInt16BE(offset - oldOffset - 2, oldOffset)
 | |
|   rnull.encode.bytes = offset - oldOffset
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rnull.encode.bytes = 0
 | |
| 
 | |
| rnull.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
|   const len = buf.readUInt16BE(offset)
 | |
| 
 | |
|   offset += 2
 | |
| 
 | |
|   const data = buf.slice(offset, offset + len)
 | |
|   offset += len
 | |
| 
 | |
|   rnull.decode.bytes = offset - oldOffset
 | |
|   return data
 | |
| }
 | |
| 
 | |
| rnull.decode.bytes = 0
 | |
| 
 | |
| rnull.encodingLength = function (data) {
 | |
|   if (!data) return 2
 | |
|   return (Buffer.isBuffer(data) ? data.length : Buffer.byteLength(data)) + 2
 | |
| }
 | |
| 
 | |
| const rhinfo = exports.hinfo = {}
 | |
| 
 | |
| rhinfo.encode = function (data, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rhinfo.encodingLength(data))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const oldOffset = offset
 | |
|   offset += 2
 | |
|   string.encode(data.cpu, buf, offset)
 | |
|   offset += string.encode.bytes
 | |
|   string.encode(data.os, buf, offset)
 | |
|   offset += string.encode.bytes
 | |
|   buf.writeUInt16BE(offset - oldOffset - 2, oldOffset)
 | |
|   rhinfo.encode.bytes = offset - oldOffset
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rhinfo.encode.bytes = 0
 | |
| 
 | |
| rhinfo.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   const data = {}
 | |
|   offset += 2
 | |
|   data.cpu = string.decode(buf, offset)
 | |
|   offset += string.decode.bytes
 | |
|   data.os = string.decode(buf, offset)
 | |
|   offset += string.decode.bytes
 | |
|   rhinfo.decode.bytes = offset - oldOffset
 | |
|   return data
 | |
| }
 | |
| 
 | |
| rhinfo.decode.bytes = 0
 | |
| 
 | |
| rhinfo.encodingLength = function (data) {
 | |
|   return string.encodingLength(data.cpu) + string.encodingLength(data.os) + 2
 | |
| }
 | |
| 
 | |
| const rptr = exports.ptr = {}
 | |
| const rcname = exports.cname = rptr
 | |
| const rdname = exports.dname = rptr
 | |
| 
 | |
| rptr.encode = function (data, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rptr.encodingLength(data))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   name.encode(data, buf, offset + 2)
 | |
|   buf.writeUInt16BE(name.encode.bytes, offset)
 | |
|   rptr.encode.bytes = name.encode.bytes + 2
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rptr.encode.bytes = 0
 | |
| 
 | |
| rptr.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const data = name.decode(buf, offset + 2)
 | |
|   rptr.decode.bytes = name.decode.bytes + 2
 | |
|   return data
 | |
| }
 | |
| 
 | |
| rptr.decode.bytes = 0
 | |
| 
 | |
| rptr.encodingLength = function (data) {
 | |
|   return name.encodingLength(data) + 2
 | |
| }
 | |
| 
 | |
| const rsrv = exports.srv = {}
 | |
| 
 | |
| rsrv.encode = function (data, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rsrv.encodingLength(data))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   buf.writeUInt16BE(data.priority || 0, offset + 2)
 | |
|   buf.writeUInt16BE(data.weight || 0, offset + 4)
 | |
|   buf.writeUInt16BE(data.port || 0, offset + 6)
 | |
|   name.encode(data.target, buf, offset + 8)
 | |
| 
 | |
|   const len = name.encode.bytes + 6
 | |
|   buf.writeUInt16BE(len, offset)
 | |
| 
 | |
|   rsrv.encode.bytes = len + 2
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rsrv.encode.bytes = 0
 | |
| 
 | |
| rsrv.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const len = buf.readUInt16BE(offset)
 | |
| 
 | |
|   const data = {}
 | |
|   data.priority = buf.readUInt16BE(offset + 2)
 | |
|   data.weight = buf.readUInt16BE(offset + 4)
 | |
|   data.port = buf.readUInt16BE(offset + 6)
 | |
|   data.target = name.decode(buf, offset + 8)
 | |
| 
 | |
|   rsrv.decode.bytes = len + 2
 | |
|   return data
 | |
| }
 | |
| 
 | |
| rsrv.decode.bytes = 0
 | |
| 
 | |
| rsrv.encodingLength = function (data) {
 | |
|   return 8 + name.encodingLength(data.target)
 | |
| }
 | |
| 
 | |
| const rcaa = exports.caa = {}
 | |
| 
 | |
| rcaa.ISSUER_CRITICAL = 1 << 7
 | |
| 
 | |
| rcaa.encode = function (data, buf, offset) {
 | |
|   const len = rcaa.encodingLength(data)
 | |
| 
 | |
|   if (!buf) buf = Buffer.alloc(rcaa.encodingLength(data))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   if (data.issuerCritical) {
 | |
|     data.flags = rcaa.ISSUER_CRITICAL
 | |
|   }
 | |
| 
 | |
|   buf.writeUInt16BE(len - 2, offset)
 | |
|   offset += 2
 | |
|   buf.writeUInt8(data.flags || 0, offset)
 | |
|   offset += 1
 | |
|   string.encode(data.tag, buf, offset)
 | |
|   offset += string.encode.bytes
 | |
|   buf.write(data.value, offset)
 | |
|   offset += Buffer.byteLength(data.value)
 | |
| 
 | |
|   rcaa.encode.bytes = len
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rcaa.encode.bytes = 0
 | |
| 
 | |
| rcaa.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const len = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
| 
 | |
|   const oldOffset = offset
 | |
|   const data = {}
 | |
|   data.flags = buf.readUInt8(offset)
 | |
|   offset += 1
 | |
|   data.tag = string.decode(buf, offset)
 | |
|   offset += string.decode.bytes
 | |
|   data.value = buf.toString('utf-8', offset, oldOffset + len)
 | |
| 
 | |
|   data.issuerCritical = !!(data.flags & rcaa.ISSUER_CRITICAL)
 | |
| 
 | |
|   rcaa.decode.bytes = len + 2
 | |
| 
 | |
|   return data
 | |
| }
 | |
| 
 | |
| rcaa.decode.bytes = 0
 | |
| 
 | |
| rcaa.encodingLength = function (data) {
 | |
|   return string.encodingLength(data.tag) + string.encodingLength(data.value) + 2
 | |
| }
 | |
| 
 | |
| const rmx = exports.mx = {}
 | |
| 
 | |
| rmx.encode = function (data, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rmx.encodingLength(data))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const oldOffset = offset
 | |
|   offset += 2
 | |
|   buf.writeUInt16BE(data.preference || 0, offset)
 | |
|   offset += 2
 | |
|   name.encode(data.exchange, buf, offset)
 | |
|   offset += name.encode.bytes
 | |
| 
 | |
|   buf.writeUInt16BE(offset - oldOffset - 2, oldOffset)
 | |
|   rmx.encode.bytes = offset - oldOffset
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rmx.encode.bytes = 0
 | |
| 
 | |
| rmx.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   const data = {}
 | |
|   offset += 2
 | |
|   data.preference = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
|   data.exchange = name.decode(buf, offset)
 | |
|   offset += name.decode.bytes
 | |
| 
 | |
|   rmx.decode.bytes = offset - oldOffset
 | |
|   return data
 | |
| }
 | |
| 
 | |
| rmx.encodingLength = function (data) {
 | |
|   return 4 + name.encodingLength(data.exchange)
 | |
| }
 | |
| 
 | |
| const ra = exports.a = {}
 | |
| 
 | |
| ra.encode = function (host, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(ra.encodingLength(host))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   buf.writeUInt16BE(4, offset)
 | |
|   offset += 2
 | |
|   ip.v4.encode(host, buf, offset)
 | |
|   ra.encode.bytes = 6
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| ra.encode.bytes = 0
 | |
| 
 | |
| ra.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   offset += 2
 | |
|   const host = ip.v4.decode(buf, offset)
 | |
|   ra.decode.bytes = 6
 | |
|   return host
 | |
| }
 | |
| 
 | |
| ra.decode.bytes = 0
 | |
| 
 | |
| ra.encodingLength = function () {
 | |
|   return 6
 | |
| }
 | |
| 
 | |
| const raaaa = exports.aaaa = {}
 | |
| 
 | |
| raaaa.encode = function (host, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(raaaa.encodingLength(host))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   buf.writeUInt16BE(16, offset)
 | |
|   offset += 2
 | |
|   ip.v6.encode(host, buf, offset)
 | |
|   raaaa.encode.bytes = 18
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| raaaa.encode.bytes = 0
 | |
| 
 | |
| raaaa.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   offset += 2
 | |
|   const host = ip.v6.decode(buf, offset)
 | |
|   raaaa.decode.bytes = 18
 | |
|   return host
 | |
| }
 | |
| 
 | |
| raaaa.decode.bytes = 0
 | |
| 
 | |
| raaaa.encodingLength = function () {
 | |
|   return 18
 | |
| }
 | |
| 
 | |
| const roption = exports.option = {}
 | |
| 
 | |
| roption.encode = function (option, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(roption.encodingLength(option))
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   const code = optioncodes.toCode(option.code)
 | |
|   buf.writeUInt16BE(code, offset)
 | |
|   offset += 2
 | |
|   if (option.data) {
 | |
|     buf.writeUInt16BE(option.data.length, offset)
 | |
|     offset += 2
 | |
|     option.data.copy(buf, offset)
 | |
|     offset += option.data.length
 | |
|   } else {
 | |
|     switch (code) {
 | |
|       // case 3: NSID.  No encode makes sense.
 | |
|       // case 5,6,7: Not implementable
 | |
|       case 8: // ECS
 | |
|         // note: do IP math before calling
 | |
|         const spl = option.sourcePrefixLength || 0
 | |
|         const fam = option.family || ip.familyOf(option.ip)
 | |
|         const ipBuf = ip.encode(option.ip, Buffer.alloc)
 | |
|         const ipLen = Math.ceil(spl / 8)
 | |
|         buf.writeUInt16BE(ipLen + 4, offset)
 | |
|         offset += 2
 | |
|         buf.writeUInt16BE(fam, offset)
 | |
|         offset += 2
 | |
|         buf.writeUInt8(spl, offset++)
 | |
|         buf.writeUInt8(option.scopePrefixLength || 0, offset++)
 | |
| 
 | |
|         ipBuf.copy(buf, offset, 0, ipLen)
 | |
|         offset += ipLen
 | |
|         break
 | |
|       // case 9: EXPIRE (experimental)
 | |
|       // case 10: COOKIE.  No encode makes sense.
 | |
|       case 11: // KEEP-ALIVE
 | |
|         if (option.timeout) {
 | |
|           buf.writeUInt16BE(2, offset)
 | |
|           offset += 2
 | |
|           buf.writeUInt16BE(option.timeout, offset)
 | |
|           offset += 2
 | |
|         } else {
 | |
|           buf.writeUInt16BE(0, offset)
 | |
|           offset += 2
 | |
|         }
 | |
|         break
 | |
|       case 12: // PADDING
 | |
|         const len = option.length || 0
 | |
|         buf.writeUInt16BE(len, offset)
 | |
|         offset += 2
 | |
|         buf.fill(0, offset, offset + len)
 | |
|         offset += len
 | |
|         break
 | |
|       // case 13:  CHAIN.  Experimental.
 | |
|       case 14: // KEY-TAG
 | |
|         const tagsLen = option.tags.length * 2
 | |
|         buf.writeUInt16BE(tagsLen, offset)
 | |
|         offset += 2
 | |
|         for (const tag of option.tags) {
 | |
|           buf.writeUInt16BE(tag, offset)
 | |
|           offset += 2
 | |
|         }
 | |
|         break
 | |
|       default:
 | |
|         throw new Error(`Unknown roption code: ${option.code}`)
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   roption.encode.bytes = offset - oldOffset
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| roption.encode.bytes = 0
 | |
| 
 | |
| roption.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
|   const option = {}
 | |
|   option.code = buf.readUInt16BE(offset)
 | |
|   option.type = optioncodes.toString(option.code)
 | |
|   offset += 2
 | |
|   const len = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
|   option.data = buf.slice(offset, offset + len)
 | |
|   switch (option.code) {
 | |
|     // case 3: NSID.  No decode makes sense.
 | |
|     case 8: // ECS
 | |
|       option.family = buf.readUInt16BE(offset)
 | |
|       offset += 2
 | |
|       option.sourcePrefixLength = buf.readUInt8(offset++)
 | |
|       option.scopePrefixLength = buf.readUInt8(offset++)
 | |
|       const padded = Buffer.alloc((option.family === 1) ? 4 : 16)
 | |
|       buf.copy(padded, 0, offset, offset + len - 4)
 | |
|       option.ip = ip.decode(padded)
 | |
|       break
 | |
|     // case 12: Padding.  No decode makes sense.
 | |
|     case 11: // KEEP-ALIVE
 | |
|       if (len > 0) {
 | |
|         option.timeout = buf.readUInt16BE(offset)
 | |
|         offset += 2
 | |
|       }
 | |
|       break
 | |
|     case 14:
 | |
|       option.tags = []
 | |
|       for (let i = 0; i < len; i += 2) {
 | |
|         option.tags.push(buf.readUInt16BE(offset))
 | |
|         offset += 2
 | |
|       }
 | |
|     // don't worry about default.  caller will use data if desired
 | |
|   }
 | |
| 
 | |
|   roption.decode.bytes = len + 4
 | |
|   return option
 | |
| }
 | |
| 
 | |
| roption.decode.bytes = 0
 | |
| 
 | |
| roption.encodingLength = function (option) {
 | |
|   if (option.data) {
 | |
|     return option.data.length + 4
 | |
|   }
 | |
|   const code = optioncodes.toCode(option.code)
 | |
|   switch (code) {
 | |
|     case 8: // ECS
 | |
|       const spl = option.sourcePrefixLength || 0
 | |
|       return Math.ceil(spl / 8) + 8
 | |
|     case 11: // KEEP-ALIVE
 | |
|       return (typeof option.timeout === 'number') ? 6 : 4
 | |
|     case 12: // PADDING
 | |
|       return option.length + 4
 | |
|     case 14: // KEY-TAG
 | |
|       return 4 + (option.tags.length * 2)
 | |
|   }
 | |
|   throw new Error(`Unknown roption code: ${option.code}`)
 | |
| }
 | |
| 
 | |
| const ropt = exports.opt = {}
 | |
| 
 | |
| ropt.encode = function (options, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(ropt.encodingLength(options))
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   const rdlen = encodingLengthList(options, roption)
 | |
|   buf.writeUInt16BE(rdlen, offset)
 | |
|   offset = encodeList(options, roption, buf, offset + 2)
 | |
| 
 | |
|   ropt.encode.bytes = offset - oldOffset
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| ropt.encode.bytes = 0
 | |
| 
 | |
| ropt.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   const options = []
 | |
|   let rdlen = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
|   let o = 0
 | |
|   while (rdlen > 0) {
 | |
|     options[o++] = roption.decode(buf, offset)
 | |
|     offset += roption.decode.bytes
 | |
|     rdlen -= roption.decode.bytes
 | |
|   }
 | |
|   ropt.decode.bytes = offset - oldOffset
 | |
|   return options
 | |
| }
 | |
| 
 | |
| ropt.decode.bytes = 0
 | |
| 
 | |
| ropt.encodingLength = function (options) {
 | |
|   return 2 + encodingLengthList(options || [], roption)
 | |
| }
 | |
| 
 | |
| const rdnskey = exports.dnskey = {}
 | |
| 
 | |
| rdnskey.PROTOCOL_DNSSEC = 3
 | |
| rdnskey.ZONE_KEY = 0x80
 | |
| rdnskey.SECURE_ENTRYPOINT = 0x8000
 | |
| 
 | |
| rdnskey.encode = function (key, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rdnskey.encodingLength(key))
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   const keydata = key.key
 | |
|   if (!Buffer.isBuffer(keydata)) {
 | |
|     throw new Error('Key must be a Buffer')
 | |
|   }
 | |
| 
 | |
|   offset += 2 // Leave space for length
 | |
|   buf.writeUInt16BE(key.flags, offset)
 | |
|   offset += 2
 | |
|   buf.writeUInt8(rdnskey.PROTOCOL_DNSSEC, offset)
 | |
|   offset += 1
 | |
|   buf.writeUInt8(key.algorithm, offset)
 | |
|   offset += 1
 | |
|   keydata.copy(buf, offset, 0, keydata.length)
 | |
|   offset += keydata.length
 | |
| 
 | |
|   rdnskey.encode.bytes = offset - oldOffset
 | |
|   buf.writeUInt16BE(rdnskey.encode.bytes - 2, oldOffset)
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rdnskey.encode.bytes = 0
 | |
| 
 | |
| rdnskey.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   var key = {}
 | |
|   var length = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
|   key.flags = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
|   if (buf.readUInt8(offset) !== rdnskey.PROTOCOL_DNSSEC) {
 | |
|     throw new Error('Protocol must be 3')
 | |
|   }
 | |
|   offset += 1
 | |
|   key.algorithm = buf.readUInt8(offset)
 | |
|   offset += 1
 | |
|   key.key = buf.slice(offset, oldOffset + length + 2)
 | |
|   offset += key.key.length
 | |
|   rdnskey.decode.bytes = offset - oldOffset
 | |
|   return key
 | |
| }
 | |
| 
 | |
| rdnskey.decode.bytes = 0
 | |
| 
 | |
| rdnskey.encodingLength = function (key) {
 | |
|   return 6 + Buffer.byteLength(key.key)
 | |
| }
 | |
| 
 | |
| const rrrsig = exports.rrsig = {}
 | |
| 
 | |
| rrrsig.encode = function (sig, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rrrsig.encodingLength(sig))
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   const signature = sig.signature
 | |
|   if (!Buffer.isBuffer(signature)) {
 | |
|     throw new Error('Signature must be a Buffer')
 | |
|   }
 | |
| 
 | |
|   offset += 2 // Leave space for length
 | |
|   buf.writeUInt16BE(types.toType(sig.typeCovered), offset)
 | |
|   offset += 2
 | |
|   buf.writeUInt8(sig.algorithm, offset)
 | |
|   offset += 1
 | |
|   buf.writeUInt8(sig.labels, offset)
 | |
|   offset += 1
 | |
|   buf.writeUInt32BE(sig.originalTTL, offset)
 | |
|   offset += 4
 | |
|   buf.writeUInt32BE(sig.expiration, offset)
 | |
|   offset += 4
 | |
|   buf.writeUInt32BE(sig.inception, offset)
 | |
|   offset += 4
 | |
|   buf.writeUInt16BE(sig.keyTag, offset)
 | |
|   offset += 2
 | |
|   name.encode(sig.signersName, buf, offset)
 | |
|   offset += name.encode.bytes
 | |
|   signature.copy(buf, offset, 0, signature.length)
 | |
|   offset += signature.length
 | |
| 
 | |
|   rrrsig.encode.bytes = offset - oldOffset
 | |
|   buf.writeUInt16BE(rrrsig.encode.bytes - 2, oldOffset)
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rrrsig.encode.bytes = 0
 | |
| 
 | |
| rrrsig.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   var sig = {}
 | |
|   var length = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
|   sig.typeCovered = types.toString(buf.readUInt16BE(offset))
 | |
|   offset += 2
 | |
|   sig.algorithm = buf.readUInt8(offset)
 | |
|   offset += 1
 | |
|   sig.labels = buf.readUInt8(offset)
 | |
|   offset += 1
 | |
|   sig.originalTTL = buf.readUInt32BE(offset)
 | |
|   offset += 4
 | |
|   sig.expiration = buf.readUInt32BE(offset)
 | |
|   offset += 4
 | |
|   sig.inception = buf.readUInt32BE(offset)
 | |
|   offset += 4
 | |
|   sig.keyTag = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
|   sig.signersName = name.decode(buf, offset)
 | |
|   offset += name.decode.bytes
 | |
|   sig.signature = buf.slice(offset, oldOffset + length + 2)
 | |
|   offset += sig.signature.length
 | |
|   rrrsig.decode.bytes = offset - oldOffset
 | |
|   return sig
 | |
| }
 | |
| 
 | |
| rrrsig.decode.bytes = 0
 | |
| 
 | |
| rrrsig.encodingLength = function (sig) {
 | |
|   return 20 +
 | |
|     name.encodingLength(sig.signersName) +
 | |
|     Buffer.byteLength(sig.signature)
 | |
| }
 | |
| 
 | |
| const rrp = exports.rp = {}
 | |
| 
 | |
| rrp.encode = function (data, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rrp.encodingLength(data))
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   offset += 2 // Leave space for length
 | |
|   name.encode(data.mbox || '.', buf, offset, { mail: true })
 | |
|   offset += name.encode.bytes
 | |
|   name.encode(data.txt || '.', buf, offset)
 | |
|   offset += name.encode.bytes
 | |
|   rrp.encode.bytes = offset - oldOffset
 | |
|   buf.writeUInt16BE(rrp.encode.bytes - 2, oldOffset)
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rrp.encode.bytes = 0
 | |
| 
 | |
| rrp.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   const data = {}
 | |
|   offset += 2
 | |
|   data.mbox = name.decode(buf, offset, { mail: true }) || '.'
 | |
|   offset += name.decode.bytes
 | |
|   data.txt = name.decode(buf, offset) || '.'
 | |
|   offset += name.decode.bytes
 | |
|   rrp.decode.bytes = offset - oldOffset
 | |
|   return data
 | |
| }
 | |
| 
 | |
| rrp.decode.bytes = 0
 | |
| 
 | |
| rrp.encodingLength = function (data) {
 | |
|   return 2 + name.encodingLength(data.mbox || '.') + name.encodingLength(data.txt || '.')
 | |
| }
 | |
| 
 | |
| const typebitmap = {}
 | |
| 
 | |
| typebitmap.encode = function (typelist, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(typebitmap.encodingLength(typelist))
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   var typesByWindow = []
 | |
|   for (var i = 0; i < typelist.length; i++) {
 | |
|     var typeid = types.toType(typelist[i])
 | |
|     if (typesByWindow[typeid >> 8] === undefined) {
 | |
|       typesByWindow[typeid >> 8] = []
 | |
|     }
 | |
|     typesByWindow[typeid >> 8][(typeid >> 3) & 0x1F] |= 1 << (7 - (typeid & 0x7))
 | |
|   }
 | |
| 
 | |
|   for (i = 0; i < typesByWindow.length; i++) {
 | |
|     if (typesByWindow[i] !== undefined) {
 | |
|       var windowBuf = Buffer.from(typesByWindow[i])
 | |
|       buf.writeUInt8(i, offset)
 | |
|       offset += 1
 | |
|       buf.writeUInt8(windowBuf.length, offset)
 | |
|       offset += 1
 | |
|       windowBuf.copy(buf, offset)
 | |
|       offset += windowBuf.length
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   typebitmap.encode.bytes = offset - oldOffset
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| typebitmap.encode.bytes = 0
 | |
| 
 | |
| typebitmap.decode = function (buf, offset, length) {
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   var typelist = []
 | |
|   while (offset - oldOffset < length) {
 | |
|     var window = buf.readUInt8(offset)
 | |
|     offset += 1
 | |
|     var windowLength = buf.readUInt8(offset)
 | |
|     offset += 1
 | |
|     for (var i = 0; i < windowLength; i++) {
 | |
|       var b = buf.readUInt8(offset + i)
 | |
|       for (var j = 0; j < 8; j++) {
 | |
|         if (b & (1 << (7 - j))) {
 | |
|           var typeid = types.toString((window << 8) | (i << 3) | j)
 | |
|           typelist.push(typeid)
 | |
|         }
 | |
|       }
 | |
|     }
 | |
|     offset += windowLength
 | |
|   }
 | |
| 
 | |
|   typebitmap.decode.bytes = offset - oldOffset
 | |
|   return typelist
 | |
| }
 | |
| 
 | |
| typebitmap.decode.bytes = 0
 | |
| 
 | |
| typebitmap.encodingLength = function (typelist) {
 | |
|   var extents = []
 | |
|   for (var i = 0; i < typelist.length; i++) {
 | |
|     var typeid = types.toType(typelist[i])
 | |
|     extents[typeid >> 8] = Math.max(extents[typeid >> 8] || 0, typeid & 0xFF)
 | |
|   }
 | |
| 
 | |
|   var len = 0
 | |
|   for (i = 0; i < extents.length; i++) {
 | |
|     if (extents[i] !== undefined) {
 | |
|       len += 2 + Math.ceil((extents[i] + 1) / 8)
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   return len
 | |
| }
 | |
| 
 | |
| const rnsec = exports.nsec = {}
 | |
| 
 | |
| rnsec.encode = function (record, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rnsec.encodingLength(record))
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   offset += 2 // Leave space for length
 | |
|   name.encode(record.nextDomain, buf, offset)
 | |
|   offset += name.encode.bytes
 | |
|   typebitmap.encode(record.rrtypes, buf, offset)
 | |
|   offset += typebitmap.encode.bytes
 | |
| 
 | |
|   rnsec.encode.bytes = offset - oldOffset
 | |
|   buf.writeUInt16BE(rnsec.encode.bytes - 2, oldOffset)
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rnsec.encode.bytes = 0
 | |
| 
 | |
| rnsec.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   var record = {}
 | |
|   var length = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
|   record.nextDomain = name.decode(buf, offset)
 | |
|   offset += name.decode.bytes
 | |
|   record.rrtypes = typebitmap.decode(buf, offset, length - (offset - oldOffset))
 | |
|   offset += typebitmap.decode.bytes
 | |
| 
 | |
|   rnsec.decode.bytes = offset - oldOffset
 | |
|   return record
 | |
| }
 | |
| 
 | |
| rnsec.decode.bytes = 0
 | |
| 
 | |
| rnsec.encodingLength = function (record) {
 | |
|   return 2 +
 | |
|     name.encodingLength(record.nextDomain) +
 | |
|     typebitmap.encodingLength(record.rrtypes)
 | |
| }
 | |
| 
 | |
| const rnsec3 = exports.nsec3 = {}
 | |
| 
 | |
| rnsec3.encode = function (record, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rnsec3.encodingLength(record))
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   const salt = record.salt
 | |
|   if (!Buffer.isBuffer(salt)) {
 | |
|     throw new Error('salt must be a Buffer')
 | |
|   }
 | |
| 
 | |
|   const nextDomain = record.nextDomain
 | |
|   if (!Buffer.isBuffer(nextDomain)) {
 | |
|     throw new Error('nextDomain must be a Buffer')
 | |
|   }
 | |
| 
 | |
|   offset += 2 // Leave space for length
 | |
|   buf.writeUInt8(record.algorithm, offset)
 | |
|   offset += 1
 | |
|   buf.writeUInt8(record.flags, offset)
 | |
|   offset += 1
 | |
|   buf.writeUInt16BE(record.iterations, offset)
 | |
|   offset += 2
 | |
|   buf.writeUInt8(salt.length, offset)
 | |
|   offset += 1
 | |
|   salt.copy(buf, offset, 0, salt.length)
 | |
|   offset += salt.length
 | |
|   buf.writeUInt8(nextDomain.length, offset)
 | |
|   offset += 1
 | |
|   nextDomain.copy(buf, offset, 0, nextDomain.length)
 | |
|   offset += nextDomain.length
 | |
|   typebitmap.encode(record.rrtypes, buf, offset)
 | |
|   offset += typebitmap.encode.bytes
 | |
| 
 | |
|   rnsec3.encode.bytes = offset - oldOffset
 | |
|   buf.writeUInt16BE(rnsec3.encode.bytes - 2, oldOffset)
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rnsec3.encode.bytes = 0
 | |
| 
 | |
| rnsec3.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   var record = {}
 | |
|   var length = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
|   record.algorithm = buf.readUInt8(offset)
 | |
|   offset += 1
 | |
|   record.flags = buf.readUInt8(offset)
 | |
|   offset += 1
 | |
|   record.iterations = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
|   const saltLength = buf.readUInt8(offset)
 | |
|   offset += 1
 | |
|   record.salt = buf.slice(offset, offset + saltLength)
 | |
|   offset += saltLength
 | |
|   const hashLength = buf.readUInt8(offset)
 | |
|   offset += 1
 | |
|   record.nextDomain = buf.slice(offset, offset + hashLength)
 | |
|   offset += hashLength
 | |
|   record.rrtypes = typebitmap.decode(buf, offset, length - (offset - oldOffset))
 | |
|   offset += typebitmap.decode.bytes
 | |
| 
 | |
|   rnsec3.decode.bytes = offset - oldOffset
 | |
|   return record
 | |
| }
 | |
| 
 | |
| rnsec3.decode.bytes = 0
 | |
| 
 | |
| rnsec3.encodingLength = function (record) {
 | |
|   return 8 +
 | |
|     record.salt.length +
 | |
|     record.nextDomain.length +
 | |
|     typebitmap.encodingLength(record.rrtypes)
 | |
| }
 | |
| 
 | |
| const rds = exports.ds = {}
 | |
| 
 | |
| rds.encode = function (digest, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rds.encodingLength(digest))
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   const digestdata = digest.digest
 | |
|   if (!Buffer.isBuffer(digestdata)) {
 | |
|     throw new Error('Digest must be a Buffer')
 | |
|   }
 | |
| 
 | |
|   offset += 2 // Leave space for length
 | |
|   buf.writeUInt16BE(digest.keyTag, offset)
 | |
|   offset += 2
 | |
|   buf.writeUInt8(digest.algorithm, offset)
 | |
|   offset += 1
 | |
|   buf.writeUInt8(digest.digestType, offset)
 | |
|   offset += 1
 | |
|   digestdata.copy(buf, offset, 0, digestdata.length)
 | |
|   offset += digestdata.length
 | |
| 
 | |
|   rds.encode.bytes = offset - oldOffset
 | |
|   buf.writeUInt16BE(rds.encode.bytes - 2, oldOffset)
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rds.encode.bytes = 0
 | |
| 
 | |
| rds.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   var digest = {}
 | |
|   var length = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
|   digest.keyTag = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
|   digest.algorithm = buf.readUInt8(offset)
 | |
|   offset += 1
 | |
|   digest.digestType = buf.readUInt8(offset)
 | |
|   offset += 1
 | |
|   digest.digest = buf.slice(offset, oldOffset + length + 2)
 | |
|   offset += digest.digest.length
 | |
|   rds.decode.bytes = offset - oldOffset
 | |
|   return digest
 | |
| }
 | |
| 
 | |
| rds.decode.bytes = 0
 | |
| 
 | |
| rds.encodingLength = function (digest) {
 | |
|   return 6 + Buffer.byteLength(digest.digest)
 | |
| }
 | |
| 
 | |
| const rsshfp = exports.sshfp = {}
 | |
| 
 | |
| rsshfp.getFingerprintLengthForHashType = function getFingerprintLengthForHashType (hashType) {
 | |
|   switch (hashType) {
 | |
|     case 1: return 20
 | |
|     case 2: return 32
 | |
|   }
 | |
| }
 | |
| 
 | |
| rsshfp.encode = function encode (record, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rsshfp.encodingLength(record))
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   offset += 2 // The function call starts with the offset pointer at the RDLENGTH field, not the RDATA one
 | |
|   buf[offset] = record.algorithm
 | |
|   offset += 1
 | |
|   buf[offset] = record.hash
 | |
|   offset += 1
 | |
| 
 | |
|   const fingerprintBuf = Buffer.from(record.fingerprint.toUpperCase(), 'hex')
 | |
|   if (fingerprintBuf.length !== rsshfp.getFingerprintLengthForHashType(record.hash)) {
 | |
|     throw new Error('Invalid fingerprint length')
 | |
|   }
 | |
|   fingerprintBuf.copy(buf, offset)
 | |
|   offset += fingerprintBuf.byteLength
 | |
| 
 | |
|   rsshfp.encode.bytes = offset - oldOffset
 | |
|   buf.writeUInt16BE(rsshfp.encode.bytes - 2, oldOffset)
 | |
| 
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rsshfp.encode.bytes = 0
 | |
| 
 | |
| rsshfp.decode = function decode (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   const record = {}
 | |
|   offset += 2 // Account for the RDLENGTH field
 | |
|   record.algorithm = buf[offset]
 | |
|   offset += 1
 | |
|   record.hash = buf[offset]
 | |
|   offset += 1
 | |
| 
 | |
|   const fingerprintLength = rsshfp.getFingerprintLengthForHashType(record.hash)
 | |
|   record.fingerprint = buf.slice(offset, offset + fingerprintLength).toString('hex').toUpperCase()
 | |
|   offset += fingerprintLength
 | |
|   rsshfp.decode.bytes = offset - oldOffset
 | |
|   return record
 | |
| }
 | |
| 
 | |
| rsshfp.decode.bytes = 0
 | |
| 
 | |
| rsshfp.encodingLength = function (record) {
 | |
|   return 4 + Buffer.from(record.fingerprint, 'hex').byteLength
 | |
| }
 | |
| 
 | |
| const rnaptr = exports.naptr = {}
 | |
| 
 | |
| rnaptr.encode = function (data, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rnaptr.encodingLength(data))
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
|   offset += 2
 | |
|   buf.writeUInt16BE(data.order || 0, offset)
 | |
|   offset += 2
 | |
|   buf.writeUInt16BE(data.preference || 0, offset)
 | |
|   offset += 2
 | |
|   string.encode(data.flags, buf, offset)
 | |
|   offset += string.encode.bytes
 | |
|   string.encode(data.services, buf, offset)
 | |
|   offset += string.encode.bytes
 | |
|   string.encode(data.regexp, buf, offset)
 | |
|   offset += string.encode.bytes
 | |
|   name.encode(data.replacement, buf, offset)
 | |
|   offset += name.encode.bytes
 | |
|   rnaptr.encode.bytes = offset - oldOffset
 | |
|   buf.writeUInt16BE(rnaptr.encode.bytes - 2, oldOffset)
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rnaptr.encode.bytes = 0
 | |
| 
 | |
| rnaptr.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
|   const data = {}
 | |
|   offset += 2
 | |
|   data.order = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
|   data.preference = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
|   data.flags = string.decode(buf, offset)
 | |
|   offset += string.decode.bytes
 | |
|   data.services = string.decode(buf, offset)
 | |
|   offset += string.decode.bytes
 | |
|   data.regexp = string.decode(buf, offset)
 | |
|   offset += string.decode.bytes
 | |
|   data.replacement = name.decode(buf, offset)
 | |
|   offset += name.decode.bytes
 | |
|   rnaptr.decode.bytes = offset - oldOffset
 | |
|   return data
 | |
| }
 | |
| 
 | |
| rnaptr.decode.bytes = 0
 | |
| 
 | |
| rnaptr.encodingLength = function (data) {
 | |
|   return string.encodingLength(data.flags) +
 | |
|     string.encodingLength(data.services) +
 | |
|     string.encodingLength(data.regexp) +
 | |
|     name.encodingLength(data.replacement) + 6
 | |
| }
 | |
| 
 | |
| const rtlsa = exports.tlsa = {}
 | |
| 
 | |
| rtlsa.encode = function (cert, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(rtlsa.encodingLength(cert))
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   const certdata = cert.certificate
 | |
|   if (!Buffer.isBuffer(certdata)) {
 | |
|     throw new Error('Certificate must be a Buffer')
 | |
|   }
 | |
| 
 | |
|   offset += 2 // Leave space for length
 | |
|   buf.writeUInt8(cert.usage, offset)
 | |
|   offset += 1
 | |
|   buf.writeUInt8(cert.selector, offset)
 | |
|   offset += 1
 | |
|   buf.writeUInt8(cert.matchingType, offset)
 | |
|   offset += 1
 | |
|   certdata.copy(buf, offset, 0, certdata.length)
 | |
|   offset += certdata.length
 | |
| 
 | |
|   rtlsa.encode.bytes = offset - oldOffset
 | |
|   buf.writeUInt16BE(rtlsa.encode.bytes - 2, oldOffset)
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| rtlsa.encode.bytes = 0
 | |
| 
 | |
| rtlsa.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   const cert = {}
 | |
|   const length = buf.readUInt16BE(offset)
 | |
|   offset += 2
 | |
|   cert.usage = buf.readUInt8(offset)
 | |
|   offset += 1
 | |
|   cert.selector = buf.readUInt8(offset)
 | |
|   offset += 1
 | |
|   cert.matchingType = buf.readUInt8(offset)
 | |
|   offset += 1
 | |
|   cert.certificate = buf.slice(offset, oldOffset + length + 2)
 | |
|   offset += cert.certificate.length
 | |
|   rtlsa.decode.bytes = offset - oldOffset
 | |
|   return cert
 | |
| }
 | |
| 
 | |
| rtlsa.decode.bytes = 0
 | |
| 
 | |
| rtlsa.encodingLength = function (cert) {
 | |
|   return 5 + Buffer.byteLength(cert.certificate)
 | |
| }
 | |
| 
 | |
| const renc = exports.record = function (type) {
 | |
|   switch (type.toUpperCase()) {
 | |
|     case 'A': return ra
 | |
|     case 'PTR': return rptr
 | |
|     case 'CNAME': return rcname
 | |
|     case 'DNAME': return rdname
 | |
|     case 'TXT': return rtxt
 | |
|     case 'NULL': return rnull
 | |
|     case 'AAAA': return raaaa
 | |
|     case 'SRV': return rsrv
 | |
|     case 'HINFO': return rhinfo
 | |
|     case 'CAA': return rcaa
 | |
|     case 'NS': return rns
 | |
|     case 'SOA': return rsoa
 | |
|     case 'MX': return rmx
 | |
|     case 'OPT': return ropt
 | |
|     case 'DNSKEY': return rdnskey
 | |
|     case 'RRSIG': return rrrsig
 | |
|     case 'RP': return rrp
 | |
|     case 'NSEC': return rnsec
 | |
|     case 'NSEC3': return rnsec3
 | |
|     case 'SSHFP': return rsshfp
 | |
|     case 'DS': return rds
 | |
|     case 'NAPTR': return rnaptr
 | |
|     case 'TLSA': return rtlsa
 | |
|   }
 | |
|   return runknown
 | |
| }
 | |
| 
 | |
| const answer = exports.answer = {}
 | |
| 
 | |
| answer.encode = function (a, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(answer.encodingLength(a))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   name.encode(a.name, buf, offset)
 | |
|   offset += name.encode.bytes
 | |
| 
 | |
|   buf.writeUInt16BE(types.toType(a.type), offset)
 | |
| 
 | |
|   if (a.type.toUpperCase() === 'OPT') {
 | |
|     if (a.name !== '.') {
 | |
|       throw new Error('OPT name must be root.')
 | |
|     }
 | |
|     buf.writeUInt16BE(a.udpPayloadSize || 4096, offset + 2)
 | |
|     buf.writeUInt8(a.extendedRcode || 0, offset + 4)
 | |
|     buf.writeUInt8(a.ednsVersion || 0, offset + 5)
 | |
|     buf.writeUInt16BE(a.flags || 0, offset + 6)
 | |
| 
 | |
|     offset += 8
 | |
|     ropt.encode(a.options || [], buf, offset)
 | |
|     offset += ropt.encode.bytes
 | |
|   } else {
 | |
|     let klass = classes.toClass(a.class === undefined ? 'IN' : a.class)
 | |
|     if (a.flush) klass |= FLUSH_MASK // the 1st bit of the class is the flush bit
 | |
|     buf.writeUInt16BE(klass, offset + 2)
 | |
|     buf.writeUInt32BE(a.ttl || 0, offset + 4)
 | |
| 
 | |
|     offset += 8
 | |
|     const enc = renc(a.type)
 | |
|     enc.encode(a.data, buf, offset)
 | |
|     offset += enc.encode.bytes
 | |
|   }
 | |
| 
 | |
|   answer.encode.bytes = offset - oldOffset
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| answer.encode.bytes = 0
 | |
| 
 | |
| answer.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const a = {}
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   a.name = name.decode(buf, offset)
 | |
|   offset += name.decode.bytes
 | |
|   a.type = types.toString(buf.readUInt16BE(offset))
 | |
|   if (a.type === 'OPT') {
 | |
|     a.udpPayloadSize = buf.readUInt16BE(offset + 2)
 | |
|     a.extendedRcode = buf.readUInt8(offset + 4)
 | |
|     a.ednsVersion = buf.readUInt8(offset + 5)
 | |
|     a.flags = buf.readUInt16BE(offset + 6)
 | |
|     a.flag_do = ((a.flags >> 15) & 0x1) === 1
 | |
|     a.options = ropt.decode(buf, offset + 8)
 | |
|     offset += 8 + ropt.decode.bytes
 | |
|   } else {
 | |
|     const klass = buf.readUInt16BE(offset + 2)
 | |
|     a.ttl = buf.readUInt32BE(offset + 4)
 | |
| 
 | |
|     a.class = classes.toString(klass & NOT_FLUSH_MASK)
 | |
|     a.flush = !!(klass & FLUSH_MASK)
 | |
| 
 | |
|     const enc = renc(a.type)
 | |
|     a.data = enc.decode(buf, offset + 8)
 | |
|     offset += 8 + enc.decode.bytes
 | |
|   }
 | |
| 
 | |
|   answer.decode.bytes = offset - oldOffset
 | |
|   return a
 | |
| }
 | |
| 
 | |
| answer.decode.bytes = 0
 | |
| 
 | |
| answer.encodingLength = function (a) {
 | |
|   const data = (a.data !== null && a.data !== undefined) ? a.data : a.options
 | |
|   return name.encodingLength(a.name) + 8 + renc(a.type).encodingLength(data)
 | |
| }
 | |
| 
 | |
| const question = exports.question = {}
 | |
| 
 | |
| question.encode = function (q, buf, offset) {
 | |
|   if (!buf) buf = Buffer.alloc(question.encodingLength(q))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   name.encode(q.name, buf, offset)
 | |
|   offset += name.encode.bytes
 | |
| 
 | |
|   buf.writeUInt16BE(types.toType(q.type), offset)
 | |
|   offset += 2
 | |
| 
 | |
|   buf.writeUInt16BE(classes.toClass(q.class === undefined ? 'IN' : q.class), offset)
 | |
|   offset += 2
 | |
| 
 | |
|   question.encode.bytes = offset - oldOffset
 | |
|   return q
 | |
| }
 | |
| 
 | |
| question.encode.bytes = 0
 | |
| 
 | |
| question.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const oldOffset = offset
 | |
|   const q = {}
 | |
| 
 | |
|   q.name = name.decode(buf, offset)
 | |
|   offset += name.decode.bytes
 | |
| 
 | |
|   q.type = types.toString(buf.readUInt16BE(offset))
 | |
|   offset += 2
 | |
| 
 | |
|   q.class = classes.toString(buf.readUInt16BE(offset))
 | |
|   offset += 2
 | |
| 
 | |
|   const qu = !!(q.class & QU_MASK)
 | |
|   if (qu) q.class &= NOT_QU_MASK
 | |
| 
 | |
|   question.decode.bytes = offset - oldOffset
 | |
|   return q
 | |
| }
 | |
| 
 | |
| question.decode.bytes = 0
 | |
| 
 | |
| question.encodingLength = function (q) {
 | |
|   return name.encodingLength(q.name) + 4
 | |
| }
 | |
| 
 | |
| exports.AUTHORITATIVE_ANSWER = 1 << 10
 | |
| exports.TRUNCATED_RESPONSE = 1 << 9
 | |
| exports.RECURSION_DESIRED = 1 << 8
 | |
| exports.RECURSION_AVAILABLE = 1 << 7
 | |
| exports.AUTHENTIC_DATA = 1 << 5
 | |
| exports.CHECKING_DISABLED = 1 << 4
 | |
| exports.DNSSEC_OK = 1 << 15
 | |
| 
 | |
| exports.encode = function (result, buf, offset) {
 | |
|   const allocing = !buf
 | |
| 
 | |
|   if (allocing) buf = Buffer.alloc(exports.encodingLength(result))
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const oldOffset = offset
 | |
| 
 | |
|   if (!result.questions) result.questions = []
 | |
|   if (!result.answers) result.answers = []
 | |
|   if (!result.authorities) result.authorities = []
 | |
|   if (!result.additionals) result.additionals = []
 | |
| 
 | |
|   header.encode(result, buf, offset)
 | |
|   offset += header.encode.bytes
 | |
| 
 | |
|   offset = encodeList(result.questions, question, buf, offset)
 | |
|   offset = encodeList(result.answers, answer, buf, offset)
 | |
|   offset = encodeList(result.authorities, answer, buf, offset)
 | |
|   offset = encodeList(result.additionals, answer, buf, offset)
 | |
| 
 | |
|   exports.encode.bytes = offset - oldOffset
 | |
| 
 | |
|   // just a quick sanity check
 | |
|   if (allocing && exports.encode.bytes !== buf.length) {
 | |
|     return buf.slice(0, exports.encode.bytes)
 | |
|   }
 | |
| 
 | |
|   return buf
 | |
| }
 | |
| 
 | |
| exports.encode.bytes = 0
 | |
| 
 | |
| exports.decode = function (buf, offset) {
 | |
|   if (!offset) offset = 0
 | |
| 
 | |
|   const oldOffset = offset
 | |
|   const result = header.decode(buf, offset)
 | |
|   offset += header.decode.bytes
 | |
| 
 | |
|   offset = decodeList(result.questions, question, buf, offset)
 | |
|   offset = decodeList(result.answers, answer, buf, offset)
 | |
|   offset = decodeList(result.authorities, answer, buf, offset)
 | |
|   offset = decodeList(result.additionals, answer, buf, offset)
 | |
| 
 | |
|   exports.decode.bytes = offset - oldOffset
 | |
| 
 | |
|   return result
 | |
| }
 | |
| 
 | |
| exports.decode.bytes = 0
 | |
| 
 | |
| exports.encodingLength = function (result) {
 | |
|   return header.encodingLength(result) +
 | |
|     encodingLengthList(result.questions || [], question) +
 | |
|     encodingLengthList(result.answers || [], answer) +
 | |
|     encodingLengthList(result.authorities || [], answer) +
 | |
|     encodingLengthList(result.additionals || [], answer)
 | |
| }
 | |
| 
 | |
| exports.streamEncode = function (result) {
 | |
|   const buf = exports.encode(result)
 | |
|   const sbuf = Buffer.alloc(2)
 | |
|   sbuf.writeUInt16BE(buf.byteLength)
 | |
|   const combine = Buffer.concat([sbuf, buf])
 | |
|   exports.streamEncode.bytes = combine.byteLength
 | |
|   return combine
 | |
| }
 | |
| 
 | |
| exports.streamEncode.bytes = 0
 | |
| 
 | |
| exports.streamDecode = function (sbuf) {
 | |
|   const len = sbuf.readUInt16BE(0)
 | |
|   if (sbuf.byteLength < len + 2) {
 | |
|     // not enough data
 | |
|     return null
 | |
|   }
 | |
|   const result = exports.decode(sbuf.slice(2))
 | |
|   exports.streamDecode.bytes = exports.decode.bytes
 | |
|   return result
 | |
| }
 | |
| 
 | |
| exports.streamDecode.bytes = 0
 | |
| 
 | |
| function encodingLengthList (list, enc) {
 | |
|   let len = 0
 | |
|   for (let i = 0; i < list.length; i++) len += enc.encodingLength(list[i])
 | |
|   return len
 | |
| }
 | |
| 
 | |
| function encodeList (list, enc, buf, offset) {
 | |
|   for (let i = 0; i < list.length; i++) {
 | |
|     enc.encode(list[i], buf, offset)
 | |
|     offset += enc.encode.bytes
 | |
|   }
 | |
|   return offset
 | |
| }
 | |
| 
 | |
| function decodeList (list, enc, buf, offset) {
 | |
|   for (let i = 0; i < list.length; i++) {
 | |
|     list[i] = enc.decode(buf, offset)
 | |
|     offset += enc.decode.bytes
 | |
|   }
 | |
|   return offset
 | |
| }
 |