import {
Composer,
CST,
Lexer,
LineCounter,
Parser,
} from 'yaml'
If you're interested only in the final output, parse()
will directly produce native JavaScript
If you'd like to retain the comments and other metadata, parseDocument()
and parseAllDocuments()
will produce Document instances that allow for further processing.
If you're looking to do something more specific, this section might be for you.
Internally, the process of turning a sequence of characters into Documents relies on three stages, each of which is also exposed to external users. First, the Lexer splits the character stream into lexical tokens, i.e. sequences of characters and control codes. Next, the Parser builds concrete syntax tree representations of each document and directive in the stream. Finally, the Composer builds a more user-friendly and accessible Document representation of each document.
Both the Lexer and Parser accept incomplete input, allowing for them and the Composer to be used with e.g. Node.js streams or other systems that handle data in chunks.
import { Lexer } from 'yaml'
const tokens = new Lexer().lex('foo: bar\nfee:\n [24,"42"]\n')
console.dir(Array.from(tokens))
> [
'\x02', '\x1F', 'foo', ':',
' ', '\x1F', 'bar', '\n',
'\x1F', 'fee', ':', '\n',
' ', '[', '\x1F', '24',
',', '"42"', ']', '\n'
]
The API for the lexer is rather minimal, and offers no configuration.
If the input stream is chunked, the lex()
method may be called separately for each chunk if the incomplete
argument is true
.
At the end of input, lex()
should be called a final time with incomplete: false
to ensure that the remaining tokens are emitted.
Internally, the lexer operates a state machine that determines how it parses its input.
Initially, the lexer is always in the stream
state.
The lexer constructor and its lex()
method should never throw an error.
All tokens are identifiable either by their exact value or their first character. In addition to slices of the input stream, a few control characters are additionally used within the output.
Value | Token | Meaning |
---|---|---|
\x02 |
doc-mode | Start of a document within the default stream context. |
\x18 |
flow-error-end | Unexpected end of a flow collection, e.g. due to an unindent. Should be considered an error. |
\x1f |
scalar | The next token after this one is a scalar value, irrespective of its value or first character. |
\n , \r\n |
newline | In certain cases (such as end of input), an empty string may also be emitted; it should also be considered as a newline. |
--- |
doc-start | Explicit marker for the start of a document. Will be preceded by a doc-mode token. |
... |
doc-end | Explicit marker for the end of a document. |
- |
seq-item-ind | Block sequence item indicator, separated by whitespace. |
? |
explicit-key-ind | Explicit block map key indicator, separated by whitespace. |
: |
map-value-ind | Block map value indicator. |
{ |
flow-map-start | |
} |
flow-map-end | |
[ |
flow-seq-start | |
] |
flow-seq-end | |
, |
comma | Separator between flow collection items. |
\u{FEFF} |
byte-order-mark | Treated as whitespace in stream & content in a document. |
If any of the control characters do show up directly in the input stream, they will be treated normally, and even when bare will be preceded by a SCALAR control token in the output.
All remaining tokens are identifiable by their first character:
First char | Token | Meaning |
---|---|---|
, \t |
space | Only contains space characters if token indicates indentation. Otherwise may contain repeats of either character. |
# |
comment | Separated from preceding by whitespace. Does not include the trailing newline. |
% |
directive-line | Only produced in a stream context. |
* |
alias | |
& |
anchor | |
! |
tag | |
' |
single-quoted-scalar | Should also include ' as a last character, if input is valid. |
" |
double-quoted-scalar | Should also include " as a last character, if input is valid. |
⎮ , > |
block-scalar-header | Expected to be followed by optional whitespace & comment, a newline, and then a scalar value. |
import { Parser } from 'yaml'
for (const token of new Parser().parse('foo: [24,"42"]\n'))
console.dir(token, { depth: null })
> {
type: 'document',
offset: 0,
start: [],
value: {
type: 'block-map',
offset: 0,
indent: 0,
items: [
{
start: [],
key: { type: 'scalar', offset: 0, indent: 0, source: 'foo' },
sep: [
{ type: 'map-value-ind', offset: 3, indent: 0, source: ':' },
{ type: 'space', offset: 4, indent: 0, source: ' ' }
],
value: {
type: 'flow-collection',
offset: 5,
indent: 0,
start: { type: 'flow-seq-start', offset: 5, indent: 0, source: '[' },
items: [
{ type: 'scalar', offset: 6, indent: 0, source: '24' },
{ type: 'comma', offset: 8, indent: 0, source: ',' },
{
type: 'double-quoted-scalar',
offset: 9,
indent: 0,
source: '"42"'
}
],
end: [
{ type: 'flow-seq-end', offset: 13, indent: 0, source: ']' },
{ type: 'newline', offset: 14, indent: 0, source: '\n' }
]
}
}
]
}
}
The parser by default uses an internal Lexer instance, and provides a similarly minimal API for producing a Concrete Syntax Tree representation of the input stream.
The tokens emitted by the parser are JavaScript objects, each of which has a type
value that's one of the following: directive-line
, document
, byte-order-mark
, space
, comment
, newline
.
Of these, only directive-line
and document
should be considered as content.
The parser does not validate its output, trying instead to produce a most YAML-ish representation of any input. It should never throw errors, but may (rarely) include error tokens in its output.
To validate a CST, you will need to compose it into a Document
.
If the document contains errors, they will be included in the document's errors
array, and each error will will contain an offset
within the source string, which you may then use to find the corresponding node in the CST.
Create a new parser.
If defined, onNewLine
is called separately with the start position of each new line (in parse()
, including the start of input).
Parse source
as a YAML stream, generating tokens for each directive, document and other structure as it is completely parsed.
If incomplete
, a part of the last line may be left as a buffer for the next call.
Errors are not thrown, but are yielded as { type: 'error', offset, message }
tokens.
Advance the parser by one lexical token.
Used internally by parser.parse()
; exposed to allow for use with an external lexer.
For debug purposes, if the LOG_TOKENS
env var is true-ish, all lexical tokens will be pretty-printed using console.log()
as they are being processed.
For a complete description of CST node interfaces, please consult the cst.ts source.
Some of the most common node properties include:
Property | Type | Description |
---|---|---|
type |
string |
The only node property that's always defined. Identifies the node type. May be used as a TS type guard. |
offset |
number |
The start index within the source string or character stream. |
source |
string |
A raw string representation of the node's value, including all newlines and indentation. |
indent |
number |
The indent level of the current line; mostly just for internal use. |
items |
Item[] |
The contents of a collection; exact shape depends on the collection type. |
start , sep , end |
SourceToken[] |
Content before, within, and after "actual" values. Includes item and collection indicators, anchors, tags, comments, as well as other things. |
Collection items contain some subset of the following properties:
Item property | Type | Description |
---|---|---|
start |
SourceToken[] |
Always defined. Content before the actual value. May include comments that are later assigned to the preceding item. |
key |
Token ⎮ null |
Set for key/value pairs only, so never used in block sequences. |
sep |
SourceToken[] |
Content between the key and the value. If defined, indicates that the key logically exists, even if its value is null . |
value |
Token ⎮ null |
The value. Normally set, but may be left out for e.g. explicit keys with no matching value. |
import { LineCounter, Parser } from 'yaml'
const lineCounter = new LineCounter()
const parser = new Parser(lineCounter.addNewLine))
const tokens = parser.parse('foo:\n- 24\n- "42"\n')
Array.from(tokens) // forces iteration
lineCounter.lineStarts
> [ 0, 5, 10, 17 ]
lineCounter.linePos(3)
> { line: 1, col: 4 }
lineCounter.linePos(5)
> { line: 2, col: 1 }
Tracks newlines during parsing in order to provide an efficient API for determining the one-indexed { line, col }
position for any offset within the input.
Adds the starting index of a new line.
Should be called in order, or the internal lineStarts
array will need to be sorted before calling linePos()
.
Bound to the instance, so may be used directly as a callback.
Performs a binary search and returns the 1-indexed { line, col }
position of offset
.
If line === 0
, addNewLine
has never been called or offset
is before the first known newline.
import { Composer, Parser } from 'yaml'
const src = 'foo: bar\nfee: [24, "42"]'
const tokens = new Parser().parse(src)
const docs = new Composer().compose(tokens)
Array.from(docs, doc => doc.toJS())
> [{ foo: 'bar', fee: [24, '42'] }]
Create a new Document composer.
Does not include an internal Parser instance, so an external one will be needed.
options
will be used during composition, and passed to the new Document
constructor.
composer.compose(tokens: Iterable<Token>, forceDoc?: boolean, endOffset?: number): Generator<Document.Parsed>
Compose tokens into documents.
Convenience wrapper combining calls to composer.next()
and composer.end()
.
Advance the composed by one CST token.
Always call at end of input to push out any remaining document.
If forceDoc
is true and the stream contains no document, still emit a final document including any comments and directives that would be applied to a subsequent document.
offset
should be set if forceDoc
is also set, to set the document range end and to indicate errors correctly.
Current stream status information. Mostly useful at the end of input for an empty stream.
import { CST } from 'yaml'
For most use cases, the Document or pure JS interfaces provided by the library are the right tool. Sometimes, though, it's important to keep the original YAML source in as pristine a condition as possible. For those cases, the concrete syntax tree (CST) representation is provided, as it retains every character of the input, including whitespace.
Create a new scalar token with the value value
.
Values that represent an actual string but may be parsed as a different type should use a type
other than 'PLAIN'
,
as this function does not support any schema operations and won't check for such conflicts.
Argument | Type | Default | Description |
---|---|---|---|
value | string |
The string representation of the value, which will have its content properly indented. Required. | |
context.end | SourceToken[] |
Comments and whitespace after the end of the value, or after the block scalar header. If undefined, a newline will be added. | |
context.implicitKey | boolean |
false |
Being within an implicit key may affect the resolved type of the token's value. |
context.indent | number |
The indent level of the token. Required. | |
context.inFlow | boolean |
false |
Is this scalar within a flow collection? This may affect the resolved type of the token's value. |
context.offset | number |
-1 |
The offset position of the token. |
context.type | Scalar.Type |
The preferred type of the scalar token. If undefined, the previous type of the token will be used, defaulting to 'PLAIN' . |
const [doc] = new Parser().parse('foo: "bar" #comment')
const item = doc.value.items[0].value
> {
type: 'double-quoted-scalar',
offset: 5,
indent: 0,
source: '"bar"',
end: [
{ type: 'space', offset: 10, indent: 0, source: ' ' },
{ type: 'comment', offset: 11, indent: 0, source: '#comment' }
]
}
YAML.resolveAsScalar(item)
> { value: 'bar', type: 'QUOTE_DOUBLE', comment: 'comment', length: 14 }
Custom type guards for detecting CST collections and scalars, in both their block and flow forms.
If token
is a CST flow or block scalar, determine its string value and a few other attributes.
Otherwise, return null
.
Set the value of token
to the given string value
, overwriting any previous contents and type that it may have.
Best efforts are made to retain any comments previously associated with the token
, though all contents within a collection's items
will be overwritten.
Values that represent an actual string but may be parsed as a different type should use a type
other than 'PLAIN'
, as this function does not support any schema operations and won't check for such conflicts.
Argument | Type | Default | Description |
---|---|---|---|
token | Token |
Any token. If it does not include an indent value, the value will be stringified as if it were an implicit key. Required. |
|
value | string |
The string representation of the value, which will have its content properly indented. Required. | |
context.afterKey | boolean |
false |
In most cases, values after a key should have an additional level of indentation. |
context.implicitKey | boolean |
false |
Being within an implicit key may affect the resolved type of the token's value. |
context.inFlow | boolean |
false |
Being within a flow collection may affect the resolved type of the token's value. |
context.type | Scalar.Type |
The preferred type of the scalar token. If undefined, the previous type of the token will be used, defaulting to 'PLAIN' . |
function findScalarAtOffset(
cst: CST.Document,
offset: number
): CST.FlowScalar | CST.BlockScalar | undefined {
let res: CST.FlowScalar | CST.BlockScalar | undefined = undefined
CST.visit(cst, ({ key, value }) => {
for (const token of [key, value])
if (CST.isScalar(token)) {
if (token.offset > offset) return CST.visit.BREAK
if (
token.offset == offset ||
(token.offset < offset && token.offset + token.source.length > offset)
) {
res = token
return CST.visit.BREAK
}
}
})
return res
}
Stringify a CST document, token, or collection item. Fair warning: This applies no validation whatsoever, and simply concatenates the sources in their logical order.
Apply a visitor to a CST document or item. Effectively, the general-purpose workhorse of navigating the CST.
Walks through the tree (depth-first) starting from cst
as the root, calling a visitor
function with two arguments when entering each item:
item
: The current item, which includes the following members:start: SourceToken[]
– Source tokens before the key or value, possibly including its anchor or tag.key?: Token | null
– Set for pair values. May then benull
, if the key before the:
separator is empty.sep?: SourceToken[]
– Source tokens between the key and the value, which should include the:
map value indicator ifvalue
is set.value?: Token
– The value of a sequence item, or of a map pair.
path
: The steps from the root to the current node, as an array of['key' | 'value', number]
tuples.
The return value of the visitor may be used to control the traversal:
undefined
(default): Do nothing and continueCST.visit.SKIP
: Do not visit the children of this token, continue with next siblingCST.visit.BREAK
: Terminate traversal completelyCST.visit.REMOVE
: Remove the current item, then continue with the next onenumber
: Set the index of the next step. This is useful especially if the index of the current token has changed.function
: Define the next visitor for this item. After the original visitor is called on item entry, next visitors are called after handling a non-emptykey
and when exiting the item.
const [doc] = new Parser().parse('[ foo, bar, baz ]')
CST.visit(doc, (item, path) => {
if (!CST.isScalar(item.value)) return
const scalar = CST.resolveAsScalar(item.value)
if (scalar?.value === 'bar') {
const parent = CST.visit.parentCollection(doc, path)
const idx = path[path.length - 1][1]
const { indent } = item.value
parent.items.splice(idx, 0, {
start: item.start.slice(),
value: CST.createScalarToken('bing', { end: [], indent })
})
return idx + 2
}
})
CST.stringify(doc)
> '[ foo, bing, bar, baz ]'
A couple of utility functions are provided for working with the path
:
CST.visit.itemAtPath(cst, path): CST.CollectionItem | undefined
– Find the item atpath
fromcst
as the root.CST.visit.parentCollection(cst, path): CST.BlockMap | CST.BlockSequence | CST.FlowCollection
– Get the immediate parent collection of the item atpath
fromcst
as the root. Throws an error if the collection is not found, which should never happen if the item itself exists.