Take advantage of the new 'filename' and 'lineno' attributes of Node and Property objects to provide more informative error and status messages by printing the actual source file reference of the corresponding object, always using the unquoted file:line format in error messages. No functional change is introduced by this commit. Signed-off-by: Luca Burelli <l.burelli@arduino.cc>
2289 lines
78 KiB
Python
2289 lines
78 KiB
Python
# Copyright (c) 2019, Nordic Semiconductor
|
|
# SPDX-License-Identifier: BSD-3-Clause
|
|
|
|
# Tip: You can view just the documentation with 'pydoc3 devicetree.dtlib'
|
|
|
|
"""
|
|
A library for extracting information from .dts (devicetree) files. See the
|
|
documentation for the DT and Node classes for more information.
|
|
|
|
The top-level entry point of the library is the DT class. DT.__init__() takes a
|
|
.dts file to parse and a list of directories to search for any /include/d
|
|
files.
|
|
"""
|
|
|
|
import collections
|
|
import enum
|
|
import errno
|
|
import os
|
|
import re
|
|
import string
|
|
import sys
|
|
import textwrap
|
|
from collections.abc import Iterable
|
|
from typing import TYPE_CHECKING, Any, NamedTuple, NoReturn, Optional, Union
|
|
|
|
# NOTE: tests/test_dtlib.py is the test suite for this library.
|
|
|
|
class DTError(Exception):
|
|
"Exception raised for devicetree-related errors"
|
|
|
|
class Node:
|
|
r"""
|
|
Represents a node in the devicetree ('node-name { ... };').
|
|
|
|
These attributes are available on Node instances:
|
|
|
|
name:
|
|
The name of the node (a string).
|
|
|
|
filename:
|
|
The name of the .dts file where the node is defined
|
|
|
|
lineno:
|
|
The line number in the .dts file where the node starts.
|
|
|
|
unit_addr:
|
|
The portion after the '@' in the node's name, or the empty string if the
|
|
name has no '@' in it.
|
|
|
|
Note that this is a string. Run int(node.unit_addr, 16) to get an
|
|
integer.
|
|
|
|
props:
|
|
A dict that maps the properties defined on the node to
|
|
their values. 'props' is indexed by property name (a string), and values
|
|
are Property objects.
|
|
|
|
To convert property values to Python numbers or strings, use
|
|
dtlib.to_num(), dtlib.to_nums(), or dtlib.to_string().
|
|
|
|
Property values are represented as 'bytes' arrays to support the full
|
|
generality of DTS, which allows assignments like
|
|
|
|
x = "foo", < 0x12345678 >, [ 9A ];
|
|
|
|
This gives x the value b"foo\0\x12\x34\x56\x78\x9A". Numbers in DTS are
|
|
stored in big-endian format.
|
|
|
|
nodes:
|
|
A dict containing the subnodes of the node, indexed by name.
|
|
|
|
labels:
|
|
A list with all labels pointing to the node, in the same order as the
|
|
labels appear, but with duplicates removed.
|
|
|
|
'label_1: label_2: node { ... };' gives 'labels' the value
|
|
["label_1", "label_2"].
|
|
|
|
parent:
|
|
The parent Node of the node. 'None' for the root node.
|
|
|
|
path:
|
|
The path to the node as a string, e.g. "/foo/bar".
|
|
|
|
dt:
|
|
The DT instance this node belongs to.
|
|
"""
|
|
|
|
#
|
|
# Public interface
|
|
#
|
|
|
|
def __init__(self, name: str, parent: Optional["Node"], dt: "DT", filename: str, lineno: int):
|
|
"""
|
|
Node constructor. Not meant to be called directly by clients.
|
|
"""
|
|
# Remember to update DT.__deepcopy__() if you change this.
|
|
|
|
self._name = name
|
|
self._filename = filename
|
|
self._lineno = lineno
|
|
self.props: dict[str, Property] = {}
|
|
self.nodes: dict[str, Node] = {}
|
|
self.labels: list[str] = []
|
|
self.parent = parent
|
|
self.dt = dt
|
|
|
|
self._omit_if_no_ref = False
|
|
self._is_referenced = False
|
|
|
|
if name.count("@") > 1:
|
|
dt._parse_error("multiple '@' in node name")
|
|
if name != "/":
|
|
for char in name:
|
|
if char not in _nodename_chars:
|
|
dt._parse_error(f"{self.path}: bad character '{char}' "
|
|
"in node name")
|
|
|
|
@property
|
|
def name(self) -> str:
|
|
"""
|
|
See the class documentation.
|
|
"""
|
|
# Converted to a property to discourage renaming -- that has to be done
|
|
# via DT.move_node.
|
|
return self._name
|
|
|
|
@property
|
|
def lineno(self) -> int:
|
|
"""
|
|
See the class documentation.
|
|
"""
|
|
return self._lineno
|
|
|
|
@property
|
|
def filename(self) -> str:
|
|
"""
|
|
See the class documentation.
|
|
"""
|
|
return self._filename
|
|
|
|
@property
|
|
def unit_addr(self) -> str:
|
|
"""
|
|
See the class documentation.
|
|
"""
|
|
return self.name.partition("@")[2]
|
|
|
|
@property
|
|
def path(self) -> str:
|
|
"""
|
|
See the class documentation.
|
|
"""
|
|
node_names = []
|
|
|
|
# This dynamic computation is required to be able to move
|
|
# nodes in the DT class.
|
|
cur = self
|
|
while cur.parent:
|
|
node_names.append(cur.name)
|
|
cur = cur.parent
|
|
|
|
return "/" + "/".join(reversed(node_names))
|
|
|
|
def node_iter(self) -> Iterable['Node']:
|
|
"""
|
|
Returns a generator for iterating over the node and its children,
|
|
recursively.
|
|
|
|
For example, this will iterate over all nodes in the tree (like
|
|
dt.node_iter()).
|
|
|
|
for node in dt.root.node_iter():
|
|
...
|
|
"""
|
|
yield self
|
|
for node in self.nodes.values():
|
|
yield from node.node_iter()
|
|
|
|
def _get_prop(self, name: str) -> 'Property':
|
|
# Returns the property named 'name' on the node, creating it if it
|
|
# doesn't already exist. Move the entry to the end of the props
|
|
# dictionary so the order is preserved in the output DTS file.
|
|
|
|
prop = self.props.pop(name, None)
|
|
if not prop:
|
|
prop = Property(self, name)
|
|
self.props[name] = prop
|
|
return prop
|
|
|
|
def _del(self) -> None:
|
|
# Removes the node from the tree. When called on the root node,
|
|
# this method will leave it empty but still part of the tree.
|
|
|
|
if self.parent is None:
|
|
self.nodes.clear()
|
|
self.props.clear()
|
|
return
|
|
self.parent.nodes.pop(self.name) # type: ignore
|
|
|
|
def __str__(self):
|
|
"""
|
|
Returns a DTS representation of the node. Called automatically if the
|
|
node is print()ed.
|
|
"""
|
|
def safe_relpath(filename):
|
|
# Returns a relative path to the file, or the absolute path if
|
|
# a relative path cannot be established (on Windows with files
|
|
# in different drives, for example).
|
|
try:
|
|
return os.path.relpath(filename, start=self.dt._base_dir)
|
|
except ValueError:
|
|
return filename
|
|
|
|
rel_filename = safe_relpath(self.filename)
|
|
s = f"\n/* node '{self.path}' defined in {rel_filename}:{self.lineno} */\n"
|
|
s += "".join(label + ": " for label in self.labels)
|
|
|
|
s += f"{self.name} {{\n"
|
|
|
|
lines = []
|
|
comments = {}
|
|
for prop in self.props.values():
|
|
prop_str = textwrap.indent(str(prop), "\t")
|
|
lines.extend(prop_str.splitlines(True))
|
|
|
|
rel_filename = safe_relpath(prop.filename)
|
|
comment = f"/* in {rel_filename}:{prop.lineno} */"
|
|
comments[len(lines)-1] = comment
|
|
|
|
if lines:
|
|
max_len = max([ len(line.rstrip()) for line in lines ])
|
|
for i, line in enumerate(lines):
|
|
if i in comments:
|
|
line += " " * (max_len - len(line) + 1) + comments[i] + "\n"
|
|
s += line
|
|
|
|
for child in self.nodes.values():
|
|
s += textwrap.indent(child.__str__(), "\t") + "\n"
|
|
|
|
s += "};"
|
|
|
|
return s
|
|
|
|
def __repr__(self):
|
|
"""
|
|
Returns some information about the Node instance. Called automatically
|
|
if the Node instance is evaluated.
|
|
"""
|
|
return f"<Node {self.path} in {self.filename}:{self.lineno}>"
|
|
|
|
# See Property.type
|
|
class Type(enum.IntEnum):
|
|
EMPTY = 0
|
|
BYTES = 1
|
|
NUM = 2
|
|
NUMS = 3
|
|
STRING = 4
|
|
STRINGS = 5
|
|
PATH = 6
|
|
PHANDLE = 7
|
|
PHANDLES = 8
|
|
PHANDLES_AND_NUMS = 9
|
|
COMPOUND = 10
|
|
|
|
class _MarkerType(enum.IntEnum):
|
|
# Types of markers in property values
|
|
|
|
# References
|
|
PATH = 0 # &foo
|
|
PHANDLE = 1 # <&foo>
|
|
LABEL = 2 # foo: <1 2 3>
|
|
|
|
# Start of data blocks of specific type
|
|
UINT8 = 3 # [00 01 02] (and also used for /incbin/)
|
|
UINT16 = 4 # /bits/ 16 <1 2 3>
|
|
UINT32 = 5 # <1 2 3>
|
|
UINT64 = 6 # /bits/ 64 <1 2 3>
|
|
STRING = 7 # "foo"
|
|
|
|
class Property:
|
|
"""
|
|
Represents a property ('x = ...').
|
|
|
|
These attributes are available on Property instances:
|
|
|
|
name:
|
|
The name of the property (a string).
|
|
|
|
filename:
|
|
The name of the .dts file where the property is defined
|
|
|
|
lineno:
|
|
The line number in the .dts file where the property starts.
|
|
|
|
value:
|
|
The value of the property, as a 'bytes' string. Numbers are stored in
|
|
big-endian format, and strings are null-terminated. Putting multiple
|
|
comma-separated values in an assignment (e.g., 'x = < 1 >, "foo"') will
|
|
concatenate the values.
|
|
|
|
See the to_*() methods for converting the value to other types.
|
|
|
|
type:
|
|
The type of the property, inferred from the syntax used in the
|
|
assignment. This is one of the following constants (with example
|
|
assignments):
|
|
|
|
Assignment | Property.type
|
|
----------------------------+------------------------
|
|
foo; | dtlib.Type.EMPTY
|
|
foo = []; | dtlib.Type.BYTES
|
|
foo = [01 02]; | dtlib.Type.BYTES
|
|
foo = /bits/ 8 <1>; | dtlib.Type.BYTES
|
|
foo = <1>; | dtlib.Type.NUM
|
|
foo = <>; | dtlib.Type.NUMS
|
|
foo = <1 2 3>; | dtlib.Type.NUMS
|
|
foo = <1 2>, <3>; | dtlib.Type.NUMS
|
|
foo = "foo"; | dtlib.Type.STRING
|
|
foo = "foo", "bar"; | dtlib.Type.STRINGS
|
|
foo = <&l>; | dtlib.Type.PHANDLE
|
|
foo = <&l1 &l2 &l3>; | dtlib.Type.PHANDLES
|
|
foo = <&l1 &l2>, <&l3>; | dtlib.Type.PHANDLES
|
|
foo = <&l1 1 2 &l2 3 4>; | dtlib.Type.PHANDLES_AND_NUMS
|
|
foo = <&l1 1 2>, <&l2 3 4>; | dtlib.Type.PHANDLES_AND_NUMS
|
|
foo = &l; | dtlib.Type.PATH
|
|
*Anything else* | dtlib.Type.COMPOUND
|
|
|
|
*Anything else* includes properties mixing phandle (<&label>) and node
|
|
path (&label) references with other data.
|
|
|
|
Data labels in the property value do not influence the type.
|
|
|
|
labels:
|
|
A list with all labels pointing to the property, in the same order as the
|
|
labels appear, but with duplicates removed.
|
|
|
|
'label_1: label2: x = ...' gives 'labels' the value
|
|
["label_1", "label_2"].
|
|
|
|
offset_labels:
|
|
A dictionary that maps any labels within the property's value to their
|
|
offset, in bytes. For example, 'x = < 0 label_1: 1 label_2: >' gives
|
|
'offset_labels' the value {"label_1": 4, "label_2": 8}.
|
|
|
|
Iteration order will match the order of the labels on Python versions
|
|
that preserve dict insertion order.
|
|
|
|
node:
|
|
The Node the property is on.
|
|
"""
|
|
|
|
#
|
|
# Public interface
|
|
#
|
|
|
|
def __init__(self, node: Node, name: str):
|
|
# Remember to update DT.__deepcopy__() if you change this.
|
|
|
|
if "@" in name:
|
|
node.dt._parse_error("'@' is only allowed in node names")
|
|
|
|
self.name = name
|
|
self.filename = ""
|
|
self.lineno = -1
|
|
self.value = b""
|
|
self.labels: list[str] = []
|
|
# We have to wait to set this until later, when we've got
|
|
# the entire tree.
|
|
self.offset_labels: dict[str, int] = {}
|
|
self.node: Node = node
|
|
|
|
self._label_offset_lst: list[tuple[str, int]] = []
|
|
|
|
# A list of [offset, label, type] lists (sorted by offset),
|
|
# giving the locations of references within the value. 'type'
|
|
# is either _MarkerType.PATH, for a node path reference,
|
|
# _MarkerType.PHANDLE, for a phandle reference, or
|
|
# _MarkerType.LABEL, for a label on/within data. Node paths
|
|
# and phandles need to be patched in after parsing.
|
|
self._markers: list[list] = []
|
|
|
|
@property
|
|
def type(self) -> Type:
|
|
"""
|
|
See the class documentation.
|
|
"""
|
|
# Data labels (e.g. 'foo = label: <3>') are irrelevant, so filter them
|
|
# out
|
|
types = [marker[1] for marker in self._markers
|
|
if marker[1] != _MarkerType.LABEL]
|
|
|
|
if not types:
|
|
return Type.EMPTY
|
|
|
|
if types == [_MarkerType.UINT8]:
|
|
return Type.BYTES
|
|
|
|
if types == [_MarkerType.UINT32]:
|
|
return Type.NUM if len(self.value) == 4 else Type.NUMS
|
|
|
|
# Treat 'foo = <1 2 3>, <4 5>, ...' as Type.NUMS too
|
|
if set(types) == {_MarkerType.UINT32}:
|
|
return Type.NUMS
|
|
|
|
if set(types) == {_MarkerType.STRING}:
|
|
return Type.STRING if len(types) == 1 else Type.STRINGS
|
|
|
|
if types == [_MarkerType.PATH]:
|
|
return Type.PATH
|
|
|
|
if (types == [_MarkerType.UINT32, _MarkerType.PHANDLE]
|
|
and len(self.value) == 4):
|
|
return Type.PHANDLE
|
|
|
|
if set(types) == {_MarkerType.UINT32, _MarkerType.PHANDLE}:
|
|
if len(self.value) == 4*types.count(_MarkerType.PHANDLE):
|
|
# Array with just phandles in it
|
|
return Type.PHANDLES
|
|
# Array with both phandles and numbers
|
|
return Type.PHANDLES_AND_NUMS
|
|
|
|
return Type.COMPOUND
|
|
|
|
def to_num(self, signed=False) -> int:
|
|
"""
|
|
Returns the value of the property as a number.
|
|
|
|
Raises DTError if the property was not assigned with this syntax (has
|
|
Property.type Type.NUM):
|
|
|
|
foo = < 1 >;
|
|
|
|
signed (default: False):
|
|
If True, the value will be interpreted as signed rather than
|
|
unsigned.
|
|
"""
|
|
if self.type is not Type.NUM:
|
|
_err(f"expected property '{self.name}' on {self.node.path} in "
|
|
f"{self.filename}:{self.lineno} to be assigned with "
|
|
f"'{self.name} = < (number) >;', not '{self}'")
|
|
|
|
return int.from_bytes(self.value, "big", signed=signed)
|
|
|
|
def to_nums(self, signed=False) -> list[int]:
|
|
"""
|
|
Returns the value of the property as a list of numbers.
|
|
|
|
Raises DTError if the property was not assigned with this syntax (has
|
|
Property.type Type.NUM or Type.NUMS):
|
|
|
|
foo = < 1 2 ... >;
|
|
|
|
signed (default: False):
|
|
If True, the values will be interpreted as signed rather than
|
|
unsigned.
|
|
"""
|
|
if self.type not in (Type.NUM, Type.NUMS):
|
|
_err(f"expected property '{self.name}' on {self.node.path} in "
|
|
f"{self.filename}:{self.lineno} to be assigned with "
|
|
f"'{self.name} = < (number) (number) ... >;', not '{self}'")
|
|
|
|
return [int.from_bytes(self.value[i:i + 4], "big", signed=signed)
|
|
for i in range(0, len(self.value), 4)]
|
|
|
|
def to_bytes(self) -> bytes:
|
|
"""
|
|
Returns the value of the property as a raw 'bytes', like
|
|
Property.value, except with added type checking.
|
|
|
|
Raises DTError if the property was not assigned with this syntax (has
|
|
Property.type Type.BYTES):
|
|
|
|
foo = [ 01 ... ];
|
|
"""
|
|
if self.type is not Type.BYTES:
|
|
_err(f"expected property '{self.name}' on {self.node.path} "
|
|
f"in {self.filename}:{self.lineno} to be assigned with "
|
|
f"'{self.name} = [ (byte) (byte) ... ];', not '{self}'")
|
|
|
|
return self.value
|
|
|
|
def to_string(self) -> str:
|
|
"""
|
|
Returns the value of the property as a string.
|
|
|
|
Raises DTError if the property was not assigned with this syntax (has
|
|
Property.type Type.STRING):
|
|
|
|
foo = "string";
|
|
|
|
This function might also raise UnicodeDecodeError if the string is
|
|
not valid UTF-8.
|
|
"""
|
|
if self.type is not Type.STRING:
|
|
_err(f"expected property '{self.name}' on {self.node.path} "
|
|
f"in {self.filename}:{self.lineno} to be assigned with "
|
|
f"'{self.name} = \"string\";', not '{self}'")
|
|
|
|
try:
|
|
ret = self.value.decode("utf-8")[:-1] # Strip null
|
|
except UnicodeDecodeError:
|
|
_err(f"value of property '{self.name}' ({self.value!r}) "
|
|
f"on {self.node.path} in {self.filename}:{self.lineno} "
|
|
"is not valid UTF-8")
|
|
|
|
return ret # The separate 'return' appeases the type checker.
|
|
|
|
def to_strings(self) -> list[str]:
|
|
"""
|
|
Returns the value of the property as a list of strings.
|
|
|
|
Raises DTError if the property was not assigned with this syntax (has
|
|
Property.type Type.STRING or Type.STRINGS):
|
|
|
|
foo = "string", "string", ... ;
|
|
|
|
Also raises DTError if any of the strings are not valid UTF-8.
|
|
"""
|
|
if self.type not in (Type.STRING, Type.STRINGS):
|
|
_err(f"expected property '{self.name}' on {self.node.path} in "
|
|
f"{self.filename}:{self.lineno} to be assigned with "
|
|
f"'{self.name} = \"string\", \"string\", ... ;', not '{self}'")
|
|
|
|
try:
|
|
ret = self.value.decode("utf-8").split("\0")[:-1]
|
|
except UnicodeDecodeError:
|
|
_err(f"value of property '{self.name}' ({self.value!r}) "
|
|
f"on {self.node.path} in {self.filename}:{self.lineno} "
|
|
"is not valid UTF-8")
|
|
|
|
return ret # The separate 'return' appeases the type checker.
|
|
|
|
def to_node(self) -> Node:
|
|
"""
|
|
Returns the Node the phandle in the property points to.
|
|
|
|
Raises DTError if the property was not assigned with this syntax (has
|
|
Property.type Type.PHANDLE).
|
|
|
|
foo = < &bar >;
|
|
"""
|
|
if self.type is not Type.PHANDLE:
|
|
_err(f"expected property '{self.name}' on {self.node.path} in "
|
|
f"{self.filename}:{self.lineno} to be assigned with "
|
|
f"'{self.name} = < &foo >;', not '{self}'")
|
|
|
|
return self.node.dt.phandle2node[int.from_bytes(self.value, "big")]
|
|
|
|
def to_nodes(self) -> list[Node]:
|
|
"""
|
|
Returns a list with the Nodes the phandles in the property point to.
|
|
|
|
Raises DTError if the property value contains anything other than
|
|
phandles. All of the following are accepted:
|
|
|
|
foo = < >
|
|
foo = < &bar >;
|
|
foo = < &bar &baz ... >;
|
|
foo = < &bar ... >, < &baz ... >;
|
|
"""
|
|
def type_ok():
|
|
if self.type in (Type.PHANDLE, Type.PHANDLES):
|
|
return True
|
|
# Also accept 'foo = < >;'
|
|
return self.type is Type.NUMS and not self.value
|
|
|
|
if not type_ok():
|
|
_err(f"expected property '{self.name}' on {self.node.path} in "
|
|
f"{self.filename}:{self.lineno} to be assigned with "
|
|
f"'{self.name} = < &foo &bar ... >;', not '{self}'")
|
|
|
|
return [self.node.dt.phandle2node[int.from_bytes(self.value[i:i + 4],
|
|
"big")]
|
|
for i in range(0, len(self.value), 4)]
|
|
|
|
def to_path(self) -> Node:
|
|
"""
|
|
Returns the Node referenced by the path stored in the property.
|
|
|
|
Raises DTError if the property was not assigned with either of these
|
|
syntaxes (has Property.type Type.PATH or Type.STRING):
|
|
|
|
foo = &bar;
|
|
foo = "/bar";
|
|
|
|
For the second case, DTError is raised if the path does not exist.
|
|
"""
|
|
if self.type not in (Type.PATH, Type.STRING):
|
|
_err(f"expected property '{self.name}' on {self.node.path} in "
|
|
f"{self.filename}:{self.lineno} to be assigned with either "
|
|
f"'{self.name} = &foo' or '{self.name} = \"/path/to/node\"', "
|
|
f"not '{self}'")
|
|
|
|
try:
|
|
path = self.value.decode("utf-8")[:-1]
|
|
except UnicodeDecodeError:
|
|
_err(f"value of property '{self.name}' ({self.value!r}) "
|
|
f"on {self.node.path} in {self.filename}:{self.lineno} "
|
|
"is not valid UTF-8")
|
|
|
|
try:
|
|
ret = self.node.dt.get_node(path)
|
|
except DTError:
|
|
_err(f"property '{self.name}' on {self.node.path} in "
|
|
f"{self.filename}:{self.lineno} points to the non-existent "
|
|
f'node "{path}"')
|
|
|
|
return ret # The separate 'return' appeases the type checker.
|
|
|
|
def __str__(self):
|
|
s = "".join(label + ": " for label in self.labels) + self.name
|
|
if not self.value:
|
|
return s + ";"
|
|
|
|
s += " ="
|
|
newline = "\n" + " " * len(s)
|
|
|
|
for i, (pos, marker_type, ref) in enumerate(self._markers):
|
|
if i < len(self._markers) - 1:
|
|
next_marker = self._markers[i + 1]
|
|
else:
|
|
next_marker = None
|
|
|
|
# End of current marker
|
|
end = next_marker[0] if next_marker else len(self.value)
|
|
|
|
if marker_type is _MarkerType.STRING:
|
|
# end - 1 to strip off the null terminator
|
|
s += f' "{_decode_and_escape(self.value[pos:end - 1])}"'
|
|
if end != len(self.value):
|
|
s += f",{newline}"
|
|
elif marker_type is _MarkerType.PATH:
|
|
s += " &" + ref
|
|
if end != len(self.value):
|
|
s += f",{newline}"
|
|
else:
|
|
# <> or []
|
|
|
|
if marker_type is _MarkerType.LABEL:
|
|
s += f" {ref}:"
|
|
elif marker_type is _MarkerType.PHANDLE:
|
|
s += " &" + ref
|
|
pos += 4
|
|
if pos != len(self.value) and len(s) - array_start > 80: # noqa: F821
|
|
s += array_newline # noqa: F821
|
|
array_start = len(s)
|
|
# Subtle: There might be more data between the phandle and
|
|
# the next marker, so we can't 'continue' here
|
|
else: # marker_type is _MarkerType.UINT*
|
|
elm_size = _TYPE_TO_N_BYTES[marker_type]
|
|
s += _N_BYTES_TO_START_STR[elm_size]
|
|
array_start = len(s)
|
|
array_newline = "\n" + " " * array_start
|
|
|
|
while pos != end:
|
|
num = int.from_bytes(self.value[pos:pos + elm_size],
|
|
"big")
|
|
if elm_size == 1:
|
|
s += f" {num:02X}"
|
|
else:
|
|
s += f" {hex(num)}"
|
|
|
|
pos += elm_size
|
|
if pos != len(self.value) and len(s) - array_start > 80:
|
|
s += array_newline
|
|
array_start = len(s)
|
|
|
|
if (pos != 0
|
|
and (not next_marker
|
|
or next_marker[1]
|
|
not in (_MarkerType.PHANDLE, _MarkerType.LABEL))):
|
|
|
|
s += _N_BYTES_TO_END_STR[elm_size]
|
|
if pos != len(self.value):
|
|
s += f",{newline}"
|
|
|
|
return s + ";"
|
|
|
|
def __repr__(self):
|
|
return (f"<Property '{self.name}' on {self.node.path} in "
|
|
f"{self.filename}:{self.lineno}>")
|
|
|
|
#
|
|
# Internal functions
|
|
#
|
|
|
|
def _add_marker(self, marker_type: _MarkerType, data: Any = None):
|
|
# Helper for registering markers in the value that are processed after
|
|
# parsing. See _fixup_props(). 'marker_type' identifies the type of
|
|
# marker, and 'data' has any optional data associated with the marker.
|
|
|
|
# len(self.value) gives the current offset. This function is called
|
|
# while the value is built. We use a list instead of a tuple to be able
|
|
# to fix up offsets later (they might increase if the value includes
|
|
# path references, e.g. 'foo = &bar, <3>;', which are expanded later).
|
|
self._markers.append([len(self.value), marker_type, data])
|
|
|
|
# For phandle references, add a dummy value with the same length as a
|
|
# phandle. This is handy for the length check in _register_phandles().
|
|
if marker_type is _MarkerType.PHANDLE:
|
|
self.value += b"\0\0\0\0"
|
|
|
|
class _T(enum.IntEnum):
|
|
# Token IDs used by the DT lexer.
|
|
|
|
# These values must be contiguous and start from 1.
|
|
INCLUDE = 1
|
|
LINE = 2
|
|
STRING = 3
|
|
DTS_V1 = 4
|
|
PLUGIN = 5
|
|
MEMRESERVE = 6
|
|
BITS = 7
|
|
DEL_PROP = 8
|
|
DEL_NODE = 9
|
|
OMIT_IF_NO_REF = 10
|
|
LABEL = 11
|
|
CHAR_LITERAL = 12
|
|
REF = 13
|
|
INCBIN = 14
|
|
SKIP = 15
|
|
EOF = 16
|
|
|
|
# These values must be larger than the above contiguous range.
|
|
NUM = 17
|
|
PROPNODENAME = 18
|
|
MISC = 19
|
|
BYTE = 20
|
|
BAD = 21
|
|
|
|
class _FileStackElt(NamedTuple):
|
|
# Used for maintaining the /include/ stack.
|
|
|
|
filename: str
|
|
lineno: int
|
|
contents: str
|
|
pos: int
|
|
|
|
_TokVal = Union[int, str]
|
|
|
|
class _Token(NamedTuple):
|
|
id: int
|
|
val: _TokVal
|
|
|
|
def __repr__(self):
|
|
id_repr = _T(self.id).name
|
|
return f'Token(id=_T.{id_repr}, val={repr(self.val)})'
|
|
|
|
class DT:
|
|
"""
|
|
Represents a devicetree parsed from a .dts file (or from many files, if the
|
|
.dts file /include/s other files). Creating many instances of this class is
|
|
fine. The library has no global state.
|
|
|
|
These attributes are available on DT instances:
|
|
|
|
root:
|
|
A Node instance representing the root (/) node.
|
|
|
|
alias2node:
|
|
A dictionary that maps maps alias strings (from /aliases) to Node
|
|
instances
|
|
|
|
label2node:
|
|
A dictionary that maps each node label (a string) to the Node instance
|
|
for the node.
|
|
|
|
label2prop:
|
|
A dictionary that maps each property label (a string) to a Property
|
|
instance.
|
|
|
|
label2prop_offset:
|
|
A dictionary that maps each label (a string) within a property value
|
|
(e.g., 'x = label_1: < 1 label2: 2 >;') to a (prop, offset) tuple, where
|
|
'prop' is a Property instance and 'offset' the byte offset (0 for label_1
|
|
and 4 for label_2 in the example).
|
|
|
|
phandle2node:
|
|
A dictionary that maps each phandle (a number) to a Node instance.
|
|
|
|
memreserves:
|
|
A list of (labels, address, length) tuples for the /memreserve/s in the
|
|
.dts file, in the same order as they appear in the file.
|
|
|
|
'labels' is a possibly empty set with all labels preceding the memreserve
|
|
(e.g., 'label1: label2: /memreserve/ ...'). 'address' and 'length' are
|
|
numbers.
|
|
|
|
filename:
|
|
The filename passed to the DT constructor.
|
|
"""
|
|
|
|
#
|
|
# Public interface
|
|
#
|
|
|
|
def __init__(self, filename: Optional[str], include_path: Iterable[str] = (),
|
|
force: bool = False, base_dir: Optional[str] = None):
|
|
"""
|
|
Parses a DTS file to create a DT instance. Raises OSError if 'filename'
|
|
can't be opened, and DTError for any parse errors.
|
|
|
|
filename:
|
|
Path to the .dts file to parse. (If None, an empty devicetree
|
|
is created; this is unlikely to be what you want.)
|
|
|
|
include_path:
|
|
An iterable (e.g. list or tuple) containing paths to search for
|
|
/include/d and /incbin/'d files. By default, files are only looked up
|
|
relative to the .dts file that contains the /include/ or /incbin/.
|
|
|
|
force:
|
|
Try not to raise DTError even if the input tree has errors.
|
|
For experimental use; results not guaranteed.
|
|
|
|
base_dir:
|
|
Path to the directory that is to be used as the reference for
|
|
the generated relative paths in comments. When not provided, the
|
|
current working directory is used.
|
|
"""
|
|
# Remember to update __deepcopy__() if you change this.
|
|
|
|
self._root: Optional[Node] = None
|
|
self.alias2node: dict[str, Node] = {}
|
|
self.label2node: dict[str, Node] = {}
|
|
self.label2prop: dict[str, Property] = {}
|
|
self.label2prop_offset: dict[str, tuple[Property, int]] = {}
|
|
self.phandle2node: dict[int, Node] = {}
|
|
self.memreserves: list[tuple[set[str], int, int]] = []
|
|
self.filename = filename
|
|
|
|
self._force = force
|
|
self._base_dir = base_dir or os.getcwd()
|
|
|
|
if filename is not None:
|
|
self._parse_file(filename, include_path)
|
|
else:
|
|
self._include_path: list[str] = []
|
|
|
|
@property
|
|
def root(self) -> Node:
|
|
"""
|
|
See the class documentation.
|
|
"""
|
|
# This is necessary because mypy can't tell that we never
|
|
# treat self._root as a non-None value until it's initialized
|
|
# properly in _parse_dt().
|
|
return self._root # type: ignore
|
|
|
|
def get_node(self, path: str) -> Node:
|
|
"""
|
|
Returns the Node instance for the node with path or alias 'path' (a
|
|
string). Raises DTError if the path or alias doesn't exist.
|
|
|
|
For example, both dt.get_node("/foo/bar") and dt.get_node("bar-alias")
|
|
will return the 'bar' node below:
|
|
|
|
/dts-v1/;
|
|
|
|
/ {
|
|
foo {
|
|
bar_label: bar {
|
|
baz {
|
|
};
|
|
};
|
|
};
|
|
|
|
aliases {
|
|
bar-alias = &bar-label;
|
|
};
|
|
};
|
|
|
|
Fetching subnodes via aliases is supported:
|
|
dt.get_node("bar-alias/baz") returns the 'baz' node.
|
|
"""
|
|
if path.startswith("/"):
|
|
return _root_and_path_to_node(self.root, path, path)
|
|
|
|
# Path does not start with '/'. First component must be an alias.
|
|
alias, _, rest = path.partition("/")
|
|
if alias not in self.alias2node:
|
|
_err(f"no alias '{alias}' found -- did you forget the leading "
|
|
"'/' in the node path?")
|
|
|
|
return _root_and_path_to_node(self.alias2node[alias], rest, path)
|
|
|
|
def has_node(self, path: str) -> bool:
|
|
"""
|
|
Returns True if the path or alias 'path' exists. See DT.get_node().
|
|
"""
|
|
try:
|
|
self.get_node(path)
|
|
return True
|
|
except DTError:
|
|
return False
|
|
|
|
def move_node(self, node: Node, new_path: str):
|
|
"""
|
|
Move a node 'node' to a new path 'new_path'. The entire subtree
|
|
rooted at 'node' is moved along with it.
|
|
|
|
You cannot move the root node or provide a 'new_path' value
|
|
where a node already exists. This method raises an exception
|
|
in both cases.
|
|
|
|
As a restriction on the current implementation, the parent node
|
|
of the new path must exist.
|
|
"""
|
|
if node is self.root:
|
|
_err("the root node can't be moved")
|
|
|
|
if self.has_node(new_path):
|
|
_err(f"can't move '{node.path}' to '{new_path}': "
|
|
'destination node exists')
|
|
|
|
if not new_path.startswith('/'):
|
|
_err(f"path '{new_path}' doesn't start with '/'")
|
|
|
|
for component in new_path.split('/'):
|
|
for char in component:
|
|
if char not in _nodename_chars:
|
|
_err(f"new path '{new_path}': bad character '{char}'")
|
|
|
|
old_name = node.name
|
|
old_path = node.path
|
|
|
|
new_parent_path, _, new_name = new_path.rpartition('/')
|
|
if new_parent_path == '':
|
|
# '/foo'.rpartition('/') is ('', '/', 'foo').
|
|
new_parent_path = '/'
|
|
if not self.has_node(new_parent_path):
|
|
_err(f"can't move '{old_path}' to '{new_path}': "
|
|
f"parent node '{new_parent_path}' doesn't exist")
|
|
new_parent = self.get_node(new_parent_path)
|
|
if TYPE_CHECKING:
|
|
assert new_parent is not None
|
|
assert node.parent is not None
|
|
|
|
del node.parent.nodes[old_name]
|
|
node._name = new_name
|
|
node.parent = new_parent
|
|
new_parent.nodes[new_name] = node
|
|
|
|
def node_iter(self) -> Iterable[Node]:
|
|
"""
|
|
Returns a generator for iterating over all nodes in the devicetree.
|
|
|
|
For example, this will print the name of each node that has a property
|
|
called 'foo':
|
|
|
|
for node in dt.node_iter():
|
|
if "foo" in node.props:
|
|
print(node.name)
|
|
"""
|
|
yield from self.root.node_iter()
|
|
|
|
def __str__(self):
|
|
"""
|
|
Returns a DTS representation of the devicetree. Called automatically if
|
|
the DT instance is print()ed.
|
|
"""
|
|
s = "/dts-v1/;\n"
|
|
|
|
if self.memreserves:
|
|
s += "\n"
|
|
for labels, address, offset in self.memreserves:
|
|
# List the labels in a consistent order to help with testing
|
|
for label in labels:
|
|
s += f"{label}: "
|
|
s += f"/memreserve/ {address:#018x} {offset:#018x};\n"
|
|
|
|
return s + str(self.root)
|
|
|
|
def __repr__(self):
|
|
"""
|
|
Returns some information about the DT instance. Called automatically if
|
|
the DT instance is evaluated.
|
|
"""
|
|
if self.filename:
|
|
return (f"DT(filename='{self.filename}', "
|
|
f"include_path={self._include_path})")
|
|
return super().__repr__()
|
|
|
|
def __deepcopy__(self, memo):
|
|
"""
|
|
Implements support for the standard library copy.deepcopy()
|
|
function on DT instances.
|
|
"""
|
|
|
|
# We need a new DT, obviously. Make a new, empty one.
|
|
ret = DT(None, (), self._force, self._base_dir)
|
|
|
|
# Now allocate new Node objects for every node in self, to use
|
|
# in the new DT. Set their parents to None for now and leave
|
|
# them without any properties. We will recursively initialize
|
|
# copies of parents before copies of children next.
|
|
path2node_copy = {
|
|
node.path: Node(node.name, None, ret, node.filename, node.lineno)
|
|
for node in self.node_iter()
|
|
}
|
|
|
|
# Point each copy of a node to the copy of its parent and set up
|
|
# copies of each property.
|
|
#
|
|
# Share data when possible. For example, Property.value has
|
|
# type 'bytes', which is immutable. We therefore don't need a
|
|
# copy and can just point to the original data.
|
|
|
|
for node in self.node_iter():
|
|
node_copy = path2node_copy[node.path]
|
|
|
|
parent = node.parent
|
|
if parent is not None:
|
|
node_copy.parent = path2node_copy[parent.path]
|
|
|
|
prop_name2prop_copy = {
|
|
prop.name: Property(node_copy, prop.name)
|
|
for prop in node.props.values()
|
|
}
|
|
for prop_name, prop_copy in prop_name2prop_copy.items():
|
|
prop = node.props[prop_name]
|
|
prop_copy.value = prop.value
|
|
prop_copy.labels = prop.labels[:]
|
|
prop_copy.offset_labels = prop.offset_labels.copy()
|
|
prop_copy._label_offset_lst = prop._label_offset_lst[:]
|
|
prop_copy._markers = [marker[:] for marker in prop._markers]
|
|
prop_copy.filename = prop.filename
|
|
prop_copy.lineno = prop.lineno
|
|
node_copy.props = prop_name2prop_copy
|
|
|
|
node_copy.nodes = {
|
|
child_name: path2node_copy[child_node.path]
|
|
for child_name, child_node in node.nodes.items()
|
|
}
|
|
|
|
node_copy.labels = node.labels[:]
|
|
|
|
node_copy._omit_if_no_ref = node._omit_if_no_ref
|
|
node_copy._is_referenced = node._is_referenced
|
|
|
|
# The copied nodes and properties are initialized, so
|
|
# we can finish initializing the copied DT object now.
|
|
|
|
ret._root = path2node_copy['/']
|
|
|
|
def copy_node_lookup_table(attr_name):
|
|
original = getattr(self, attr_name)
|
|
copy = {
|
|
key: path2node_copy[original[key].path]
|
|
for key in original
|
|
}
|
|
setattr(ret, attr_name, copy)
|
|
|
|
copy_node_lookup_table('alias2node')
|
|
copy_node_lookup_table('label2node')
|
|
copy_node_lookup_table('phandle2node')
|
|
|
|
ret_label2prop = {}
|
|
for label, prop in self.label2prop.items():
|
|
node_copy = path2node_copy[prop.node.path]
|
|
prop_copy = node_copy.props[prop.name]
|
|
ret_label2prop[label] = prop_copy
|
|
ret.label2prop = ret_label2prop
|
|
|
|
ret_label2prop_offset = {}
|
|
for label, prop_offset in self.label2prop_offset.items():
|
|
prop, offset = prop_offset
|
|
node_copy = path2node_copy[prop.node.path]
|
|
prop_copy = node_copy.props[prop.name]
|
|
ret_label2prop_offset[label] = (prop_copy, offset)
|
|
ret.label2prop_offset = ret_label2prop_offset
|
|
|
|
ret.memreserves = [
|
|
(set(memreserve[0]), memreserve[1], memreserve[2])
|
|
for memreserve in self.memreserves
|
|
]
|
|
|
|
ret.filename = self.filename
|
|
|
|
return ret
|
|
|
|
#
|
|
# Parsing
|
|
#
|
|
|
|
def _parse_file(self, filename: str, include_path: Iterable[str]):
|
|
self._include_path = list(include_path)
|
|
|
|
with open(filename, encoding="utf-8") as f:
|
|
self._file_contents = f.read()
|
|
|
|
self._tok_i = self._tok_end_i = 0
|
|
self._filestack: list[_FileStackElt] = []
|
|
|
|
self._lexer_state: int = _DEFAULT
|
|
self._saved_token: Optional[_Token] = None
|
|
|
|
self._lineno: int = 1
|
|
|
|
self._parse_header()
|
|
self._parse_memreserves()
|
|
self._parse_dt()
|
|
|
|
self._register_phandles()
|
|
self._fixup_props()
|
|
self._register_aliases()
|
|
self._remove_unreferenced()
|
|
self._register_labels()
|
|
|
|
def _parse_header(self):
|
|
# Parses /dts-v1/ (expected) and /plugin/ (unsupported) at the start of
|
|
# files. There may be multiple /dts-v1/ at the start of a file.
|
|
|
|
has_dts_v1 = False
|
|
|
|
while self._peek_token().id == _T.DTS_V1:
|
|
has_dts_v1 = True
|
|
self._next_token()
|
|
self._expect_token(";")
|
|
# /plugin/ always comes after /dts-v1/
|
|
if self._peek_token().id == _T.PLUGIN:
|
|
self._parse_error("/plugin/ is not supported")
|
|
|
|
if not has_dts_v1:
|
|
self._parse_error("expected '/dts-v1/;' at start of file")
|
|
|
|
def _parse_memreserves(self):
|
|
# Parses /memreserve/, which appears after /dts-v1/
|
|
|
|
while True:
|
|
# Labels before /memreserve/
|
|
labels = []
|
|
while self._peek_token().id == _T.LABEL:
|
|
_append_no_dup(labels, self._next_token().val)
|
|
|
|
if self._peek_token().id == _T.MEMRESERVE:
|
|
self._next_token()
|
|
self.memreserves.append(
|
|
(labels, self._eval_prim(), self._eval_prim()))
|
|
self._expect_token(";")
|
|
elif labels:
|
|
self._parse_error("expected /memreserve/ after labels at "
|
|
"beginning of file")
|
|
else:
|
|
return
|
|
|
|
def _parse_dt(self):
|
|
# Top-level parsing loop
|
|
|
|
while True:
|
|
tok = self._next_token()
|
|
|
|
if tok.val == "/":
|
|
# '/ { ... };', the root node
|
|
if not self._root:
|
|
self._root = Node(
|
|
name="/", parent=None, dt=self, filename=self.filename, lineno=self._lineno
|
|
)
|
|
self._parse_node(self.root)
|
|
|
|
elif tok.id in (_T.LABEL, _T.REF):
|
|
# '&foo { ... };' or 'label: &foo { ... };'. The C tools only
|
|
# support a single label here too.
|
|
|
|
if tok.id == _T.LABEL:
|
|
label = tok.val
|
|
tok = self._next_token()
|
|
if tok.id != _T.REF:
|
|
self._parse_error("expected label reference (&foo)")
|
|
else:
|
|
label = None
|
|
|
|
try:
|
|
node = self._ref2node(tok.val)
|
|
except DTError as e:
|
|
self._parse_error(e)
|
|
self._parse_node(node)
|
|
|
|
if label:
|
|
_append_no_dup(node.labels, label)
|
|
|
|
elif tok.id == _T.DEL_NODE:
|
|
self._next_ref2node()._del()
|
|
self._expect_token(";")
|
|
|
|
elif tok.id == _T.OMIT_IF_NO_REF:
|
|
self._next_ref2node()._omit_if_no_ref = True
|
|
self._expect_token(";")
|
|
|
|
elif tok.id == _T.EOF:
|
|
if not self._root:
|
|
self._parse_error("no root node defined")
|
|
return
|
|
|
|
else:
|
|
self._parse_error("expected '/' or label reference (&foo)")
|
|
|
|
def _parse_node(self, node):
|
|
# Parses the '{ ... };' part of 'node-name { ... };'.
|
|
|
|
# We need to track which child nodes were defined in this set
|
|
# of curly braces in order to reject duplicate node names.
|
|
current_child_names = set()
|
|
|
|
self._expect_token("{")
|
|
while True:
|
|
labels, omit_if_no_ref = self._parse_propnode_labels()
|
|
tok = self._next_token()
|
|
|
|
if tok.id == _T.PROPNODENAME:
|
|
if self._peek_token().val == "{":
|
|
# '<tok> { ...', expect node
|
|
|
|
# Fetch the existing node if it already exists. This
|
|
# happens when overriding nodes.
|
|
child = node.nodes.get(tok.val)
|
|
if child:
|
|
if child.name in current_child_names:
|
|
self._parse_error(f'{child.path}: duplicate node name')
|
|
else:
|
|
child = Node(
|
|
name=tok.val,
|
|
parent=node,
|
|
dt=self,
|
|
filename=self.filename,
|
|
lineno=self._lineno,
|
|
)
|
|
current_child_names.add(tok.val)
|
|
|
|
for label in labels:
|
|
_append_no_dup(child.labels, label)
|
|
|
|
if omit_if_no_ref:
|
|
child._omit_if_no_ref = True
|
|
|
|
node.nodes[child.name] = child
|
|
self._parse_node(child)
|
|
|
|
else:
|
|
# Not '<tok> { ...', expect property assignment
|
|
|
|
if omit_if_no_ref:
|
|
self._parse_error(
|
|
"/omit-if-no-ref/ can only be used on nodes")
|
|
|
|
prop = node._get_prop(tok.val)
|
|
prop.filename = self.filename
|
|
prop.lineno = self._lineno
|
|
|
|
if self._check_token("="):
|
|
self._parse_assignment(prop)
|
|
elif not self._check_token(";"):
|
|
# ';' is for an empty property, like 'foo;'
|
|
self._parse_error("expected '{', '=', or ';'")
|
|
|
|
for label in labels:
|
|
_append_no_dup(prop.labels, label)
|
|
|
|
elif tok.id == _T.DEL_NODE:
|
|
tok2 = self._next_token()
|
|
if tok2.id != _T.PROPNODENAME:
|
|
self._parse_error("expected node name")
|
|
if tok2.val in node.nodes:
|
|
node.nodes[tok2.val]._del()
|
|
self._expect_token(";")
|
|
|
|
elif tok.id == _T.DEL_PROP:
|
|
tok2 = self._next_token()
|
|
if tok2.id != _T.PROPNODENAME:
|
|
self._parse_error("expected property name")
|
|
node.props.pop(tok2.val, None)
|
|
self._expect_token(";")
|
|
|
|
elif tok.val == "}":
|
|
self._expect_token(";")
|
|
return
|
|
|
|
else:
|
|
self._parse_error("expected node name, property name, or '}'")
|
|
|
|
def _parse_propnode_labels(self):
|
|
# _parse_node() helpers for parsing labels and /omit-if-no-ref/s before
|
|
# nodes and properties. Returns a (<label list>, <omit-if-no-ref bool>)
|
|
# tuple.
|
|
|
|
labels = []
|
|
omit_if_no_ref = False
|
|
while True:
|
|
tok = self._peek_token()
|
|
if tok.id == _T.LABEL:
|
|
_append_no_dup(labels, tok.val)
|
|
elif tok.id == _T.OMIT_IF_NO_REF:
|
|
omit_if_no_ref = True
|
|
elif (labels or omit_if_no_ref) and tok.id != _T.PROPNODENAME:
|
|
# Got something like 'foo: bar: }'
|
|
self._parse_error("expected node or property name")
|
|
else:
|
|
return labels, omit_if_no_ref
|
|
|
|
self._next_token()
|
|
|
|
def _parse_assignment(self, prop):
|
|
# Parses the right-hand side of property assignment
|
|
#
|
|
# prop:
|
|
# 'Property' instance being assigned
|
|
|
|
# Remove any old value, path/phandle references, and in-value labels,
|
|
# in case the property value is being overridden
|
|
prop.value = b""
|
|
prop._markers = []
|
|
|
|
while True:
|
|
# Parse labels before the value (e.g., '..., label: < 0 >')
|
|
self._parse_value_labels(prop)
|
|
|
|
tok = self._next_token()
|
|
|
|
if tok.val == "<":
|
|
self._parse_cells(prop, 4)
|
|
|
|
elif tok.id == _T.BITS:
|
|
n_bits = self._expect_num()
|
|
if n_bits not in {8, 16, 32, 64}:
|
|
self._parse_error("expected 8, 16, 32, or 64")
|
|
self._expect_token("<")
|
|
self._parse_cells(prop, n_bits//8)
|
|
|
|
elif tok.val == "[":
|
|
self._parse_bytes(prop)
|
|
|
|
elif tok.id == _T.STRING:
|
|
prop._add_marker(_MarkerType.STRING)
|
|
prop.value += self._unescape(tok.val.encode("utf-8")) + b"\0"
|
|
|
|
elif tok.id == _T.REF:
|
|
prop._add_marker(_MarkerType.PATH, tok.val)
|
|
|
|
elif tok.id == _T.INCBIN:
|
|
self._parse_incbin(prop)
|
|
|
|
else:
|
|
self._parse_error("malformed value")
|
|
|
|
# Parse labels after the value (e.g., '< 0 > label:, ...')
|
|
self._parse_value_labels(prop)
|
|
|
|
tok = self._next_token()
|
|
if tok.val == ";":
|
|
return
|
|
if tok.val == ",":
|
|
continue
|
|
self._parse_error("expected ';' or ','")
|
|
|
|
def _parse_cells(self, prop, n_bytes):
|
|
# Parses '<...>'
|
|
|
|
prop._add_marker(_N_BYTES_TO_TYPE[n_bytes])
|
|
|
|
while True:
|
|
tok = self._peek_token()
|
|
if tok.id == _T.REF:
|
|
self._next_token()
|
|
if n_bytes != 4:
|
|
self._parse_error("phandle references are only allowed in "
|
|
"arrays with 32-bit elements")
|
|
prop._add_marker(_MarkerType.PHANDLE, tok.val)
|
|
|
|
elif tok.id == _T.LABEL:
|
|
prop._add_marker(_MarkerType.LABEL, tok.val)
|
|
self._next_token()
|
|
|
|
elif self._check_token(">"):
|
|
return
|
|
|
|
else:
|
|
# Literal value
|
|
num = self._eval_prim()
|
|
try:
|
|
prop.value += num.to_bytes(n_bytes, "big")
|
|
except OverflowError:
|
|
try:
|
|
# Try again as a signed number, in case it's negative
|
|
prop.value += num.to_bytes(n_bytes, "big", signed=True)
|
|
except OverflowError:
|
|
self._parse_error(
|
|
f"{num} does not fit in {8*n_bytes} bits")
|
|
|
|
def _parse_bytes(self, prop):
|
|
# Parses '[ ... ]'
|
|
|
|
prop._add_marker(_MarkerType.UINT8)
|
|
|
|
while True:
|
|
tok = self._next_token()
|
|
if tok.id == _T.BYTE:
|
|
prop.value += tok.val.to_bytes(1, "big")
|
|
|
|
elif tok.id == _T.LABEL:
|
|
prop._add_marker(_MarkerType.LABEL, tok.val)
|
|
|
|
elif tok.val == "]":
|
|
return
|
|
|
|
else:
|
|
self._parse_error("expected two-digit byte or ']'")
|
|
|
|
def _parse_incbin(self, prop):
|
|
# Parses
|
|
#
|
|
# /incbin/ ("filename")
|
|
#
|
|
# and
|
|
#
|
|
# /incbin/ ("filename", <offset>, <size>)
|
|
|
|
prop._add_marker(_MarkerType.UINT8)
|
|
|
|
self._expect_token("(")
|
|
|
|
tok = self._next_token()
|
|
if tok.id != _T.STRING:
|
|
self._parse_error("expected quoted filename")
|
|
filename = tok.val
|
|
|
|
tok = self._next_token()
|
|
if tok.val == ",":
|
|
offset = self._eval_prim()
|
|
self._expect_token(",")
|
|
size = self._eval_prim()
|
|
self._expect_token(")")
|
|
else:
|
|
if tok.val != ")":
|
|
self._parse_error("expected ',' or ')'")
|
|
offset = None
|
|
|
|
try:
|
|
with self._open(filename, "rb") as f:
|
|
if offset is None:
|
|
prop.value += f.read()
|
|
else:
|
|
f.seek(offset)
|
|
prop.value += f.read(size)
|
|
except OSError as e:
|
|
self._parse_error(f"could not read '{filename}': {e}")
|
|
|
|
def _parse_value_labels(self, prop):
|
|
# _parse_assignment() helper for parsing labels before/after each
|
|
# comma-separated value
|
|
|
|
while True:
|
|
tok = self._peek_token()
|
|
if tok.id != _T.LABEL:
|
|
return
|
|
prop._add_marker(_MarkerType.LABEL, tok.val)
|
|
self._next_token()
|
|
|
|
def _node_phandle(self, node, src_prop):
|
|
# Returns the phandle for Node 'node', creating a new phandle if the
|
|
# node has no phandle, and fixing up the value for existing
|
|
# self-referential phandles (which get set to b'\0\0\0\0' initially).
|
|
# Self-referential phandles must be rewritten instead of recreated, so
|
|
# that labels are preserved.
|
|
|
|
if "phandle" in node.props:
|
|
phandle_prop = node.props["phandle"]
|
|
else:
|
|
phandle_prop = Property(node, "phandle")
|
|
phandle_prop._add_marker(_MarkerType.UINT32) # For displaying
|
|
phandle_prop.value = b'\0\0\0\0'
|
|
|
|
if phandle_prop.value == b'\0\0\0\0':
|
|
phandle_i = 1
|
|
while phandle_i in self.phandle2node:
|
|
phandle_i += 1
|
|
self.phandle2node[phandle_i] = node
|
|
|
|
phandle_prop.filename = src_prop.filename
|
|
phandle_prop.lineno = src_prop.lineno
|
|
phandle_prop.value = phandle_i.to_bytes(4, "big")
|
|
node.props["phandle"] = phandle_prop
|
|
|
|
return phandle_prop.value
|
|
|
|
# Expression evaluation
|
|
|
|
def _eval_prim(self):
|
|
tok = self._peek_token()
|
|
if tok.id in (_T.NUM, _T.CHAR_LITERAL):
|
|
return self._next_token().val
|
|
|
|
tok = self._next_token()
|
|
if tok.val != "(":
|
|
self._parse_error("expected number or parenthesized expression")
|
|
val = self._eval_ternary()
|
|
self._expect_token(")")
|
|
return val
|
|
|
|
def _eval_ternary(self):
|
|
val = self._eval_or()
|
|
if self._check_token("?"):
|
|
if_val = self._eval_ternary()
|
|
self._expect_token(":")
|
|
else_val = self._eval_ternary()
|
|
return if_val if val else else_val
|
|
return val
|
|
|
|
def _eval_or(self):
|
|
val = self._eval_and()
|
|
while self._check_token("||"):
|
|
val = 1 if self._eval_and() or val else 0
|
|
return val
|
|
|
|
def _eval_and(self):
|
|
val = self._eval_bitor()
|
|
while self._check_token("&&"):
|
|
val = 1 if self._eval_bitor() and val else 0
|
|
return val
|
|
|
|
def _eval_bitor(self):
|
|
val = self._eval_bitxor()
|
|
while self._check_token("|"):
|
|
val |= self._eval_bitxor()
|
|
return val
|
|
|
|
def _eval_bitxor(self):
|
|
val = self._eval_bitand()
|
|
while self._check_token("^"):
|
|
val ^= self._eval_bitand()
|
|
return val
|
|
|
|
def _eval_bitand(self):
|
|
val = self._eval_eq()
|
|
while self._check_token("&"):
|
|
val &= self._eval_eq()
|
|
return val
|
|
|
|
def _eval_eq(self):
|
|
val = self._eval_rela()
|
|
while True:
|
|
if self._check_token("=="):
|
|
val = 1 if val == self._eval_rela() else 0
|
|
elif self._check_token("!="):
|
|
val = 1 if val != self._eval_rela() else 0
|
|
else:
|
|
return val
|
|
|
|
def _eval_rela(self):
|
|
val = self._eval_shift()
|
|
while True:
|
|
if self._check_token("<"):
|
|
val = 1 if val < self._eval_shift() else 0
|
|
elif self._check_token(">"):
|
|
val = 1 if val > self._eval_shift() else 0
|
|
elif self._check_token("<="):
|
|
val = 1 if val <= self._eval_shift() else 0
|
|
elif self._check_token(">="):
|
|
val = 1 if val >= self._eval_shift() else 0
|
|
else:
|
|
return val
|
|
|
|
def _eval_shift(self):
|
|
val = self._eval_add()
|
|
while True:
|
|
if self._check_token("<<"):
|
|
val <<= self._eval_add()
|
|
elif self._check_token(">>"):
|
|
val >>= self._eval_add()
|
|
else:
|
|
return val
|
|
|
|
def _eval_add(self):
|
|
val = self._eval_mul()
|
|
while True:
|
|
if self._check_token("+"):
|
|
val += self._eval_mul()
|
|
elif self._check_token("-"):
|
|
val -= self._eval_mul()
|
|
else:
|
|
return val
|
|
|
|
def _eval_mul(self):
|
|
val = self._eval_unary()
|
|
while True:
|
|
if self._check_token("*"):
|
|
val *= self._eval_unary()
|
|
elif self._check_token("/"):
|
|
denom = self._eval_unary()
|
|
if not denom:
|
|
self._parse_error("division by zero")
|
|
val //= denom
|
|
elif self._check_token("%"):
|
|
denom = self._eval_unary()
|
|
if not denom:
|
|
self._parse_error("division by zero")
|
|
val %= denom
|
|
else:
|
|
return val
|
|
|
|
def _eval_unary(self):
|
|
if self._check_token("-"):
|
|
return -self._eval_unary()
|
|
if self._check_token("~"):
|
|
return ~self._eval_unary()
|
|
if self._check_token("!"):
|
|
return 0 if self._eval_unary() else 1
|
|
return self._eval_prim()
|
|
|
|
#
|
|
# Lexing
|
|
#
|
|
|
|
def _check_token(self, val):
|
|
if self._peek_token().val == val:
|
|
self._next_token()
|
|
return True
|
|
return False
|
|
|
|
def _peek_token(self):
|
|
if not self._saved_token:
|
|
self._saved_token = self._next_token()
|
|
return self._saved_token
|
|
|
|
def _next_token(self):
|
|
if self._saved_token:
|
|
tmp = self._saved_token
|
|
self._saved_token = None
|
|
return tmp
|
|
|
|
while True:
|
|
tok_id = None
|
|
|
|
match = _token_re.match(self._file_contents, self._tok_end_i)
|
|
if match:
|
|
tok_id = match.lastindex
|
|
if tok_id == _T.CHAR_LITERAL:
|
|
val = self._unescape(match.group(tok_id).encode("utf-8"))
|
|
if len(val) != 1:
|
|
self._parse_error("character literals must be length 1")
|
|
tok_val = ord(val)
|
|
else:
|
|
tok_val = match.group(tok_id)
|
|
|
|
elif self._lexer_state is _DEFAULT:
|
|
match = _num_re.match(self._file_contents, self._tok_end_i)
|
|
if match:
|
|
tok_id = _T.NUM
|
|
num_s = match.group(1)
|
|
tok_val = int(num_s,
|
|
16 if num_s.startswith(("0x", "0X")) else
|
|
8 if num_s[0] == "0" else
|
|
10)
|
|
|
|
elif self._lexer_state is _EXPECT_PROPNODENAME:
|
|
match = _propnodename_re.match(self._file_contents,
|
|
self._tok_end_i)
|
|
if match:
|
|
tok_id = _T.PROPNODENAME
|
|
tok_val = match.group(1)
|
|
self._lexer_state = _DEFAULT
|
|
|
|
else: # self._lexer_state is _EXPECT_BYTE
|
|
match = _byte_re.match(self._file_contents, self._tok_end_i)
|
|
if match:
|
|
tok_id = _T.BYTE
|
|
tok_val = int(match.group(), 16)
|
|
|
|
if not tok_id:
|
|
match = _misc_re.match(self._file_contents, self._tok_end_i)
|
|
if match:
|
|
tok_id = _T.MISC
|
|
tok_val = match.group()
|
|
else:
|
|
self._tok_i = self._tok_end_i
|
|
# Could get here due to a node/property naming appearing in
|
|
# an unexpected context as well as for bad characters in
|
|
# files. Generate a token for it so that the error can
|
|
# trickle up to some context where we can give a more
|
|
# helpful error message.
|
|
return _Token(_T.BAD, "<unknown token>")
|
|
|
|
self._tok_i = match.start()
|
|
self._tok_end_i = match.end()
|
|
|
|
if tok_id == _T.SKIP:
|
|
self._lineno += tok_val.count("\n")
|
|
continue
|
|
|
|
# /include/ is handled in the lexer in the C tools as well, and can
|
|
# appear anywhere
|
|
if tok_id == _T.INCLUDE:
|
|
# Can have newlines between /include/ and the filename
|
|
self._lineno += tok_val.count("\n")
|
|
# Do this manual extraction instead of doing it in the regex so
|
|
# that we can properly count newlines
|
|
filename = tok_val[tok_val.find('"') + 1:-1]
|
|
self._enter_file(filename)
|
|
continue
|
|
|
|
if tok_id == _T.LINE:
|
|
# #line directive
|
|
self._lineno = int(tok_val.split()[0]) - 1
|
|
self.filename = tok_val[tok_val.find('"') + 1:-1]
|
|
continue
|
|
|
|
if tok_id == _T.EOF:
|
|
if self._filestack:
|
|
self._leave_file()
|
|
continue
|
|
return _Token(_T.EOF, "<EOF>")
|
|
|
|
# State handling
|
|
|
|
if (tok_id in (_T.DEL_PROP, _T.DEL_NODE, _T.OMIT_IF_NO_REF)
|
|
or tok_val in ("{", ";")):
|
|
|
|
self._lexer_state = _EXPECT_PROPNODENAME
|
|
|
|
elif tok_val == "[":
|
|
self._lexer_state = _EXPECT_BYTE
|
|
|
|
elif tok_id in (_T.MEMRESERVE, _T.BITS) or tok_val == "]":
|
|
self._lexer_state = _DEFAULT
|
|
|
|
return _Token(tok_id, tok_val)
|
|
|
|
def _expect_token(self, tok_val):
|
|
# Raises an error if the next token does not have the string value
|
|
# 'tok_val'. Returns the token.
|
|
|
|
tok = self._next_token()
|
|
if tok.val != tok_val:
|
|
self._parse_error(f"expected '{tok_val}', not '{tok.val}'")
|
|
|
|
return tok
|
|
|
|
def _expect_num(self):
|
|
# Raises an error if the next token is not a number. Returns the token.
|
|
|
|
tok = self._next_token()
|
|
if tok.id != _T.NUM:
|
|
self._parse_error("expected number")
|
|
return tok.val
|
|
|
|
def _parse_error(self, s):
|
|
# This works out for the first line of the file too, where rfind()
|
|
# returns -1
|
|
column = self._tok_i - self._file_contents.rfind("\n", 0,
|
|
self._tok_i + 1)
|
|
_err(f"{self.filename}:{self._lineno} (column {column}): "
|
|
f"parse error: {s}")
|
|
|
|
def _enter_file(self, filename):
|
|
# Enters the /include/d file 'filename', remembering the position in
|
|
# the /include/ing file for later
|
|
|
|
self._filestack.append(
|
|
_FileStackElt(self.filename, self._lineno,
|
|
self._file_contents, self._tok_end_i))
|
|
|
|
# Handle escapes in filenames, just for completeness
|
|
filename = self._unescape(filename.encode("utf-8"))
|
|
try:
|
|
filename = filename.decode("utf-8")
|
|
except UnicodeDecodeError:
|
|
self._parse_error("filename is not valid UTF-8")
|
|
|
|
with self._open(filename, encoding="utf-8") as f:
|
|
try:
|
|
self._file_contents = f.read()
|
|
except OSError as e:
|
|
self._parse_error(e)
|
|
|
|
# Check for recursive /include/
|
|
for i, parent in enumerate(self._filestack):
|
|
if filename == parent[0]:
|
|
self._parse_error("recursive /include/:\n" + " ->\n".join(
|
|
[f"{parent[0]}:{parent[1]}"
|
|
for parent in self._filestack[i:]] +
|
|
[filename]))
|
|
|
|
self.filename = f.name
|
|
self._lineno = 1
|
|
self._tok_end_i = 0
|
|
|
|
def _leave_file(self):
|
|
# Leaves an /include/d file, returning to the file that /include/d it
|
|
|
|
self.filename, self._lineno, self._file_contents, self._tok_end_i = (
|
|
self._filestack.pop())
|
|
|
|
def _next_ref2node(self):
|
|
# Checks that the next token is a label/path reference and returns the
|
|
# Node it points to. Only used during parsing, so uses _parse_error()
|
|
# on errors to save some code in callers.
|
|
|
|
label = self._next_token()
|
|
if label.id != _T.REF:
|
|
self._parse_error(
|
|
"expected label (&foo) or path (&{/foo/bar}) reference")
|
|
try:
|
|
return self._ref2node(label.val)
|
|
except DTError as e:
|
|
self._parse_error(e)
|
|
|
|
def _ref2node(self, s):
|
|
# Returns the Node the label/path reference 's' points to
|
|
|
|
if s[0] == "{":
|
|
# Path reference (&{/foo/bar})
|
|
path = s[1:-1]
|
|
if not path.startswith("/"):
|
|
_err(f"node path '{path}' does not start with '/'")
|
|
# Will raise DTError if the path doesn't exist
|
|
return _root_and_path_to_node(self.root, path, path)
|
|
|
|
# Label reference (&foo).
|
|
|
|
# label2node hasn't been filled in yet, and using it would get messy
|
|
# when nodes are deleted
|
|
for node in self.node_iter():
|
|
if s in node.labels:
|
|
return node
|
|
|
|
_err(f"undefined node label '{s}'")
|
|
|
|
#
|
|
# Post-processing
|
|
#
|
|
|
|
def _register_phandles(self):
|
|
# Registers any manually-inserted phandle properties in
|
|
# self.phandle2node, so that we can avoid allocating any phandles from
|
|
# that set. Also checks the format of the phandles and does misc.
|
|
# sanity checking.
|
|
|
|
for node in self.node_iter():
|
|
phandle = node.props.get("phandle")
|
|
if phandle:
|
|
if len(phandle.value) != 4:
|
|
_err(f"{node.path}: bad phandle length "
|
|
f"({len(phandle.value)}), expected 4 bytes")
|
|
|
|
is_self_referential = False
|
|
for marker in phandle._markers:
|
|
_, marker_type, ref = marker
|
|
if marker_type is _MarkerType.PHANDLE:
|
|
# The phandle's value is itself a phandle reference
|
|
if self._ref2node(ref) is node:
|
|
# Alright to set a node's phandle equal to its own
|
|
# phandle. It'll force a new phandle to be
|
|
# allocated even if the node is otherwise
|
|
# unreferenced.
|
|
is_self_referential = True
|
|
break
|
|
|
|
_err(f"{node.path}: {phandle.name} "
|
|
"refers to another node")
|
|
|
|
# Could put on else on the 'for' above too, but keep it
|
|
# somewhat readable
|
|
if not is_self_referential:
|
|
phandle_val = int.from_bytes(phandle.value, "big")
|
|
|
|
if phandle_val in {0, 0xFFFFFFFF}:
|
|
_err(f"{node.path}: bad value {phandle_val:#010x} "
|
|
f"for {phandle.name}")
|
|
|
|
if phandle_val in self.phandle2node:
|
|
_err(f"{node.path}: duplicated phandle {phandle_val:#x} "
|
|
"(seen before at "
|
|
f"{self.phandle2node[phandle_val].path})")
|
|
|
|
self.phandle2node[phandle_val] = node
|
|
|
|
def _fixup_props(self):
|
|
# Fills in node path and phandle references in property values, and
|
|
# registers labels within values. This must be done after parsing,
|
|
# since forwards references are allowed and nodes and properties might
|
|
# be deleted.
|
|
|
|
for node in self.node_iter():
|
|
# The tuple() avoids a 'dictionary changed size during iteration'
|
|
# error
|
|
for prop in tuple(node.props.values()):
|
|
# 'prev_pos' and 'pos' are indices in the unpatched
|
|
# property value. The result is built up in 'res'.
|
|
prev_pos = 0
|
|
res = b""
|
|
|
|
for marker in prop._markers:
|
|
pos, marker_type, ref = marker
|
|
|
|
# Add data before the marker, reading from the unpatched
|
|
# property value
|
|
res += prop.value[prev_pos:pos]
|
|
|
|
# Fix the marker offset so that it's correct for the
|
|
# patched property value, for later (not used in this
|
|
# function). The offset might change due to path
|
|
# references, which expand to something like "/foo/bar".
|
|
marker[0] = len(res)
|
|
|
|
if marker_type is _MarkerType.LABEL:
|
|
# This is a temporary format so that we can catch
|
|
# duplicate references. prop._label_offset_lst is changed
|
|
# to a dictionary that maps labels to offsets in
|
|
# _register_labels().
|
|
_append_no_dup(prop._label_offset_lst, (ref, len(res)))
|
|
elif marker_type in (_MarkerType.PATH, _MarkerType.PHANDLE):
|
|
# Path or phandle reference
|
|
try:
|
|
ref_node = self._ref2node(ref)
|
|
except DTError as e:
|
|
_err(f"{prop.node.path}: {e}")
|
|
|
|
# For /omit-if-no-ref/
|
|
ref_node._is_referenced = True
|
|
|
|
if marker_type is _MarkerType.PATH:
|
|
res += ref_node.path.encode("utf-8") + b'\0'
|
|
else: # marker_type is PHANDLE
|
|
res += self._node_phandle(ref_node, prop)
|
|
# Skip over the dummy phandle placeholder
|
|
pos += 4
|
|
|
|
prev_pos = pos
|
|
|
|
# Store the final fixed-up value. Add the data after the last
|
|
# marker.
|
|
prop.value = res + prop.value[prev_pos:]
|
|
|
|
def _register_aliases(self):
|
|
# Registers aliases from the /aliases node in self.alias2node. Also
|
|
# checks the format of the alias properties.
|
|
|
|
# We copy this to self.alias2node at the end to avoid get_node()
|
|
# looking up paths via other aliases while verifying aliases
|
|
alias2node = {}
|
|
|
|
alias_re = re.compile("[0-9a-z-]+$")
|
|
|
|
aliases = self.root.nodes.get("aliases")
|
|
if aliases:
|
|
for prop in aliases.props.values():
|
|
if not alias_re.match(prop.name):
|
|
_err(f"/aliases: alias property name '{prop.name}' "
|
|
"should include only characters from [0-9a-z-]")
|
|
|
|
# Property.to_path() checks that the node exists, has
|
|
# the right type, etc. Swallow errors for invalid
|
|
# aliases with self._force.
|
|
try:
|
|
alias2node[prop.name] = prop.to_path()
|
|
except DTError:
|
|
if self._force:
|
|
continue
|
|
raise
|
|
|
|
self.alias2node = alias2node
|
|
|
|
def _remove_unreferenced(self):
|
|
# Removes any unreferenced nodes marked with /omit-if-no-ref/ from the
|
|
# tree
|
|
|
|
# tuple() is to avoid 'RuntimeError: dictionary changed size during
|
|
# iteration' errors
|
|
for node in tuple(self.node_iter()):
|
|
if node._omit_if_no_ref and not node._is_referenced:
|
|
node._del()
|
|
|
|
def _register_labels(self):
|
|
# Checks for duplicate labels and registers labels in label2node,
|
|
# label2prop, and label2prop_offset
|
|
|
|
label2things = collections.defaultdict(set)
|
|
|
|
# Register all labels and the nodes/props they point to in label2things
|
|
for node in self.node_iter():
|
|
for label in node.labels:
|
|
label2things[label].add(node)
|
|
self.label2node[label] = node
|
|
|
|
for prop in node.props.values():
|
|
for label in prop.labels:
|
|
label2things[label].add(prop)
|
|
self.label2prop[label] = prop
|
|
|
|
for label, offset in prop._label_offset_lst:
|
|
label2things[label].add((prop, offset))
|
|
self.label2prop_offset[label] = (prop, offset)
|
|
|
|
# See _fixup_props()
|
|
prop.offset_labels = dict(prop._label_offset_lst)
|
|
|
|
for label, things in label2things.items():
|
|
if len(things) > 1:
|
|
strings = []
|
|
for thing in things:
|
|
if isinstance(thing, Node):
|
|
strings.append(f"on {thing.path}")
|
|
elif isinstance(thing, Property):
|
|
strings.append(f"on property '{thing.name}' "
|
|
f"of node {thing.node.path}")
|
|
else:
|
|
# Label within property value
|
|
strings.append("in the value of property "
|
|
f"'{thing[0].name}' of node "
|
|
f"{thing[0].node.path}")
|
|
|
|
# Give consistent error messages to help with testing
|
|
strings.sort()
|
|
|
|
_err(f"Label '{label}' appears " + " and ".join(strings))
|
|
|
|
|
|
#
|
|
# Misc.
|
|
#
|
|
|
|
def _unescape(self, b):
|
|
# Replaces backslash escapes in the 'bytes' array 'b'. We can't do this at
|
|
# the string level, because the result might not be valid UTF-8 when
|
|
# octal/hex escapes are involved.
|
|
|
|
def sub(match):
|
|
esc = match.group(1)
|
|
if esc == b"a":
|
|
return b"\a"
|
|
if esc == b"b":
|
|
return b"\b"
|
|
if esc == b"t":
|
|
return b"\t"
|
|
if esc == b"n":
|
|
return b"\n"
|
|
if esc == b"v":
|
|
return b"\v"
|
|
if esc == b"f":
|
|
return b"\f"
|
|
if esc == b"r":
|
|
return b"\r"
|
|
|
|
if esc[0] in b"01234567":
|
|
# Octal escape
|
|
try:
|
|
return int(esc, 8).to_bytes(1, "big")
|
|
except OverflowError:
|
|
self._parse_error("octal escape out of range (> 255)")
|
|
|
|
if esc[0] == ord("x") and len(esc) > 1:
|
|
# Hex escape
|
|
return int(esc[1:], 16).to_bytes(1, "big")
|
|
|
|
# Return <char> as-is for other \<char>
|
|
return esc[0].to_bytes(1, "big")
|
|
|
|
return _unescape_re.sub(sub, b)
|
|
|
|
def _open(self, filename, mode="r", **kwargs):
|
|
# Wrapper around standard Python open(), accepting the same params.
|
|
# But searches for a 'filename' file in the directory of the current
|
|
# file and the include path.
|
|
|
|
# The C tools support specifying stdin with '-' too
|
|
if filename == "-":
|
|
return sys.stdin.buffer if "b" in mode else sys.stdin
|
|
|
|
# Try the directory of the current file first
|
|
dirname = os.path.dirname(self.filename)
|
|
try:
|
|
return open(os.path.join(dirname, filename), mode, **kwargs)
|
|
except OSError as e:
|
|
if e.errno != errno.ENOENT:
|
|
self._parse_error(e)
|
|
|
|
# Try each directory from the include path
|
|
for path in self._include_path:
|
|
try:
|
|
return open(os.path.join(path, filename), mode, **kwargs)
|
|
except OSError as e:
|
|
if e.errno != errno.ENOENT:
|
|
self._parse_error(e)
|
|
continue
|
|
|
|
self._parse_error(f"'{filename}' could not be found")
|
|
|
|
#
|
|
# Public functions
|
|
#
|
|
|
|
def to_num(data: bytes, length: Optional[int] = None,
|
|
signed: bool = False) -> int:
|
|
"""
|
|
Converts the 'bytes' array 'data' to a number. The value is expected to be
|
|
in big-endian format, which is standard in devicetree.
|
|
|
|
length (default: None):
|
|
The expected length of the value in bytes, as a simple type check. If
|
|
None, the length check is skipped.
|
|
|
|
signed (default: False):
|
|
If True, the value will be interpreted as signed rather than unsigned.
|
|
"""
|
|
_check_is_bytes(data)
|
|
if length is not None:
|
|
_check_length_positive(length)
|
|
if len(data) != length:
|
|
_err(f"{data!r} is {len(data)} bytes long, expected {length}")
|
|
|
|
return int.from_bytes(data, "big", signed=signed)
|
|
|
|
def to_nums(data: bytes, length: int = 4, signed: bool = False) -> list[int]:
|
|
"""
|
|
Like Property.to_nums(), but takes an arbitrary 'bytes' array. The values
|
|
are assumed to be in big-endian format, which is standard in devicetree.
|
|
"""
|
|
_check_is_bytes(data)
|
|
_check_length_positive(length)
|
|
|
|
if len(data) % length:
|
|
_err(f"{data!r} is {len(data)} bytes long, "
|
|
f"expected a length that's a multiple of {length}")
|
|
|
|
return [int.from_bytes(data[i:i + length], "big", signed=signed)
|
|
for i in range(0, len(data), length)]
|
|
|
|
#
|
|
# Private helpers
|
|
#
|
|
|
|
def _check_is_bytes(data):
|
|
if not isinstance(data, bytes):
|
|
_err(f"'{data}' has type '{type(data).__name__}', expected 'bytes'")
|
|
|
|
def _check_length_positive(length):
|
|
if length < 1:
|
|
_err("'length' must be greater than zero, was " + str(length))
|
|
|
|
def _append_no_dup(lst, elm):
|
|
# Appends 'elm' to 'lst', but only if it isn't already in 'lst'. Lets us
|
|
# preserve order, which a set() doesn't.
|
|
|
|
if elm not in lst:
|
|
lst.append(elm)
|
|
|
|
def _decode_and_escape(b):
|
|
# Decodes the 'bytes' array 'b' as UTF-8 and backslash-escapes special
|
|
# characters
|
|
|
|
# Hacky but robust way to avoid double-escaping any '\' spit out by
|
|
# 'backslashreplace' bytes.translate() can't map to more than a single
|
|
# byte, but str.translate() can map to more than one character, so it's
|
|
# nice here. There's probably a nicer way to do this.
|
|
return (b.decode("utf-8", "surrogateescape")
|
|
.translate(_escape_table)
|
|
.encode("utf-8", "surrogateescape")
|
|
.decode("utf-8", "backslashreplace"))
|
|
|
|
def _root_and_path_to_node(cur, path, fullpath):
|
|
# Returns the node pointed at by 'path', relative to the Node 'cur'. For
|
|
# example, if 'cur' has path /foo/bar, and 'path' is "baz/qaz", then the
|
|
# node with path /foo/bar/baz/qaz is returned. 'fullpath' is the path as
|
|
# given in the .dts file, for error messages.
|
|
|
|
for component in path.split("/"):
|
|
# Collapse multiple / in a row, and allow a / at the end
|
|
if not component:
|
|
continue
|
|
|
|
if component not in cur.nodes:
|
|
_err(f"component '{component}' in path '{fullpath}' "
|
|
"does not exist")
|
|
|
|
cur = cur.nodes[component]
|
|
|
|
return cur
|
|
|
|
def _err(msg) -> NoReturn:
|
|
raise DTError(msg)
|
|
|
|
_escape_table = str.maketrans({
|
|
"\\": "\\\\",
|
|
'"': '\\"',
|
|
"\a": "\\a",
|
|
"\b": "\\b",
|
|
"\t": "\\t",
|
|
"\n": "\\n",
|
|
"\v": "\\v",
|
|
"\f": "\\f",
|
|
"\r": "\\r"})
|
|
|
|
# Lexer states
|
|
_DEFAULT = 0
|
|
_EXPECT_PROPNODENAME = 1
|
|
_EXPECT_BYTE = 2
|
|
|
|
_num_re = re.compile(r"(0[xX][0-9a-fA-F]+|[0-9]+)(?:ULL|UL|LL|U|L)?")
|
|
|
|
# A leading \ is allowed property and node names, probably to allow weird node
|
|
# names that would clash with other stuff
|
|
_propnodename_re = re.compile(r"\\?([a-zA-Z0-9,._+*#?@-]+)")
|
|
|
|
# Node names are more restrictive than property names.
|
|
_nodename_chars = set(string.ascii_letters + string.digits + ',._+-@')
|
|
|
|
# Misc. tokens that are tried after a property/node name. This is important, as
|
|
# there's overlap with the allowed characters in names.
|
|
_misc_re = re.compile(
|
|
"|".join(re.escape(pat) for pat in (
|
|
"==", "!=", "!", "=", ",", ";", "+", "-", "*", "/", "%", "~", "?", ":",
|
|
"^", "(", ")", "{", "}", "[", "]", "<<", "<=", "<", ">>", ">=", ">",
|
|
"||", "|", "&&", "&")))
|
|
|
|
_byte_re = re.compile(r"[0-9a-fA-F]{2}")
|
|
|
|
# Matches a backslash escape within a 'bytes' array. Captures the 'c' part of
|
|
# '\c', where c might be a single character or an octal/hex escape.
|
|
_unescape_re = re.compile(br'\\([0-7]{1,3}|x[0-9A-Fa-f]{1,2}|.)')
|
|
|
|
def _init_tokens():
|
|
# Builds a (<token 1>)|(<token 2>)|... regex and returns it. The
|
|
# way this is constructed makes the token's value as an int appear
|
|
# in match.lastindex after a match.
|
|
|
|
# Each pattern must have exactly one capturing group, which can capture any
|
|
# part of the pattern. This makes match.lastindex match the token type.
|
|
# _Token.val is based on the captured string.
|
|
token_spec = {
|
|
_T.INCLUDE: r'(/include/\s*"(?:[^\\"]|\\.)*")',
|
|
# #line directive or GCC linemarker
|
|
_T.LINE:
|
|
r'^#(?:line)?[ \t]+([0-9]+[ \t]+"(?:[^\\"]|\\.)*")(?:[ \t]+[0-9]+){0,4}',
|
|
|
|
_T.STRING: r'"((?:[^\\"]|\\.)*)"',
|
|
_T.DTS_V1: r"(/dts-v1/)",
|
|
_T.PLUGIN: r"(/plugin/)",
|
|
_T.MEMRESERVE: r"(/memreserve/)",
|
|
_T.BITS: r"(/bits/)",
|
|
_T.DEL_PROP: r"(/delete-property/)",
|
|
_T.DEL_NODE: r"(/delete-node/)",
|
|
_T.OMIT_IF_NO_REF: r"(/omit-if-no-ref/)",
|
|
_T.LABEL: r"([a-zA-Z_][a-zA-Z0-9_]*):",
|
|
_T.CHAR_LITERAL: r"'((?:[^\\']|\\.)*)'",
|
|
_T.REF: r"&([a-zA-Z_][a-zA-Z0-9_]*|{[a-zA-Z0-9,._+*#?@/-]*})",
|
|
_T.INCBIN: r"(/incbin/)",
|
|
# Whitespace, C comments, and C++ comments
|
|
_T.SKIP: r"(\s+|(?:/\*(?:.|\n)*?\*/)|//.*$)",
|
|
# Return a token for end-of-file so that the parsing code can
|
|
# always assume that there are more tokens when looking
|
|
# ahead. This simplifies things.
|
|
_T.EOF: r"(\Z)",
|
|
}
|
|
|
|
# MULTILINE is needed for C++ comments and #line directives
|
|
return re.compile("|".join(token_spec[tok_id] for tok_id in
|
|
range(1, _T.EOF + 1)),
|
|
re.MULTILINE | re.ASCII)
|
|
|
|
_token_re = _init_tokens()
|
|
|
|
_TYPE_TO_N_BYTES = {
|
|
_MarkerType.UINT8: 1,
|
|
_MarkerType.UINT16: 2,
|
|
_MarkerType.UINT32: 4,
|
|
_MarkerType.UINT64: 8,
|
|
}
|
|
|
|
_N_BYTES_TO_TYPE = {
|
|
1: _MarkerType.UINT8,
|
|
2: _MarkerType.UINT16,
|
|
4: _MarkerType.UINT32,
|
|
8: _MarkerType.UINT64,
|
|
}
|
|
|
|
_N_BYTES_TO_START_STR = {
|
|
1: " [",
|
|
2: " /bits/ 16 <",
|
|
4: " <",
|
|
8: " /bits/ 64 <",
|
|
}
|
|
|
|
_N_BYTES_TO_END_STR = {
|
|
1: " ]",
|
|
2: " >",
|
|
4: " >",
|
|
8: " >",
|
|
}
|