Network Working Group G. Trewitt Request for Comments: 1076 Stanford University Obsoletes: RFC 1023 C. Partridge BBN/NNSC November 1988 HEMS Monitoring and Control Language TABLE OF CONTENTS 1. Status of This Memo 1 Introduction 2 2. Overview and Scope 2 3. Overview of Query Processor Operation 4 4. Encoding of Queries and Responses 5 4.1 Notation Used in This Proposal 5 5. Data Organization 6 5.1 Example Data Tree 7 5.2 Arrays 8 6. Components of a Query 9 7. Reply to a Query 10 8. Query Language 12 8.1 Moving Around in the Data Tree 14 8.2 Retrieving Data 15 8.3 Data Attributes 16 8.4 Examining Memory 18 8.5 Control Operations: Modifying the Data Tree 19 8.6 Associative Data Access: Filters 21 8.7 Terminating a Query 26 9. Extending the Set of Values 27 10. Authorization 27 11. Errors 28 I. ASN.1 Descriptions of Query Language Components 29 I.1 Operation Codes 30 I.2 Error Returns 31 I.3 Filters 33 I.4 Attributes 34 I.5 VendorSpecific 36 II. Implementation Hints 36 III. Obtaining a Copy of the ASN.1 Specification 42 1. STATUS OF THIS MEMO This RFC specifies a query language for monitoring and control of network entities. This RFC supercedes RFC-1023, extending the query language and providing more discussion of the underlying issues.
This language is a component of the High-Level Entity Monitoring System (HEMS) described in RFC-1021 and RFC-1022. Readers may wish to consult these RFCs when reading this memo. RFC-1024 contains detailed assignments of numbers and structures used in this system. Portions of RFC-1024 that define query language structures are superceded by definitions in this memo. This memo assumes a knowledge of the ISO data encoding standard, ASN.1. Distribution of this memo is unlimited. INTRODUCTION This RFC specifies the design of a general-purpose, yet efficient, monitoring and control language for managing network entities. The data in the entity is modeled as a hierarchy and specific items are named by giving the path from the root of the tree. Most items are read-only, but some can be "set" in order to perform control operations. Both requests and responses are represented using the ISO ASN.1 data encoding rules. 2. OVERVIEW AND SCOPE The basic model of monitoring and control used in this memo is that a query is sent to a monitored entity and the entity sends back a response. The term query is used in the database sense -- it may request information, modify data, or both. We will use gateway- oriented examples, but it should be understood that this query- response mechanism is applicable to any IP entity. In particular, there is no notion of an interactive "conversation" as in SMTP [RFC-821] or FTP [RFC-959]. A query is a complete request that stands on its own and elicits a complete response. In order to design the query language, we had to define a model for the data to be retrieved by the queries, which required some understanding of and assumptions to be made about the data. We ended up with a fairly flexible data model, which places few limits on the type or size of the data. Wherever possible, we give motivations for the design decisions or assumptions that led to particular features or definitions. Some of the important global considerations and assumptions are: - The query processor should place as little computational burden on the monitored entity as possible. - It should not be necessary for a monitored entity to store the complete query. Nothing in the query language should
preclude an implementation from being able to process the query on the fly, producing portions of the response while the query is still being read and parsed. There may be other constraints that require large amounts of data to be buffered, but the query language design must not be one. - It is assumed that there is some mechanism to transport a sequence of octets to a query processor within the monitored entity and that there is some mechanism to return a sequence of octets to the entity making the query. In HEMS, this is provided by HEMP and its underlying transport layer. The query language design is independent of these details, however, and could be grafted onto some other protocol. - The data model must provide organization for the data, so that it can be conveniently named. - Much of the data to be monitored will be contained in tables. Some tables may contain other tables. The query language should be able to deal with such tables. - We don't provide capabilities for data reduction in the query language. We will provide for data selection, for example, only retrieving certain table entries, but we will not provide general facilities for processing data, such as computing averages. - Because one monitoring center may be querying many (possibly hetrogenous) hosts, it must be possible to write generic queries that can be sent to all hosts, and have the query elicit as much information as is available from each host. i.e., queries must not be aborted just because they requested non-existent data. There were some assumptions that we specifically did not make: - It is up to the implementation to choose what degree of concurrency will be allowed when processing queries. By locking only portions of the database, it should be possible to achieve good concurrency while still preventing deadlock. - This specification makes no statement about the use of the "definite" and "indefinite" length forms in ASN.1. There is currently some debate about this usage in the ISO community; implementors should note the recommendations in the ASN.1 specification.
Other RFCs associated with HEMS are: RFC-1021 Overview; RFC-1022 Transport protocol and message encapsulation; RFC-1024 Precise data definitions. The rest of this report is organized as follows: Section 3 Gives a brief overview of the data model and the operation of the query processor. Section 4 Describes the encoding used for queries and responses, and the notation used to represent them in this report. Section 5 Describes how the data is organized in the monitored entity, and the view provided of it by the query processor. Section 6 Describes the basic data types that may be given to the query processor as input. Section 7 Describes how a reply to a query is organized. Section 8 Describes the operations available in the query language. Section 9 Describes how the set of data in the tree may be extended. Section 10 Describes how authorization issues affect the execution of a query. Section 11 Describes how errors are reported, and their effect on the processing of the query. Appendix I Gives precise ASN.1 definitions of the data types used by the query processor. Appendix II Gives extensive implementation hints for the core of the query processor. 3. OVERVIEW OF QUERY PROCESSOR OPERATION In this section, we give an overview of the operation of the query processor, to provide a framework for the later sections. The query language models the manageable data as a tree, with each
branch representing a different aspect of the entity, such as different layers of protocols. Subtrees are further divided to provide additional structure to the data. The leaves of the tree contain the actual data. Given this data representation, the task of the query processor is to traverse this tree and retrieve (or modify) data specified in a query. A query consists of instructions to move around in the tree and to retrieve (or modify) named data. The result of a query is an exact image of the parts of the tree that the query processor visited. The query processor is very simple -- it only understands eight commands, most of which share the same structure. It is helpful to think of the query processor as an automaton that walks around in the tree, directed by commands in the query. As it moves around, it copies the tree structure it traverses to the query result. Data that is requested by the query is copied into the result as well. Data that is changed by a query is copied into the result after the modification is made. 4. ENCODING OF QUERIES AND RESPONSES Both queries and responses are encoded using the representation defined in ISO Standard ASN.1 (Abstract Syntax Notation 1). ASN.1 represents data as sequences of <tag,length,contents> triples that are encoded as a stream of octets. The data tuples may be recursively nested to represent structured data such as arrays or records. For a full description, see the ISO standards IS 8824 and IS 8825. See appendix for information about obtaining these documents. 4.1 Notation Used in This Proposal The notation used in this memo is similar to that used in ASN.1, but less formal, smaller, and (hopefully) easier to read. We will refer to a <tag,length,contents> tuple as a "data object". In this RFC, we will not be concerned with the details of the object lengths. They exist in the actual ASN.1 encoding, but will be omitted in the examples here. Data objects that have no internal ASN.1 structure such as integer or octet string are referred to as "simple types" or "simple objects". Objects which are constructed out of other ASN.1 data objects will be referred to as "composite types" or "composite objects".
The notation ID(value) represents a simple object whose tag is "ID" with the given value. A composite object is represented as ID{ ... contents ... } where contents is a sequence of data objects. The contents may include both simple and structured types, so the structure is fully recursive. The difference between simple and composite types is close to the meaning of the "constructor" bit in ASN.1. For the uses here, the distinction is made based upon the semantics of the data, not the representation. Therefore, even though an OctetString can be represented in ASN.1 using either constructed or non-constructed forms, it is conceptually a simple type, with no internal structure, and will always be written as ID("some arbitrary string") in this RFC. There are situations where it is necessary to specify a type but give no value, such as when referring to the name of the data. In this situation, the same notation is used, but with the value omitted: ID or ID() or ID{} Such objects have zero length and no contents. The latter two forms are used when a distinction is being made between simple and composite data, but the difference is just notation -- the representation is the same. ASN.1 distinguishes between four "classes" of tags: universal, application-specific, context-dependent, and reserved. HEMS and this query language use the first three. Universal tags are assigned in the ASN.1 standard and its addendums for common types, and are understood by any application using ASN.1. Application-specific tags are limited in scope to a particular application. These are used for "well-known" identifiers that must be recognizable in any context, such as derived data types. Finally, context-dependent tags are used for objects whose meaning is dependent upon where they are encountered. Most tags that identify data are context-dependent. 5. DATA ORGANIZATION Data in a monitored entity is modeled as a hierarchy. Implementations are not required to organize the data internally as a hierarchy, but they must provide this view of the data through the query language. A hierarchy offers useful structure for the following operations:
Organization A hierarchy allows related data to be grouped together in a natural way. Naming The name of a piece of data is just the path from the root to the data of interest. Mapping onto ASN.1 ASN.1 can easily represent a hierarchy by using a "constructor" type as an envelope for an entire subtree. Efficient Representation Hierarchical structures are compact and can be traversed quickly. Safe Locking If it is necessary to lock part of the hierarchy (for example, when doing an update), locking an entire subtree can be done efficiently and safely, with no danger of deadlock. We will use the term "data tree" to refer to this entire structure. Note that this internal model is completely independent of the external ASN.1 representation -- any other suitable representation would do. For the sake of efficiency, we do make a one-to-one mapping between ASN.1 tags and the (internal) names of the nodes. The same could be done for any other external representation. Each node in the hierarchy must have names for its component parts. Although we would normally think of names as being ASCII strings such as "input errors", the actual name is just an ASN.1 tag. Such names are small integers (typically, less than 30) and so can easily be mapped by the monitored entity onto its internal representation. We use the term "dictionary" to mean an internal node in the hierarchy. Leaf nodes contain the actual data. A dictionary may contain both leaf nodes and other dictionaries. 5.1 Example Data Tree Here is a possible organization of the hierarchy in an entity that has several network interfaces and does IP routing. The exact organization of data in entities is specified in RFC-1024. This skeletal data tree will be used throughout this RFC in query examples. System { name -- host name clock-msec -- msec since boot
interfaces -- # of interfaces memory } Interfaces { -- one per interface InterfaceData{ address, mtu, netMask, ARP{...}, ... } InterfaceData{ address, mtu, netMask, ARP{...}, ... } : } IPRouting { Entry{ ip-addr, interface, cost, ... } Entry{ ip-addr, interface, cost, ... } : } There are three top-level dictionaries in this hierarchy (System, Interfaces, and IPRouting) and three other dictionary types (InterfaceData, Entry, and ARP), each with multiple instances. The "name" of the clock in this entity would be: system{ clock-msec } and the name of a routing table entry's IP address would be: IPRouting{ Entry{ ip-addr } }. More than one piece of data can be named by a single ASN.1 object. The entire collection of system information is named by: system and the name of a routing table's IP address and cost would be: IPRouting{ Entry{ ip-addr, cost } }. 5.2 Arrays There is one sub-type of a dictionary that is used as the basis for tables of objects with identical types. We call these dictionaries arrays. In the example above, the dictionaries for interfaces, routing tables, and ARP tables are all arrays. In the examples above, the "ip-addr" and "cost" fields are named. In fact, these names refer to the field values for ALL of the routing table entries -- the name doesn't (and can't) specify which routing table entry is intended. This ambiguity is a problem wherever data is organized in tables. If there was a meaningful index for such tables (e.g., "routing table entry #1"), there would be no problem. Unfortunately, there usually isn't such an index. The solution to this problem requires that the data be accessed on the basis of some of its content. Filters, discussed in section 8.6, provide this mechanism. The primary difference between arrays and plain dictionaries is that
arrays may contain only one type of item, while dictionaries, in general, will contain many different types of items. For example, the dictionary IPRouting (which is an array) will contain only items of type Entry. The fact that these objects are viewed externally as arrays or tables does not mean that they are represented in an implementation as linear lists of objects. Any collection of same-typed objects is viewed as an array, even though it might be stored internally in some other format, for example, as a hash table. 6. COMPONENTS OF A QUERY A HEMS query consists of a sequence of ASN.1 objects, interpreted by a simple stack-based interpreter. [Although we define the query language in terms of the operations of a stack machine, the language does not require an implementation to use a stack machine. This is a well-understood model, and is easy to implement.] One ASN.1 tag is reserved for operation codes; all other tags indicate data that will eventually be used by an operation. These objects are pushed onto the stack when received. Opcodes are immediately executed and may remove or add items to the stack. Because ASN.1 itself provides tags, very little needs to be done to the incoming ASN.1 objects to make them suitable for use by the query interpreter. Each ASN.1 object in a query will fit into one of the following categories: Opcode An opcode tells the query interpreter to perform an action. They are described in detail in section 8. Opcodes are represented by an application-specific type whose value determines the operation. Template These are objects that name one or more items in the data tree. Named items may be either simple items (leaf nodes) or entire dictionaries, in which case the entire subtree "underneath" the dictionary is understood. Templates are used to select specific data to be retrieved from the data tree. A template may be either simple or structured, depending upon what it is naming. A template only names the data -- there are no values contained in it. Therefore the leaf objects in a template will all have a length of zero. Examples of very simple templates are: name() or System{} Each of these is just one ASN.1 data object, with zero length. The first names a single data item in the "System"
dictionary (and must appear in that context), and the second names the entire "System" dictionary. A more complex template such as: Interfaces{ InterfaceData{ address, netMask, ARP } } names two simple data items and a dictionary, iterated over all occurrences of "InterfaceData" within the Interfaces array. Path A path is a special case of a template that names only a single node in the tree. It specifies a path down into the dictionary tree and names exactly one node in the dictionary tree. Value These are used to give data values when needed in a query, for example, when changing a value in the data tree. A value can be thought of as either a filled-in template or as the ASN.1 representation some part of the data tree. Filter A boolean expression that can be executed in the context of a particular dictionary that is used to select or not select items in the dictionary. The expressions consist of the primitives "equal", "greater-or-equal", "less-or-equal", and "present" possibly joined by "and", "or", and "not". (See section 8.6.) Values, Paths, and Templates usually have names in the context- dependent class, except for a few special cases, which are in the application-specific class. 7. REPLY TO A QUERY The data returned to the monitoring entity is a sequence of ASN.1 data items. Conceptually, the reply is a subset of the data tree, where the query selects which portions are to be included. This is exactly true for data retrieval requests, and essentially true for data modification requests -- the reply contains the data after it has been modified. The key point is that the data in a reply represents the state of the data tree immediately after the query was executed. The sequence of the data is determined by the sequence of query language operations and the order of data items within Templates and Values given as input to these operations. If a query requests data from two of the top-level dictionaries in the data tree, by giving two templates such as: System{ name, interfaces } Interfaces{
InterfaceData { address, netMask, mtu } } then the response will consist of two ASN.1 data objects, as follows: System { name("system name"), interfaces(2) } Interfaces { InterfaceData { address(36.8.0.1), netMask(FFFF0000), mtu(1500) } InterfaceData { address(10.1.0.1), mtu(1008), netMask(FF000000) } } With few exceptions, each of the data items in the hierarchy is named in the context-specific ASN.1 type space. Because of this, the returned objects must be fully qualified. For example, the name of the entity must always be returned encapsulated inside an ASN.1 object for "System". If it were not, there would be no way to tell if the object that was returned was "name" inside the "System" dictionary or "address" inside the "interfaces" dictionary (assuming in this case that "name" and "address" were assigned the same integer as their ASN.1 tags). Having fully-qualified data simplifies decoding of the data at the receiving end and allows the tags to be locally chosen. Definitions for tags within routing tables won't conflict with definitions for tags within interfaces. Therefore, the people doing the name assignments are less constrained. In addition, most of the identifiers will be fairly small integers, which is an advantage because ASN.1 can fit tag numbers up to 30 in a one-octet tag field. Larger numbers require a second octet. If data is requested that doesn't exist, either because the tag is not defined, or because an implementation doesn't provide that data (such as when the data is optional), the response will contain an ASN.1 object that is empty. The tag will be the same as in the query, and the object will have a length of zero. The same response is given if the requested data does exist, but the invoker of the query does not have authorization to access it. See section 10 for more discussion of authorization mechanisms.
This allows completely generic queries to be composed without regard to whether the data is defined or implemented at all of the entities that will receive the query. All of the available data will be returned, without generating errors that might otherwise terminate the processing of the query. 8. QUERY LANGUAGE The query language is designed to be expressive enough to write useful queries with, yet simple enough to be easy to implement. The query processor should be as simple and fast as possible, in order to avoid placing a burden on the monitored entity, which may be a critical node such as a gateway. Although queries are formed in a flexible way using what we term a "language", this is not a programming language. There are operations that operate on data, but most other features of programming languages are not present. In particular: - Programs are not stored in the query processor. - The only form of temporary storage is a stack, of limited depth. - There are no subroutines. - There are no explicit control structures defined in the language. The central element of the language is the stack. It may contain templates, (and therefore paths), values, and filters taken from the query. In addition, it can contain dictionaries (and therefore arrays) from the data tree. At the beginning of a query, it contains one item, the root dictionary. The overall operation consists of reading ASN.1 objects from the input stream. All objects that aren't opcodes are pushed onto the stack as soon as they are read. Each opcode is executed immediately and may remove items from the stack, may generate ASN.1 objects and send them to the output stream, and may leave items on the stack. Because each input object is dealt with immediately, portions of the response may be generated while the query is still being received. In the descriptions below, operator names are in capital letters, preceded by the arguments used from the stack and followed by results left on the stack. For example:
OP a b OP a t means that the OP operator takes <a> and <b> off of the stack and leaves <t> on the stack. Most of the operators in the query language leave the first operand (<a> in this example) on the stack for future use. If both <a> and <b> were received as part of the query (as opposed to being calculated by previous operations), then this part of the query would have consisted of the sequence: <a> <b> OP So, like other stack-based languages, the arguments and operators must be presented in postfix order, with an operator following its operands. Here is a summary of all of the operators defined in the query language. Most of the operators can take several different sets of operands and behave differently based upon the operand types. Details and examples are given later. BEGIN dict1 path BEGIN dict1 dict array path filter BEGIN array dict Move down in the data tree, establishing a context for future operations. END dict END -- Undo the most recent BEGIN. GET dict GET dict dict template GET dict array template filter GET array Retrieve data from the data tree. GET-ATTRIBUTES dict GET-ATTRIBUTES dict dict template GET-ATTRIBUTES dict array template filter GET-ATTRIBUTES array Retrieve attribute information about data in the data tree. GET-RANGE dict path start length GET-RANGE dict Retrieve a subrange of an OctetString. Used for reading memory. SET dict value SET dict array value filter SET array Change values in the data tree, possibly performing control functions.
CREATE array value CREATE dict Create new table entries. DELETE array filter DELETE array Delete table entries. These operators are defined so that it is impossible to generate an invalid query response. Since a response is supposed to be a snapshot of a portion (or portions) of the data tree, it is important that only data that is actually in the tree be put in the response. Two features of the language help guarantee this: - Data is put in the response directly from the tree (by GET-*). Data does not go from the tree to the stack and then into the response. - Dictionaries on the stack are all derived from the initial, root dictionary. The operations that manipulate dictionaries (BEGIN and END) also update the response with the new location in the tree. 8.1 Moving Around in the Data Tree The initial point of reference in the data tree is the root. That is, operators name data starting at the root of the tree. It is useful to be able to move to some other dictionary in the tree and then name data from that point. The BEGIN operator moves down in the tree and END undoes the last unmatched BEGIN. BEGIN is used for two purposes: - By moving to a dictionary closer to the data of interest, the name of the data can be shorter than if the full name (from the root) were given. - It is used to establish a context for filtered operations to operate in. Filters are discussed in section 8.6. BEGIN dict1 path BEGIN dict1 dict Follow <path> down the dictionary starting from <dict1>. Push the final dictionary named by <path> onto the stack. <path> must name a dictionary (not a leaf node). At the same time, produce the beginning octets of an ASN.1 object corresponding to the new dictionary. It is up to the implementation to choose between using the "indefinite length" representation or the "definite length" form and going back and filling the length in later.
END dict END -- Pop <dict> off of the stack and terminate the open ASN.1 object(s) started by the matching BEGIN. Must be paired with a BEGIN. If an END operation pops the root dictionary off of the stack, the query is terminated. <path> must point to a regular dictionary. If any part of it refers to a non-existent node, if it points to a leaf node, or if it refers to a node inside an array-type dictionary, then it is in error, and the query is terminated immediately. An additional form of BEGIN, which takes a filter argument, is described later. 8.2 Retrieving Data The basic model that all of the data retrieval operations follow is that they take a template and fill in the leaf nodes of the template with the appropriate data values. GET dict template GET dict Emit an ASN.1 object with the same "shape" as the given template, except with values filled in for each node. The first ASN.1 tag of <template> should refer to an object in <dict>. If a dictionary tag is supplied anywhere in <template>, the entire dictionary contents are emitted to the response. Any items in the template that are not in <dictionary> (or its components) are represented as objects with a length of zero. dict GET dict If there is no template, get all of the items in the dictionary. This is equivalent to providing a template that lists all of the items in the dictionary. An additional form of GET, which takes a filter argument, is described later. Here is an example of using the BEGIN operator to move down the data tree to the TCP dictionary and then using the GET operator to retrieve 5 data values from the TCP Stats dictionary: IPTransport{ TCP } BEGIN Stats{ octetsIn, octetsOut, inputPkts, outputPkts, badtag } GET END
This might return: IPTransport{ TCP Stats{ octetsIn(13255), octetsOut(82323), inputPkts(9213), outputPkts(12425), badtag() } } "badtag" is a tag value that is undefined. No value is returned for it, indicating that there is no data value associated with it. 8.3 Data Attributes Although ASN.1 "self-describes" the structure and syntax of the data, it gives no information about what the data means. For example, by looking at the raw data, it is possible to tell that an item is of type [context 5] and is 4 octets long. That does not tell how to interpret the data (is this an integer, an IP address, or a 4- character string?) or what the data means (IP address of what?). Even if the data were "tagged", in ASN.1 parlance, that would only give the base type (e.g., IP-address or counter) and not the meaning. Most of the time, this information will come from RFC-1024, which defines the ASN.1 tags and their precise meaning. When extensions have been made, it may not be possible to get documentation on the extensions. (Extensions are discussed in section 9.) The GET-ATTRIBUTES operator is similar to the GET operator, but returns a set of attributes describing the data rather than the data itself. This information is intended to be sufficient to let a human understand the meaning of the data and to let a sophisticated application treat the data appropriately. Such an application could use the attribute information to format the data on a display and decide whether it is appropriate to subtract one sample from another. Some of the attributes are textual descriptions to help a human understand the nature of the data and provide meaningful labels for it. Extensive descriptions of standard data are optional, since they are defined in RFC-1024. Complete descriptions of extensions must be available, even if they are documented in a user's manual. Network firefighters may not have a current manual handy when the network is broken. The format of the attributes is not as simple as the format of the data itself. It isn't possible to use the data's tag, since that would look exactly like the data itself. The format is: Attributes ::= [APPLICATION 3] IMPLICIT SEQUENCE { tagASN1 [0] IMPLICIT INTEGER,
valueFormat [1] IMPLICIT INTEGER, longDesc [2] IMPLICIT IA5String OPTIONAL, shortDesc [3] IMPLICIT IA5String OPTIONAL, unitsDesc [4] IMPLICIT IA5String OPTIONAL, precision [5] IMPLICIT INTEGER OPTIONAL, properties [6] IMPLICIT BITSTRING OPTIONAL, valueSet [7] IMPLICIT SET OF valueDesc OPTIONAL } The GET-ATTRIBUTES operator is similar to the GET operator. The major difference is that dictionaries named in the template do not elicit data for the entire subtree. GET-ATTRIBUTES dict template GET-ATTRIBUTES dict Emit a single ASN.1 Attributes object for each of the objects named in <template>. For each of these, the tagASN1 field will be set to the corresponding tag from the template. The rest of the fields are set as appropriate for the data object. Any items in the template that are not in <dictionary> (or its components) elicit an Attributes object with a valueFormat of NULL, and no other descriptive information. or dict GET-ATTRIBUTES dict If there is no template, emit Attribute objects for all of the items in the dictionary. This is equivalent to providing a template that lists all of the items in the dictionary. This allows a complete list of a dictionary's contents to be obtained. An additional form of GET-ATTRIBUTES, which takes a filter argument, is described later. Here is an example of using the GET-ATTRIBUTES operator to request attributes for three objects in the System dictionary: System{ name, badtag, clock-msec } GET-ATTRIBUTES "badtag" is some unknown tag. The result might be: System{ Attributes{ tagASN1(name), valueFormat(IA5String), longDesc("The primary hostname."),
shortDesc("hostname") }, Attributes{ tagASN1(badtag), valueFormat(NULL) } Attributes{ tagASN1(clock-msec), valueFormat(Integer), longDesc("milliseconds since boot"), shortDesc("uptime"), unitsDesc("ms") precision(4294967296), properties(1) } Note that in this example "name" and "clock-msec" are integer values for the ASN.1 tags for the two data items. "badtag" is an integer value that has no corresponding name in this context. There will always be exactly as many Attributes items in the result as there are objects in the template. Attributes objects for items which do not exist in the entity will have a valueFormat of NULL and none of the optional elements will appear. [ A much cleaner method would be to store the attributes as sub-components of the data item of interest. For example, requesting System{ clock-msec } GET would normally just get the value of the data. Asking for an additional layer down the tree would now get its attributes: System{ clock-msec{ shortDesc, unitsDesc } GET would get the named attributes. (The attributes would be named with application-specific tags.) Unfortunately, ASN.1 doesn't provide a notation to describe this type of organization. So, we're stuck with the GET-ATTRIBUTES operator. However, if a cleaner organization were possible, this decision would have been made differently. ] 8.4 Examining Memory Even with the ability to symbolically access all of this information in an entity, there will still be times when it is necessary to get to very low levels and examine memory, as in remote debugging operations. The building blocks outlined so far can easily be extended to allow memory to be examined. Memory is modeled as an ordinary object in the data tree, with an ASN.1 representation of OctetString. Because of the variety of addressing architectures in existence, the conversion from the
internal memory model to OctetString is very machine-dependent. The only simple case is for byte-addressed machines with 8 bits per byte. Each address space in an entity is represented by one "memory" data item. In a one-address-space situation, this dictionary will probably be in "System" dictionary. If each process has its own address space, then one "memory" item might exist for each process. Again, this is very machine-dependent. Although the GET-RANGE operator is provided primarily for the purpose of retrieving the contents of memory, it can be used on any object whose base type is OctetString. GET-RANGE dict path start length GET-RANGE dict Get <length> elements of the OctetString, starting at <start>. <start> and <length> are both ASN.1 INTEGER type. <path>, starting from <dict>, must specify a node representing memory, or some other OctetString. The returned data may not be <length> octets long, since it may take more than one octet to represent one memory location. Memory items in the data tree are special in that they will not automatically be returned when the entire contents of a dictionary are requested. e.g., If memory is part of the "System" dictionary, then the query System GET will emit the entire contents of the System dictionary, but not the memory item. 8.5 Control Operations: Modifying the Data Tree All of the operators defined so far only allow data in an entity to be retrieved. By replacing the template argument used in the GET operators with a value, data in the entity can be changed. Very few items in the data tree can be changed; those that can are noted in RFC-1024. Values in the data tree can modified in order to change configuration parameters, patch routing tables, etc. Control functions, such as bringing an interface "down" or "up", do not usually map directly to changing a value. In such cases, an item in the tree can be defined to have arbitrary side-effects when a value is assigned to it. Control operations then consist of "setting" this item to an appropriate command code. Reading the value of such an item might return the current status. Again, details of such data tree items are given in RFC-1024.
This "virtual command-and-status register" model is very powerful, and can be extended by an implementation to provide whatever controls are needed. It has the advantage that the control function is associated with the controlled object in the data tree. In addition, no additional language features are required to support control functions, and the same operations used to locate data for retrieval are used to describe what is being controlled. For all of the control and data modification operations, the fill- in-the-blank model used for data retrieval is extended: the response to an operation is the affected part of the data tree, after the operation has been executed. Therefore, for normal execution, SET and CREATE will return the object given as an argument, and DELETE will return nothing (because the affected portion was deleted). SET dict value SET dict Set the value(s) of data in the entity to the value(s) given in <value>. The result will be the value of the data after the SET. Attempting to set a non-settable item will not produce an error, but will yield a result in the reply different from what was sent. CREATE array value CREATE dict Insert a new entry into <array>. Depending upon the context, there may be severe restrictions about what constitutes a valid <value>. The result will be the actual item added to the <array>. Note that only one item can be added per CREATE operation. DELETE array filter DELETE array Delete the entry(s) in <array> that match <filter>. Filters are described later in this document. Normally, no data items will be produced in the response, but if any of the items that matched the filter could not be deleted, they will be returned in the response. An additional form of SET, which takes a filter argument, is described later. Here is an example of attempting to use SET to change the number of interfaces in an entity: System{ interfaces(5) } SET Since that is not a settable parameter, the result would be: System{ interfaces(2) } giving the old value. Here is an example of how CREATE would be used to add a routing table entry for net for 128.89.0.0.
IPRouting BEGIN -- get dictionary Entries{ DestAddr(128.89.0.0), ... } -- entry to insert CREATE END -- finished with dict The result would be what was added: IPRouting{ Entries{ DestAddr(128.89.0.0), ... } } The results in the response of these operators is consistent of the global model of the response: it contains a subset of the data in the tree immediately after the query is executed. Note that CREATE and DELETE only operate on arrays, and then only on arrays that are specifically intended for it. For example, it is quite reasonable to add and remove entries from routing tables or ARP tables, both of which are arrays. However, it doesn't make sense to add or remove entries in the "Interfaces" dictionary, since the contents of that array is dictated by the hardware. For each array in the data tree, RFC-1024 indicates whether CREATE and DELETE are valid. CREATE and DELETE are always invalid in non-array contexts. If DELETE were allowed on monitored data, then the deleted data would become unmonitorable to the entire world. Conversely, if it were possible to CREATE arbitrary dictionary entries, there would be no way to give such entries any meaning. Even with the data in place, there is nothing that would couple the data to the operation of the monitored entity. [Creation and deletion would also add considerable complication to an implementation, because without them, all of the data structures that represent the data tree are essentially static, with the exception of dynamic tables such as the ones mentioned, which already have mechanisms in place for adding and removing entries.] 8.6 Associative Data Access: Filters One problem that has not been dealt with was alluded to earlier: When dealing with array data, how do you specify one or more entries based upon some value in the array entries? Consider the situation where there are several interfaces. The data might be organized as: Interfaces { -- one per interface InterfaceData { address, mtu, netMask, ARP{...}, ... } InterfaceData { address, mtu, netMask, ARP{...}, ... } : } If you only want information about one interface (perhaps because
there is an enormous amount of data about each), then you have to have some way to name it. One possibility would be to just number the interfaces and refer to the desired interface as InterfaceData(3) for the third one. But this is not sufficient, because interface numbers may change over time, perhaps from one reboot to the next. It is even worse when dealing with arrays with many elements, such as routing tables, TCP connections, etc. Large, changing arrays are probably the more common case, in fact. Because of the lack of utility of indexing in this context, there is no general mechanism provided in the language for indexing. A better scheme is to select objects based upon some value contained in them, such as the IP address. The query language uses filters to select specific table entries that an operator will operate on. The operators BEGIN, GET, GET-ATTRIBUTES, SET, and DELETE can take a filter argument that restricts their operation to entries that match the filter. A filter is a boolean expression that is executed for each element in an array. If an array entry "matches" the filter (i.e., if the filter produces a "true" result), then it is used by the operation. A filter expression is very restricted: it can only compare data contained in the array element and the comparisons are only against constants. Comparisons may be connected by AND, OR and NOT operators. The ASN.1 definition of a filter is: Filter ::= [APPLICATION 2] CHOICE { present [0] DataPath, equal [1] DataValue, greaterOrEqual [2] DataValue, lessOrEqual [3] DataValue, and [4] SEQUENCE OF Filter, or [5] SEQUENCE OF Filter, not [6] Filter } DataPath ::= ANY -- Path with no value DataValue ::= ANY -- Single data value This definition is similar to the filters used in the ISO monitoring protocol (CMIP) and was derived from that specification.
"DataPath" is the name of a single data item; "DataValue" is the value of a single data item. The three comparisons are all of the form "data OP constant", where "data" is the value from the tree, "constant" is the value from the filter expression, and "OP" is one of equal, greater-than-or-equal, or less-than-or-equal. The last operation, "present", tests to see if the named item exists in the data tree. By its nature, it requires no value, so only a path needs to be given. Here is an example of a filter that matches an Interface whose IP address is 10.1.0.1: Filter{ equal{ address(10.0.0.51) } } Note that the name of the data to be compared is relative to the "InterfaceData" dictionary. Each operator, when given a filter argument, takes an array (dictionary containing only one type of item) as its first argument. In the current example, this would be "Interfaces". The items in it are all of type "InterfaceData". This tag is referred to as the "iteration tag". Before a filtered operation is used, BEGIN must be used to put the array (dictionary) on top of the stack, to establish it as the context that the filter iterates over. The general operation of a filtered operation is then: 1. Iterate over the items in the array. 2. For each element in the array, execute the filter. 3. If the filter succeeds, do the requested operation (GET/SET/etc.) on the matched element, using the template/value/path as input to the operation. At this point, the execution of the operation is the same as in the non-filtered case. This is a model of operation; actual implementations may take advantage of whatever lookup techniques are available for the particular table (array) involved. Therefore, there are three inputs to a filtered operation: 1. The "current dictionary" on the stack. This is the array-type dictionary to be searched, set by an earlier BEGIN. 2. A filter, to test each item in the array. Each path or value mentioned in the filter must be named in the context
of an item in the array, as if it was the current dictionary. For example, in the case where a filtered operation iterates over the set of "InterfaceData" items in the "Interfaces" array, each value or path in the filter should name an item in the "InterfaceData" dictionary, such as "address". 3. A template, path, or value associated with the operation to be performed. The leading ASN.1 tag in this must match the iteration tag. In the current example where the filter is searching the "Interfaces" dictionary, the first tag in the template/tag/value must be "InterfaceData". The operators which take filters as arguments are: BEGIN array path filter BEGIN array dict Find a dictionary in <array> that matches <filter>. Use that as the starting point for <path> and push the dictionary corresponding to <path> onto the stack. If more than one dictionary matches <filter>, then any of the matches may be used. This specification does not state how the choice is made. At least one dictionary must match; it is an error if there are no matches. (Perhaps it should be an error for there to be multiple matches; actual experience is needed to decide.) GET array template filter GET array For each item in <array> that matches <filter>, fill in the template with values from the data tree and emit the result. The first tag of <template> must be equal to the iteration tag. Selected parts of matched items are emitted based upon <template>, just as in a non-filtered GET operation. GET-ATTRIBUTES array template filter GET-ATTRIBUTES array Same as GET, except emit attributes rather than data values. SET array value filter SET array Same as GET, except set the values in <value> rather than retrieving values. Several values in the data tree will be changed if the filter matches more than one item in the array. DELETE array filter DELETE array Delete the entry(s) in <array> that match <filter>.
Notes about filter execution: - Expressions are executed by inorder tree traversal. - Since the filter operations are all GETs and comparisons, there are no side-effects to filter execution, so an implementation is free to execute only as much of the filter as required to produce a result (e.g., don't execute the rest of an AND if the first comparison turns out to be false). - It is not an error for a filter to test a data item that isn't in the data tree. In this situation, the comparison just fails (is false). This means that filters don't need to test for the existence of optional data before attempting to compare it. Here is an example of how filtering would be used to obtain the input and output packet counts for the interface with IP address 10.0.0.51. Interfaces BEGIN -- dictionary InterfaceData{ pktsIn, pktsOut } -- template Filter{ equal{ address(10.0.0.51) } } GET END -- finished with dict The returned value would be something like: Interfaces{ -- BEGIN InterfaceData{ pktsIn(1345134), pktsOut(1023729) } -- GET } -- END The annotations indicate which part of the response is generated by the different operators in the query. Here is an example of accessing a table contained within some other table. Suppose we want to get at the ARP table for the interface with IP address 36.8.0.1 and retrieve the entire ARP entry for the host with IP address 36.8.0.23. In order to retrieve a single entry in the ARP table (using a filtered GET), a BEGIN must be used to get down to the ARP table. Since the ARP table is contained within the Interfaces dictionary (an array), a filtered BEGIN must be used. Interfaces BEGIN -- dictionary InterfaceData{ ARP } -- path Filter{ equal{ address(36.8.0.1) } } -- filter BEGIN -- filtered BEGIN
-- Now in ARP table for 38.0.0.1; get entry for 38.8.0.23. addrMap -- whole entry Filter{ equal{ ipAddr(36.8.0.23) } } -- filter GET -- filtered GET END END The result would be: Interfaces{ -- first BEGIN InterfaceData{ ARP{ -- second BEGIN addrMap{ ipAddr(36.8.0.23), physAddr(..) } -- from GET } } -- first END } -- second END Note which parts of the output are generated by different parts of the query. Here is an example of how the SET operator would be used to shut down the interface with ip-address 10.0.0.51 by changing its status to "down". Interfaces BEGIN -- get dictionary Interface{ Status(down) } -- value to set Filter{ equal{ IP-addr(10.0.0.51) } } SET END If the SET is successful, the result would be: Interfaces{ -- BEGIN Interface{ Status(down) } -- from SET } -- END 8.7 Terminating a Query A query is implicitly terminated when there are no more ASN.1 objects to be processed by the interpreter. For a perfectly-formed query, the interpreter would be back in the state it was when it started: the stack would have only the root dictionary on it, and all of the ASN.1 objects in the result would be terminated. If there are still "open" ASN.1 objects in the result (caused by leaving ENDs off of the query), then these are closed, as if a sufficient number of ENDs were provided. This condition would be indicated by the existence of dictionaries other than the root dictionary on the stack.
If an extra END is received that would pop the root dictionary off of the stack, the query is terminated immediately. No error is generated. 9. EXTENDING THE SET OF VALUES There are two ways to extend the set of values understood by the query language. The first is to register the data and its meaning and get an ASN.1 tag assigned for it. This is the preferred method because it makes that data specification available for everyone to use. The second method is to use the VendorSpecific application type to "wrap" the vendor-specific data. Wherever an implementation defines data that is not in RFC-1024, the "VendorSpecific" tag should be used to label a dictionary containing the vendor-specific data. For example, if a vendor had some data associated with interfaces that was too strange to get standard numbers assigned for, they could, instead represent the data like this: interfaces { interface { in-pkts, out-pkts, ... VendorSpecific { ephemeris, declination } } } In this case, ephemeris and declination correspond to two context- dependent tags assigned by the vendor for their non-standard data. If the vendor-specific method is chosen, the private data MUST have descriptions available through the GET-ATTRIBUTES operator. Even with this descriptive ability, the preferred method is to get standard numbers assigned if possible. 10. AUTHORIZATION This specification does not state what type of authorization system is used, if any. Different systems may have needs for different mechanisms (authorization levels, capability sets, etc.), and some systems may not care about authorization at all. The only effect that an authorization system has on a query is to restrict what data items in the tree may be retrieved or modified. Therefore, there are no explicit query language features that deal with protection. Instead, protection mechanisms are implicit and may make some of the data invisible (for GET) or non-writable (for SET):
- Each query runs with some level of authorization or set of capabilities, determined by its environment (HEMS and the HEMP header). - Associated with each data item in the data tree is some sort of test to determine if a query's authorization should grant it access to the item. Authorization tests are only applied to query language operations that retrieve information (GET, GET-ATTRIBUTES, and GET-RANGE) or modify it (SET, CREATE, DELETE). An authorization system must not affect the operation of BEGIN and END. In particular, the authorization must not hide entire dictionaries, because that would make a BEGIN on such a dictionary fail, terminating the entire query. 11. ERRORS If some particular information is requested but is not available, it will be returned as "no-value" by giving the ASN.1 length as 0. When there is any other kind of error, such as having improper arguments on the top of the stack or trying to execute BEGIN when the path doesn't refer to a dictionary, an ERROR object is emitted. The contents of this object identify the exact nature of the error: Error ::= [APPLICATION 0] IMPLICIT SEQUENCE { errorCode INTEGER, errorInstance INTEGER, errorOffset INTEGER errorDescription IA5String, errorOp INTEGER, } errorCode identifies what the error was, and errorInstance is an implementation-dependent code that gives a more precise indication of where the error occured. errorOffset is the location within the query where the error occurred. If an operation was being executed, errorOp contains its operation code, otherwise zero. errorDescription is a text string that can be printed that gives some description of the error. It will at least describe the errorCode, but may also give details implied by errorInstance. Detailed definitions of all of the fields are given in appendix I.2. Since there may be several unterminated ASN.1 objects in progress at the time the error occurs, each one must be terminated. Each unterminated object will be closed with a copy of the ERROR object. Depending upon the type of length encoding used for this object, this
will involve filling the value for the length (definite length form) or emitting two zero octets (indefinite length form). After all objects are terminated, a final copy of the ERROR object will be emitted. This structure guarantees that the error will be noticed at every level of interpretation on the receiving end. In summary, if there was an error before any ASN.1 objects were generated, then the result would simply be: error{...} If a couple of ASN.1 objects were unterminated when the error occurred, the result might look like: interfaces{ interface { name(...) type(...) error{...} } error{...} } error{...} It would be possible to define a "WARNING" object that has a similar (or same) format as ERROR, but that would be used to annotate responses when a non-fatal "error" occurs, such as attempting to SET/CREATE/DELETE and the operation is denied. This would be an additional complication, and we left it out in the interests of simplicity. I. ASN.1 DESCRIPTIONS OF QUERY LANGUAGE COMPONENTS A query consists of a sequence of ASN.1 objects, as follows: Query := IMPLICIT SEQUENCE of QueryElement; QueryElement ::= CHOICE { Operation, Filter, Template, Path, InputValue } Operation and Filter are defined below. The others are: Template ::= any Path ::= any InputValue ::= any These three are all similar, but have different restrictions on their structure:
Template Specifies a portion of the tree, naming one or more values, but not containing any values. Path Specifies a single path from one point in the tree to another, naming exactly one value, but not containing a value. InputValue Gives a value to be used by a query language operator. A query response consists of a sequence of ASN.1 objects, as follows: Response := IMPLICIT SEQUENCE of ResponseElement; ResponseElement ::= CHOICE { ResultValue, Error } Error is defined below. The others are: ResultValue ::= any ResultValue is similar to Template, above: ResultValue Specifies a portion of the tree, naming and containing one or more values. The distinctions between these are elaborated in section 6. I.1 Operation Codes Operation codes are all encoded in a single application-specific type, whose value determines the operation to be performed. The definition is: Operation ::= [APPLICATION 1] IMPLICIT INTEGER { reserved(0), begin(1), end(2), get(3), get-attributes(4), get-range(5), set(6),
create(7), delete(8) } I.2 Error Returns An Error object is returned within a reply when an error is encountered during the processing of a query. Note that the definition this object is similar to that of the HEMP protocol error structure. The error codes have been selected to keep the code spaces distinct between the two. This is intended to ease the processing of error messages. See section 11 for more information. Error ::= [APPLICATION 0] IMPLICIT SEQUENCE { errorCode INTEGER, errorInstance INTEGER, errorOffset INTEGER errorDescription IA5String, errorOp INTEGER, } The fields are defined as follows: errorCode Identifies the general cause of the error. errorInstance An implementation-dependent code that gives a more precise indication of where the error occured in the query processor. This is most useful when internal errors are reported. errorOffset The location within the query where the error was detected. The first octet of the query is numbered zero. errorOp If an operation was being executed, this contains its operation code, otherwise zero. errorDescription A text string that can be printed that gives some description of the error. It will at least describe the errorCode, but may also give details implied by errorInstance. Some errors are associated with the execution of specific operations, and others with the overall operation of the query interpreter. The errorCodes are split into two groups. The first group deals with overall interpreter operation. Except for
"unknown operation", these do not set errorOp. 100 Other error. Any error not listed below. 101 Format error. An error has been detected in the format of the input stream, preventing further interpretation of the query. 102 System error. The query processor has failed in some way due to an internal error. 103 Stack overflow. Too many items were pushed on the stack. 104 Unknown operation. The operation code is invalid. errorOp is set. The second group is errors that are associated with the execution of particular operations. errorOp will always be set for these. 200 Other operation error. Any error, associated with an operation, not listed below. 201 Stack underflow. An operation expected to see some number of operands on the stack, and there were fewer items on the stack. 202 Operand error. An operation expected to see certain operand types on the stack, and something else was there. 203 Invalid path for BEGIN. A path given for BEGIN was invalid, because some element in the path didn't exist. 204 Non-dictionary for BEGIN. A path given for BEGIN was invalid, because the given node was a leaf node, not a dictionary. 205 BEGIN on array element. The path specified an array element. The path must point at a single, unique, node. A filtered BEGIN should have been used.
206 Empty filter for BEGIN. The filter for a BEGIN didn't match any array element. 207 Filtered operation on non-array. A filtered operation was attempted on a regular dictionary. Filters can only be used on arrays. 208 Index out of bounds. The starting address or length for a GET-RANGE operation went outside the bounds for the given object. 209 Bad object for GET-RANGE. GET-RANGE can only be applied to objects whose base type is OctetString. This list is probably not quite complete, and would need to be extended, based upon implementation experience. I.3 Filters Many of the operations can take a filter argument to select among elements in an array. They are discussed in section 8.6. Filter ::= [APPLICATION 2] CHOICE { present [0] DataPath, equal [1] DataValue, greaterOrEqual [2] DataValue, lessOrEqual [3] DataValue, and [4] SEQUENCE OF Filter, or [5] SEQUENCE OF Filter, not [6] Filter } DataPath ::= ANY -- Path with no value DataValue ::= ANY -- Single data value A filter is executed by inorder traversal of its ASN.1 structure. The basic filter operations are: present tests for the existence of a particular data item in the data tree
equal tests to see if the named data item is equal to the given value. greaterOrEqual tests to see if the named data item is greater than or equal to the given value. lessOrEqual tests to see if the named data item is less than or equal to the given value. These may be combined with "and", "or", and "not" operators to form arbitrary boolean expressions. The "and" and "or" operators will take any number of terms. Terms are only evaluated up to the point where the outcome of the expression is determined (i.e., an "and" term's value is false or an "or" term's value is true). I.4 Attributes One or more Attributes structure is returned by the GET-ATTRIBUTES operator. This structure provides descriptive information about items in the data tree. See the discussion in section 8.3. Attributes ::= [APPLICATION 3] IMPLICIT SEQUENCE { tagASN1 [0] IMPLICIT INTEGER, valueFormat [1] IMPLICIT INTEGER, longDesc [2] IMPLICIT IA5String OPTIONAL, shortDesc [3] IMPLICIT IA5String OPTIONAL, unitsDesc [4] IMPLICIT IA5String OPTIONAL, precision [5] IMPLICIT INTEGER OPTIONAL, properties [6] IMPLICIT BITSTRING OPTIONAL, valueSet [7] IMPLICIT SET OF valueDesc OPTIONAL } valueDesc ::= IMPLICIT SEQUENCE { value [0] ANY, -- Single data value desc [1] IA5String } The meanings of the various attributes are given below. tagASN1 The ASN.1 tag for this object. This attribute is required. valueFormat The underlying ASN.1 type of the object (e.g., SEQUENCE or OCTETSTRING or Counter). This is not just the tag number, but the entire tag, as it would appear in an ASN.1 object. As such, it includes the class, which should be either UNIVERSAL or APPLICATION. Applications receiving this should
ignore the constructor bit. This attribute is required. longDesc A potentially lengthy text description which fully defines the object. This attribute is optional for objects defined in this memo and required for entity-specific objects. shortDesc A short mnemonic string of less than 15 characters, suitable for labeling the value on a display. This attribute is optional. unitsDesc A short string used for integer values to indicate the units in which the value is measured (e.g., "ms", "sec", "pkts", etc.). This attribute is optional. precision For Counter objects, the value at which the Counter will roll-over. Required for all Counter objects. properties A bitstring of boolean properties of the object. If the bit is on, it has the given property. This attribute is optional. The bits currently defined are: 0 If true, the difference between two values of this object is significant. For example, the changes of a packet count is always significant, it always conveys information. In this case, the 0 bit would be set. On the other hand, the difference between two readings of a queue length may be meaningless. 1 If true, the value may be modified with SET, CREATE, and DELETE. Applicability of CREATE and DELETE depends upon whether the object is in an array. 2 If true, the object is a dictionary, and a BEGIN may be used on it. If false, the object is leaf node in the data tree. 3 If true, the object is an array-type dictionary, and filters may be used to traverse it. (Bit 2 will be true also.) valueSet For data that is defined as an ASN.1 CHOICE type (an enumerated type), this gives descriptions for each of the possible values that the data object may assume.
Each valueDesc is a <value,description> pair. This information is especially important for control items, which are very likely to appear in VendorSpecific dictionaries, exactly the situation where descriptive information is needed. I.5 VendorSpecific See the discussion in section 9. VendorSpecific ::= [APPLICATION 4] IMPLICIT SET of ANY II. IMPLEMENTATION HINTS Although it is not normally in the spirit of RFCs to define an implementation, the authors feel that some suggestions will be useful to implementors of the query language. This list is not meant to be complete, but merely to give some hints about how the authors imagine that the query processor might be implemented efficiently. - It should be understood that the stack is of very limited depth. Because of the nature of the query language, it can get only about 4 entries (for arguments) plus the depth of the tree (up to one BEGIN per level in the tree). This comes out to about a dozen entries in the stack, a modest requirement. - The stack is an abstraction -- it should be implemented with pointers, not by copying dictionaries, etc. - An object-oriented approach should make implementation fairly easy. Changes to the "shape" if the data items (which will certainly occur, early on) will also be easier to make. - Only a few "messages" need to be understood by objects. By having pointers to action routines for each basic operation (GET,SET,...) associated with each node in the tree, common routines (e.g., emit a long integer located at address X) can be shared, and special routines (e.g., set the interface state for interface X) can be implemented in a common framework. Higher levels need know nothing about what data is being dealt with. - Most interesting objects are dictionaries, each of which can be implemented using pointers to the data and procedure "hooks" to perform specific operations such as GET, SET,
filtering, etc. - The hardest part is actually extracting the data from existing TCP/IP implementations that weren't designed with detailed monitoring in mind. Query processors interfacing to a UNIX kernel will have to make many system calls in order to extract some of the more intricate structures, such as routing tables. This should be less of a problem if a system is designed with easy monitoring as a goal. A Skeletal Implementation This section gives a rather detailed example of the core of a query processor. This code has not been tested, and is intended only to give implementors ideas about how to tackle some aspects of query processor implementation with finesse, rather than brute force. The suggested architecture is for each dictionary to have a "traverse" routine associated with it, which is called when any sort of operation has to be done on that dictionary. Most nodes will share the same traversal routine, but array dictionaries will usually have routines that know about whatever special lookup mechanisms are required. Non-dictionary nodes would have two routines, "action", and "compare", which implement query language operations and filter comparisons, respectively. Most nodes would share these routines. For example, there should be one "action" routine that does query language operations on 32-bit integers, and another that works on 16-bit integers, etc. Any traversal procedure would take arguments like: traverse(node, mask, op, filter) Treenode node; /* generic node-in-tree */ ASN mask; /* internal ASN.1 form */ enum opset op; /* what to do */ Filter filter; /* zero if no filter */ enum opset { begin, get, set, create, delete, geta, c_le, c_ge, c_eq, c_exist }; The traversal procedure is called whenever anything must be done within a dictionary. The arguments are: node the current dictionary.
mask is either the template, path, or value, depending upon the operation being performed. The top-level identifier of this object will be looked up in the context of <node>. op is the operation to be performed, either one of the basic operations, or a filter operation. filter is the filter to be applied, or zero if none. There will be no filter when <op> is a filter-comparison operation. The general idea is that the traversal proc associated with a node has all of the knowledge about how to get around in this subtree encoded within it. Hopefully, this will be the only place this knowledge is coded. Here is a skeleton of the "standard" traversal proc, written mostly in C. When the query processor needs to execute a "GET" operation, it would just call: traverse(current, template, GET, 0) Notes about this example: - This traversal routine handles either query language operations (GET, SET, etc.) or low-level filter operations. Separate routines could be defined for the two classes of operations, but they do much of the same work. - Dictionary nodes have a <traversal> proc defined. - Leaf nodes have an <action> proc, which implement GET, SET, GET-ATTRIBUTES, CREATE, and DELETE, and a <compare> proc, which performs low-level filter comparisons. - In the generic routine, the filter argument is unused, because the generic routine isn't used for array dictionaries, and only array dictionaries use filters. - An ASN type contains the top level tag and a list of sub-components. - size(mask) takes an ASN.1 object and tells how many sub-items are in it. Zero means that this is a simple object. - lookup(node, tag) looks up a tag in the given (tree)node, returning a pointer to the node. If the tag doesn't exist
in that node, a pointer to a special node "NullItem" is returned. NullItem looks like a leaf node and has procs that perform the correct action for non-existent data. - This example does not do proper error handling, or ASN.1 generation, both of which would require additional code in this routine. /* * For op = GET/SET/etc, return: * true on error, otherwise false. * When op is a filter operation, return: * the result of the comparison. */ int std_traverse(node, mask, op, filter) Treenode node; /* current node */ ASN mask; /* internal ASN.1 form */ enum opset op; /* what to do */ Filter filter; /* unused in this routine */ { ASN item; Treenode target; boolean rv = false; extern Treenode NullItem; if (filter != null) { error(...); return true; } target = lookup(node, mask.tag); /* We are at the leaf of the template/path/value. */ if (size(mask) == 0) switch (op) { case BEGIN: /* non-existent node, or leaf node */ if (target == NullItem || target.traverse == 0) { error(...); return true; } else { begin(node, mask.tag); return false; } case GET: case SET: case GETA:
case GETR: case CREATE: case DELETE: /* A leaf in the mask specifies entire directory. For GET, traverse the entire subtree. */ if (target.traverse) if (op == GET) { foreach subnode in target /* Need to test to not GET memory. */ rv |= (*target.traverse) (target, subnode.tag, op, 0); return rv; } else if (op == SET) /* no-op */ return false; else if (op != GETA) { error(...); return true; } /* We're at a leaf in both the mask and the tree. Just execute the operation. */ else { if (op == BEGIN) { /* Can't begin on leaf */ error(...); return true; else return (*target.action)(target, mask, op); } } /* else */ default: /* Comparison ops. */ return (*target.compare)(target, mask, op); } /* switch */ /* We only get here if mask has structure. */ /* can't have multiple targets for BEGIN */ if (op == BEGIN && size(mask) != 1) { error(...); return true; } /* or for a single filter operation. */ if (op is comparison && size(mask) != 1) { error(...); return false; } /* Iterate over the components in mask */ foreach item in mask {
if (target.traverse) /* traverse subtree. */ rv |= (*component.traverse)(component, item, op, 0); else /* leaf node, at last. */ if (op is comparison) return (*target.compare)(target, mask, op); else return (*target.action)(target, mask, op); } /* foreach */ return rv; } /* std_traverse */ Here is a bare skeleton of an array-type dictionary's traversal proc. int array_traverse(node, mask, op, filter) Treenode node; /* current node */ ASN mask; /* internal ASN.1 form */ enum opset op; /* what to do */ Filter filter; /* unused in this routine */ { Treenode target; boolean rv = false; extern Treenode NullItem; /* Didn't find that key. */ if (mask.tag != this array's iteration tag) return false; if (op == BEGIN && filter == null) { error(...); return 1; } /* The implementation of this loop is the major trick! */ /* Needs to stop after first filter success on BEGIN. */ foreach target in node { if (filter == null || /* if no filter, or */ ExecFilter(target, filter)) /* if it succeeds */ rv |= (target.traverse*)(target, mask, op, 0); } } /* array_traverse */ Object-oriented programming languages, such as C++, Modula, and Ada, are well suited to this style of implementation. There should be no particular difficulty with using a conventional language such as C or Pascal, however.
III. OBTAINING A COPY OF THE ASN.1 SPECIFICATION Copies of ISO Standard ASN.1 (Abstract Syntax Notation 1) are available from the following source. It comes in two parts; both are needed: IS 8824 -- Specification (meaning, notation) IS 8825 -- Encoding Rules (representation) They are available from: Omnicom Inc. 115 Park St, S.E. (new address as of March, 1987) Vienna, VA 22180 (703) 281-1135