1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
|
*******************************
Application Layer Frame Support
*******************************
.. contents:: Table of Contents
Baseline
========
- `Suricata rules format <https://docs.suricata.io/en/latest/rules/intro.html>`_
General Concepts
================
Frame support was introduced with Suricata 7.0. Up until 6.0.x, Suricata's architecture and state of parsers meant that the network traffic available to the detection engine was just a stream of data, without detail about higher level parsers.
.. note:: For Suricata, *Frame* is a generic term that can represent any unit of network data we are interested in, which could be comprised of one or several records of other, lower level protocol(s). Frames work as "stream annotations", allowing Suricata to tell the detection engine what type of record exists at a specific offset in the stream.
The normal pipeline of detection in Suricata implied that:
- Certain rules could be quite costly performance-wise. This happened because the same stream could be inspected several times for different rules, since for certain signatures the detection is done when Suricata is still inspecting a lower level stream, not the application layer protocol (e.g., *TCP* traffic, in place of *SMB* one);
- Rules could be difficult and tedious to write (and read), requiring that writers went in byte-detail to express matching on specific payload patterns.
What the Frame support offers is the ability to "point" to a specific portion of stream and identify what type of traffic Suricata is looking at. Then, as the engine reassembles the stream, one can have "read access" to that portion of the stream, aggregating concepts like what type of application layer protocol that is, and differentiating between ``header``, ``data`` or even protocol versions (*SMB1*, *SMB2*...).
The goal of the stream *Frame* is to expose application layer protocol `PDUs <https://en.wikipedia.org/wiki/Protocol_data_unit>`_ and other such arbitrary elements to the detection engine directly, instead of relying on Transactions. The main purpose is to bring *TCP data* processing times down by specialising/ filtering down traffic detection.
Adding Frame Support to a Parser
================================
The application layer parser exposes frames it supports to the detect engine, by tagging them as they're parsed. The rest works automatically.
In order to allow the engine to identify frames for records of a given application layer parser, thought must be given as to which frames make sense for the specific protocol you are handling. Some parsers may have clear ``header`` and ``data`` fields that form its *protocol data unit* (pdu). For others, the distinction might be between ``request`` and ``response``, only. Whereas for others it may make sense to have specific types of data. This is better understood by seeing the different types of frame keywords, which vary on a per-protocol basis.
It is also important to keep follow naming conventions when defining Frame Types. While a protocol may have strong naming standards for certain structures, do compare those with what Suricata already has registered:
- ``hdr``: used for the record header portion
- ``data``: is used for the record data portion
- ``pdu``: unless documented otherwise, means the whole record, comprising ``hdr`` and ``data``
- ``request``: a message from a client to a server
- ``response``: a message from a server to a client
Basic steps
~~~~~~~~~~~
Once the frame types that make sense for a given protocol are defined, the basic steps for adding them are:
- create an enum with the frame types;
- identify the parsing function(s) where application layer records are parsed;
- identify the correct moment to register the frames;
- use the Frame API calls directly or build upon them and use your functions to register the frames;
- register the relevant frame callbacks when registering the parser.
Once these are done, you can enable frame eve-output to confirm that your frames are being properly registered. It is important to notice that some hard coded limits could influence what you see on the logs (max size of log output; type of logging for the payload, cf. https://redmine.openinfosecfoundation.org/issues/4988).
If all the steps are successfully followed, you should be able to write a rule using the *frame* keyword and the frame types you have registered with the application layer parser.
Using the *SMB* parser as example, before frame support, a rule would look like::
alert tcp ... flow:to_server; content:"|ff|SMB"; content:"some smb 1 issue";
With frame support, one is able to do::
alert smb ... flow:to_server; frame:smb1.data; content:"some smb 1 issue";
Implementation Examples & API Callbacks
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Though the steps are the same, there are a few differences when implementing frame support in Rust or in C. The following sections elaborate on that, as well as on the process itself. (Note that the code snippets have omitted portions of code that weren't so relevant to this document).
Rust
----
This section shows how Frame support is added in Rust, using examples from the `SIP parser <https://github.com/OISF/suricata/blob/master/rust/src/sip/sip.rs>`_, and the `telnet parser <https://github.com/OISF/suricata/blob/master/rust/src/telnet/telnet.rs>`_.
**Define the frame types**. The frame types are defined as an enum. In Rust, make sure to derive from the ``AppLayerFrameType``:
.. literalinclude:: ../../../../../rust/src/sip/sip.rs
:caption: rust/src/sip/sip.rs
:language: rust
:start-after: // app-layer-frame-documentation tag start: FrameType enum
:end-before: // app-layer-frame-documentation tag end: FrameType enum
**Frame registering**. Some understanding of the parser will be needed in order to find where the frames should be registered. It makes sense that it will happen when the input stream is being parsed into records. See when some pdu and request frames are created for SIP:
.. literalinclude:: ../../../../../rust/src/sip/sip.rs
:caption: rust/src/sip/sip.rs
:language: rust
:start-after: // app-layer-frame-documentation tag start: parse_request
:end-before: // app-layer-frame-documentation tag end: parse_request
:dedent: 4
.. note:: when to create PDU frames
The standard approach we follow for frame registration is that a frame ``pdu`` will always be created, regardless of parser status (in practice, before the parser is called). The other frames are then created when and if only the parser succeeds.
**Use the Frame API or build upon them as needed**. These are the frame registration functions highlighted above:
.. literalinclude:: ../../../../../rust/src/sip/sip.rs
:caption: rust/src/sip/sip.rs
:language: rust
:start-after: // app-layer-frame-documentation tag start: function to add frames
:end-before: // app-layer-frame-documentation tag end: function to add frames
**Register relevant frame callbacks.** As these are inferred from the ``#[derive(AppLayerFrameType)]`` statement, all that is needed is:
.. literalinclude:: ../../../../../rust/src/sip/sip.rs
:caption: rust/src/sip/sip.rs
:language: rust
:start-at: get_frame_id_by_name
:end-at: ffi_name_from_id),
:dedent: 8
.. note:: on frame_len
For protocols which search for an end of frame char, like telnet, indicate unknown length by passing ``-1``. Once the length is known, it must be updated. For those where length is a field in the record (e.g. *SIP*), the frame is set to match said length, even if that is bigger than the current input
The telnet parser has examples of using the Frame API directly for registering telnet frames, and also illustrates how that is done when length is not yet known:
.. literalinclude:: ../../../../../rust/src/telnet/telnet.rs
:caption: rust/src/telnet/telnet.rs
:language: rust
:start-after: // app-layer-frame-documentation tag start: parse_request
:end-before: // app-layer-frame-documentation tag end: parse_request
:lines: 1-3, 22-49
:dedent: 4
We then update length later on (note especially lines 3 and 10):
.. literalinclude:: ../../../../../rust/src/telnet/telnet.rs
:caption: rust/src/telnet/telnet.rs
:language: rust
:start-after: // app-layer-frame-documentation tag start: update frame_len
:end-before: // app-layer-frame-documentation tag end: update frame_len
:linenos:
:dedent: 12
The Frame API calls parameters represent:
- ``flow``: dedicated data type, carries specific flow-related data
- ``stream_slice``: dedicated data type, carries stream data, shown further bellow
- ``frame_start``: a pointer to the start of the frame buffer in the stream (``cur_i`` in the SMB code snippet)
- ``frame_len``: what we expect the frame length to be (the engine may need to wait until it has enough data. See what is done in the telnet snippet request frames registering)
- ``frame_type``: type of frame it's being registering (defined in an enum, as shown further above)
``StreamSlice`` contains the input data to the parser, alongside other Stream-related data important in parsing context. Definition is found in *applayer.rs*:
.. literalinclude:: ../../../../../rust/src/applayer.rs
:caption: rust/src/applayer.rs
:language: rust
:start-at: pub struct StreamSlice
:end-before: impl StreamSlice
C code
------
Implementing Frame support in C involves a bit more manual work, as one cannot make use of the Rust derives. Code snippets from the *HTTP* parser:
Defining the frame types with the enum means:
.. literalinclude:: ../../../../../src/app-layer-htp.c
:caption: src/app-layer-htp.c
:start-after: /* app-layer-frame-documentation tag start: HttpFrameTypes
:end-before: /* app-layer-frame-documentation tag end: HttpFrameTypes
:lines: 1-16
The HTTP parser uses the Frame registration functions from the C API (``app-layer-frames.c``) directly for registering request Frames. Here we also don't know the length yet. The ``0`` indicates flow direction: ``toserver``, and ``1`` would be used for ``toclient``:
.. literalinclude:: ../../../../../src/app-layer-htp.c
:caption: src/app-layer-htp.c
:start-after: /* app-layer-frame-documentation tag start: frame registration http request
:end-before: /* app-layer-frame-documentation tag end: frame registration http request
:dedent: 4
Updating ``frame->len`` later:
.. literalinclude:: ../../../../../src/app-layer-htp.c
:caption: src/app-layer-htp.c
:start-after: /* app-layer-frame-documentation tag start: updating frame->len
:end-before: /* app-layer-frame-documentation tag end: updating frame->len
:dedent: 4
Register relevant callbacks (note that the actual functions will also have to be written, for C):
.. literalinclude:: ../../../../../src/app-layer-htp.c
:caption: src/app-layer-htp.c
:language: c
:start-after: /* app-layer-frame-documentation tag start: registering relevant callbacks
:end-before: /* app-layer-frame-documentation tag end: registering relevant callbacks
:dedent: 8
.. note:: The ``GetFrameIdByName`` functions can be "probed", so they should not generate any output or that could be misleading (for instance, Suricata generating a log message stating that a valid frame type is unknown).
Visual context
==============
``input`` and ``input_len`` are used to calculate the proper offset, for storing the frame. The stream buffer slides forward, so frame offsets/frames have to be updated. The `relative offset` (``rel_offset``) reflects that:
.. code-block:: c
Start:
[ stream ]
[ frame ...........]
rel_offset: 2
len: 19
Slide:
[ stream ]
[ frame .... .]
rel_offset: -10
len: 19
Slide:
[ stream ]
[ frame ........... ]
rel_offset: -16
len: 19
The way the engine handles stream frames can be illustrated as follows:
.. image:: img/StreamFrames.png
:scale: 80
|