blob: d134c07ce7dc6b298db616803ee7cf1551e78a8d [file] [log] [blame]
Fredrik Svedberg8d0f4892021-02-16 21:59:50 +01001# Copyright (C) 2020-2021 Arm Limited or its affiliates. All rights reserved.
Tim Hall79d07d22020-04-27 18:20:16 +01002#
3# SPDX-License-Identifier: Apache-2.0
4#
5# Licensed under the Apache License, Version 2.0 (the License); you may
6# not use this file except in compliance with the License.
7# You may obtain a copy of the License at
8#
9# www.apache.org/licenses/LICENSE-2.0
10#
11# Unless required by applicable law or agreed to in writing, software
12# distributed under the License is distributed on an AS IS BASIS, WITHOUT
13# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14# See the License for the specific language governing permissions and
15# limitations under the License.
Tim Hall79d07d22020-04-27 18:20:16 +010016# Description:
17# Functions used to write to a TensorFlow Lite format file. Supports adding in file identifiers.
Tim Hall79d07d22020-04-27 18:20:16 +010018import flatbuffers
Diego Russoe8a10452020-04-21 17:39:10 +010019import flatbuffers.number_types as N
20import numpy as np
21from flatbuffers import encode
Diego Russoea6111a2020-04-14 18:41:58 +010022from flatbuffers.builder import UOffsetTFlags
23
Michael McGeagh7a6f8432020-12-02 15:29:22 +000024from .errors import VelaError
Diego Russoe8a10452020-04-21 17:39:10 +010025from .nn_graph import PassPlacement
Louis Verhaardaee5d752020-09-30 09:01:52 +020026from .operation import Op
Patrik Gustavsson5e26eda2021-06-30 09:07:16 +020027from .reader_util import align_inputs_indices
Patrik Gustavssoneca2e952020-05-27 09:15:11 +020028from .tensor import MemType
Samuel Panijel6f4955a2021-06-10 13:40:03 +030029from .tensor import TensorPurpose
Tim Hall79d07d22020-04-27 18:20:16 +010030from .tflite import Buffer
31from .tflite import Metadata
Diego Russoe8a10452020-04-21 17:39:10 +010032from .tflite import Model
33from .tflite import Operator
34from .tflite import OperatorCode
35from .tflite import QuantizationParameters
36from .tflite import SubGraph
37from .tflite import Tensor
38from .tflite_mapping import builtin_operator_inv_map
39from .tflite_mapping import BuiltinOperator
Diego Russoe8a10452020-04-21 17:39:10 +010040from .tflite_mapping import datatype_inv_map
41
Tim Hallffe8e282021-06-24 18:29:53 +010042# the python flatbuffer interface is missing a method to add in file identifier. patching it in here:
Tim Hall79d07d22020-04-27 18:20:16 +010043
44tflite_version = 3
45tflite_file_identifier = "TFL" + str(tflite_version)
46
47
Tim Hall79d07d22020-04-27 18:20:16 +010048def FinishWithFileIdentifier(self, rootTable, fid):
49 if fid is None or len(fid) != 4:
Michael McGeagh7a6f8432020-12-02 15:29:22 +000050 raise VelaError("FileIdentifier must be 4 chars")
Tim Hall79d07d22020-04-27 18:20:16 +010051
52 flags = N.Uint8Flags
53 prepSize = 4
54 self.Prep(self.minalign, prepSize + len(fid))
55 for i in range(3, -1, -1):
56 self.head = self.head - flags.bytewidth
57 encode.Write(flags.packer_type, self.Bytes, self.Head(), ord(fid[i]))
58
59 return self.Finish(rootTable)
60
61
62flatbuffers.Builder.FinishWithFileIdentifier = FinishWithFileIdentifier
63
64
65def make_vector(v):
66 try:
67 len(v)
68 return v
69 except TypeError:
70 return [v]
71
72
73class TFLiteSerialiser:
74 def __init__(self, nng):
75 self.builder = flatbuffers.Builder(0)
76 self.nng = nng
77
78 self.scratch_buf_id = 0 # Always assign scratch to buffer 0
Patrik Gustavssoneca2e952020-05-27 09:15:11 +020079 self.scratch_fast_buf_id = 1 # Always assign scratch_fast to buffer 1
Tim Hall79d07d22020-04-27 18:20:16 +010080 self.buffers_to_write = [] # have an empty array there
81
Michael McGeaghf3e3ad72020-12-02 12:39:03 +000082 self.ops_to_ignore = (Op.Const, Op.Placeholder, Op.SubgraphInput)
Tim Hall79d07d22020-04-27 18:20:16 +010083
84 self.tensors_to_reshape = {}
85
86 self.subgraphs_to_write = [sg for sg in self.nng.subgraphs if sg.placement == PassPlacement.Cpu]
87
88 all_ops = []
89 for sg in self.subgraphs_to_write:
90 for ps in sg.passes:
91 for op in ps.ops:
92 if op.type not in self.ops_to_ignore:
Patrik Gustavsson5e26eda2021-06-30 09:07:16 +020093 # swap from nng input indexing to TensorFlow Lite input indexing
94 self.align_nng_inputs_to_tflite(op)
Tim Hall79d07d22020-04-27 18:20:16 +010095 all_ops.append(op)
Louis Verhaardaee5d752020-09-30 09:01:52 +020096 if op.type.is_conv2d_op() or op.type.is_depthwise_conv2d_op():
Andreas Nevalainend8c032d2020-09-11 10:25:09 +020097 # If values are None op has non-constant weights
98 if op.inputs[1].values is not None:
99 self.tensors_to_reshape[op.inputs[1]] = (3, 0, 1, 2)
Louis Verhaardaee5d752020-09-30 09:01:52 +0200100 if op.type == Op.FullyConnected:
Andreas Nevalainend8c032d2020-09-11 10:25:09 +0200101 # If values are None op has non-constant weights
102 if op.inputs[1].values is not None:
103 self.tensors_to_reshape[op.inputs[1]] = (1, 0)
Tim Hall79d07d22020-04-27 18:20:16 +0100104
Louis Verhaardaee5d752020-09-30 09:01:52 +0200105 # list of tuple(Op, string); the custom code is only used for 3rd party custom operators
106 self.operator_codes = sorted(set((op.type, op.attrs.get("custom_code", "")) for op in all_ops))
Tim Hall79d07d22020-04-27 18:20:16 +0100107 self.operator_code_map = {}
108
Patrik Gustavsson5e26eda2021-06-30 09:07:16 +0200109 def align_nng_inputs_to_tflite(self, op):
110 from_indices = op.type.info.indices
111 _, _, to_indices = builtin_operator_inv_map[op.type]
112 op.inputs = align_inputs_indices(from_indices, to_indices, op.inputs)
113
Tim Hall79d07d22020-04-27 18:20:16 +0100114 def write_byte_vector(self, v, alignment=1):
115 builder = self.builder
116 builder.StartVector(1, len(v), alignment)
117 for e in v[::-1]:
118 builder.PrependByte(e)
119 return builder.EndVector(len(v))
120
121 def write_int_vector(self, v):
122 builder = self.builder
123 builder.StartVector(4, len(v), 4)
124 for e in v[::-1]:
125 builder.PrependInt32(e)
126 return builder.EndVector(len(v))
127
128 def write_long_vector(self, v):
129 builder = self.builder
130 builder.StartVector(8, len(v), 8)
131 for e in v[::-1]:
132 builder.PrependInt64(e)
133 return builder.EndVector(len(v))
134
135 def write_float_vector(self, v):
136 builder = self.builder
137 builder.StartVector(4, len(v), 4)
138 for e in v[::-1]:
139 builder.PrependFloat32(e)
140 return builder.EndVector(len(v))
141
142 def write_offset_vector(self, v):
143 builder = self.builder
144 builder.StartVector(4, len(v), 4)
145 for e in v[::-1]:
146 builder.PrependUOffsetTRelative(e)
147 return builder.EndVector(len(v))
148
Tim Hallc8310b12020-06-17 14:53:11 +0100149 def assign_buffers_to_tensors(self, tensors, scratch_tensor):
150 if scratch_tensor is not None:
151 scratch_tensor_mem_area = scratch_tensor.mem_area
Tim Hall25f605c2020-05-18 18:04:26 +0100152 else:
153 scratch_tensor_mem_area = None # all tensors are initialised to MemArea.Unknown
154
Tim Hall79d07d22020-04-27 18:20:16 +0100155 buffer_map = {}
Patrik Gustavssoneca2e952020-05-27 09:15:11 +0200156
Patrik Gustavsson3ab94522020-06-29 17:36:55 +0200157 buf_idx = 2
Tim Hall79d07d22020-04-27 18:20:16 +0100158
159 for tens in tensors:
Patrik Gustavssoneca2e952020-05-27 09:15:11 +0200160 # Set buffer ids depending on allocation
161 if tens.is_allocated_in_tensor_arena(scratch_tensor_mem_area):
Tim Hall79d07d22020-04-27 18:20:16 +0100162 buffer_map[tens] = self.scratch_buf_id
Patrik Gustavssoneca2e952020-05-27 09:15:11 +0200163 elif tens.mem_type == MemType.Scratch_fast:
164 # For Scratch_fast when not co-allocated with scratch in the TensorArena:
165 buffer_map[tens] = self.scratch_fast_buf_id
Tim Hall79d07d22020-04-27 18:20:16 +0100166 else:
167 buffer_map[tens] = buf_idx
168 buf_idx += 1
169
Tim Hallc8310b12020-06-17 14:53:11 +0100170 # Initialize buffers_to_write to a length equal to number of buffers so
Tim Hall79d07d22020-04-27 18:20:16 +0100171 # they can be appended at the correct index during tensor serialization
172 self.buffers_to_write = [None] * (buf_idx)
173
174 return buffer_map
175
Louis Verhaardaee5d752020-09-30 09:01:52 +0200176 def serialise_operator_code(self, idx, op_type, custom_code):
Tim Hall79d07d22020-04-27 18:20:16 +0100177 builder = self.builder
178 custom_code_offset = None
Louis Verhaardaee5d752020-09-30 09:01:52 +0200179 if op_type == Op.Custom:
Patrik Gustavsson5e26eda2021-06-30 09:07:16 +0200180 tf_code, opt_serializer, _ = builtin_operator_inv_map[op_type]
Louis Verhaardaee5d752020-09-30 09:01:52 +0200181 custom_code_offset = builder.CreateString(custom_code)
Tim Hall79d07d22020-04-27 18:20:16 +0100182 else:
Tim Halle9194df2020-08-04 20:37:01 +0100183 assert (
Louis Verhaardaee5d752020-09-30 09:01:52 +0200184 op_type in builtin_operator_inv_map
185 ), "Vela does not contain a mapping to serialise {} operator to a TensorFlow Lite operator".format(op_type)
Patrik Gustavsson5e26eda2021-06-30 09:07:16 +0200186 tf_code, opt_serializer, _ = builtin_operator_inv_map[op_type]
Tim Hall79d07d22020-04-27 18:20:16 +0100187
Tim Hallb2183762021-01-25 21:42:56 +0000188 if op_type == Op.CustomNpuOp:
Tim Halle9194df2020-08-04 20:37:01 +0100189 assert (
Tim Hallb2183762021-01-25 21:42:56 +0000190 tf_code == BuiltinOperator.CUSTOM
Tim Halle9194df2020-08-04 20:37:01 +0100191 ), "Vela only supports serialising NpuOp operators as TensorFlow Lite Custom operators"
Tim Hall79d07d22020-04-27 18:20:16 +0100192 custom_code_offset = builder.CreateString("ethos-u")
193
Tim Hallb2183762021-01-25 21:42:56 +0000194 # there can be multiple different types of 3rd party custom operators (i.e. non-"ethos-u" ones). therefore we
195 # need to add an extra level of indirection to this particular entry in the operator_code_map to allow for the
196 # correct lookup later on
197 if op_type == Op.Custom:
198 if op_type not in self.operator_code_map:
199 self.operator_code_map[op_type] = {}
200 self.operator_code_map[op_type][custom_code] = (idx, tf_code, opt_serializer)
201 else:
202 self.operator_code_map[op_type] = (idx, tf_code, opt_serializer)
Tim Hall79d07d22020-04-27 18:20:16 +0100203
204 OperatorCode.OperatorCodeStart(builder)
Tim Hall42abec12021-02-04 21:31:57 +0000205 OperatorCode.OperatorCodeAddDeprecatedBuiltinCode(builder, tf_code if tf_code < 127 else 127)
Tim Hall79d07d22020-04-27 18:20:16 +0100206 OperatorCode.OperatorCodeAddBuiltinCode(builder, tf_code)
207 if custom_code_offset is not None:
208 OperatorCode.OperatorCodeAddCustomCode(builder, custom_code_offset)
209
210 return OperatorCode.OperatorCodeEnd(builder)
211
212 def serialise_quantization_parameters(self, quant):
213 builder = self.builder
214
Fredrik Svedberg8d0f4892021-02-16 21:59:50 +0100215 qp = None
Tim Hall79d07d22020-04-27 18:20:16 +0100216 min = None
217 max = None
218 scale = None
219 zero_point = None
220 if quant is not None:
221 if quant.min is not None:
222 min = self.write_float_vector(make_vector(quant.min))
223 if quant.max is not None:
224 max = self.write_float_vector(make_vector(quant.max))
225 if quant.scale_f32 is not None:
226 scale = self.write_float_vector(make_vector(quant.scale_f32))
227 if quant.zero_point is not None:
228 zero_point = self.write_long_vector(make_vector(quant.zero_point))
229
Fredrik Svedberg8d0f4892021-02-16 21:59:50 +0100230 QuantizationParameters.QuantizationParametersStart(builder)
231 if min is not None:
232 QuantizationParameters.QuantizationParametersAddMin(builder, min)
233 if max is not None:
234 QuantizationParameters.QuantizationParametersAddMax(builder, max)
235 if scale is not None:
236 QuantizationParameters.QuantizationParametersAddScale(builder, scale)
237 if zero_point is not None:
238 QuantizationParameters.QuantizationParametersAddZeroPoint(builder, zero_point)
Fredrik Svedbergcc8569f2021-11-01 14:25:29 +0100239 if quant.quant_dim is not None:
240 QuantizationParameters.QuantizationParametersAddQuantizedDimension(builder, quant.quant_dim)
Fredrik Svedberg8d0f4892021-02-16 21:59:50 +0100241 qp = QuantizationParameters.QuantizationParametersEnd(builder)
242
243 return qp
Tim Hall79d07d22020-04-27 18:20:16 +0100244
245 def serialise_tensor(self, tens):
246 builder = self.builder
247 tens_shape = tens.shape
James Peet7519d502021-07-19 16:47:58 +0100248 values = tens.values
Tim Hall79d07d22020-04-27 18:20:16 +0100249
250 if values is None:
251 values = np.empty(shape=(0), dtype=np.uint8)
252
253 if tens in self.tensors_to_reshape:
254 reorder = self.tensors_to_reshape[tens]
255 tens_shape = [tens_shape[idx] for idx in reorder]
256 values = values.transpose(reorder)
257
Tim Hall79d07d22020-04-27 18:20:16 +0100258 buf_id = self.buffer_map[tens]
Patrik Gustavssoneca2e952020-05-27 09:15:11 +0200259 self.buffers_to_write[buf_id] = values.flatten().view(np.uint8)
Tim Hall79d07d22020-04-27 18:20:16 +0100260
261 shape = self.write_int_vector(tens_shape)
262
263 name = builder.CreateString(tens.name)
264 quant = self.serialise_quantization_parameters(tens.quantization)
265
266 Tensor.TensorStart(builder)
267 Tensor.TensorAddShape(builder, shape)
268 Tensor.TensorAddType(builder, datatype_inv_map[tens.dtype])
269 # All tensors must have a valid backing buffer, even if it is empty.
270 # Empty buffers should be kept unique for TensorFlow Lite Micro
271 Tensor.TensorAddBuffer(builder, buf_id)
272 Tensor.TensorAddName(builder, name)
Fredrik Svedberg8d0f4892021-02-16 21:59:50 +0100273 if quant is not None:
274 Tensor.TensorAddQuantization(builder, quant)
275 Tensor.TensorAddIsVariable(builder, tens.is_variable)
Tim Hall79d07d22020-04-27 18:20:16 +0100276
277 res = Tensor.TensorEnd(builder)
278 return res
279
280 def serialise_operator(self, op):
281 builder = self.builder
282
Fredrik Svedberg8d0f4892021-02-16 21:59:50 +0100283 inputs_offset = self.write_int_vector(
284 [self.tensor_map[tens] if tens in self.tensor_map else -1 for tens in op.inputs]
285 )
Michael McGeaghbb1b09e2020-08-19 11:24:17 +0100286 outputs_offset = self.write_int_vector(
287 [self.tensor_map[tens] for tens in op.outputs if tens in self.tensor_map]
288 )
Fredrik Svedberg8d0f4892021-02-16 21:59:50 +0100289 intermediates_offset = self.write_int_vector(
290 [self.tensor_map[tens] for tens in op.intermediates if tens in self.tensor_map]
291 )
Tim Hall79d07d22020-04-27 18:20:16 +0100292
Tim Hallb2183762021-01-25 21:42:56 +0000293 if op.type == Op.Custom:
294 op_idx, tflop, opt_serializer = self.operator_code_map[op.type][op.attrs.get("custom_code", "")]
295 else:
296 op_idx, tflop, opt_serializer = self.operator_code_map[op.type]
Tim Hall79d07d22020-04-27 18:20:16 +0100297
298 builtin_opt_offset = None
299 custom_opt_offset = None
300 if opt_serializer is not None:
301 attrs = dict(op.attrs)
302 if "strides" in attrs:
303 attrs["stride_h"] = attrs["strides"][1]
304 attrs["stride_w"] = attrs["strides"][2]
305 if "ksize" in attrs:
306 attrs["filter_height"] = attrs["ksize"][1]
307 attrs["filter_width"] = attrs["ksize"][2]
308 if "dilation" in attrs:
309 attrs["dilation_h_factor"] = attrs["dilation"][1]
310 attrs["dilation_w_factor"] = attrs["dilation"][2]
311 if "channel_multiplier" in attrs:
312 attrs["depth_multiplier"] = attrs["channel_multiplier"]
Louis Verhaardc86a9d22020-11-02 18:04:27 +0100313 attrs["fused_activation_function"] = op.activation.op_type if op.activation is not None else None
Tim Hall79d07d22020-04-27 18:20:16 +0100314
315 builtin_opt_offset, custom_opt_offset = opt_serializer.serialize(builder, attrs)
316
317 mutating_variable_inputs_offset = self.write_byte_vector([])
318 Operator.OperatorStart(builder)
319 Operator.OperatorAddOpcodeIndex(builder, op_idx)
320 Operator.OperatorAddInputs(builder, inputs_offset)
321 Operator.OperatorAddOutputs(builder, outputs_offset)
Fredrik Svedberg8d0f4892021-02-16 21:59:50 +0100322 Operator.OperatorAddIntermediates(builder, intermediates_offset)
Tim Hall79d07d22020-04-27 18:20:16 +0100323
324 if builtin_opt_offset is not None:
325 Operator.OperatorAddBuiltinOptionsType(builder, opt_serializer.builtin_opt_type)
326 Operator.OperatorAddBuiltinOptions(builder, builtin_opt_offset)
327 if custom_opt_offset is not None:
328 Operator.OperatorAddCustomOptions(builder, custom_opt_offset)
329 Operator.OperatorAddCustomOptionsFormat(builder, opt_serializer.custom_opt_format)
330
331 Operator.OperatorAddMutatingVariableInputs(builder, mutating_variable_inputs_offset)
332 return Operator.OperatorEnd(builder)
333
334 def serialise_subgraph(self, sg):
335 builder = self.builder
336 tensor_set = set()
Tim Hall79d07d22020-04-27 18:20:16 +0100337 all_ops = []
Michael McGeagh515c9562020-09-02 15:52:43 +0100338 placeholder_ops = []
339
Tim Hall79d07d22020-04-27 18:20:16 +0100340 for ps in sg.passes:
341 for op in ps.ops:
342 if op.type not in self.ops_to_ignore:
343 all_ops.append(op)
Louis Verhaardaee5d752020-09-30 09:01:52 +0200344 elif op.type == Op.Placeholder:
Michael McGeagh515c9562020-09-02 15:52:43 +0100345 placeholder_ops.append(op)
Tim Hall79d07d22020-04-27 18:20:16 +0100346
Michael McGeagh515c9562020-09-02 15:52:43 +0100347 # Add the tensors from all valid ops, as well as the tensors from placeholder ops
348 # This allows us to serialise tensors which arent attached to any specific ops,
349 # e.g. due to an empty graph containing no ops
350 for op in all_ops + placeholder_ops:
Fredrik Svedberg8d0f4892021-02-16 21:59:50 +0100351 for tens in op.inputs + op.outputs + op.intermediates:
Andreas Nevalainend8c032d2020-09-11 10:25:09 +0200352 if tens is not None:
353 tensor_set.add(tens)
Tim Hall79d07d22020-04-27 18:20:16 +0100354
355 all_tensors = [tens for nm, idx, tens in sorted((tens.name, idx, tens) for idx, tens in enumerate(tensor_set))]
356
Samuel Panijel6f4955a2021-06-10 13:40:03 +0300357 scratch_tensors = [tens for tens in all_tensors if tens.purpose is TensorPurpose.Scratch]
Patrik Gustavsson3ab94522020-06-29 17:36:55 +0200358
Tim Hallc8310b12020-06-17 14:53:11 +0100359 if len(scratch_tensors) == 0:
360 scratch_tensor = None
361 else:
362 assert len(scratch_tensors) == 1, "Multiple scratch tensors"
363 scratch_tensor = scratch_tensors[0]
364
Tim Hall79d07d22020-04-27 18:20:16 +0100365 self.tensor_map = {tens: idx for idx, tens in enumerate(all_tensors)}
Tim Hallc8310b12020-06-17 14:53:11 +0100366 self.buffer_map = self.assign_buffers_to_tensors(all_tensors, scratch_tensor)
Tim Hall79d07d22020-04-27 18:20:16 +0100367
368 tensors_offset = self.write_offset_vector([self.serialise_tensor(tens) for tens in all_tensors])
369
Tim Hall79d07d22020-04-27 18:20:16 +0100370 # Make sure the input_tensors haven't been modified
371 assert all(inp in sg.original_inputs for inp in sg.input_tensors)
Michael McGeaghbb1b09e2020-08-19 11:24:17 +0100372 inputs = [self.tensor_map[tens] for tens in sg.original_inputs if tens in self.tensor_map]
Tim Hallc8310b12020-06-17 14:53:11 +0100373
Tim Hallc8310b12020-06-17 14:53:11 +0100374 inputs_offset = self.write_int_vector(inputs)
Michael McGeaghbb1b09e2020-08-19 11:24:17 +0100375 outputs_offset = self.write_int_vector(
376 [self.tensor_map[tens] for tens in sg.output_tensors if tens in self.tensor_map]
377 )
Tim Hall79d07d22020-04-27 18:20:16 +0100378
379 operators_offset = self.write_offset_vector([self.serialise_operator(op) for op in all_ops])
380
381 SubGraph.SubGraphStart(builder)
382 SubGraph.SubGraphAddTensors(builder, tensors_offset)
383 SubGraph.SubGraphAddInputs(builder, inputs_offset)
384 SubGraph.SubGraphAddOutputs(builder, outputs_offset)
385
386 SubGraph.SubGraphAddOperators(builder, operators_offset)
387
388 return SubGraph.SubGraphEnd(builder)
389
390 def write_aligned_bytes(self, buf):
391 builder = self.builder
392 builder.nested = True
393 data = bytes(buf)
394 length_bytes = UOffsetTFlags.py_type(len(data))
395 builder.Prep(16, length_bytes) # Reserve aligned storage
396 builder.head = UOffsetTFlags.py_type(builder.Head() - length_bytes) # Update FlatBuffer internal pointer
397 builder.Bytes[builder.Head() : builder.Head() + length_bytes] = data # Assign bytes to aligned area
398 return builder.EndVector(length_bytes)
399
400 def serialise_buffer(self, buf):
401 builder = self.builder
402 data = None
403 if buf is not None:
404 data = self.write_aligned_bytes(buf)
405 Buffer.BufferStart(builder)
406 if data is not None:
407 Buffer.BufferAddData(builder, data)
408 return Buffer.BufferEnd(builder)
409
410 def serialise_metadata(self, metadata):
411 builder = self.builder
412 name = builder.CreateString(metadata[0])
413
414 Metadata.MetadataStart(builder)
415 Metadata.MetadataAddName(builder, name)
416 Metadata.MetadataAddBuffer(builder, metadata[1])
417
418 return Metadata.MetadataEnd(builder)
419
420 def serialise_model(self):
421 builder = self.builder
422 operator_code_offset = self.write_offset_vector(
Louis Verhaardaee5d752020-09-30 09:01:52 +0200423 [self.serialise_operator_code(idx, optype, code) for idx, (optype, code) in enumerate(self.operator_codes)]
Tim Hall79d07d22020-04-27 18:20:16 +0100424 )
425
426 description = builder.CreateString("Vela Optimised")
427
428 subgraph_offset = self.write_offset_vector([self.serialise_subgraph(sg) for sg in self.subgraphs_to_write])
429
430 # Fill the metadata buffer
431 version = np.int32(0)
432 subgraph_idx = np.int32(len(self.subgraphs_to_write)) # Only 1 supported currently
433 nbr_tensors = np.int32(len(self.tensor_map))
434
Fredrik Svedberge22ba8c2021-01-27 16:53:41 +0100435 if not any([name == b"OfflineMemoryAllocation" for name, _ in self.nng.metadata]):
436 # An offset of -1 indicates that the tensor will be allocated online by Tensorflow Lite Micro
437 offsets = [np.int32(-1)] * nbr_tensors
Tim Hall79d07d22020-04-27 18:20:16 +0100438
Fredrik Svedberge22ba8c2021-01-27 16:53:41 +0100439 # Ensure that the order of the offsets match the order of the tensors
440 for tens, idx in self.tensor_map.items():
441 # Set offsets for tensor allocated in Tensor Arena or in the scratch_fast area
442 if tens.mem_type in (MemType.Scratch, MemType.Scratch_fast):
443 offsets[idx] = np.int32(tens.address) if tens.address is not None else np.int32(0)
Tim Hall79d07d22020-04-27 18:20:16 +0100444
Fredrik Svedberge22ba8c2021-01-27 16:53:41 +0100445 self.nng.metadata.append(
446 ("OfflineMemoryAllocation", np.array([version, subgraph_idx, nbr_tensors] + offsets))
447 )
Michael McGeagh22f74e12020-08-07 16:21:03 +0100448
449 metadata_list = []
450 for name, buffer in self.nng.metadata:
451 self.buffers_to_write.append(buffer)
452 metadata_list.append((name, len(self.buffers_to_write) - 1))
Tim Hall79d07d22020-04-27 18:20:16 +0100453
454 buffers_offset = self.write_offset_vector([self.serialise_buffer(buf) for buf in self.buffers_to_write])
Tim Hall79d07d22020-04-27 18:20:16 +0100455 metadata_offset = self.write_offset_vector([self.serialise_metadata(metadata) for metadata in metadata_list])
456
457 Model.ModelStart(builder)
458 Model.ModelAddVersion(builder, tflite_version)
459 Model.ModelAddOperatorCodes(builder, operator_code_offset)
460 Model.ModelAddSubgraphs(builder, subgraph_offset)
461 Model.ModelAddDescription(builder, description)
462 Model.ModelAddBuffers(builder, buffers_offset)
463 Model.ModelAddMetadata(builder, metadata_offset)
464 return Model.ModelEnd(builder)
465
466 def serialise(self):
467
468 model = self.serialise_model()
469
470 self.builder.FinishWithFileIdentifier(model, tflite_file_identifier)
471
472 return self.builder.Output()
473
474 def write(self, filename):
475 with open(self.filename, "wb") as f:
476 f.write(self.serialised_buf)
477
478
479def write_tflite(nng, filename):
480 writer = TFLiteSerialiser(nng)
481 buf = writer.serialise()
482
483 with open(filename, "wb") as f:
484 f.write(buf)