tf-metadata.py 8.7 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263
  1. #!/usr/bin/env python
  2. from __future__ import unicode_literals
  3. import json
  4. import io
  5. import sys
  6. import os
  7. from tensorflow.core.framework import api_def_pb2
  8. from tensorflow.core.framework import op_def_pb2
  9. from google.protobuf import text_format
  10. categories = {
  11. 'Const': 'Constant',
  12. 'Conv2D': 'Layer',
  13. 'BiasAdd': 'Layer',
  14. 'DepthwiseConv2dNative': 'Layer',
  15. 'Relu': 'Activation',
  16. 'Relu6': 'Activation',
  17. 'Softmax': 'Activation',
  18. 'Sigmoid': 'Activation',
  19. 'LRN': 'Normalization',
  20. 'MaxPool': 'Pool',
  21. 'MaxPoolV2': 'Pool',
  22. 'AvgPool': 'Pool',
  23. 'Reshape': 'Shape',
  24. 'Squeeze': 'Shape',
  25. 'ConcatV2': 'Tensor',
  26. 'Split': 'Tensor',
  27. 'Dequantize': 'Tensor',
  28. 'Identity': 'Control',
  29. 'Variable': 'Control',
  30. 'VariableV2': 'Control',
  31. 'Assign': 'Control',
  32. 'BatchNormWithGlobalNormalization': 'Normalization',
  33. 'FusedBatchNorm': 'Normalization',
  34. # 'VariableV2':
  35. # 'Assign':
  36. # 'BiasAdd':
  37. }
  38. def find_multiline(line, colon):
  39. if colon == -1:
  40. return None
  41. line = line[colon+1:]
  42. while line.startswith(' '):
  43. line = line[1:]
  44. if line.startswith('<<'):
  45. line = line[2:]
  46. return line
  47. return None
  48. def str_escape(text):
  49. return text.replace('\n', '\\n').replace('\"', '\\\"')
  50. def pbtxt_from_multiline(multiline_pbtxt):
  51. pbtxt = ''
  52. while len(multiline_pbtxt) > 0:
  53. index = multiline_pbtxt.find('\n')
  54. if index == -1:
  55. pbtxt = pbtxt + multiline_pbtxt
  56. multiline_pbtxt = ''
  57. break
  58. line = multiline_pbtxt[0:index]
  59. multiline_pbtxt = multiline_pbtxt[index+1:]
  60. colon = line.find(':')
  61. end = find_multiline(line, colon)
  62. if end == None:
  63. pbtxt = pbtxt + line + '\n'
  64. continue
  65. pbtxt = pbtxt + line[0:colon+1]
  66. unescaped = ''
  67. newline = False
  68. line = ''
  69. while len(multiline_pbtxt) > 0:
  70. index = multiline_pbtxt.find('\n')
  71. line = multiline_pbtxt[0:index]
  72. multiline_pbtxt = multiline_pbtxt[index+1:]
  73. if line.startswith(end):
  74. line = line[len(end):]
  75. break
  76. if newline:
  77. unescaped = unescaped + '\n'
  78. newline = True
  79. unescaped = unescaped + line
  80. line = ''
  81. pbtxt = pbtxt + '\"' + str_escape(unescaped) + '\"' + line + '\n'
  82. return pbtxt
  83. def read_api_def_map(folder):
  84. api_def_map = {}
  85. file_list = os.listdir(folder)
  86. file_list = sorted(file_list)
  87. for filename in file_list:
  88. api_defs = api_def_pb2.ApiDefs()
  89. filename = folder + '/' + filename
  90. with open(filename) as handle:
  91. multiline_pbtxt = handle.read()
  92. pbtxt = pbtxt_from_multiline(multiline_pbtxt)
  93. text_format.Merge(pbtxt, api_defs)
  94. for api_def in api_defs.op:
  95. api_def_map[api_def.graph_op_name] = api_def
  96. return api_def_map
  97. def convert_type(type):
  98. return { 'type': 'type', 'value': type }
  99. def convert_tensor(tensor):
  100. return { 'type': 'tensor', 'value': '?' }
  101. def convert_shape(shape):
  102. return { 'type': 'shape', 'value': '?' }
  103. def convert_number(number):
  104. if number == float('inf'):
  105. return 'NaN'
  106. if number == float('-inf'):
  107. return '-NaN'
  108. return number
  109. def convert_attr_value(attr_value):
  110. if attr_value.HasField('list'):
  111. list = []
  112. attr_value_list = attr_value.list
  113. if len(attr_value_list.s) > 0:
  114. for s in attr_value_list.s:
  115. list.append(s)
  116. if len(attr_value_list.i) > 0:
  117. for i in attr_value_list.i:
  118. list.append(i)
  119. if len(attr_value_list.f) > 0:
  120. for f in attr_value_list.f:
  121. list.append(convert_number(f))
  122. if len(attr_value_list.type) > 0:
  123. for type in attr_value_list.type:
  124. list.append(convert_type(type))
  125. if len(list) == 0:
  126. for _, value in attr_value_list.ListFields():
  127. if len(value) > 0:
  128. raise Exception()
  129. return list
  130. if attr_value.HasField('s'):
  131. return attr_value.s
  132. if attr_value.HasField('i'):
  133. return attr_value.i
  134. if attr_value.HasField('f'):
  135. return convert_number(attr_value.f)
  136. if attr_value.HasField('b'):
  137. return attr_value.b
  138. if attr_value.HasField('type'):
  139. return convert_type(attr_value.type)
  140. if attr_value.HasField('tensor'):
  141. return convert_tensor(attr_value.tensor)
  142. if attr_value.HasField('shape'):
  143. return convert_shape(attr_value.shape)
  144. raise Exception()
  145. api_def_map = read_api_def_map('../third_party/tensorflow/tensorflow/core/api_def/base_api')
  146. input_file = '../third_party/tensorflow/tensorflow/core/ops/ops.pbtxt';
  147. ops_list = op_def_pb2.OpList()
  148. with open(input_file) as input_handle:
  149. text_format.Merge(input_handle.read(), ops_list)
  150. json_root = []
  151. for op in ops_list.op:
  152. # print(op.name)
  153. json_schema = {}
  154. if op.name in categories:
  155. json_schema['category'] = categories[op.name]
  156. api_def = api_def_pb2.ApiDef()
  157. if op.name in api_def_map:
  158. api_def = api_def_map[op.name]
  159. # if op.deprecation.version != 0:
  160. # print('[' + op.name + ']')
  161. # print(op.deprecation.version)
  162. # print(op.deprecation.explanation)
  163. api_def_attr_map = {}
  164. for attr in api_def.attr:
  165. api_def_attr_map[attr.name] = attr
  166. api_def_in_arg_map = {}
  167. for in_arg in api_def.in_arg:
  168. api_def_in_arg_map[in_arg.name] = in_arg
  169. api_def_out_arg_map = {}
  170. for out_arg in api_def.out_arg:
  171. api_def_out_arg_map[out_arg.name] = out_arg
  172. if api_def.summary:
  173. json_schema['summary'] = api_def.summary
  174. if api_def.description:
  175. json_schema['description'] = api_def.description
  176. for attr in op.attr:
  177. if not 'attributes' in json_schema:
  178. json_schema['attributes'] = []
  179. json_attribute = {}
  180. json_attribute['name'] = attr.name
  181. if attr.type:
  182. json_attribute['type'] = attr.type
  183. if attr.name in api_def_attr_map:
  184. api_def_attr = api_def_attr_map[attr.name]
  185. if api_def_attr.description:
  186. json_attribute['description'] = api_def_attr.description
  187. if attr.has_minimum:
  188. json_attribute['minimum'] = attr.minimum
  189. if attr.HasField('allowed_values'):
  190. json_attribute['allowedValues'] = convert_attr_value(attr.allowed_values)
  191. if attr.HasField('default_value'):
  192. json_attribute['defaultValue'] = convert_attr_value(attr.default_value)
  193. json_schema['attributes'].append(json_attribute)
  194. for input_arg in op.input_arg:
  195. if not 'inputs' in json_schema:
  196. json_schema['inputs'] = []
  197. json_input = {}
  198. json_input['name'] = input_arg.name
  199. if input_arg.name in api_def_in_arg_map:
  200. api_def_in_arg = api_def_in_arg_map[input_arg.name]
  201. if api_def_in_arg.description:
  202. json_input['description'] = api_def_in_arg.description
  203. if input_arg.number_attr:
  204. json_input['numberAttr'] = input_arg.number_attr
  205. if input_arg.type:
  206. json_input['type'] = input_arg.type
  207. if input_arg.type_attr:
  208. json_input['typeAttr'] = input_arg.type_attr
  209. if input_arg.type_list_attr:
  210. json_input['typeListAttr'] = input_arg.type_list_attr
  211. if input_arg.is_ref:
  212. json_input['isRef'] = True
  213. json_schema['inputs'].append(json_input)
  214. for output_arg in op.output_arg:
  215. if not 'outputs' in json_schema:
  216. json_schema['outputs'] = []
  217. json_output = {}
  218. json_output['name'] = output_arg.name
  219. if output_arg.name in api_def_out_arg_map:
  220. api_def_out_arg = api_def_out_arg_map[output_arg.name]
  221. if api_def_out_arg.description:
  222. json_output['description'] = api_def_out_arg.description
  223. if output_arg.type:
  224. json_output['type'] = output_arg.type
  225. elif output_arg.type_attr:
  226. json_output['typeAttr'] = output_arg.type_attr
  227. elif output_arg.type_list_attr:
  228. json_output['typeListAttr'] = output_arg.type_list_attr
  229. if output_arg.is_ref:
  230. json_output['isRef'] = True
  231. json_schema['outputs'].append(json_output)
  232. json_root.append({
  233. 'name': op.name,
  234. 'schema': json_schema
  235. })
  236. json_file = '../src/tf-metadata.json'
  237. with io.open(json_file, 'w', newline='') as fout:
  238. json_data = json.dumps(json_root, sort_keys=True, indent=2)
  239. for line in json_data.splitlines():
  240. line = line.rstrip()
  241. if sys.version_info[0] < 3:
  242. line = unicode(line)
  243. fout.write(line)
  244. fout.write('\n')