From 1e85c429ceaad860aba16d1f518160d263c094c0 Mon Sep 17 00:00:00 2001
From: lyg <1543117173@qq.com>
Date: 星期二, 08 四月 2025 11:48:52 +0800
Subject: [PATCH] 生成指令帧和包格式结构

---
 db_struct_flow.py | 1218 +++++++++++++++++++++++++++++++++++++++++++++------------
 1 files changed, 961 insertions(+), 257 deletions(-)

diff --git a/db_struct_flow.py b/db_struct_flow.py
index ade1e21..bf2817a 100644
--- a/db_struct_flow.py
+++ b/db_struct_flow.py
@@ -1,20 +1,52 @@
 import os
+import time
 from datetime import datetime
 
 from openai import OpenAI
-from pathlib import Path
 import re
 import json
-import copy
 
-from datas import pkt_vc, pkt_datas, dev_pkt, proj_data
-from db.db_generate import create_project, create_device, create_data_stream
-from db.models import TProject, TDevice
+import data_templates
+from knowledgebase.db.db_helper import create_project, create_device, create_data_stream, \
+    update_rule_enc, create_extend_info, create_ref_ds_rule_stream, create_ins_format
+from knowledgebase.db.data_creator import create_prop_enc, create_enc_pkt, get_data_ty, create_any_pkt
+
+from knowledgebase.db.models import TProject
+
+file_map = {
+    "鏂囨。鍚堝苟": "./doc/鏂囨。鍚堝苟.md",
+    "閬ユ祴婧愬寘璁捐鎶ュ憡": "./doc/XA-5D鏃犱汉鏈哄垎绯荤粺鎺㈡祴婧愬寘璁捐鎶ュ憡锛堝叕寮�锛�.md",
+    "閬ユ祴澶х翰": "./doc/XA-5D鏃犱汉鏈烘帰娴嬪ぇ绾诧紙鍏紑锛�.md",
+    "鎬荤嚎浼犺緭閫氫俊甯у垎閰�": "./doc/XA-5D鏃犱汉鏈�1314A鎬荤嚎浼犺緭閫氫俊甯у垎閰嶏紙鍏紑锛�.md",
+    "搴旂敤杞欢鐢ㄦ埛闇�姹�": "./doc/XA-5D鏃犱汉鏈鸿蒋浠剁敤鎴烽渶姹傦紙鍏紑锛�.docx.md",
+    "鎸囦护鏍煎紡": "./doc/ZL鏍煎紡(鍏紑).docx.md"
+}
+# file_map = {
+#     "閬ユ祴婧愬寘璁捐鎶ュ憡": "./docs/HY-4A鏁扮鍒嗙郴缁熼仴娴嬫簮鍖呰璁℃姤鍛� Z 240824 鏇存敼3(鍐呴儴) .docx.md",
+#     "閬ユ祴澶х翰": "./docs/HY-4A鍗槦閬ユ祴澶х翰 Z 240824 鏇存敼3锛堝唴閮級.docx.md",
+#     "鎬荤嚎浼犺緭閫氫俊甯у垎閰�": "./docs/HY-4A鍗槦1553B鎬荤嚎浼犺緭閫氫俊甯у垎閰� Z 240824 鏇存敼3锛堝唴閮級.docx.md",
+#     "搴旂敤杞欢鐢ㄦ埛闇�姹�": "./docs/HY-4A鏁扮鍒嗙郴缁熷簲鐢ㄨ蒋浠剁敤鎴烽渶姹傦紙鏄熷姟绠$悊鍒嗗唽锛� Z 240831 鏇存敼4锛堝唴閮級.docx.md"
+# }
+# file_map = {
+#     "鏂囨。鍚堝苟": "./doc/鏂囨。鍚堝苟.md",
+#     "閬ユ祴婧愬寘璁捐鎶ュ憡": "./doc/XA-5D鏃犱汉鏈哄垎绯荤粺鎺㈡祴婧愬寘璁捐鎶ュ憡锛堝叕寮�锛�.md",
+#     "閬ユ祴澶х翰": "./doc/XA-5D鏃犱汉鏈烘帰娴嬪ぇ绾诧紙鍏紑锛�.md",
+#     "鎬荤嚎浼犺緭閫氫俊甯у垎閰�": "./doc/XA-5D鏃犱汉鏈�1314A鎬荤嚎浼犺緭閫氫俊甯у垎閰嶏紙鍏紑锛�.md"
+# }
 
 BASE_URL = 'https://dashscope.aliyuncs.com/compatible-mode/v1'
 API_KEY = 'sk-15ecf7e273ad4b729c7f7f42b542749e'
-MODEL_NAME = 'qwen-long'
+MODEL_NAME = 'qwen2.5-14b-instruct-1m'
 
+# BASE_URL = 'http://10.74.15.164:11434/v1/'
+# API_KEY = 'ollama'
+# MODEL_NAME = 'qwen2.5:32b-128k'
+
+# BASE_URL = 'http://10.74.15.164:1001/api'
+# API_KEY = 'sk-a909385bc14d4491a718b6ee264c3227'
+# MODEL_NAME = 'qwen2.5:32b-128k'
+
+USE_CACHE = True
 assistant_msg = """
 # 瑙掕壊
 浣犳槸涓�涓笓涓氱殑鏂囨。閫氫俊鍒嗘瀽甯堬紝鎿呴暱杩涜鏂囨。鍒嗘瀽鍜岄�氫俊鍗忚鍒嗘瀽锛屽悓鏃惰兘澶熻В鏋� markdown 绫诲瀷鐨勬枃妗c�傛嫢鏈夋垚鐔熷噯纭殑鏂囨。闃呰涓庡垎鏋愯兘鍔涳紝鑳藉濡ュ杽澶勭悊澶氭枃妗i棿瀛樺湪寮曠敤鍏崇郴鐨勫鏉傛儏鍐点��
@@ -25,64 +57,259 @@
 2. 鍒嗘瀽鏂囨。鐨勭粨鏋勩�佷富棰樺拰閲嶇偣鍐呭锛屽悓鏍峰彧渚濇嵁鏂囨。杩涜琛ㄨ堪銆�
 3. 濡傛灉鏂囨。闂村瓨鍦ㄥ紩鐢ㄥ叧绯伙紝姊崇悊寮曠敤鑴夌粶锛屾槑纭悇鏂囨。涔嬮棿鐨勫叧鑱旓紝涓斾粎鍛堢幇鏂囨。涓綋鐜扮殑鍐呭銆�
 
-
 ### 鎶�鑳� 2锛氶�氫俊鍗忚鍒嗘瀽
 1. 鎺ユ敹閫氫俊鍗忚鐩稿叧淇℃伅锛岀悊瑙e崗璁殑瑙勫垯鍜屾祦绋嬶紝浠呬緷鎹墍缁欎俊鎭繘琛屽垎鏋愩��
 
+## 鑳屾櫙鐭ヨ瘑
+###杞欢涓昏鍔熻兘涓庤繍琛屾満鍒舵�荤粨濡備笅锛�
+1. 鏁版嵁閲囬泦鍜屽鐞嗭細
+   DIU璐熻矗鏍规嵁鍗槦鐨勫伐浣滅姸鎬佹垨妯″紡鎻愪緵閬ユ祴鏁版嵁锛屽寘鎷ā鎷熼噺锛圓N锛夈�佹�荤嚎淇″彿锛圔L锛変互鍙婃俯搴︼紙TH锛夊拰鏁板瓧閲忥紙DS锛夛紝骞跺皢杩欎簺淇℃伅鎵撳寘锛岄�氳繃鎬荤嚎鍙戦�佺粰SMU銆�
+   SMU鍒欐敹闆嗙‖閫氶亾涓婄殑閬ユ祴鍙傛暟锛屽苟閫氳繃鎬荤嚎鎺ユ敹DIU閲囬泦鐨勪俊鎭��
+2. 澶氳矾澶嶇敤涓庢暟鎹紶杈擄細
+   閬ユ祴婧愬寘琚粍缁囨垚E-PDU锛岃繘涓�姝ュ鐢ㄤ负M-PDU锛屽苟濉厖鍒癡CDU涓瀯鎴愰仴娴嬪抚銆�
+   鍒╃敤CCSDS AOS CADU鏍煎紡杩涜閬ユ祴鏁版嵁鐨勫璺鐢ㄥ拰浼犺緭銆�
+3. 铏氭嫙淇¢亾锛圴C锛夎皟搴︽満鍒讹細
+   閫氳繃甯歌閬ユ祴VC銆佺獊鍙戞暟鎹甐C銆佸欢鏃堕仴娴媀C銆佽褰曟暟鎹甐C浠ュ強鍥炴斁VC瀹炵幇涓嶅悓绫诲瀷鐨勬暟鎹笅浼犮��
+4. 閬ユ帶鎸囦护澶勭悊锛�
+   涓婅閬ユ帶鍖呮嫭鐩存帴鎸囦护鍜岄棿鎺ユ寚浠わ紝闇�缁忚繃鏍煎紡楠岃瘉鍚庤浆鍙戠粰鐩稿簲鍗曟満鎵ц銆�
+   閬ユ帶甯ч�氳繃鐗瑰畾鐨勮櫄鎷熶俊閬擄紙VC锛夎繘琛屼紶杈撱��
+杩欎簺鐭ヨ瘑闇�瑕佷綘璁颁綇锛屽啀鍚庣画鐨勫鐞嗕腑鍙互甯姪浣犵悊瑙h澶勭悊鐨勬暟鎹��
+
 ## 鐩爣瀵煎悜
 1. 閫氳繃瀵规枃妗e拰閫氫俊鍗忚鐨勫垎鏋愶紝涓虹敤鎴锋彁渚涙竻鏅般�佸噯纭殑鏁版嵁缁撴瀯锛屽府鍔╃敤鎴锋洿濂藉湴鐞嗚В鍜屼娇鐢ㄧ浉鍏充俊鎭��
-2. 浠� JSON 鏍煎紡缁勭粐杈撳嚭鍐呭锛岀‘淇濇暟鎹粨鏋勭殑瀹屾暣鎬у拰鍙鎬с��
 
 ## 瑙勫垯
-1. 姣忎竴涓瀷鍙烽兘浼氭湁涓�濂楁枃妗o紝闇�鍑嗙‘鍒ゆ柇鏄惁涓哄悓涓�涓瀷鍙风殑鏂囨。鍚庡啀杩涜鏁翠綋鍒嗘瀽銆�
-2. 姣忔鍙垎鏋愬悓涓�涓瀷鍙枫��
-3. 澶у鏁版枃妗g粨鏋勪负锛氬瀷鍙蜂笅鍖呭惈璁惧锛岃澶囦笅鍖呭惈鏁版嵁娴侊紝鏁版嵁娴佷笅鍖呭惈鏁版嵁甯э紝鏁版嵁甯т腑鏈変竴鍧楁槸鍖呭煙锛屽寘鍩熶腑浼氭寕杞藉悇绉嶇被鍨嬬殑鏁版嵁鍖呫��
-4. 杩欎簺鏂囨。閮芥槸鏁版嵁浼犺緭鍗忚鐨勬弿杩帮紝鍦ㄦ暟鎹祦銆佹暟鎹抚銆佹暟鎹寘绛変紶杈撳疄浣撲腑閮芥弿杩颁簡鍚勪釜瀛楁鐨勫垎甯冨拰姣忎釜瀛楁鐨勫ぇ灏忥紝涓斿ぇ灏忓崟浣嶄笉缁熶竴锛岄渶鐞嗚В杩欎簺鍗曚綅锛屽苟灏嗘墍鏈夎緭鍑哄崟浣嶇粺涓�涓� bits锛岀粺涓�浣跨敤length琛ㄧず銆�
-5. 濡傛灉鏈夊眰绾э紝浣跨敤鏍戝舰 JSON 杈撳嚭锛屽瓙鑺傜偣 key 浣跨敤children锛涢渶淇濊瘉鐩稿悓绫诲瀷鐨勬暟鎹粨鏋勭粺涓�锛屽苟涓斿垽鏂瘡涓眰绾ф槸浠�涔堢被鍨嬶紝杈撳嚭绫诲瀷瀛楁锛岀被鍨嬪瓧娈电殑 key 浣跨敤 type 锛涗緥濡傚綋鍓嶅眰绾т负瀛楁鏃朵娇鐢細type:"field"锛涘綋鍓嶅眰绾т负璁惧鏃朵娇鐢細type:"device"
-6.鍚嶇О鐩稿叧鐨勫瓧娈电殑 key 浣跨敤name锛涗唬鍙锋垨鑰呭敮涓�鏍囪瘑鐩稿叧鐨勫瓧娈电殑key浣跨敤id锛涘簭鍙风浉鍏崇殑瀛楁鐨刱ey浣跨敤number锛涘叾浠栨病鏈変妇渚嬬殑瀛楁浣跨敤绮剧畝鐨勭炕璇戜綔涓哄瓧娈电殑key锛�
-7.鎺㈡祴甯т负CADU锛屽叾涓寘鍚悓姝ュご鍜孷CDU锛屾寜鐓т範鎯渶瑕佷娇鐢╒CDU灞傜骇鍖呭惈涓嬩竴灞傜骇涓紶杈撳抚涓诲澶淬�佷紶杈撳抚鎻掑叆鍩熴�佷紶杈撳抚鏁版嵁鍩熴�佷紶杈撳抚灏剧殑缁撴瀯
-
+1. 姣忎竴涓瀷鍙烽兘浼氭湁涓�濂楁枃妗o紝闇�鍑嗙‘鍒ゆ柇鏄惁涓哄悓涓�涓瀷鍙风殑鏂囨。鍚庡啀杩涜鏁翠綋鍒嗘瀽锛屾瘡娆″彧鍒嗘瀽鍚屼竴涓瀷鍙风殑鏂囨。銆�
+2. 澶у鏁版枃妗g粨鏋勪负锛氬瀷鍙蜂笅鍖呭惈璁惧锛岃澶囦笅鍖呭惈鏁版嵁娴侊紝鏁版嵁娴佷笅鍖呭惈鏁版嵁甯э紝鏁版嵁甯т腑鏈変竴鍧楁槸鍖呭煙锛屽寘鍩熶腑浼氭寕杞藉悇绉嶇被鍨嬬殑鏁版嵁鍖呫��
+3. 鏂囨。閮芥槸瀵逛簬鏁版嵁浼犺緭鍗忚鐨勬弿杩帮紝鍦ㄦ暟鎹祦銆佹暟鎹抚銆佹暟鎹寘绛変紶杈撳疄浣撲腑閮芥弿杩颁簡鍚勪釜瀛楁鐨勫垎甯冦�佸悇涓瓧娈电殑澶у皬鍜屼綅缃瓑淇℃伅锛屼笖澶у皬鍗曚綅涓嶇粺涓�锛岄渶鐞嗚В杩欎簺鍗曚綅锛屽苟灏嗘墍鏈夎緭鍑哄崟浣嶇粺涓�涓� bits锛岄暱搴﹀瓧娈典娇鐢� length 琛ㄧず锛屼綅缃瓧娈典娇鐢� pos 琛ㄧず锛屽鏋滀负鍙橀暱浣跨敤鈥�"鍙橀暱"鈥濊〃绀恒��
+4. 濡傛灉鏈夊眰绾э紝浣跨敤鏍戝舰 JSON 杈撳嚭锛屽鏋滄湁瀛愯妭鐐癸紝瀛愯妭鐐� key 浣跨敤children锛涢渶淇濊瘉涓�娆¤緭鍑虹殑鏁版嵁缁撴瀯缁熶竴锛屽苟涓斿垽鏂瘡涓眰绾ф槸浠�涔堢被鍨嬶紝杈撳嚭绫诲瀷瀛楁锛坱ype锛夛紝绫诲瀷瀛楁鐨� key 浣跨敤 type锛岀被鍨嬪寘鎷細鍨嬪彿锛坧roject锛夈�佽澶囷紙dev锛夈�佸皝瑁呭寘锛坋nc锛夈�佺嚎鎬у寘锛坙inear锛夈�佸弬鏁帮紙para锛夛紝灏佽鍖呭瓙绾ф湁鏁版嵁鍖咃紝鎵�浠ype涓篹nc锛岀嚎鎬у寘瀛愮骇鍙湁鍙傛暟锛屾墍浠ype涓簂inear锛涙瘡涓眰绾ч兘鍖呭惈鍋忕Щ浣嶇疆锛坧os锛夛紝姣忎釜灞傜骇鐨勫亸绉讳綅缃粠0寮�濮嬨��
+5. 鍚嶇О鐩稿叧鐨勫瓧娈电殑 key 浣跨敤name锛涗唬鍙枫�佺紪鍙锋垨鑰呭敮涓�鏍囪瘑鐩稿叧鐨勫瓧娈电殑key浣跨敤id锛宨d鐢辨暟瀛椼�佽嫳鏂囧瓧姣嶃�佷笅鍒掔嚎缁勬垚涓斾互鑻辨枃瀛楁瘝寮�澶达紝闀垮害灏介噺绠�鐭紱搴忓彿鐩稿叧鐨勫瓧娈电殑key浣跨敤number锛涘亸绉讳綅缃浉鍏冲瓧娈电殑key浣跨敤pos锛涘叾浠栨病鏈変妇渚嬬殑瀛楁浣跨敤绮剧畝鐨勭炕璇戜綔涓哄瓧娈电殑key锛涙瘡涓粨鏋勫繀椤诲寘鍚玭ame鍜宨d銆�
+6. 閬ユ祴甯т负CADU锛屽叾涓寘鍚悓姝ュご鍜孷CDU锛屾寜鐓т範鎯渶瑕佷娇鐢╒CDU灞傜骇宓屽浼犺緭甯т富瀵煎ご銆佷紶杈撳抚鎻掑叆鍩熴�佷紶杈撳抚鏁版嵁鍩熴�佷紶杈撳抚灏剧殑缁撴瀯銆�
+7. 鏁版嵁鍖呭瓧娈靛寘鎷細name銆乮d銆乼ype銆乸os銆乴ength銆乧hildren锛涘弬鏁板瓧娈靛寘鎷細name銆乮d銆乸os銆乼ype銆乴ength锛涘繀椤诲寘鍚玴os鍜宭ength瀛楁銆�
+8. 甯哥敤id鍙傝�冿細閬ユ祴锛圱M锛夈�侀仴鎺э紙TC锛夈�佹�荤嚎锛圔US锛夈�佺増鏈彿锛圴er锛夈�佸簲鐢ㄨ繃绋嬫爣璇嗭紙APID锛夈��
+9. 娉ㄦ剰锛氫竴瀹氳璁板緱morkdown鏂囨。涓細灏嗕竴浜涚壒娈婂瓧绗﹁繘琛岃浆涔夛紝浠ユ鏉ヤ繚璇佹枃妗g殑姝g‘鎬э紝杩欎簺杞箟绗﹀彿锛堜篃灏辨槸鍙嶆枩鏉犫�榎鈥欙級涓嶉渶瑕佸湪缁撴灉涓緭鍑恒��
+10. 浠� JSON 鏍煎紡缁勭粐杈撳嚭鍐呭锛岀‘淇濇暟鎹粨鏋勭殑瀹屾暣鎬у拰鍙鎬э紝娉ㄦ剰锛氱敓鎴愮殑JSON璇硶鏍煎紡蹇呴』绗﹀悎json瑙勮寖锛岄伩鍏嶅嚭鐜伴敊璇��
+    
 ## 闄愬埗锛�
 - 鎵�杈撳嚭鐨勫唴瀹瑰繀椤绘寜鐓SON鏍煎紡杩涜缁勭粐锛屼笉鑳藉亸绂绘鏋惰姹傦紝涓斾弗鏍奸伒寰枃妗e唴瀹硅繘琛岃緭鍑猴紝鍙緭鍑� JSON 锛屼笉瑕佽緭鍑哄叾瀹冩枃瀛椼��
-- 涓嶈緭鍑轰换浣曟敞閲婄瓑鎻忚堪鎬т俊鎭�
-
+- 涓嶈緭鍑轰换浣曟敞閲婄瓑鎻忚堪鎬т俊鎭��
 """
+
+g_completion = None
+
+
+def read_from_file(cache_file):
+    with open(cache_file, 'r', encoding='utf-8') as f:
+        text = f.read()
+    return text
+
+
+def save_to_file(text, file_path):
+    if USE_CACHE:
+        with open(file_path, 'w', encoding='utf-8') as f:
+            f.write(text)
+
+
+json_pat = re.compile(r'```json(.*?)```', re.DOTALL)
+
+
+def remove_markdown(text):
+    # 浣跨敤姝e垯琛ㄨ揪寮忔彁鍙杍son鏂囨湰
+    try:
+        return json_pat.findall(text)[0]
+    except IndexError:
+        return text
+
+
+def rt_pkt_map_gen(pkt, trans_ser, rt_pkt_map, pkt_id, vals):
+    # 閫昏緫灏佽鍖咃紝鏁版嵁鍧椾紶杈撶殑鍙湁涓�涓紝鍙栨暟鐨勬牴鎹甊T鍦板潃銆佸瓙鍦板潃鍜屽抚鍙峰垝鍒�
+    frame_num = pkt['frameNum']
+    if trans_ser == '鏁版嵁鍧椾紶杈�':
+        # 鏁版嵁鍧椾紶杈撴牴鎹甊T鍦板潃鍜屽瓙鍦板潃鍒掑垎
+        key = f'{pkt["rt"]}_{pkt["subAddr"]}'
+        name = f'{pkt["rt"]}_{pkt["subAddr"]}_{trans_ser}'
+    else:
+        # 鍙栨暟鏍规嵁RT鍦板潃銆佸瓙鍦板潃鍜屽抚鍙峰垝鍒�
+        key = f'{pkt["rt"]}_{pkt["subAddr"]}_{pkt["frameNum"]}'
+        name = f'{pkt["rt"]}_{pkt["subAddr"]}_甯у彿{frame_num}_{trans_ser}'
+    #
+    if key not in rt_pkt_map:
+        rt_pkt_map[key] = {
+            "name": name,
+            "id": pkt_id,
+            "type": "logic",
+            "pos": 0,
+            "content": "CYCLEBUFFER,Message,28,0xFFFF",
+            "length": "",
+            "vals": vals,
+            "children": []
+        }
+    frame = f'{pkt["frameNum"]}'
+
+    interval = f'{pkt["interval"]}'.replace(".", "_")
+    if trans_ser == '鍙栨暟':
+        _key = f'RT{pkt["rtAddr"]}Frame{frame.replace("|", "_")}_Per{interval}'
+    else:
+        # 鏁版嵁鍧椾紶杈�
+        if pkt['burst']:
+            _key = f'RT{pkt["rtAddr"]}FrameALL'
+        else:
+            _key = f'RT{pkt["rtAddr"]}Frame{frame}Per{interval}'
+
+    _pkt = next(filter(lambda it: it['name'] == _key, rt_pkt_map[key]['children']), None)
+    if _pkt is None:
+        ext_info = None
+        if trans_ser == '鏁版嵁鍧椾紶杈�' and not pkt['burst']:
+            # 鏁版嵁鍧椾紶杈撲笖鏈夊懆鏈熺殑鍖呴渶瑕�
+            ext_info = [{"id": "PeriodTriger", "name": "鏃跺垎澶嶇敤鎬荤嚎瑙﹀彂灞炴��", "val": f"{pkt['interval']}"},
+                        {"id": "FrameNumber", "name": "鏃跺垎澶嶇敤鍗忚甯у彿", "val": frame}]
+        _pkt = {
+            "name": _key,
+            "id": _key,
+            "type": "enc",
+            "pos": 0,
+            "content": "1:N;EPDU",
+            "length": "length",
+            "extInfo": ext_info,
+            "children": [
+                {
+                    "id": "C02_ver",
+                    "name": "閬ユ祴鐗堟湰",
+                    "type": "para",
+                    "pos": 0,
+                    "length": 3,
+                    "dataTy": "INVAR",
+                    "content": "0"
+                },
+                {
+                    "id": "C02_type",
+                    "name": "绫诲瀷",
+                    "type": "para",
+                    "pos": 3,
+                    "length": 1,
+                    "dataTy": "INVAR",
+                    "content": "0"
+                },
+                {
+                    "id": "C02_viceHead",
+                    "name": "鍓澶存爣璇�",
+                    "type": "para",
+                    "pos": 4,
+                    "length": 1,
+                    "content": "1",
+                    "dataTy": "INVAR"
+                },
+                {
+                    "id": "C02_PackSign",
+                    "name": "APID",
+                    "type": "para",
+                    "pos": 5,
+                    "length": 11,
+                    "is_key": True,
+                    "dataTy": "ENUM"
+                },
+                {
+                    "id": "C02_SerCtr_1",
+                    "name": "搴忓垪鏍囪",
+                    "type": "para",
+                    "pos": 16,
+                    "length": 2,
+                    "content": "3"
+                },
+                {
+                    "id": "C02_SerCtr_2",
+                    "name": "鍖呭簭璁℃暟",
+                    "type": "para",
+                    "pos": 18,
+                    "length": 14,
+                    "content": "0:167772:1",
+                    "dataTy": "INCREASE"
+                },
+                {
+                    "id": "C02_PackLen",
+                    "name": "鍖呴暱",
+                    "type": "para",
+                    "pos": 32,
+                    "length": 16,
+                    "content": "1Bytes/C02_Data.length+1",
+                    "dataTy": "LEN"
+                },
+                {
+                    "id": "C02_Ser",
+                    "name": "鏈嶅姟",
+                    "type": "para",
+                    "pos": 48,
+                    "length": 8,
+                    "is_key": True,
+                    "dataTy": "ENUM"
+                },
+                {
+                    "id": "C02_SubSer",
+                    "name": "瀛愭湇鍔�",
+                    "type": "para",
+                    "pos": 56,
+                    "length": 8,
+                    "is_key": True,
+                    "dataTy": "ENUM"
+                },
+                {
+                    "id": "C02_Data",
+                    "name": "鏁版嵁鍖�",
+                    "type": "linear",
+                    "pos": 64,
+                    "length": 'length-current',
+                    "children": []
+                },
+            ]
+        }
+        rt_pkt_map[key]['children'].append(_pkt)
+    # 鏁版嵁鍖轰笅闈㈢殑鍖�
+    data_area = next(filter(lambda it: it['name'] == '鏁版嵁鍖�', _pkt['children']), None)
+    ser_sub_ser: str = pkt['service']
+    ser = ''
+    sub_ser = ''
+    if ser_sub_ser:
+        nums = re.findall(r'\d+', ser_sub_ser)
+        if len(nums) == 2:
+            ser = nums[0]
+            sub_ser = nums[1]
+    if 'children' not in pkt:
+        pkt['children'] = []
+    p_name = pkt['id'] + '_' + pkt['name']
+
+    data_area['children'].append({
+        "name": p_name,
+        "id": pkt["id"],
+        "type": "linear",
+        "pos": 0,
+        "length": pkt["length"],
+        "vals": f"0x{pkt['apid']}/{ser}/{sub_ser}/",
+        "children": pkt['children'],
+    })
+
+
+def build_vcid_content(vcs):
+    _vcs = []
+    for vc in vcs:
+        _vcs.append(vc['name'] + ',' + vc['VCID'])
+    return ' '.join(_vcs)
 
 
 class DbStructFlow:
-    files = []
-    file_objects = []
+    # 宸ョ▼
+    proj: TProject = None
+    # 閬ユ祴婧愬寘鍒楄〃锛屼粎鍖呭悕绉般�佸寘id鍜宧asParams
+    tm_pkts = []
+    # vc婧愬寘
+    vc_pkts = []
 
-    def __init__(self, doc_files):
+    def __init__(self):
         self.client = OpenAI(
             api_key=API_KEY,
             base_url=BASE_URL,
             # api_key="ollama",
             # base_url="http://192.168.1.48:11434/v1/",
         )
-        if doc_files:
-            self.files = doc_files
-        self.load_file_objs()
-        self.delete_all_files()
-        self.upload_files()
-
-    def load_file_objs(self):
-        file_stk = self.client.files.list()
-        self.file_objects = file_stk.data
-
-    def delete_all_files(self):
-        for file_object in self.file_objects:
-            self.client.files.delete(file_object.id)
-
-    def upload_file(self, file_path):
-        file_object = self.client.files.create(file=Path(file_path), purpose="file-extract")
-        return file_object
-
-    def upload_files(self):
-        self.file_objects = []
-        for file_path in self.files:
-            file_object = self.upload_file(file_path)
-            self.file_objects.append(file_object)
 
     def run(self):
         # 鐢熸垚鍨嬪彿缁撴瀯
@@ -90,78 +317,24 @@
         # 鐢熸垚鏁版嵁娴佺粨鏋� CADU
         # 鐢熸垚VCDU缁撴瀯
         # 鐢熸垚閬ユ祴鏁版嵁鍖呯粨鏋�
-        proj = self.gen_project([])
-        # proj = TProject(C_PROJECT_PK='2e090a487c1a4f7f741be3a437374e2f')
+        self.proj = self.gen_project()
 
-        devs = self.gen_device([], proj)
-        # with open('datas/璁惧鍒楄〃.json', 'w', encoding='utf8') as f:
-        #     json.dump(devs, f, ensure_ascii=False, indent=4)
-        #
-        # proj['devices'] = devs
-        #
-        # messages = []
-        # cadu = self.gen_tm_frame(messages)
-        # with open("datas/鎺㈡祴甯�.json", 'w', encoding='utf8') as f:
-        #     json.dump(cadu, f, ensure_ascii=False, indent=4)
-        #
-        # messages = []
-        # vcs = self.gen_vc(messages)
-        # with open('datas/铏氭嫙淇¢亾.json', 'w', encoding='utf8') as f:
-        #     json.dump(vcs, f, ensure_ascii=False, indent=4)
-        #
-        # messages = []
-        # pkt_vcs = self.gen_pkt_vc(messages)
-        # with open('datas/VC婧愬寘.json', 'w', encoding='utf8') as f:
-        #     json.dump(pkt_vcs, f, ensure_ascii=False, indent=4)
-        #
-        # messages = []
-        # dev_pkts = self.gen_dev_pkts(messages)
-        # with open('datas/璁惧婧愬寘.json', 'w', encoding='utf8') as f:
-        #     json.dump(dev_pkts, f, ensure_ascii=False, indent=4)
-        #
-        # messages = []
-        # _pkts = self.gen_pkts()
-        # pkts = []
-        # for pkt in _pkts:
-        #     _pkt = self.gen_pkt_details(pkt['name'])
-        #     pkts.append(_pkt)
-        # with open('datas/婧愬寘鍒楄〃.json', 'w', encoding='utf8') as f:
-        #     json.dump(pkts, f, ensure_ascii=False, indent=4)
-        #
-        # for dev in devs:
-        #     ds = dev['data_streams'][0]
-        #     _cadu = copy.deepcopy(cadu)
-        #     ds['cadu'] = _cadu
-        #     _vcdu = next(filter(lambda it: it['name'] == '浼犺緭甯�', _cadu['children']))
-        #     vcdu_data = next(filter(lambda it: it['name'] == '浼犺緭甯ф暟鎹煙', _vcdu['children']))
-        #     _vcs = copy.deepcopy(vcs)
-        #     vcdu_data['children'] = _vcs
-        #     dev_pkt = next(filter(lambda it: it['name'] == dev['name'], dev_pkts), None)
-        #     if dev_pkt is None:
-        #         continue
-        #     for pkt in dev_pkt['pkts']:
-        #         for vc in _vcs:
-        #             _pkt = next(
-        #                 filter(lambda it: it['name'] == pkt['name'] and it['vcs'].__contains__(vc['code']), pkt_vcs),
-        #                 None)
-        #             if _pkt:
-        #                 if vc.__contains__('pkts') is False:
-        #                     vc['pkts'] = []
-        #                 _pkt = next(filter(lambda it: it['name'] == _pkt['name'], pkts), None)
-        #                 if _pkt:
-        #                     vc['pkts'].append(_pkt)
-        #
-        # with open("datas/鍨嬪彿.json", 'w', encoding='utf8') as f:
-        #     json.dump(proj, f, ensure_ascii=False, indent=4)
+        devs = self.gen_device(self.proj)
+
+        # self.gen_tc()
         return ''
 
-    def _gen(self, msgs, msg):
+    def _gen(self, msgs, msg, files=None):
+        if files is None:
+            files = [file_map['鏂囨。鍚堝苟']]
         messages = [] if msgs is None else msgs
+        doc_text = ''
+        for file in files:
+            doc_text += '\n' + read_from_file(file)
         if len(messages) == 0:
-            # 濡傛灉鏄涓�娆℃彁闂姞鍏ユ枃妗�
+            # 濡傛灉鏄涓�娆℃彁闂姞鍏ystem娑堟伅
             messages.append({'role': 'system', 'content': assistant_msg})
-            for file_object in self.file_objects:
-                messages.append({'role': 'system', 'content': 'fileid://' + file_object.id})
+            messages.append({'role': 'user', 'content': "浠ヤ笅鏄枃妗e唴瀹癸細\n" + doc_text})
         messages.append({'role': 'user', 'content': msg})
 
         completion = self.client.chat.completions.create(
@@ -171,34 +344,71 @@
             temperature=0.0,
             top_p=0,
             timeout=30 * 60000,
-            max_completion_tokens=1000000
+            max_completion_tokens=1000000,
+            seed=0
             # stream_options={"include_usage": True}
         )
-
+        g_completion = completion
         text = ''
         for chunk in completion:
             if chunk.choices[0].delta.content is not None:
                 text += chunk.choices[0].delta.content
                 print(chunk.choices[0].delta.content, end="")
         print("")
+        g_completion = None
         return text
 
-    def gen_project(self, messages):
-        _msg = f"""
-鏍规嵁鏂囨。杈撳嚭鍨嬪彿淇℃伅锛屽瀷鍙峰瓧娈靛寘鎷細鍚嶇О鍜屼唬鍙凤紝浠呰緭鍑哄瀷鍙风殑灞炴�э紝涓嶈緭鍑哄叾浠栧眰绾ф暟鎹�
-        """
-        print('鍨嬪彿淇℃伅锛�')
-        text = self._gen(messages, _msg)
-        messages.append({'role': 'assistant', 'content': text})
-        text = self.remove_markdown(text)
-        proj_dict = json.loads(text)
-        # return proj_dict
+    def generate_text(self, msg, cache_file, msgs=None, files=None, validation=None, try_cnt=5):
+        if msgs is None:
+            msgs = []
+        if USE_CACHE and os.path.isfile(cache_file):
+            text = read_from_file(cache_file)
+        else:
+            s = time.time()
+            text = self._gen(msgs, msg, files)
+            text = remove_markdown(text)
+            if validation:
+                try:
+                    validation(text)
+                except BaseException as e:
+                    print(e)
+                    if try_cnt <= 0:
+                        raise RuntimeError('鐢熸垚澶辫触锛岄噸璇曟鏁板お澶氾紝寮哄埗缁撴潫锛�')
+                    return self.generate_text(msg, cache_file, msgs, files, validation, try_cnt - 1)
+            save_to_file(text, cache_file)
+            print(f'鑰楁椂锛歿time.time() - s}')
+        return text
+
+    def generate_tc_text(self, msg, cache_file, messages=None, files=None, validation=None, try_cnt=5):
+        if messages is None:
+            messages = []
+        doc_text = ''
+        for file in files:
+            doc_text += '\n' + read_from_file(file)
+        if len(messages) == 0:
+            # 濡傛灉鏄涓�娆℃彁闂姞鍏ystem娑堟伅
+            messages.append({'role': 'user', 'content': "浠ヤ笅鏄枃妗e唴瀹癸細\n" + doc_text})
+        return self.generate_text(msg, cache_file, messages, files, validation, try_cnt)
+
+    def gen_project(self):
+        #         _msg = """
+        # 鏍规嵁鏂囨。杈撳嚭鍨嬪彿淇℃伅锛屽瀷鍙峰瓧娈靛寘鎷細鍚嶇О鍜屼唬鍙枫�備粎杈撳嚭鍨嬪彿杩欎竴绾с��
+        # 渚嬪锛歿"name":"xxx","id":"xxx"}
+        # """
+        #         print('鍨嬪彿淇℃伅锛�')
+        #         text = self.generate_text(_msg, 'out/鍨嬪彿淇℃伅.json', files=[file_map['搴旂敤杞欢鐢ㄦ埛闇�姹�']])
+        #         proj_dict = json.loads(text)
+        # 宸ョ▼淇℃伅浠庣郴缁熻幏鍙�
+        proj_dict = {
+            "id": "JB200001",
+            "name": "HY-4A"
+        }
         code = proj_dict['id']
         name = proj_dict['name']
         proj = create_project(code, name, code, name, "", datetime.now())
         return proj
 
-    def gen_device(self, messages, proj):
+    def gen_device(self, proj):
         """
         璁惧鍒楄〃鐢熸垚瑙勫垯锛�
         1.濡傛枃妗d腑鏈�1553鍗忚鎻忚堪锛屽姞鍏�1553璁惧
@@ -208,28 +418,29 @@
 
         璁惧绫诲瀷锛氬伐鎺ф満[0]銆�1553B[1]
 
-        :param messages:
+        :param proj:
         :return:
         """
         proj_pk = proj.C_PROJECT_PK
         devices = []
 
         _msg = f"""
-杈撳嚭鎵�鏈夎澶囧垪琛紝璁惧瀛楁鍖呮嫭鍚嶇О锛坣ame)銆佷唬鍙凤紙code锛夛紝濡傛灉娌℃湁浠e彿鍒欎娇鐢ㄥ悕绉扮殑鑻辨枃缈昏瘧缂╁啓浠f浛涓旂缉鍐欓暱搴︿笉瓒呰繃5涓瓧绗︼紝JSON鏍煎紡锛屽苟涓旂粰姣忎釜璁惧澧炲姞涓変釜瀛楁锛岀涓�涓瓧娈礹asTcTm鈥滄槸鍚﹀寘鍚仴鎺ч仴娴嬧�濓紝鍒ゆ柇璇ヨ澶囨槸鍚﹀寘鍚仴鎺ч仴娴嬬殑鍔熻兘锛涚浜屼釜瀛楁hasTemperatureAnalog鈥滄槸鍚﹀寘鍚俯搴﹂噺銆佹ā鎷熼噺绛夋暟鎹殑閲囬泦鈥濓紝鍒ゆ柇璇ヨ澶囨槸鍚﹀寘鍚俯搴﹂噺绛変俊鎭殑閲囬泦鍔熻兘锛涚涓変釜瀛楁hasBus鈥滄槸鍚︽槸鎬荤嚎璁惧鈥濓紝鍒ゆ柇璇ヨ澶囨槸鍚﹀睘浜庢�荤嚎璁惧锛屾槸鍚︽湁RT鍦板潃锛涙瘡涓瓧娈电殑鍊奸兘浣跨敤true鎴杅alse鏉ヨ〃绀恒��
-浠呰緭鍑篔SON锛屼笉瑕佽緭鍑篔SON浠ュ鐨勪换浣曞瓧绗︺��
+杈撳嚭鍒嗙郴缁熶笅鐨勭‖浠朵骇鍝侊紙璁惧锛夊垪琛紝瀛楁鍖呮嫭锛氬悕绉�(name)銆佷唬鍙�(code)锛岀‖浠朵骇鍝佸悕绉颁竴鑸細鍖呭惈鈥滅鐞嗗崟鍏冣�濇垨鑰呪�滄帴鍙e崟鍏冣�濓紝濡傛灉娌℃湁浠e彿鍒欎娇鐢ㄥ悕绉扮殑鑻辨枃缂╁啓浠f浛缂╁啓闀垮害涓嶈秴杩�5涓瓧绗�;
+骞朵笖缁欐瘡涓‖浠朵骇鍝佸鍔犱笁涓瓧娈碉細绗竴涓瓧娈礹asTcTm鈥滄槸鍚﹀寘鍚仴鎺ч仴娴嬧�濓紝鍒ゆ柇璇ョ‖浠朵骇鍝佹槸鍚﹀寘鍚仴鎺ч仴娴嬬殑鍔熻兘銆�
+绗簩涓瓧娈礹asTemperatureAnalog鈥滄槸鍚﹀寘鍚俯搴﹂噺銆佹ā鎷熼噺绛夋暟鎹殑閲囬泦鈥濓紝鍒ゆ柇璇ョ‖浠朵骇鍝佹槸鍚﹀寘鍚俯搴﹂噺绛変俊鎭殑閲囬泦鍔熻兘銆�
+绗笁涓瓧娈礹asBus鈥滄槸鍚︽槸鎬荤嚎纭欢浜у搧鈥濓紝鍒ゆ柇璇ヨ澶囨槸鍚﹀睘浜庢�荤嚎纭欢浜у搧锛屾槸鍚︽湁RT鍦板潃锛涙瘡涓瓧娈电殑鍊奸兘浣跨敤true鎴杅alse鏉ヨ〃绀恒��
+浠呰緭鍑篔SON锛岀粨鏋勬渶澶栧眰涓烘暟缁勶紝鏁扮粍鍏冪礌涓鸿澶囦俊鎭紝涓嶈杈撳嚭JSON浠ュ鐨勪换浣曞瓧绗︺��
         """
         print('璁惧鍒楄〃锛�')
-        text = self._gen(messages, _msg)
-        text = self.remove_markdown(text)
+        cache_file = 'out/璁惧鍒楄〃.json'
+
+        def validation(gen_text):
+            _devs = json.loads(gen_text)
+            assert isinstance(_devs, list), '鏁版嵁缁撴瀯鏈�澶栧眰涓嶆槸鏁扮粍'
+            assert next(filter(lambda it: it['name'].endswith('绠$悊鍗曞厓'), _devs), None), '鐢熸垚鐨勮澶囧垪琛ㄤ腑娌℃湁绠$悊鍗曞厓'
+
+        text = self.generate_text(_msg, cache_file, files=[file_map['搴旂敤杞欢鐢ㄦ埛闇�姹�']], validation=validation)
         devs = json.loads(text)
-        hasBus = any(d['hasBus'] for d in devs)
-        if hasBus:
-            # 鎬荤嚎璁惧
-            dev = create_device("B1553", "1553鎬荤嚎", '1', 'StandardProCommunicationDev', proj_pk)
-            devices.append(dev)
-            # 鍒涘缓鏁版嵁娴�
-            ds_u153 = create_data_stream(proj_pk, dev.C_DEV_PK, 'ECSS涓婅鎬荤嚎鏁版嵁', 'U153', 'B153', '0', 'E153', '001')
-            ds_d153 = create_data_stream(proj_pk, dev.C_DEV_PK, 'ECSS涓嬭鎬荤嚎鏁版嵁', 'D153', 'B153', '1', 'E153', '001')
 
         # 绫籗MU璁惧锛屽寘鍚仴娴嬪拰閬ユ帶鍔熻兘锛屽悕绉扮粨灏句负鈥滅鐞嗗崟鍏冣��
         like_smu_devs = list(filter(lambda it: it['hasTcTm'] and it['name'].endswith('绠$悊鍗曞厓'), devs))
@@ -237,13 +448,31 @@
             dev = create_device(dev['code'], dev['name'], '0', 'StandardProCommunicationDev', proj.C_PROJECT_PK)
             devices.append(dev)
             # 鍒涘缓鏁版嵁娴�
-            ds_tmfl = create_data_stream(proj_pk, dev.C_DEV_PK, 'AOS閬ユ祴', 'TMFL', 'TMFL', '1', 'TMFL', '001')
-            ds_tcfl = create_data_stream(proj_pk, dev.C_DEV_PK, '閬ユ帶鎸囦护', 'TCFL', 'TCFL', '0', 'TCFL', '006')
+            ds_tmfl, rule_stream, _ = create_data_stream(proj_pk, dev.C_DEV_PK, 'AOS閬ユ祴', 'TMF1', 'TMFL', '1', 'TMF1',
+                                                         '001')
+            self.gen_tm_frame(proj_pk, rule_stream.C_RULE_PK, ds_tmfl, rule_stream.C_PATH)
+            # ds_tcfl, rule_stream, _ = create_data_stream(proj_pk, dev.C_DEV_PK, '閬ユ帶鎸囦护', 'TCFL', 'TCFL', '0', 'TCFL',
+            #                                              '006')
 
+        hasBus = any(d['hasBus'] for d in devs)
+        if hasBus:
+            # 鎬荤嚎璁惧
+            dev = create_device("1553", "1553鎬荤嚎", '1', 'StandardProCommunicationDev', proj_pk)
+            create_extend_info(proj_pk, "BusType", "鎬荤嚎绫诲瀷", "ECSS_Standard", dev.C_DEV_PK)
+            devices.append(dev)
+            # 鍒涘缓鏁版嵁娴�
+            ds_u153, rs_u153, rule_enc = create_data_stream(proj_pk, dev.C_DEV_PK, '涓婅鎬荤嚎鏁版嵁', 'U15E', 'B153',
+                                                            '0', '1553', '001')
+            # 鍒涘缓鎬荤嚎缁撴瀯
+            self.gen_bus(proj_pk, rule_enc, '1553', ds_u153, rs_u153.C_PATH, dev.C_DEV_NAME)
+            ds_d153, rule_stream, rule_enc = create_data_stream(proj_pk, dev.C_DEV_PK, '涓嬭鎬荤嚎鏁版嵁', 'D15E', 'B153',
+                                                                '1', '1553', '001', rs_u153.C_RULE_PK)
+            create_ref_ds_rule_stream(proj_pk, rule_stream.C_STREAM_PK, rule_stream.C_STREAM_ID,
+                                      rule_stream.C_STREAM_NAME, rule_stream.C_STREAM_DIR, rs_u153.C_STREAM_PK)
         # 绫籖TU璁惧锛屽寘鍚俯搴﹂噺鍜屾ā鎷熼噺鍔熻兘锛屽悕绉扮粨灏句负鈥滄帴鍙e崟鍏冣��
-        like_rtu_devs = list(filter(lambda it: it['hasTemperatureAnalog'] and it['name'].endswith('鎺ュ彛鍗曞厓'), devs))
-        for dev in like_rtu_devs:
-            dev = create_device(dev['code'], dev['name'], '0', 'StandardProCommunicationDev', proj.C_PROJECT_PK)
+        # like_rtu_devs = list(filter(lambda it: it['hasTemperatureAnalog'] and it['name'].endswith('鎺ュ彛鍗曞厓'), devs))
+        # for dev in like_rtu_devs:
+        #     dev = create_device(dev['code'], dev['name'], '0', 'StandardProCommunicationDev', proj.C_PROJECT_PK)
 
         # for dev in like_rtu_devs:
         #     dev = create_device(dev['code'], dev['name'], '0', '', proj.C_PROJECT_PK)
@@ -252,151 +481,626 @@
         #     ds_tmfl = create_data_stream(proj.C_PROJECT_PK, '娓╁害閲�', 'TMFL', 'TMFL', '1', 'TMFL', '001')
         #     ds_tcfl = create_data_stream(proj.C_PROJECT_PK, '妯℃嫙閲�', 'TCFL', 'TCFL', '0', 'TCFL', '006')
 
-        print()
-        # 鎬荤嚎璁惧
-        # print('鏄惁鏈夋�荤嚎璁惧锛�', end='')
-        # _msg = "鏂囨。涓弿杩扮殑鏈夋�荤嚎鐩稿叧鍐呭鍚楋紵浠呭洖绛旓細鈥滄湁鈥濇垨鈥滄棤鈥濓紝涓嶈杈撳嚭鍏朵粬鏂囨湰銆�"
-        # text = self._gen([], _msg)
-        # if text == "鏈�":
-        #     _msg = f"""
-        #     鏂囨。涓弿杩扮殑鎬荤嚎鍨嬪彿鏄灏戯紝浠呰緭鍑烘�荤嚎鍨嬪彿涓嶈杈撳嚭鍨嬪彿浠ュ鐨勫叾浠栦换浣曟枃鏈紝鎬荤嚎鍨嬪彿鐢辨暟瀛楀拰鑻辨枃瀛楁瘝缁勬垚銆�
-        #     """
-        #     print('璁惧ID锛�')
-        #     dev_code = self._gen([], _msg)
-        #     dev = create_device(dev_code, dev_code, '1', '', proj.C_PROJECT_PK)
-        #     devices.append(dev)
-
-        # 绫籗MU杞欢
-        # print('鏄惁鏈夌被SMU璁惧锛�', end='')
-        # _msg = "鏂囨。涓湁鎻忚堪閬ユ祴鍜岄仴鎺у姛鑳藉悧锛熶粎鍥炵瓟锛氣�滄湁鈥濇垨鈥滄棤鈥濓紝涓嶈杈撳嚭鍏朵粬鏂囨湰銆�"
-        # text = self._gen([], _msg)
-        # if text == "鏈�":
-        #     # 绯荤粺绠$悊鍗曞厓
-        #     print('鏄惁鏈夌郴缁熺鐞嗗崟鍏冿紙SMU锛夛細', end='')
-        #     _msg = f"鏂囨。涓湁鎻忚堪绯荤粺绠$悊鍗曞厓锛圫MU锛夊悧锛熶粎鍥炵瓟鈥滄湁鈥濇垨鈥滄棤鈥濓紝涓嶈杈撳嚭鍏朵粬鏂囨湰銆�"
-        #     text = self._gen([], _msg)
-        #     if text == "鏈�":
-        #         dev = create_device("SMU", "绯荤粺绠$悊鍗曞厓", '0', '', proj.C_PROJECT_PK)
-        #         devices.append(dev)
-        #     # 涓績鎺у埗鍗曞厓锛圕TU锛�
-        #     print('鏄惁鏈変腑蹇冩帶鍒跺崟鍏冿紙CTU锛夛細', end='')
-        #     _msg = f"鏂囨。涓湁鎻忚堪涓績鎺у埗鍗曞厓锛圕TU锛夊悧锛熶粎鍥炵瓟鈥滄湁鈥濇垨鈥滄棤鈥濓紝涓嶈杈撳嚭鍏朵粬鏂囨湰銆�"
-        #     text = self._gen([], _msg)
-        #     if text == "鏈�":
-        #         dev = create_device("CTU", "涓績鎺у埗鍗曞厓", '0', '', proj.C_PROJECT_PK)
-        #         devices.append(dev)
-        #
-        # # 绫籖TU
-        # print('鏄惁鏈夌被RTU璁惧锛�', end='')
-        # _msg = "鏂囨。涓湁鎻忚堪妯℃嫙閲忛噰闆嗗拰娓╁害閲忛噰闆嗗姛鑳藉悧锛熶粎鍥炵瓟锛氣�滄湁鈥濇垨鈥滄棤鈥濓紝涓嶈杈撳嚭鍏朵粬鏂囨湰銆�"
-        # text = self._gen([], _msg)
-        # if text == "鏈�":
-        #     dev = create_device("RTU", "杩滅疆鍗曞厓", '0', '', proj.C_PROJECT_PK)
-        #     devices.append(dev)
-        # device_dicts = json.loads(text)
-        # for device_dict in device_dicts:
-        #     data_stream = {'name': '鏁版嵁娴�', 'code': 'DS'}
-        #     device_dict['data_streams'] = [data_stream]
-        #
-        # return device_dicts
         return devices
 
-    def gen_tm_frame(self, messages):
-        _msg = f"""
-杈撳嚭鎺㈡祴甯х殑缁撴瀯锛屾帰娴嬪抚瀛楁鍖呮嫭锛氭帰娴嬪抚浠e彿(id)銆佹帰娴嬪抚鍚嶇О(name)銆侀暱搴�(length)銆佷笅绾ф暟鎹崟鍏冨垪琛紙children锛夈�備唬鍙峰鏋滄病鏈夊垯鐢ㄥ悕绉扮殑鑻辨枃缈昏瘧锛屽寘鎷笅绾ф暟鎹崟鍏冦��
+    def gen_insert_domain_params(self):
+        _msg = """
+鍒嗘瀽鏂囨。锛岃緭鍑烘彃鍏ュ煙鐨勫弬鏁板垪琛紝灏嗘墍鏈夊弬鏁板叏閮ㄨ緭鍑猴紝涓嶈鏈夐仐婕忋��
+鏁版嵁缁撴瀯鏈�澶栧眰涓烘暟缁勶紝鏁扮粍鍏冪礌涓哄弬鏁颁俊鎭璞★紝鍙傛暟淇℃伅瀛楁鍖呮嫭锛歯ame銆乮d銆乸os銆乴ength銆乼ype銆�
+1涓瓧鑺傜殑闀垮害涓�8浣嶏紝浣跨敤B0-B7鏉ヨ〃绀猴紝璇疯鐪熻绠楀弬鏁伴暱搴︺��
+鏂囨。涓綅缃弿杩颁俊鎭彲鑳藉瓨鍦ㄨ法瀛楄妭鐨勬儏鍐碉紝锛屼緥濡傦細"Byte1_B6~Byte2_B0":琛ㄧず浠庣1涓瓧鑺傜殑绗�7浣嶅埌绗�2涓瓧鑺傜殑绗�1浣嶏紝闀垮害鏄�3;"Byte27_B7~Byte28_B0":琛ㄧず浠庣27涓瓧鑺傜殑绗�8浣嶅埌绗�28涓瓧鑺傜殑绗�1浣嶏紝闀垮害鏄�2銆�
+"""
+        print('鎻掑叆鍩熷弬鏁板垪琛細')
+        files = [file_map['閬ユ祴澶х翰']]
+
+        def validation(gen_text):
+            params = json.loads(gen_text)
+            assert isinstance(params, list), '鎻掑叆鍩熷弬鏁板垪琛ㄦ暟鎹粨鏋勬渶澶栧眰蹇呴』鏄暟缁�'
+            assert len(params), '鎻掑叆鍩熷弬鏁板垪琛ㄤ笉鑳戒负绌�'
+
+        text = self.generate_text(_msg, './out/鎻掑叆鍩熷弬鏁板垪琛�.json', files=files, validation=validation)
+        return json.loads(text)
+
+    def gen_tm_frame_data(self):
+        _msg = """
         """
-        print('鎺㈡祴甯т俊鎭細')
-        text = self._gen(messages, _msg)
-        messages.append({'role': 'assistant', 'content': text})
-        text = self.remove_markdown(text)
-        cadu = json.loads(text)
+        files = [file_map['閬ユ祴澶х翰']]
+
+        def validation(gen_text):
+            pass
+
+    def gen_tm_frame(self, proj_pk, rule_pk, ds, name_path):
+        # 鎻掑叆鍩熷弬鏁板垪琛�
+        insert_domain = self.gen_insert_domain_params()
+
+        # VC婧愬寘鏍煎紡
+        vc_pkt_fields = data_templates.vc_pkt_fields  # self.gen_pkt_format()
+
+        # 鑾峰彇铏氭嫙淇¢亾 vc
+        vcs = self.gen_vc()
+        for vc in vcs:
+            vc['children'] = []
+            vc['VCID'] = str(int(vc['VCID'], 2))
+            for field in vc_pkt_fields:
+                if field['name'] == '鏁版嵁鍩�':
+                    field['children'] = []
+                vc['children'].append(dict(field))
+
+        # VCID 瀛楁鍐呭
+        vcid_content = build_vcid_content(vcs)
+
+        # 閬ユ祴甯х粨鏋勭敱妯℃澘鐢熸垚锛屽彧闇�鎻愪緵鐗瑰畾鍙傛暟
+        tm_data = {
+            "vcidContent": vcid_content,
+            'insertDomain': insert_domain,
+        }
+        cadu = data_templates.get_tm_frame(tm_data)
+
+        # VC婧愬寘
+        self.vc_pkts = self.gen_pkt_vc()
+        # 閬ユ祴婧愬寘璁捐涓殑婧愬寘鍒楄〃
+        self.tm_pkts = self.gen_pkts()
+
+        # 澶勭悊VC涓嬮潰鐨勯仴娴嬪寘鏁版嵁
+        for vc in vcs:
+            # 姝C涓嬬殑閬ユ祴鍖呰繃婊�
+            _vc_pkts = filter(lambda it: it['vcs'].__contains__(vc['id']), self.vc_pkts)
+            for _pkt in _vc_pkts:
+                # 鍒ゆ柇閬ユ祴鍖呮槸鍚︽湁璇︾粏瀹氫箟
+                if not next(filter(lambda it: it['name'] == _pkt['name'] and it['hasParams'], self.tm_pkts), None):
+                    continue
+                # 鑾峰彇鍖呰鎯�
+                _pkt = self.gen_pkt_details(_pkt['name'], _pkt['id'])
+                epdu = next(filter(lambda it: it['name'] == '鏁版嵁鍩�', vc['children']), None)
+                if epdu and _pkt:
+                    _pkt['children'] = _pkt['datas']
+                    _last_par = _pkt['children'][len(_pkt['children']) - 1]
+                    _pkt['length'] = (_last_par['pos'] + _last_par['length'])
+                    _pkt['pos'] = 0
+                    if 'children' not in epdu:
+                        epdu['children'] = []
+                    # 娣诲姞瑙f瀽瑙勫垯鍚庣紑闃叉閲嶅
+                    _pkt['id'] = _pkt['id'] + '_' + vc['VCID']
+                    # 缁欏寘鍚嶅姞浠e彿鍓嶇紑
+                    if not _pkt['name'].startswith(_pkt['id']):
+                        _pkt['name'] = _pkt['id'] + '_' + _pkt['name']
+                    epdu['children'].append(_pkt)
+                    apid_node = next(filter(lambda it: it['name'].__contains__('搴旂敤杩囩▼'), _pkt['headers']), None)
+                    ser_node = next(filter(lambda it: it['name'] == '鏈嶅姟', _pkt['headers']), None)
+                    sub_ser_node = next(filter(lambda it: it['name'] == '瀛愭湇鍔�', _pkt['headers']), None)
+                    _pkt['vals'] = \
+                        f"{apid_node['content']}/{int(ser_node['content'], 16)}/{int(sub_ser_node['content'], 16)}/"
+
+        # 閲嶆柊璁℃暟璧峰鍋忕Щ
+        self.compute_length_pos(cadu['children'])
+
+        # 灏嗘暟鎹彃鍏ユ暟鎹簱
+        seq = 1
+        for cadu_it in cadu['children']:
+            if cadu_it['name'] == 'VCDU':
+                # VCDU
+                # 灏嗕俊閬撴浛鎹㈠埌鏁版嵁鍩熶綅缃�
+                vc_data = next(filter(lambda it: it['name'].__contains__('鏁版嵁鍩�'), cadu_it['children']), None)
+                if vc_data:
+                    idx = cadu_it['children'].index(vc_data)
+                    cadu_it['children'].pop(idx)
+                    for vc in vcs:
+                        # 澶勭悊铏氭嫙淇¢亾灞炴��
+                        vc['type'] = 'logic'
+                        vc['length'] = vc_data['length']
+                        vc['pos'] = vc_data['pos']
+                        vc['content'] = 'CCSDSMPDU'
+                        vcid = vc['VCID']
+                        vc['condition'] = f'VCID=={vcid}'
+                        # 灏嗚櫄鎷熶俊閬撴彃鍏ュ埌VCDU
+                        cadu_it['children'].insert(idx, vc)
+                        idx += 1
+                for vc in vcs:
+                    self.compute_length_pos(vc['children'])
+
+                # 璁剧疆VCID鐨刢ontent
+                vcid_node = next(filter(lambda it: it['name'].__contains__('VCID'), cadu_it['children']), None)
+                if vcid_node:
+                    vcid_node['content'] = vcid_content
+
+                create_enc_pkt(proj_pk, rule_pk, cadu_it, rule_pk, seq, name_path, ds, '001', 'ENC')
+            else:
+                # 鍙傛暟
+                create_prop_enc(proj_pk, rule_pk, cadu_it, get_data_ty(cadu_it), seq)
+                seq += 1
 
         return cadu
 
-    def gen_vc(self, messages):
-        _msg = f"""
-杈撳嚭鎺㈡祴铏氭嫙淇¢亾鐨勫垝鍒嗭紝涓嶉渶瑕佹弿杩颁俊鎭紝浣跨敤涓�涓暟缁勮緭鍑猴紝瀛楁鍖呮嫭锛氫唬鍙�(code)銆乿cid銆佸悕绉�(name)銆�
-        """
+    def gen_vc(self):
+        _msg = """
+璇峰垎鏋愭枃妗d腑鐨勯仴娴嬪寘鏍煎紡锛岃緭鍑洪仴娴嬭櫄鎷熶俊閬撶殑鍒掑垎锛屾暟鎹粨鏋勬渶澶栧眰涓烘暟缁勶紝鏁扮粍鍏冪礌涓鸿櫄鎷熶俊閬撲俊鎭瓧鍏革紝瀛楀吀鍖呭惈浠ヤ笅閿�煎锛�
+id: 铏氭嫙淇¢亾浠e彿
+name: 铏氭嫙淇¢亾鍚嶇О
+VCID: 铏氭嫙淇¢亾VCID锛堜簩杩涘埗锛�
+format: 鏍规嵁铏氭嫙淇¢亾绫诲瀷鑾峰彇瀵瑰簲鐨勬暟鎹寘鐨勬牸寮忕殑鍚嶇О
+娣卞叆鐞嗚В鏂囨。涓弿杩扮殑鍏崇郴锛屼緥濡傦細鏂囨。涓弿杩颁簡甯歌閬ユ祴鏄父瑙勬暟鎹殑涓嬩紶淇¢亾锛屽苟涓旇繕鎻忚堪浜嗗垎绯荤粺甯歌閬ユ祴鍙傛暟鍖呭氨鏄疄鏃堕仴娴嬪弬鏁板寘锛屽苟涓旀枃妗d腑瀵瑰疄鏃堕仴娴嬪弬鏁板寘鐨勬牸寮忚繘琛屼簡鎻忚堪锛屾墍浠ュ父瑙勯仴娴媀C搴旇杈撳嚭涓猴細{"id": "1", "name": "甯歌閬ユ祴VC", "VCID": "0", "format": "瀹炴椂閬ユ祴鍙傛暟鍖�"}
+"""
+
+        def validation(gen_text):
+            vcs = json.loads(gen_text)
+            assert next(filter(lambda it: re.match('^[0-1]+$', it['VCID']), vcs)), '鐢熸垚鐨刅CID蹇呴』鏄簩杩涘埗'
+
         print('铏氭嫙淇¢亾锛�')
-        text = self._gen(messages, _msg)
-        messages.append({'role': 'assistant', 'content': text})
-        text = self.remove_markdown(text)
+        text = self.generate_text(_msg, "out/铏氭嫙淇¢亾.json", files=[file_map['閬ユ祴澶х翰']], validation=validation)
         vcs = json.loads(text)
         return vcs
 
-    def gen_dev_pkts(self, messages):
+    def gen_dev_pkts(self):
         _msg = f"""
-杈撳嚭鏂囨。涓帰娴嬫簮鍖呯被鍨嬪畾涔夋弿杩扮殑璁惧浠ュ強璁惧涓嬮潰鐨勬帰娴嬪寘锛屾暟鎹粨鏋勶細鏈�澶栧眰涓鸿澶囧垪琛� > 鎺㈡祴鍖呭垪琛�(pkts)锛岃澶囧瓧娈靛寘鎷細鍚嶇О(name)銆佷唬鍙�(id)锛屾簮鍖呭瓧娈靛寘鎷細鍚嶇О(name)銆佷唬鍙�(id)
+杈撳嚭鏂囨。涓仴娴嬫簮鍖呯被鍨嬪畾涔夋弿杩扮殑璁惧浠ュ強璁惧涓嬮潰鐨勯仴娴嬪寘锛屾暟鎹粨鏋勶細鏈�澶栧眰涓烘暟缁� > 璁惧 > 閬ユ祴鍖呭垪琛�(pkts)锛岃澶囧瓧娈靛寘鎷細鍚嶇О(name)銆佷唬鍙�(id)锛屾簮鍖呭瓧娈靛寘鎷細鍚嶇О(name)銆佷唬鍙�(id)
         """
-        print('璁惧鎺㈡祴婧愬寘淇℃伅锛�')
-        file = next(filter(lambda it: it.filename == 'XA-5D鏃犱汉鏈哄垎绯荤粺鎺㈡祴婧愬寘璁捐鎶ュ憡锛堝叕寮�锛�.md', self.file_objects),
-                    None)
-        messages = [{'role': 'system', 'content': assistant_msg}, {'role': 'system', 'content': 'fileid://' + file.id}]
-        text = self._gen(messages, _msg)
-        messages.append({'role': 'assistant', 'content': text})
-        text = self.remove_markdown(text)
+        print('璁惧閬ユ祴婧愬寘淇℃伅锛�')
+        files = [file_map["閬ユ祴婧愬寘璁捐鎶ュ憡"]]
+        text = self.generate_text(_msg, 'out/璁惧鏁版嵁鍖�.json', [], files)
         dev_pkts = json.loads(text)
         return dev_pkts
 
-    def gen_pkt_details(self, pkt_name):
+    def pkt_in_tm_pkts(self, pkt_name):
+        cache_file = f'out/鏁版嵁鍖�-{pkt_name}.json'
+        if os.path.isfile(cache_file):
+            return True
+        files = [file_map['閬ユ祴婧愬寘璁捐鎶ュ憡']]
+        print(f'鏂囨。涓湁鏃犫�渰pkt_name}鈥濈殑瀛楁鎻忚堪锛�', end='')
         _msg = f"""
-杈撳嚭鏂囨。涓弿杩扮殑鈥渰pkt_name}鈥濇帰娴嬪寘銆�
-鎺㈡祴鍖呭瓧娈靛寘鎷細鍚嶇О(name)銆佷唬鍙�(id)銆佸寘澶村睘鎬у垪琛�(headers)銆佹暟鎹煙鍙傛暟鍒楄〃(datas)锛�
-鍖呭ご灞炴�у瓧娈靛寘鎷細浣嶇疆(pos)銆佸悕绉�(name)銆佷唬鍙�(id)銆佸畾涔�(val)锛�
-鏁版嵁鍩熷弬鏁板瓧娈靛寘鎷細浣嶇疆(pos)銆佸悕绉�(name)銆佷唬鍙�(id)銆佸瓧鑺傞『搴�(byteOrder)锛�
-濡傛灉娌℃湁浠e彿鐢ㄥ悕绉扮殑鑻辨枃缈昏瘧浠f浛锛屽鏋滄病鏈夊悕绉扮敤浠e彿浠f浛锛�
-杈撳嚭鍐呭浠呰緭鍑簀son锛屼笉瑕佽緭鍑轰换浣曞叾浠栧唴瀹癸紒
-        """
-        print(f'鎺㈡祴婧愬寘鈥渰pkt_name}鈥濅俊鎭細')
-        file = next(filter(lambda it: it.filename == 'XA-5D鏃犱汉鏈哄垎绯荤粺鎺㈡祴婧愬寘璁捐鎶ュ憡锛堝叕寮�锛�.md', self.file_objects),
-                    None)
-        messages = [{'role': 'system', 'content': assistant_msg}, {'role': 'system', 'content': 'fileid://' + file.id}]
-        text = self._gen(messages, _msg)
-        messages.append({'role': 'assistant', 'content': text})
-        text = self.remove_markdown(text)
-        pkt = json.loads(text)
+鏂囨。涓湁閬ユ祴鍖呪�渰pkt_name}鈥濈殑瀛楁琛ㄦ弿杩板悧锛熼仴娴嬪寘鍚嶇О蹇呴』瀹屽叏鍖归厤銆傝緭鍑猴細鈥滄棤鈥濇垨鈥滄湁鈥濓紝涓嶈杈撳嚭鍏朵粬浠讳綍鍐呭銆�
+娉ㄦ剰锛氶仴娴嬪寘鐨勫瓧娈佃〃绱ф帴鐫�閬ユ祴鍖呯珷鑺傛爣棰橈紝濡傛灉绔犺妭鏍囬鍚庨潰鐪佺暐浜嗘垨鑰呰瑙亁xx鍒欐槸娌℃湁瀛楁琛ㄦ弿杩般��
+鏍规嵁鏂囨。鍐呭杈撳嚭銆�"""
+        text = self.generate_text(_msg, f'out/pkts/鏈夋棤鏁版嵁鍖�-{pkt_name}.txt', [], files)
+        return text == '鏈�'
+
+    def gen_pkt_details(self, pkt_name, pkt_id):
+        cache_file = f'out/鏁版嵁鍖�-{pkt_name}.json'
+        files = [file_map['閬ユ祴婧愬寘璁捐鎶ュ憡']]
+        if not os.path.isfile(cache_file):
+            _msg = f"""
+杈撳嚭鏂囨。涓弿杩扮殑鍚嶇О涓衡�渰pkt_name}鈥濅唬鍙蜂负鈥渰pkt_id}鈥濋仴娴嬪寘锛�
+閬ユ祴鍖呭瓧娈靛寘鎷細鍚嶇О(name)銆佷唬鍙�(id)銆佺被鍨�(type)銆佸寘澶村睘鎬у垪琛�(headers)銆佹暟鎹煙鍙傛暟鍒楄〃(datas)锛岀被鍨嬩负 linear锛�
+鍖呭ご灞炴�у瓧娈靛寘鎷細鍚嶇О(name)銆佷唬鍙�(id)銆佷綅缃�(pos)銆佸畾涔�(content)銆侀暱搴�(length)銆佺被鍨�(type)锛岀被鍨嬩负 para锛�
+鏁版嵁鍩熷弬鏁板瓧娈靛寘鎷細鍙傛暟鍚嶇О(name)銆佸弬鏁颁唬鍙�(id)銆佷綅缃�(pos)銆侀暱搴�(length)銆佸瓧鑺傞『搴�(byteOrder)锛岀被鍨嬩负 para锛�
+濡傛灉娌℃湁鍚嶇О鐢ㄤ唬鍙蜂唬鏇匡紝濡傛灉娌℃湁浠e彿鐢ㄥ悕绉扮殑鑻辨枃缈昏瘧浠f浛锛岀炕璇戝敖閲忕畝鐭紱
+浣犻渶瑕佺悊瑙f暟鎹寘鐨勪綅缃俊鎭紝骞朵笖灏嗘墍鏈夎緭鍑哄崟浣嶇粺涓�杞崲涓� bits锛屼綅缃瓧娈电殑杈撳嚭鏍煎紡蹇呴』涓烘暟鍊肩被鍨�;
+鏁版嵁缁撴瀯浠呭彧鍖呭惈閬ユ祴鍖咃紝浠呰緭鍑簀son锛屼笉瑕佽緭鍑轰换浣曞叾浠栧唴瀹广��"""
+            print(f'閬ユ祴婧愬寘鈥渰pkt_name}鈥濅俊鎭細')
+
+            def validation(gen_text):
+                _pkt = json.loads(gen_text)
+                assert 'headers' in _pkt, '鍖呯粨鏋勪腑蹇呴』鍖呭惈headers瀛楁'
+                assert 'datas' in _pkt, '鍖呯粨鏋勪腑蹇呴』鍖呭惈datas瀛楁'
+
+            text = self.generate_text(_msg, cache_file, [], files, validation)
+            pkt = json.loads(text)
+        else:
+            pkt = json.loads(read_from_file(cache_file))
+        pkt_len = 0
+        for par in pkt['datas']:
+            par['pos'] = pkt_len
+            pkt_len += par['length']
+        pkt['length'] = pkt_len
         return pkt
 
     def gen_pkts(self):
         _msg = f"""
-杈撳嚭鏂囨。涓弿杩扮殑鎺㈡祴鍖呫��
-鎺㈡祴鍖呭瓧娈靛寘鎷細鍚嶇О(name)銆佷唬鍙�(id)锛�
+杈撳嚭鏂囨。涓弿杩扮殑閬ユ祴鍖呫��
+閬ユ祴鍖呭瓧娈靛寘鎷細鍚嶇О(name)銆佷唬鍙�(id)銆乭asParams锛�
+鍚嶇О涓笉瑕佸寘鍚唬鍙凤紝
+hasParams琛ㄧず褰撳墠閬ユ祴鍖呮槸鍚︽湁鍙傛暟鍒楄〃锛岄仴娴嬪寘鐨勫弬鏁拌〃绱ф帴鐫�閬ユ祴鍖呯珷鑺傛爣棰橈紝濡傛灉绔犺妭鏍囬鍚庨潰鐪佺暐浜嗘垨鑰呰瑙亁xx鍒欐槸娌℃湁鍙傛暟琛紝
 濡傛灉娌℃湁浠e彿鐢ㄥ悕绉扮殑鑻辨枃缈昏瘧浠f浛锛屽鏋滄病鏈夊悕绉扮敤浠e彿浠f浛锛�
-椤剁骇缁撴瀯鐩存帴浠庢帰娴嬪寘寮�濮嬶紝涓嶅寘鎷帰娴嬪寘涓嬮潰鐨勫弬鏁般��
-            """
-        print(f'鎺㈡祴婧愬寘鍒楄〃锛�')
-        file = next(
-            filter(lambda it: it.filename == 'XA-5D鏃犱汉鏈哄垎绯荤粺鎺㈡祴婧愬寘璁捐鎶ュ憡锛堝叕寮�锛�.md', self.file_objects),
-            None)
-        messages = [{'role': 'system', 'content': assistant_msg},
-                    {'role': 'system', 'content': 'fileid://' + file.id}]
-        text = self._gen(messages, _msg)
-        messages.append({'role': 'assistant', 'content': text})
-        text = self.remove_markdown(text)
+鏁版嵁缁撴瀯鏈�澶栧眰涓烘暟缁勬暟缁勫厓绱犱负閬ユ祴鍖咃紝涓嶅寘鎷仴娴嬪寘涓嬮潰鐨勫弬鏁般��
+"""
+        print(f'閬ユ祴婧愬寘鍒楄〃锛�')
+        files = [file_map['閬ユ祴婧愬寘璁捐鎶ュ憡']]
+        text = self.generate_text(_msg, 'out/婧愬寘鍒楄〃.json', [], files)
         pkt = json.loads(text)
         return pkt
 
-    def gen_pkt_vc(self, messages):
+    def gen_pkt_vc(self):
         _msg = f"""
-鏍规嵁鎺㈡祴婧愬寘涓嬩紶鏃舵満瀹氫箟锛岃緭鍑哄悇涓帰娴嬫簮鍖呬俊鎭垪琛紝椤剁骇缁撴瀯涓烘暟缁勫厓绱犱负鎺㈡祴婧愬寘锛屾簮鍖呭瓧娈靛寘鎷細鍖呬唬鍙�(id)锛屽悕绉�(name)锛屾墍灞炶櫄鎷熶俊閬�(vcs)锛屼笅浼犳椂鏈猴紙timeTags锛�
+鏍规嵁閬ユ祴婧愬寘涓嬩紶鏃舵満瀹氫箟锛岃緭鍑哄悇涓仴娴嬫簮鍖呬俊鎭垪琛紝椤剁骇缁撴瀯涓烘暟缁勫厓绱犱负閬ユ祴婧愬寘锛屾簮鍖呭瓧娈靛寘鎷細鍖呬唬鍙�(id)锛屽悕绉�(name)锛屾墍灞炶櫄鎷熶俊閬�(vcs)锛屼笅浼犳椂鏈猴紙timeTags锛�
         """
-        print('鎺㈡祴婧愬寘鎵�灞炶櫄鎷熶俊閬擄細')
-        text = self._gen(messages, _msg)
-        messages.append({'role': 'assistant', 'content': text})
-        text = self.remove_markdown(text)
+        files = [file_map['閬ユ祴澶х翰']]
+        print('閬ユ祴婧愬寘鎵�灞炶櫄鎷熶俊閬擄細')
+
+        def validation(gen_text):
+            pkts = json.loads(gen_text)
+            assert len(pkts), 'VC婧愬寘鍒楄〃涓嶈兘涓虹┖'
+
+        text = self.generate_text(_msg, 'out/閬ユ祴VC婧愬寘.json', files=files, validation=validation)
         pkt_vcs = json.loads(text)
         return pkt_vcs
 
-    def remove_markdown(self, text):
-        # 鍘绘帀寮�澶寸殑```json
-        text = re.sub(r'^```json', '', text)
-        # 鍘绘帀缁撳熬鐨刞``json
-        text = re.sub(r'```$', '', text)
-        return text
+    def gen_pkt_format(self):
+        _msg = f"""
+璇蜂粩缁嗗垎绯绘枃妗o紝杈撳嚭鍚勪釜鏁版嵁鍖呯殑鏍煎紡锛屾暟鎹粨鏋勬渶澶栧眰涓烘暟缁勶紝鏁扮粍鍏冪礌涓烘暟鎹寘鏍煎紡锛屽皢涓诲澶寸殑瀛愮骇鎻愬崌鍒颁富瀵煎ご杩欎竴绾у苟涓斿幓闄や富瀵煎ご锛屾暟鎹寘type涓簂ogic锛屽寘鏁版嵁鍩焧ype涓篴ny銆�
+鍖呮牸寮廲hildren鍖呮嫭锛氱増鏈彿(id:Ver)銆佺被鍨�(id:TM_Type)銆佸壇瀵煎ご鏍囧織(id:Vice_Head)銆佸簲鐢ㄨ繃绋嬫爣璇嗙(id:Proc_Sign)銆佸垎缁勬爣蹇�(id:Group_Sign)銆佸寘搴忓垪璁℃暟(id:Package_Count)銆佸寘闀�(id:Pack_Len)銆佹暟鎹煙(id:EPDU_DATA)銆�
+children鍏冪礌鐨勫瓧娈靛寘鎷細name銆乮d銆乸os銆乴ength銆乼ype
+娉ㄦ剰锛氱敓鎴愮殑JSON璇硶鏍煎紡瑕佸悎娉曘��
+"""
+        print('閬ユ祴鍖呮牸寮忥細')
+        text = self.generate_text(_msg, 'out/鏁版嵁鍖呮牸寮�.json', files=[file_map['閬ユ祴澶х翰']])
+        pkt_formats = json.loads(text)
+        return pkt_formats
+
+    def compute_length_pos(self, items: list):
+        length = 0
+        pos = 0
+        for child in items:
+            if 'children' in child:
+                self.compute_length_pos(child['children'])
+            child['pos'] = pos
+            if 'length' in child and isinstance(child['length'], int):
+                length = length + child['length']
+                pos = pos + child['length']
+        # node['length'] = length
+
+    def gen_bus(self, proj_pk, rule_enc, rule_id, ds, name_path, dev_name):
+        _msg = f"""
+璇锋瀽鏂囨。锛屽垪鍑烘�荤嚎閫氫俊鍖呬紶杈撶害瀹氫腑鎻忚堪鐨勬墍鏈夋暟鎹寘鍒楄〃锛�
+鏁版嵁鍖呭瓧娈靛寘鎷細id銆乶ame銆乤pid(16杩涘埗瀛楃涓�)銆乻ervice(鏈嶅姟瀛愭湇鍔�)銆乴ength(bit闀垮害)銆乮nterval(浼犺緭鍛ㄦ湡)銆乻ubAddr(瀛愬湴鍧�/妯″紡)銆乫rameNum(閫氫俊甯у彿)銆�
+transSer(浼犺緭鏈嶅姟)銆乶ote(澶囨敞)銆乺tAddr(鎵�灞濺T鐨勫湴鍧�鍗佽繘鍒�)銆乺t(鎵�灞瀝t鍚嶇О)銆乼hroughBus(鏄惁缁忚繃鎬荤嚎)銆乥urst(鏄惁绐佸彂)銆乼ransDirect(浼犺緭鏂瑰悜)锛�
+鏁版嵁缁撴瀯鏈�澶栧眰鏄暟缁勶紝鏁扮粍鍏冪礌涓烘暟鎹寘锛屼互JSON鏍煎紡杈撳嚭锛屼笉瑕佽緭鍑篔SON浠ュ鐨勪换浣曟枃鏈��
+閫氫俊甯у彿锛氫娇鐢ㄦ枃妗d腑鐨勬枃鏈笉瑕佸仛浠讳綍杞崲銆�
+subAddr锛氬�间负鈥滄繁搴︹�濄�佲�滃钩閾衡�濄�佲�滄暟瀛椻�濇垨null銆�
+鏄惁缁忚繃鎬荤嚎鐨勫垽鏂緷鎹細鈥滃娉ㄢ�濆垪濉啓浜嗗唴瀹圭被浼尖�滀笉缁忚繃鎬荤嚎鈥濈殑鏂囧瓧琛ㄧず涓嶇粡杩囨�荤嚎鍚﹀垯缁忚繃鎬荤嚎銆�
+浼犺緭鏈嶅姟鍒嗕笁绉嶏細SetData(缃暟)銆丟etData(鍙栨暟)銆丏ataBlock(鏁版嵁鍧椾紶杈�)銆�
+浼犺緭鏂瑰悜鍒嗭細鈥濇敹鈥滃拰鈥濆彂鈥滐紝浼犺緭鏈嶅姟濡傛灉鏄�濆彇鏁扳�滄槸鈥濇敹鈥滐紝濡傛灉鏄�濇暟鎹潡浼犺緭鈥滃垯鏍规嵁鍖呮墍鍦ㄧ殑鍒嗙郴缁熶互鍙婅〃鏍肩殑鈥濅紶杈撴柟鍚戔�滃垪杩涜鍒ゆ柇锛屽垽鏂浜嶴MU鏉ヨ鏄敹杩樻槸鍙戙��
+鏄惁绐佸彂鐨勫垽鏂緷鎹細鏍规嵁琛ㄦ牸涓殑鈥濅紶杈撳懆鏈熲�滃垪杩涜鍒ゆ柇锛屽鏋滃~鍐欎簡绫讳技鈥濈獊鍙戔�滅殑鏂囧瓧琛ㄧず鏄獊鍙戝惁鍒欒〃绀轰笉鏄獊鍙戙��
+"""
+        print('鎬荤嚎鏁版嵁鍖咃細')
+
+        def validation(gen_text):
+            json.loads(gen_text)
+
+        text = self.generate_text(_msg, 'out/鎬荤嚎.json', files=[file_map['鎬荤嚎浼犺緭閫氫俊甯у垎閰�']], validation=validation)
+        pkts = json.loads(text)
+        # 绛涢�夌粡鎬荤嚎鐨勬暟鎹寘
+        pkts = list(filter(lambda it: it['throughBus'], pkts))
+        no_apid_pkts = list(filter(lambda it: not it['apid'], pkts))
+        # 绛涢�夋湁apid鐨勬暟鎹寘
+        pkts = list(filter(lambda it: it['apid'], pkts))
+
+        pkts2 = []
+        for pkt in pkts:
+            if self.pkt_in_tm_pkts(pkt["name"]):
+                pkts2.append(pkt)
+        for pkt in pkts2:
+            _pkt = self.gen_pkt_details(pkt['name'], pkt['id'])
+            if _pkt:
+                pkt['children'] = []
+                pkt['children'].extend(_pkt['datas'])
+                pkt['length'] = _pkt['length']
+        rt_pkt_map = {}
+        for pkt in pkts:
+            # 鏍规嵁鏁版嵁鍧椾紶杈撳拰鍙栨暟鍒嗙粍
+            # 閫昏緫灏佽鍖呯殑瑙f瀽瑙勫垯ID锛歊T[rt鍦板潃]SUB[瀛愬湴鍧�]S(S浠h〃鍙栨暟锛屾柟鍚戞槸AA琛ㄧず鍙戦�侊紱R浠h〃缃暟锛屾柟鍚戞槸BB琛ㄧず鎺ュ彈)
+            # 鍙栨暟锛氶�昏緫灏佽鍖呮牴鎹瓙鍦板潃鍜屽抚鍙风粍鍚堝垱寤猴紝鏈夊嚑涓粍鍚堝氨鍒涘缓鍑犱釜閫昏緫灏佽鍖�
+            # 鏁版嵁鍧楋細鍙湁涓�涓�昏緫灏佽鍖�
+
+            # 澶勭悊瀛愬湴鍧�
+            if pkt['burst']:
+                # 绐佸彂鍖呭瓙鍦板潃鏄�18~26
+                pkt['subAddr'] = 26
+            elif pkt['subAddr'] == '骞抽摵' or pkt['subAddr'] is None:
+                # 骞抽摵锛�11~26锛屾病鏈夊~鍐欑殑榛樿涓哄钩閾�
+                pkt['subAddr'] = 26
+            elif pkt['subAddr'] == '娣卞害':
+                # 娣卞害锛�11
+                pkt['subAddr'] = 11
+
+            # 澶勭悊甯у彿
+            if pkt['burst']:
+                # 绐佸彂锛欰LL
+                pkt['frameNum'] = 'ALL'
+            elif not pkt['frameNum']:
+                # 鏈�
+                pkt['frameNum'] = ''
+
+            # todo: 澶勭悊浼犺緭鏂瑰悜
+
+            rt_addr = pkt['rtAddr']
+            sub_addr = pkt['subAddr']
+            trans_ser = pkt['transSer']
+
+            frame_no = pkt['frameNum'].replace('|', ',')
+
+            if trans_ser == 'GetData':
+                # 鍙栨暟
+                pkt_id = f"RT{rt_addr}SUB{sub_addr}"
+                vals = f"{rt_addr}/{sub_addr}/0xAA/{frame_no}/"
+                rt_pkt_map_gen(pkt, '鍙栨暟', rt_pkt_map, pkt_id, vals)
+            elif trans_ser == 'DataBlock':
+                # 鏁版嵁鍧�
+                direct = '0xAA'
+                rt_pkt_map_gen(pkt, '鏁版嵁鍧椾紶杈�', rt_pkt_map, f"RT{rt_addr}SUB{sub_addr}{direct}",
+                               f"{rt_addr}/{sub_addr}/{direct}/ALL/")
+        _pkts = []
+        for k in rt_pkt_map:
+            _pkts.append(rt_pkt_map[k])
+
+        bus_items = data_templates.get_bus_datas(_pkts)
+        seq = 1
+        sub_key_nodes = list(filter(lambda it: 'is_key' in it, bus_items))
+        has_key = any(sub_key_nodes)
+        rule_pk = rule_enc.C_ENC_PK
+        sub_key = ''
+        key_items = []
+        self.compute_length_pos(bus_items)
+        for item in bus_items:
+            if item['type'] == 'enc':
+                if has_key:
+                    _prop_enc = create_any_pkt(proj_pk, rule_pk, item, seq, name_path, ds, 'ENC', sub_key_nodes,
+                                               key_items)
+                else:
+                    _prop_enc, rule_stream, _ = create_enc_pkt(proj_pk, rule_pk, item, rule_enc.C_ENC_PK, seq,
+                                                               name_path, ds, '001', 'ENC')
+            else:
+                # 鍙傛暟
+                _prop_enc = create_prop_enc(proj_pk, rule_pk, item, get_data_ty(item), seq)
+                if item.__contains__('is_key'):
+                    sub_key += _prop_enc.C_ENCITEM_PK + '/'
+                    key_items.append(
+                        {"pk": _prop_enc.C_ENCITEM_PK,
+                         'id': _prop_enc.C_SEGMENT_ID,
+                         'name': _prop_enc.C_NAME,
+                         'val': ''})
+            seq += 1
+        if sub_key:
+            rule_enc.C_KEY = sub_key
+            update_rule_enc(rule_enc)
+
+    def gen_tc(self):
+        # 鏁版嵁甯ф牸寮�
+        frame = self.gen_tc_transfer_frame()
+        # 鏁版嵁鍖呮牸寮�
+        pkt_format = self.gen_tc_transfer_pkt()
+        # 鏁版嵁鍖呭垪琛�
+        pkts = self.gen_tc_transfer_pkts()
+        for pkt in pkts:
+            pf = json.loads(json.dumps(pkt_format))
+            pf['name'] = pkt['name']
+            ph = next(filter(lambda x: x['name'] == '涓诲澶�', pf['children']), None)
+            apid = next(filter(lambda x: x['name'] == '搴旂敤杩涚▼鏍囪瘑绗�(APID)', ph['children']), None)
+            apid['value'] = pkt['apid']
+            apid['type'] = 'const'
+            sh = next(filter(lambda x: x['name'] == '鍓澶�', pf['children']), None)
+            ser = next(filter(lambda x: x['name'] == '鏈嶅姟绫诲瀷', sh['children']), None)
+            sub_ser = next(filter(lambda x: x['name'] == '鏈嶅姟瀛愮被鍨�', sh['children']), None)
+            ser['value'] = pkt['server']
+            ser['type'] = 'const'
+            sub_ser['value'] = pkt['subServer']
+            sub_ser['type'] = 'const'
+            frame['subPkts'].append(pf)
+        self.order = 0
+
+        def build_def(item: dict):
+            if item['type'] == 'enum':
+                return json.dumps({"EnumItems": item['enums'], "CanInput": True})
+            elif item['type'] == 'length':
+                return None
+            elif item['type'] == 'checkSum':
+                return json.dumps({"ChecksumType": "CRC-CCITT"})
+            elif item['type'] == 'subPkt':
+                return json.dumps({"CanInput": False})
+            elif item['type'] == 'combPkt':
+                return None
+            elif 'value' in item:
+                return item['value']
+
+        def create_tc_format(parent_pk, field):
+            field['order'] = self.order
+            self.order += 1
+            field['def'] = build_def(field)
+            if 'length' in field:
+                field['bitWidth'] = field['length']
+            field['bitOrder'] = None
+            field['attr'] = 0
+            if field['type'] == 'length':
+                val = field['value']
+                field['range'] = val['start'] + "~" + val['end']
+                field['formula'] = val['formula']
+            ins_format = create_ins_format(self.proj.C_PROJECT_PK, parent_pk, field)
+            if 'children' in field:
+                autocode = 1
+                if field['type'] == 'pkt':
+                    ins_format = create_ins_format(self.proj.C_PROJECT_PK, ins_format.C_INS_FORMAT_PK,
+                                                   {'order': self.order, 'type': 'subPkt',
+                                                    'def': json.dumps({"CanInput": False})})
+                    self.order += 1
+                for child in field['children']:
+                    child['autocode'] = autocode
+                    autocode += 1
+                    create_tc_format(ins_format.C_INS_FORMAT_PK, child)
+            # if 'subPkts' in field:
+            #     for pkt in field['subPkts']:
+            #         ins_format = create_ins_format(self.proj.C_PROJECT_PK, ins_format.C_INS_FORMAT_PK,
+            #                                        {'order': self.order, 'type': 'subPkt',
+            #                                         'def': json.dumps({"CanInput": False})})
+            #         create_tc_format(ins_format.C_INS_FORMAT_PK, pkt)
+
+        create_tc_format(None, frame)
+
+    def gen_tc_transfer_frame(self):
+        _msg = '''
+鍒嗘瀽YK浼犻�佸抚鏍煎紡锛屾彁鍙朰K浼犻�佸抚鐨勬暟鎹粨鏋勶紝涓嶅寘鎷暟鎹寘鐨勬暟鎹粨鏋勩��
+## 缁忛獙锛�
+瀛楁绫诲瀷鍖呮嫭锛�
+1.缁勫悎鍖咃細combPkt锛�
+2.鍥哄畾鐮佸瓧锛歝onst锛�
+3.闀垮害锛歭ength锛�
+4.鏋氫妇鍊硷細enum锛�
+5.鏍¢獙鍜岋細checkSum锛�
+6.鏁版嵁鍖猴細subPkt銆�
+
+鏍规嵁瀛楁鎻忚堪鍒嗘瀽瀛楁鐨勭被鍨嬶紝鍒嗘瀽鏂规硶锛�
+1.瀛楁鎻忚堪涓槑纭寚瀹氫簡瀛楁鍊肩殑锛岀被鍨嬩负const锛�
+2.瀛楁涓病鏈夋槑纭寚瀹氬瓧娈靛�硷紝浣嗘槸缃楀垪浜嗗彇鍊艰寖鍥寸殑锛岀被鍨嬩负enum锛�
+3.瀛楁鎻忚堪涓鏋滃瓨鍦ㄥ灞傜骇鎻忚堪鍒欑埗绾у瓧娈电殑绫诲瀷涓篶ombPkt锛�
+4.瀛楁濡傛灉鏄拰鈥滈暱搴︹�濇湁鍏筹紝绫诲瀷涓簂ength锛�
+5.濡傛灉鍜屾暟鎹煙鏈夊叧锛岀被鍨嬩负subPkt锛�
+6.瀛楁濡傛灉鍜屾牎楠屽拰鏈夊叧锛岀被鍨嬩负checkSum銆�
+
+瀛楁鍊兼彁鍙栨柟娉曪細
+1.瀛楁鎻忚堪涓槑纭寚瀹氫簡瀛楁鍊硷紝
+2.闀垮害瀛楁鐨勫�艰鏍规嵁鎻忚堪纭畾璧锋瀛楁鑼冨洿浠ュ強璁$畻鍏紡锛寁alue鏍煎紡渚嬪锛歿"start":"<code>","end":"<code>","formula":"N-1"}锛屾敞鎰忥細start鍜宔nd鐨勫�间负瀛楁code銆�
+
+## 闄愬埗锛�
+- length 鑷姩杞崲涓篵it闀垮害銆�
+- value 鏍规嵁瀛楁鎻忚堪鎻愬彇銆�
+- enums 鏈変簺瀛楁鏄灇涓惧�硷紝鏍规嵁瀛楁鎻忚堪鎻愬彇锛屾灇涓惧厓绱犵殑鏁版嵁缁撴瀯涓簕"n":"","v":"","c":""}銆�
+- 杈撳嚭鍐呭蹇呴』涓轰弗鏍肩殑json锛屼笉鑳借緭鍑洪櫎json浠ュ鐨勪换浣曞唴瀹广��
+
+瀛楁鏁版嵁缁撴瀯锛�
+涓诲澶�
+    鐗堟湰鍙枫�侀�氳繃鏍囧織銆佹帶鍒跺懡浠ゆ爣蹇椼�佺┖闂蹭綅銆丠TQ鏍囪瘑銆佽櫄鎷熶俊閬撴爣璇嗐�佸抚闀裤�佸抚搴忓垪鍙�
+浼犻�佸抚鏁版嵁鍩�
+甯у樊閿欐帶鍒跺煙銆�
+
+# 杈撳嚭鍐呭渚嬪瓙锛�
+{
+    "name": "YK甯�",
+    "type": "pkt"
+    "children":[
+        {
+            "name": "涓诲澶�",
+            "code": "primaryHeader",
+            "length": 2,
+            "value": "00",
+            "type": "combPkt",
+            "children": [
+                {
+                    "name": "鐗堟湰鍙�",
+                    "code": "verNum"
+                    "length": 1,
+                    "value": "00"
+                }
+            ]
+        }
+    ],
+    "subPkts":[]
+}
+'''
+
+        def validation(gen_text):
+            json.loads(gen_text)
+
+        text = self.generate_tc_text(_msg, 'out/tc_transfer_frame.json', files=[file_map['鎸囦护鏍煎紡']],
+                                     validation=validation)
+        frame = json.loads(text)
+        return frame
+
+    def gen_tc_transfer_pkt(self):
+        _msg = '''
+浠呭垎鏋怸K鍖呮牸寮忥紝鎻愬彇YK鍖呮暟鎹粨鏋勩��
+## 缁忛獙锛�
+
+瀛楁绫诲瀷鍖呮嫭锛�
+1.缁勫悎鍖咃細combPkt锛�
+2.鍥哄畾鐮佸瓧锛歝onst锛�
+3.闀垮害锛歭ength锛�
+4.鏋氫妇鍊硷細enum锛�
+5.鏍¢獙鍜岋細checkSum锛�
+6.鏁版嵁鍖猴細subPkt銆�
+
+鏍规嵁瀛楁鎻忚堪鍒嗘瀽瀛楁鐨勭被鍨嬶紝鍒嗘瀽鏂规硶锛�
+1.瀛楁鎻忚堪涓槑纭寚瀹氫簡瀛楁鍊肩殑锛岀被鍨嬩负const锛�
+2.瀛楁涓病鏈夋槑纭寚瀹氬瓧娈靛�硷紝浣嗘槸缃楀垪浜嗗彇鍊艰寖鍥寸殑锛岀被鍨嬩负enum锛�
+3.瀛楁鎻忚堪涓鏋滃瓨鍦ㄥ灞傜骇鎻忚堪鍒欑埗绾у瓧娈电殑绫诲瀷涓篶ombPkt锛�
+4.瀛楁濡傛灉鏄拰鈥滈暱搴︹�濇湁鍏筹紝绫诲瀷涓簂ength锛�
+5.濡傛灉鍜屾暟鎹煙鏈夊叧锛岀被鍨嬩负subPkt锛�
+6.瀛楁濡傛灉鍜屾牎楠屽拰鏈夊叧锛岀被鍨嬩负checkSum銆�
+
+瀛楁鍊兼彁鍙栨柟娉曪細
+1.瀛楁鎻忚堪涓槑纭寚瀹氫簡瀛楁鍊硷紝
+2.闀垮害瀛楁鐨勫�艰鏍规嵁鎻忚堪纭畾璧锋瀛楁鑼冨洿浠ュ強璁$畻鍏紡锛寁alue鏍煎紡渚嬪锛歿"start":"<code>","end":"<code>","formula":"N-1"}锛屾敞鎰忥細start鍜宔nd鐨勫�间负瀛楁code銆�
+
+## 闄愬埗锛�
+- length 鑷姩杞崲涓篵it闀垮害銆�
+- value 鏍规嵁瀛楁鎻忚堪鎻愬彇銆�
+- enums 鏈変簺瀛楁鏄灇涓惧�硷紝鏍规嵁瀛楁鎻忚堪鎻愬彇锛屾灇涓惧厓绱犵殑鏁版嵁缁撴瀯涓簕"n":"","v":"","c":""}銆�
+- 杈撳嚭鍐呭蹇呴』涓轰弗鏍肩殑json锛屼笉鑳借緭鍑洪櫎json浠ュ鐨勪换浣曞唴瀹广��
+
+瀛楁鏁版嵁缁撴瀯锛�
+涓诲澶�
+    鍖呰瘑鍒�
+        鍖呯増鏈彿銆佸寘绫诲瀷銆佹暟鎹尯澶存爣蹇椼�佸簲鐢ㄨ繘绋嬫爣璇嗙(APID)
+    鍖呭簭鍒楁帶鍒�
+        搴忓垪鏍囧織
+        鍖呭簭鍒楄鏁�
+    鍖呴暱
+鍓澶�
+    CCSDS鍓澶存爣蹇�
+    YK鍖呯増鏈彿
+    鍛戒护姝g‘搴旂瓟锛圓ck锛�
+    鏈嶅姟绫诲瀷
+    鏈嶅姟瀛愮被鍨�
+    婧愬湴鍧�
+搴旂敤鏁版嵁鍖�
+甯у樊閿欐帶鍒跺煙銆�
+
+# 杈撳嚭鍐呭渚嬪瓙锛�
+{
+    "name": "YK鍖�",
+    "type": "pkt"
+    "children":[
+        {
+            "name": "涓诲澶�",
+            "code": "primaryHeader",
+            "length": 2,
+            "value": "00",
+            "type": "combPkt",
+            "children": [
+                {
+                    "name": "鐗堟湰鍙�",
+                    "code": "verNum"
+                    "length": 1,
+                    "value": "00"
+                }
+            ]
+        }
+    ],
+    "subPkts":[]
+}
+'''
+
+        def validation(gen_text):
+            json.loads(gen_text)
+
+        text = self.generate_tc_text(_msg, 'out/tc_transfer_pkt.json', files=[file_map['鎸囦护鏍煎紡']],
+                                     validation=validation)
+        pkt_format = json.loads(text)
+        return pkt_format
+
+    def gen_tc_transfer_pkts(self):
+        _msg = '''
+鍒嗘瀽鏂囨。鍒楀嚭鎵�鏈夌殑閬ユ帶婧愬寘銆�
+## 鏁版嵁缁撴瀯濡備笅锛�
+[{
+"name": "xxx",
+"code":"pkt",
+"apid":"0xAA",
+"server":"0x1",
+"subServer":"0x2"
+}]
+'''
+
+        def validation(gen_text):
+            json.loads(gen_text)
+
+        text = self.generate_tc_text(_msg, 'out/tc_transfer_pkts.json', files=[file_map['鎸囦护鏍煎紡']],
+                                     validation=validation)
+        pkts = json.loads(text)
+        return pkts
 
 
 if __name__ == '__main__':
-    md_file = 'D:\\workspace\\PythonProjects\\KnowledgeBase\\doc\\鏂囨。鍚堝苟.md'
-    md_file2 = 'D:\\workspace\\PythonProjects\\KnowledgeBase\\doc\\XA-5D鏃犱汉鏈哄垎绯荤粺鎺㈡祴婧愬寘璁捐鎶ュ憡锛堝叕寮�锛�.md'
-    # 鍚姩澶фā鍨嬪鐞嗘祦绋�
-    ret_text = DbStructFlow([md_file, md_file2]).run()
+    try:
+        os.makedirs("./out/pkts", exist_ok=True)
+        # 鍚姩澶фā鍨嬪鐞嗘祦绋�
+        ret_text = DbStructFlow().run()
+    except KeyboardInterrupt:
+        if g_completion:
+            g_completion.close()

--
Gitblit v1.9.1