Added CSV output back in, fixed bugs in refactoring

This commit is contained in:
Maximilian Schmeller 2022-10-14 00:15:33 +09:00
parent b1dc01b101
commit 65c21fb6ce
4 changed files with 120 additions and 33 deletions

View file

@ -8,7 +8,7 @@ from tqdm import tqdm
from latency_graph.latency_graph_structure import LatencyGraph
from matching.subscriptions import sanitize
from message_tree.message_tree_structure import DepTree, E2EBreakdownItem
from message_tree.message_tree_structure import DepTree, E2EBreakdownItem, depth, size
from tracing_interop.tr_types import TrCallbackInstance, TrPublishInstance, TrPublisher, TrCallbackObject, TrContext, \
TrSubscriptionObject, TrTimer, TrNode, TrTopic
@ -104,7 +104,7 @@ def build_dep_trees(end_topics, lat_graph, tr, excluded_path_patterns, time_limi
pubs = end_topic.publishers
for pub in pubs:
msgs = list(pub.instances)
for msg in tqdm(msgs, desc="Processing output messages"):
for msg in tqdm(msgs, mininterval=5.0, desc="Processing output messages"):
msg: TrPublishInstance
tree = get_dep_tree(msg, lat_graph, tr, excluded_path_patterns, time_limit_s)
all_trees.append(tree)
@ -233,8 +233,17 @@ def e2e_paths_sorted_desc(tree: DepTree, input_topic_patterns):
`input_topic_patterns`. The paths are sorted by length in a descending manner (element 0 is longest).
"""
def _collect_all_paths(t: DepTree):
return [(*_collect_all_paths(d), t.head) for d in t.deps]
# TODO: Make this a Dijkstra/similar implementation instead. This is so slow it's funny.
def _collect_all_paths(t: DepTree, lvl=0):
"""
Returns a flat list of all paths that lead from t's root to a leaf.
"""
if lvl > 30 or not t.deps:
return [(t.head, )]
# List of flat lists of paths
deps_paths = [_collect_all_paths(d, lvl + 1) for d in t.deps]
return [(*path, t.head) for dep_paths in deps_paths for path in dep_paths]
def _trim_path(path):
valid_input = False
@ -250,14 +259,17 @@ def e2e_paths_sorted_desc(tree: DepTree, input_topic_patterns):
if not valid_input:
return None
if i == 0 or not isinstance(path[i - 1], TrCallbackInstance):
print(f"[WARN] Message has no publishing callback in dep tree.")
return path[i:] # Return path from first message that fits an input topic pattern
if i == 0:
#print(end=".")
return None
if not isinstance(path[i - 1], TrCallbackInstance):
#print(end="#")
return None
return path[i - 1:] # Return path from its publishing callback if it exists
paths = _collect_all_paths(tree)
paths = list(filter(lambda p: p is not None, map(_trim_path, tqdm(paths, desc="_trim_path"))))
paths = list(filter(None, map(_trim_path, paths)))
paths.sort(key=lambda path: path[-1].timestamp - path[0].timestamp, reverse=True)
return paths
@ -328,7 +340,7 @@ def _repr_path(path: List[TrPublishInstance | TrCallbackInstance]):
def aggregate_e2e_paths(paths: List[List[TrPublishInstance | TrCallbackInstance]]):
path_cohorts = defaultdict(list)
for path in paths:
for path in tqdm(paths, mininterval=5.0, desc="Aggregating E2E path cohorts"):
key = _repr_path(path)
path_cohorts[key].append(path)

View file

@ -40,7 +40,7 @@ def e2e_breakdown_inst_stack(*paths: List[E2EBreakdownItem]):
fig.suptitle("Detailed E2E Latency Path Breakdown")
bottom = 0
for i in range(len(paths)):
for i in range(len(paths[0])):
e2e_items = [path[i] for path in paths]
durations = np.array([item.duration for item in e2e_items])
ax.bar(range(len(paths)), durations, bottom=bottom)

View file

@ -5,12 +5,16 @@ E2EBreakdownItem = namedtuple("E2EBreakdownItem", ("type", "duration", "location
DepTree = namedtuple("DepTree", ("head", "deps"))
def depth(tree: DepTree):
return 1 + max(map(depth, tree.deps), default=0)
def depth(tree: DepTree, lvl=0):
if lvl > 1000:
return 0
return 1 + max(map(lambda d: depth(d, lvl + 1), tree.deps), default=0)
def size(tree: DepTree):
return 1 + sum(map(size, tree.deps))
def size(tree: DepTree, lvl=0):
if lvl > 1000:
return 0
return 1 + sum(map(lambda d: size(d, lvl + 1), tree.deps))
def fanout(tree: DepTree):

View file

@ -25,6 +25,7 @@
"\n",
"from misc.utils import cached, parse_as\n",
"\n",
"# TODO: This is useless for some reason (goal is to make graphs in notebook large and hi-res)\n",
"plt.rcParams['figure.dpi'] = 300\n",
"\n",
"%matplotlib inline"
@ -134,10 +135,10 @@
"\n",
"# E2E paths are uniquely identified by a string like \"/topic/1 -> void(Node1)(args1) -> /topic/2 -> void(Node2)(args2) -> void(Node2)(args3) -> ...\".\n",
"# Certain patterns only occur in initial setup or in scenario switching and can be excluded via RegEx patterns here.\n",
"E2E_EXCLUDED_PATH_PATTERNS = [r\"NDTScanMatcher\"]\n",
"E2E_EXCLUDED_PATH_PATTERNS = [r\"NDTScanMatcher\", r\"^/parameter_events\"]\n",
"\n",
"\n",
"DDD = False\n",
"DEBUG = False\n",
"\n",
"# This code overrides the above constants with environment variables, do not edit.\n",
"for env_key, env_value in os.environ.items():\n",
@ -443,7 +444,7 @@
"source": [
"%%skip_if_false E2E_ENABLED\n",
"%%skip_if_false E2E_PLOT\n",
"%%skip_if_false DDD\n",
"%%skip_if_false DEBUG\n",
"\n",
"\n",
"fig, ax = plt.subplots(num=\"e2e_plot\")\n",
@ -486,33 +487,53 @@
"collapsed": false
}
},
{
"cell_type": "code",
"execution_count": null,
"outputs": [],
"source": [
"%%skip_if_false DEBUG\n",
"\n",
"import pickle\n",
"with open(\"trees.pkl\", \"rb\") as f:\n",
" trees = pickle.load(f)"
],
"metadata": {
"collapsed": false
}
},
{
"cell_type": "code",
"execution_count": null,
"outputs": [],
"source": [
"%%skip_if_false E2E_ENABLED\n",
"%%skip_if_false E2E_PLOT\n",
"\n",
"%load_ext line_profiler\n",
"\n",
"\n",
"from message_tree.message_tree_algorithms import e2e_paths_sorted_desc\n",
"from message_tree.message_tree_plots import e2e_breakdown_type_hist\n",
"\n",
"def _map_tree_to_e2es(tree):\n",
" return e2e_paths_sorted_desc(tree, E2E_INPUT_TOPIC_PATTERNS)\n",
"trees_paths = [e2e_paths_sorted_desc(tree, E2E_INPUT_TOPIC_PATTERNS) for tree in tqdm(trees, mininterval=10.0,\n",
" desc=\"Extracting E2E paths\")]\n",
"all_paths = [p for paths in trees_paths for p in paths]\n",
"#all_e2e_items = [i for p in all_paths for i in p]"
],
"metadata": {
"collapsed": false
}
},
{
"cell_type": "code",
"execution_count": null,
"outputs": [],
"source": [
"%%skip_if_false E2E_ENABLED\n",
"%%skip_if_false E2E_PLOT\n",
"\n",
"def ppp():\n",
" trees_paths = [e2e_paths_sorted_desc(tree, E2E_INPUT_TOPIC_PATTERNS) for tree in tqdm(trees, desc=\"Extracting E2E paths\")]\n",
" all_paths = [p for paths in trees_paths for p in paths]\n",
" all_e2e_items = [i for p in all_paths for i in p]\n",
"from message_tree.message_tree_algorithms import e2e_latency_breakdown\n",
"\n",
" return all_e2e_items\n",
"\n",
"%lprun -f e2e_paths_sorted_desc ppp()\n",
"\n",
"e2e_breakdown_type_hist(ppp())\n",
"conv_items = [i for p in tqdm(all_paths, mininterval=5.0, desc=\"Calculating E2E latency breakdowns\")\n",
" for i in e2e_latency_breakdown(p)]\n",
"e2e_breakdown_type_hist(conv_items)\n",
"\n",
"None\n"
],
@ -525,10 +546,60 @@
"execution_count": null,
"outputs": [],
"source": [
"%%skip_if_false E2E_ENABLED\n",
"\n",
"from message_tree.message_tree_algorithms import aggregate_e2e_paths\n",
"\n",
"cohorts = aggregate_e2e_paths(all_paths)\n",
"cohort_pairs = [(k, v) for k, v in cohorts.items()]\n",
"cohort_pairs.sort(key=lambda kv: len(kv[1]), reverse=True)\n",
"\n",
"path_records = [{\"path\": path_key,\n",
" \"timestamp\": path[-1].timestamp,\n",
" \"e2e_latency\": path[-1].timestamp - path[0].timestamp} \\\n",
" for path_key, paths in cohort_pairs for path in paths if path]\n",
"\n",
"out_df = pd.DataFrame.from_records(path_records)\n",
"out_df.to_csv(os.path.join(OUT_PATH, \"e2e.csv\"), sep=\"\\t\", index=False)\n",
"\n",
"mode_cohort_key, mode_cohort = cohort_pairs[0]"
],
"metadata": {
"collapsed": false
}
},
{
"cell_type": "code",
"execution_count": null,
"outputs": [],
"source": [
"%%skip_if_false E2E_ENABLED\n",
"%%skip_if_false E2E_PLOT\n",
"\n",
"from message_tree.message_tree_plots import e2e_breakdown_inst_stack\n",
"\n",
"\n",
"mode_cohort_breakdown = [e2e_latency_breakdown(p) for p in mode_cohort[:200]]\n",
"print(len(mode_cohort))\n",
"print(mode_cohort_key.replace(\" -> \", \"\\n -> \"))\n",
"e2e_breakdown_inst_stack(*mode_cohort_breakdown)"
],
"metadata": {
"collapsed": false
}
},
{
"cell_type": "code",
"execution_count": null,
"outputs": [],
"source": [
"%%skip_if_false E2E_ENABLED\n",
"%%skip_if_false DEBUG\n",
"\n",
"import pickle\n",
"\n",
"with open(\"trees.pkl\", \"wb\") as f:\n",
" pickle.dump(trees, f)\n"
" pickle.dump(trees, f)"
],
"metadata": {
"collapsed": false