From 7d6808cbcfbb0f61f93e536d36968eeda5bd302c Mon Sep 17 00:00:00 2001 From: Luck Date: Wed, 6 Jun 2018 15:39:16 +0100 Subject: Convert to Gradle --- .../java/me/lucko/spark/profiler/StackNode.java | 141 +++++++++++++++++++++ 1 file changed, 141 insertions(+) create mode 100644 spark-common/src/main/java/me/lucko/spark/profiler/StackNode.java (limited to 'spark-common/src/main/java/me/lucko/spark/profiler/StackNode.java') diff --git a/spark-common/src/main/java/me/lucko/spark/profiler/StackNode.java b/spark-common/src/main/java/me/lucko/spark/profiler/StackNode.java new file mode 100644 index 0000000..575400a --- /dev/null +++ b/spark-common/src/main/java/me/lucko/spark/profiler/StackNode.java @@ -0,0 +1,141 @@ +/* + * WarmRoast + * Copyright (C) 2013 Albert Pham + * + * This program is free software: you can redistribute it and/or modify + * it under the terms of the GNU General Public License as published by + * the Free Software Foundation, either version 3 of the License, or + * (at your option) any later version. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. + * + * You should have received a copy of the GNU General Public License + * along with this program. If not, see . +*/ + +package me.lucko.spark.profiler; + +import com.google.gson.stream.JsonWriter; + +import java.io.IOException; +import java.util.ArrayList; +import java.util.Collection; +import java.util.Collections; +import java.util.List; +import java.util.Map; +import java.util.concurrent.ConcurrentHashMap; +import java.util.concurrent.atomic.LongAdder; + +/** + * Represents a node in the overall sampling stack. + * + *

The base implementation of this class is only used for the root of node structures. The + * {@link StackTraceNode} class is used for representing method calls in the structure.

+ */ +public class StackNode implements Comparable { + + private static final int MAX_STACK_DEPTH = 300; + + /** + * The name of this node + */ + private final String name; + + /** + * A map of this nodes children + */ + private final Map children = new ConcurrentHashMap<>(); + + /** + * The accumulated sample time for this node + */ + private final LongAdder totalTime = new LongAdder(); + + public StackNode(String name) { + this.name = name; + } + + public String getName() { + return this.name; + } + + public Collection getChildren() { + if (this.children.isEmpty()) { + return Collections.emptyList(); + } + + List list = new ArrayList<>(this.children.values()); + list.sort(null); + return list; + } + + private StackNode resolveChild(String name) { + return this.children.computeIfAbsent(name, StackNode::new); + } + + private StackNode resolveChild(String className, String methodName) { + return this.children.computeIfAbsent(StackTraceNode.formName(className, methodName), name -> new StackTraceNode(className, methodName)); + } + + public long getTotalTime() { + return this.totalTime.longValue(); + } + + public void accumulateTime(long time) { + this.totalTime.add(time); + } + + private void log(StackTraceElement[] elements, int skip, long time) { + accumulateTime(time); + + if (skip >= MAX_STACK_DEPTH) { + return; + } + + if (elements.length - skip == 0) { + return; + } + + StackTraceElement bottom = elements[elements.length - (skip + 1)]; + resolveChild(bottom.getClassName(), bottom.getMethodName()).log(elements, skip + 1, time); + } + + public void log(StackTraceElement[] elements, long time) { + log(elements, 0, time); + } + + @Override + public int compareTo(StackNode o) { + return getName().compareTo(o.getName()); + } + + public void serializeTo(JsonWriter writer) throws IOException { + writer.beginObject(); + + // append metadata about this node + appendMetadata(writer); + + // include the total time recorded for this node + writer.name("totalTime").value(getTotalTime()); + + // append child nodes, if any are present + Collection childNodes = getChildren(); + if (!childNodes.isEmpty()) { + writer.name("children").beginArray(); + for (StackNode child : childNodes) { + child.serializeTo(writer); + } + writer.endArray(); + } + + writer.endObject(); + } + + protected void appendMetadata(JsonWriter writer) throws IOException { + writer.name("name").value(getName()); + } + +} -- cgit