aboutsummaryrefslogtreecommitdiffhomepage
path: root/patches/server/1046-Bundle-spark.patch
diff options
context:
space:
mode:
Diffstat (limited to 'patches/server/1046-Bundle-spark.patch')
-rw-r--r--patches/server/1046-Bundle-spark.patch401
1 files changed, 401 insertions, 0 deletions
diff --git a/patches/server/1046-Bundle-spark.patch b/patches/server/1046-Bundle-spark.patch
new file mode 100644
index 0000000000..84624b4705
--- /dev/null
+++ b/patches/server/1046-Bundle-spark.patch
@@ -0,0 +1,401 @@
+From 0000000000000000000000000000000000000000 Mon Sep 17 00:00:00 2001
+From: Riley Park <[email protected]>
+Date: Tue, 16 Jul 2024 14:55:23 -0700
+Subject: [PATCH] Bundle spark
+
+
+diff --git a/build.gradle.kts b/build.gradle.kts
+index 9e6c2a4630ce75e4115f76b5e7a1e0b50e8b3197..faf3e3fd72e8c915e7a4803dacbe1bb576c6663e 100644
+--- a/build.gradle.kts
++++ b/build.gradle.kts
+@@ -76,6 +76,10 @@ dependencies {
+ implementation("io.papermc:reflection-rewriter-runtime:$reflectionRewriterVersion")
+ implementation("io.papermc:reflection-rewriter-proxy-generator:$reflectionRewriterVersion")
+ // Paper end - Remap reflection
++ // Paper start - spark
++ implementation("me.lucko:spark-api:0.1-20240720.200737-2")
++ implementation("me.lucko:spark-paper:1.10.119-SNAPSHOT")
++ // Paper end - spark
+ }
+
+ paperweight {
+diff --git a/src/main/java/io/papermc/paper/SparksFly.java b/src/main/java/io/papermc/paper/SparksFly.java
+new file mode 100644
+index 0000000000000000000000000000000000000000..62e2d5704c348955bc8284dc2d54c933b7bcdd06
+--- /dev/null
++++ b/src/main/java/io/papermc/paper/SparksFly.java
+@@ -0,0 +1,211 @@
++package io.papermc.paper;
++
++import io.papermc.paper.configuration.GlobalConfiguration;
++import io.papermc.paper.plugin.entrypoint.classloader.group.PaperPluginClassLoaderStorage;
++import io.papermc.paper.plugin.provider.classloader.ConfiguredPluginClassLoader;
++import io.papermc.paper.plugin.provider.classloader.PaperClassLoaderStorage;
++import io.papermc.paper.util.MCUtil;
++import java.util.Collection;
++import java.util.List;
++import java.util.concurrent.ConcurrentLinkedQueue;
++import java.util.logging.Level;
++import java.util.logging.Logger;
++import me.lucko.spark.paper.api.Compatibility;
++import me.lucko.spark.paper.api.PaperClassLookup;
++import me.lucko.spark.paper.api.PaperScheduler;
++import me.lucko.spark.paper.api.PaperSparkModule;
++import net.kyori.adventure.text.Component;
++import net.kyori.adventure.text.format.TextColor;
++import net.minecraft.util.ExceptionCollector;
++import org.bukkit.Server;
++import org.bukkit.command.Command;
++import org.bukkit.command.CommandSender;
++import org.bukkit.craftbukkit.CraftServer;
++
++// It's like electricity.
++public final class SparksFly {
++ public static final String ID = "spark";
++ public static final String COMMAND_NAME = "spark";
++
++ private static final String PREFER_SPARK_PLUGIN_PROPERTY = "paper.preferSparkPlugin";
++
++ private static final int SPARK_YELLOW = 0xffc93a;
++
++ private final Logger logger;
++ private final PaperSparkModule spark;
++ private final ConcurrentLinkedQueue<Runnable> mainThreadTaskQueue;
++
++ private boolean enabled;
++ private boolean disabledInConfigurationWarningLogged;
++
++ public SparksFly(final Server server) {
++ this.mainThreadTaskQueue = new ConcurrentLinkedQueue<>();
++ this.logger = Logger.getLogger(ID);
++ this.logger.log(Level.INFO, "This server bundles the spark profiler. For more information please visit https://docs.papermc.io/paper/profiling");
++ this.spark = PaperSparkModule.create(Compatibility.VERSION_1_0, server, this.logger, new PaperScheduler() {
++ @Override
++ public void executeAsync(final Runnable runnable) {
++ MCUtil.scheduleAsyncTask(this.catching(runnable, "asynchronous"));
++ }
++
++ @Override
++ public void executeSync(final Runnable runnable) {
++ SparksFly.this.mainThreadTaskQueue.offer(this.catching(runnable, "synchronous"));
++ }
++
++ private Runnable catching(final Runnable runnable, final String type) {
++ return () -> {
++ try {
++ runnable.run();
++ } catch (final Throwable t) {
++ SparksFly.this.logger.log(Level.SEVERE, "An exception was encountered while executing a " + type + " spark task", t);
++ }
++ };
++ }
++ }, new PaperClassLookup() {
++ @Override
++ public Class<?> lookup(final String className) throws Exception {
++ final ExceptionCollector<ClassNotFoundException> exceptions = new ExceptionCollector<>();
++ try {
++ return Class.forName(className);
++ } catch (final ClassNotFoundException e) {
++ exceptions.add(e);
++ for (final ConfiguredPluginClassLoader loader : ((PaperPluginClassLoaderStorage) PaperClassLoaderStorage.instance()).getGlobalGroup().getClassLoaders()) {
++ try {
++ final Class<?> loadedClass = loader.loadClass(className, true, false, true);
++ if (loadedClass != null) {
++ return loadedClass;
++ }
++ } catch (final ClassNotFoundException exception) {
++ exceptions.add(exception);
++ }
++ }
++ exceptions.throwIfPresent();
++ return null;
++ }
++ }
++ });
++ }
++
++ public void executeMainThreadTasks() {
++ Runnable task;
++ while ((task = this.mainThreadTaskQueue.poll()) != null) {
++ task.run();
++ }
++ }
++
++ public void enableEarlyIfRequested() {
++ if (!isPluginPreferred() && shouldEnableImmediately()) {
++ this.enable();
++ }
++ }
++
++ public void enableBeforePlugins() {
++ if (!isPluginPreferred()) {
++ this.enable();
++ }
++ }
++
++ public void enableAfterPlugins(final Server server) {
++ final boolean isPluginPreferred = isPluginPreferred();
++ final boolean isPluginEnabled = isPluginEnabled(server);
++ if (!isPluginPreferred || !isPluginEnabled) {
++ if (isPluginPreferred && !this.enabled) {
++ this.logger.log(Level.INFO, "The spark plugin has been preferred but was not loaded. The bundled spark profiler will enabled instead.");
++ }
++ this.enable();
++ }
++ }
++
++ private void enable() {
++ if (!this.enabled) {
++ if (GlobalConfiguration.get().spark.enabled) {
++ this.enabled = true;
++ this.spark.enable();
++ } else {
++ if (!this.disabledInConfigurationWarningLogged) {
++ this.logger.log(Level.INFO, "The spark profiler will not be enabled because it is currently disabled in the configuration.");
++ this.disabledInConfigurationWarningLogged = true;
++ }
++ }
++ }
++ }
++
++ public void disable() {
++ if (this.enabled) {
++ this.spark.disable();
++ this.enabled = false;
++ }
++ }
++
++ public void registerCommandBeforePlugins(final Server server) {
++ if (!isPluginPreferred()) {
++ this.registerCommand(server);
++ }
++ }
++
++ public void registerCommandAfterPlugins(final Server server) {
++ if ((!isPluginPreferred() || !isPluginEnabled(server)) && server.getCommandMap().getCommand(COMMAND_NAME) == null) {
++ this.registerCommand(server);
++ }
++ }
++
++ private void registerCommand(final Server server) {
++ server.getCommandMap().register(COMMAND_NAME, "paper", new CommandImpl(COMMAND_NAME, this.spark.getPermissions()));
++ }
++
++ public void tickStart() {
++ this.spark.onServerTickStart();
++ }
++
++ public void tickEnd(final double duration) {
++ this.spark.onServerTickEnd(duration);
++ }
++
++ void executeCommand(final CommandSender sender, final String[] args) {
++ this.spark.executeCommand(sender, args);
++ }
++
++ List<String> tabComplete(final CommandSender sender, final String[] args) {
++ return this.spark.tabComplete(sender, args);
++ }
++
++ public static boolean isPluginPreferred() {
++ return Boolean.getBoolean(PREFER_SPARK_PLUGIN_PROPERTY);
++ }
++
++ private static boolean isPluginEnabled(final Server server) {
++ return server.getPluginManager().isPluginEnabled(ID);
++ }
++
++ private static boolean shouldEnableImmediately() {
++ return GlobalConfiguration.get().spark.enableImmediately;
++ }
++
++ public static final class CommandImpl extends Command {
++ CommandImpl(final String name, final Collection<String> permissions) {
++ super(name);
++ this.setPermission(String.join(";", permissions));
++ }
++
++ @Override
++ public boolean execute(final CommandSender sender, final String commandLabel, final String[] args) {
++ final SparksFly spark = ((CraftServer) sender.getServer()).spark;
++ if (spark.enabled) {
++ spark.executeCommand(sender, args);
++ } else {
++ sender.sendMessage(Component.text("The spark profiler is currently disabled.", TextColor.color(SPARK_YELLOW)));
++ }
++ return true;
++ }
++
++ @Override
++ public List<String> tabComplete(final CommandSender sender, final String alias, final String[] args) throws IllegalArgumentException {
++ final SparksFly spark = ((CraftServer) sender.getServer()).spark;
++ if (spark.enabled) {
++ return spark.tabComplete(sender, args);
++ }
++ return List.of();
++ }
++ }
++}
+diff --git a/src/main/java/io/papermc/paper/plugin/provider/source/FileProviderSource.java b/src/main/java/io/papermc/paper/plugin/provider/source/FileProviderSource.java
+index 6b8ed8a0baaf4a57d20e57cec3400af5561ddd79..48604e7f96adc9e226e034054c5e2bad0b024eb5 100644
+--- a/src/main/java/io/papermc/paper/plugin/provider/source/FileProviderSource.java
++++ b/src/main/java/io/papermc/paper/plugin/provider/source/FileProviderSource.java
+@@ -1,6 +1,9 @@
+ package io.papermc.paper.plugin.provider.source;
+
++import com.mojang.logging.LogUtils;
++import io.papermc.paper.SparksFly;
+ import io.papermc.paper.plugin.PluginInitializerManager;
++import io.papermc.paper.plugin.configuration.PluginMeta;
+ import io.papermc.paper.plugin.entrypoint.EntrypointHandler;
+ import io.papermc.paper.plugin.provider.type.PluginFileType;
+ import org.bukkit.plugin.InvalidPluginException;
+@@ -17,12 +20,14 @@ import java.nio.file.attribute.BasicFileAttributes;
+ import java.util.Set;
+ import java.util.function.Function;
+ import java.util.jar.JarFile;
++import org.slf4j.Logger;
+
+ /**
+ * Loads a plugin provider at the given plugin jar file path.
+ */
+ public class FileProviderSource implements ProviderSource<Path, Path> {
+
++ private static final Logger LOGGER = LogUtils.getClassLogger();
+ private final Function<Path, String> contextChecker;
+ private final boolean applyRemap;
+
+@@ -82,6 +87,12 @@ public class FileProviderSource implements ProviderSource<Path, Path> {
+ );
+ }
+
++ final PluginMeta config = type.getConfig(file);
++ if ((config.getName().equals("spark") && config.getMainClass().equals("me.lucko.spark.bukkit.BukkitSparkPlugin")) && !SparksFly.isPluginPreferred()) {
++ LOGGER.info("The spark plugin will not be loaded as this server bundles the spark profiler.");
++ return;
++ }
++
+ type.register(entrypointHandler, file, context);
+ }
+
+diff --git a/src/main/java/net/minecraft/server/MinecraftServer.java b/src/main/java/net/minecraft/server/MinecraftServer.java
+index 4ac8bc8dc326ef12c4ffdfdf8325f3111ca5b665..ca70815b73199835b88c9d68c8a01699536d320f 100644
+--- a/src/main/java/net/minecraft/server/MinecraftServer.java
++++ b/src/main/java/net/minecraft/server/MinecraftServer.java
+@@ -764,6 +764,8 @@ public abstract class MinecraftServer extends ReentrantBlockableEventLoop<TickTa
+ // Paper end - Configurable player collision
+
+ this.server.enablePlugins(org.bukkit.plugin.PluginLoadOrder.POSTWORLD);
++ this.server.spark.registerCommandBeforePlugins(this.server); // Paper - spark
++ this.server.spark.enableAfterPlugins(this.server); // Paper - spark
+ if (io.papermc.paper.plugin.PluginInitializerManager.instance().pluginRemapper != null) io.papermc.paper.plugin.PluginInitializerManager.instance().pluginRemapper.pluginsEnabled(); // Paper - Remap plugins
+ io.papermc.paper.command.brigadier.PaperCommands.INSTANCE.setValid(); // Paper - reset invalid state for event fire below
+ io.papermc.paper.plugin.lifecycle.event.LifecycleEventRunner.INSTANCE.callReloadableRegistrarEvent(io.papermc.paper.plugin.lifecycle.event.types.LifecycleEvents.COMMANDS, io.papermc.paper.command.brigadier.PaperCommands.INSTANCE, org.bukkit.plugin.Plugin.class, io.papermc.paper.plugin.lifecycle.event.registrar.ReloadableRegistrarEvent.Cause.INITIAL); // Paper - call commands event for regular plugins
+@@ -1051,6 +1053,7 @@ public abstract class MinecraftServer extends ReentrantBlockableEventLoop<TickTa
+ Commands.COMMAND_SENDING_POOL.shutdownNow(); // Paper - Perf: Async command map building; Shutdown and don't bother finishing
+ // CraftBukkit start
+ if (this.server != null) {
++ this.server.spark.disable(); // Paper - spark
+ this.server.disablePlugins();
+ this.server.waitForAsyncTasksShutdown(); // Paper - Wait for Async Tasks during shutdown
+ }
+@@ -1244,6 +1247,7 @@ public abstract class MinecraftServer extends ReentrantBlockableEventLoop<TickTa
+ // tasks are default scheduled at -1 + delay, and first tick will tick at 1
+ final long actualDoneTimeMs = System.currentTimeMillis() - org.bukkit.craftbukkit.Main.BOOT_TIME.toEpochMilli(); // Paper - Add total time
+ LOGGER.info("Done ({})! For help, type \"help\"", String.format(java.util.Locale.ROOT, "%.3fs", actualDoneTimeMs / 1000.00D)); // Paper - Add total time
++ this.server.spark.enableBeforePlugins(); // Paper - spark
+ org.spigotmc.WatchdogThread.tick();
+ // Paper end - Improved Watchdog Support
+ org.spigotmc.WatchdogThread.hasStarted = true; // Paper
+@@ -1626,17 +1630,21 @@ public abstract class MinecraftServer extends ReentrantBlockableEventLoop<TickTa
+ }
+
+ if (this.emptyTicks >= j) {
++ this.server.spark.tickStart(); // Paper - spark
+ if (this.emptyTicks == j) {
+ MinecraftServer.LOGGER.info("Server empty for {} seconds, pausing", this.pauseWhileEmptySeconds());
+ this.autoSave();
+ }
+
+ this.server.getScheduler().mainThreadHeartbeat(); // CraftBukkit
++ this.server.spark.executeMainThreadTasks(); // Paper - spark
+ this.tickConnection();
++ this.server.spark.tickEnd(((double)(System.nanoTime() - lastTick) / 1000000D)); // Paper - spark
+ return;
+ }
+ }
+
++ this.server.spark.tickStart(); // Paper - spark
+ new com.destroystokyo.paper.event.server.ServerTickStartEvent(this.tickCount+1).callEvent(); // Paper - Server Tick Events
+ ++this.tickCount;
+ this.tickRateManager.tick();
+@@ -1654,11 +1662,13 @@ public abstract class MinecraftServer extends ReentrantBlockableEventLoop<TickTa
+ ProfilerFiller gameprofilerfiller = Profiler.get();
+
+ this.runAllTasks(); // Paper - move runAllTasks() into full server tick (previously for timings)
++ this.server.spark.executeMainThreadTasks(); // Paper - spark
+ // Paper start - Server Tick Events
+ long endTime = System.nanoTime();
+ long remaining = (TICK_TIME - (endTime - lastTick)) - catchupTime;
+ new com.destroystokyo.paper.event.server.ServerTickEndEvent(this.tickCount, ((double)(endTime - lastTick) / 1000000D), remaining).callEvent();
+ // Paper end - Server Tick Events
++ this.server.spark.tickEnd(((double)(endTime - lastTick) / 1000000D)); // Paper - spark
+ gameprofilerfiller.push("tallying");
+ long k = Util.getNanos() - i;
+ int l = this.tickCount % 100;
+diff --git a/src/main/java/net/minecraft/server/dedicated/DedicatedServer.java b/src/main/java/net/minecraft/server/dedicated/DedicatedServer.java
+index c06863578c5d654706d93e73059d89c12ae502a5..17a158ff6ce6520b69a5a0032ba4c05449dd0cf8 100644
+--- a/src/main/java/net/minecraft/server/dedicated/DedicatedServer.java
++++ b/src/main/java/net/minecraft/server/dedicated/DedicatedServer.java
+@@ -223,6 +223,7 @@ public class DedicatedServer extends MinecraftServer implements ServerInterface
+ this.paperConfigurations.initializeGlobalConfiguration(this.registryAccess());
+ this.paperConfigurations.initializeWorldDefaultsConfiguration(this.registryAccess());
+ // Paper end - initialize global and world-defaults configuration
++ this.server.spark.enableEarlyIfRequested(); // Paper - spark
+ // Paper start - fix converting txt to json file; convert old users earlier after PlayerList creation but before file load/save
+ if (this.convertOldUsers()) {
+ this.getProfileCache().save(false); // Paper
+@@ -232,6 +233,7 @@ public class DedicatedServer extends MinecraftServer implements ServerInterface
+ org.spigotmc.WatchdogThread.doStart(org.spigotmc.SpigotConfig.timeoutTime, org.spigotmc.SpigotConfig.restartOnCrash); // Paper - start watchdog thread
+ thread.start(); // Paper - Enhance console tab completions for brigadier commands; start console thread after MinecraftServer.console & PaperConfig are initialized
+ io.papermc.paper.command.PaperCommands.registerCommands(this); // Paper - setup /paper command
++ this.server.spark.registerCommandBeforePlugins(this.server); // Paper - spark
+ com.destroystokyo.paper.Metrics.PaperMetrics.startMetrics(); // Paper - start metrics
+ com.destroystokyo.paper.VersionHistoryManager.INSTANCE.getClass(); // Paper - load version history now
+
+diff --git a/src/main/java/org/bukkit/craftbukkit/CraftServer.java b/src/main/java/org/bukkit/craftbukkit/CraftServer.java
+index 1df118625aeb57f3353c40f9552f00d2f34fc655..11ff5f3b5cd25f0ad6ca944d59bca8434f8510d8 100644
+--- a/src/main/java/org/bukkit/craftbukkit/CraftServer.java
++++ b/src/main/java/org/bukkit/craftbukkit/CraftServer.java
+@@ -312,6 +312,7 @@ public final class CraftServer implements Server {
+ public static Exception excessiveVelEx; // Paper - Velocity warnings
+ private final io.papermc.paper.logging.SysoutCatcher sysoutCatcher = new io.papermc.paper.logging.SysoutCatcher(); // Paper
+ private final io.papermc.paper.potion.PaperPotionBrewer potionBrewer; // Paper - Custom Potion Mixes
++ public final io.papermc.paper.SparksFly spark; // Paper - spark
+
+ // Paper start - Folia region threading API
+ private final io.papermc.paper.threadedregions.scheduler.FallbackRegionScheduler regionizedScheduler = new io.papermc.paper.threadedregions.scheduler.FallbackRegionScheduler();
+@@ -489,6 +490,7 @@ public final class CraftServer implements Server {
+ }
+ this.potionBrewer = new io.papermc.paper.potion.PaperPotionBrewer(console); // Paper - custom potion mixes
+ datapackManager = new io.papermc.paper.datapack.PaperDatapackManager(console.getPackRepository()); // Paper
++ this.spark = new io.papermc.paper.SparksFly(this); // Paper - spark
+ }
+
+ public boolean getCommandBlockOverride(String command) {
+@@ -1115,6 +1117,7 @@ public final class CraftServer implements Server {
+ this.reloadData();
+ org.spigotmc.SpigotConfig.registerCommands(); // Spigot
+ io.papermc.paper.command.PaperCommands.registerCommands(this.console); // Paper
++ this.spark.registerCommandBeforePlugins(this); // Paper - spark
+ this.overrideAllCommandBlockCommands = this.commandsConfiguration.getStringList("command-block-overrides").contains("*");
+ this.ignoreVanillaPermissions = this.commandsConfiguration.getBoolean("ignore-vanilla-permissions");
+
+@@ -1143,6 +1146,7 @@ public final class CraftServer implements Server {
+ this.loadPlugins();
+ this.enablePlugins(PluginLoadOrder.STARTUP);
+ this.enablePlugins(PluginLoadOrder.POSTWORLD);
++ this.spark.registerCommandAfterPlugins(this); // Paper - spark
+ if (io.papermc.paper.plugin.PluginInitializerManager.instance().pluginRemapper != null) io.papermc.paper.plugin.PluginInitializerManager.instance().pluginRemapper.pluginsEnabled(); // Paper - Remap plugins
+ // Paper start - brigadier command API
+ io.papermc.paper.command.brigadier.PaperCommands.INSTANCE.setValid(); // to clear invalid state for event fire below