File size: 12,019 Bytes
6369972 cd2e83b 6369972 cd2e83b 6369972 cd2e83b |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 |
"""
PROMPT> python -m src.plan.app_text2plan
"""
import gradio as gr
import os
import subprocess
import time
import sys
import threading
from datetime import datetime
from math import ceil
from src.llm_factory import get_available_llms
from src.plan.filenames import FilenameEnum
from src.plan.plan_file import PlanFile
from src.plan.speedvsdetail import SpeedVsDetailEnum
from src.prompt.prompt_catalog import PromptCatalog
from src.utils.get_env_as_string import get_env_as_string
from src.utils.time_since_last_modification import time_since_last_modification
# Global variables
active_proc = None
stop_event = threading.Event()
latest_run_id = None
latest_run_dir = None
MODULE_PATH_PIPELINE = "src.plan.run_plan_pipeline"
DEFAULT_PROMPT_UUID = "4dc34d55-0d0d-4e9d-92f4-23765f49dd29"
# The pipeline outputs are stored in a log.txt file in the run directory.
# So we don't need to relay the process output to the console.
# Except for debugging purposes, you can set this to True.
RELAY_PROCESS_OUTPUT = False
prompt_catalog = PromptCatalog()
prompt_catalog.load(os.path.join(os.path.dirname(__file__), 'data', 'simple_plan_prompts.jsonl'))
# Prefill the input box with the default prompt
default_prompt_item = prompt_catalog.find(DEFAULT_PROMPT_UUID)
if default_prompt_item:
gradio_default_example = default_prompt_item.prompt
else:
raise ValueError("DEFAULT_PROMPT_UUID prompt not found.")
# Show all prompts in the catalog as examples
all_prompts = prompt_catalog.all()
gradio_examples = []
for prompt_item in all_prompts:
gradio_examples.append([prompt_item.prompt])
available_models = get_available_llms()
available_model_names = [model for model in available_models]
def has_pipeline_complete_file(path_dir: str):
"""
Checks if the pipeline has completed by looking for the completion file.
"""
files = os.listdir(path_dir)
return FilenameEnum.PIPELINE_COMPLETE.value in files
class MarkdownBuilder:
"""
Helper class to build Markdown-formatted strings.
"""
def __init__(self):
self.rows = []
def add_line(self, line: str):
self.rows.append(line)
def add_code_block(self, code: str):
self.rows.append("```\n" + code + "\n```")
def status(self, status_message: str):
self.add_line(f"### Status")
self.add_line(status_message)
def path_to_run_dir(self, absolute_path_to_run_dir: str):
self.add_line(f"### Output dir")
self.add_code_block(absolute_path_to_run_dir)
def list_files(self, path_dir: str):
self.add_line(f"### Output files")
files = os.listdir(path_dir)
files.sort()
filenames = "\n".join(files)
self.add_code_block(filenames)
def to_markdown(self):
return "\n".join(self.rows)
def run_planner(submit_or_retry_button: str, plan_prompt: str, llm_model: str, speedvsdetail: str):
"""
This generator function:
1. Prepare the run directory.
2. Updates environment with RUN_ID and LLM_MODEL.
3. Launches the Python pipeline script.
5. Polls the output dir (every second) to yield a Markdown-formatted file listing.
If a stop is requested or if the pipeline-complete file is detected, it exits.
"""
global active_proc, stop_event, latest_run_dir, latest_run_id
# Clear any previous stop signal.
stop_event.clear()
submit_or_retry = submit_or_retry_button.lower()
if submit_or_retry == "retry":
if latest_run_id is None:
raise ValueError("No previous run to retry. Please submit a plan first.")
run_id = latest_run_id
run_path = os.path.join("run", run_id)
absolute_path_to_run_dir = latest_run_dir
print(f"Retrying the run with ID: {run_id}")
if not os.path.exists(run_path):
raise Exception(f"The run path is supposed to exist from an earlier run. However the no run path exists: {run_path}")
elif submit_or_retry == "submit":
run_id = datetime.now().strftime("%Y%m%d_%H%M%S")
run_path = os.path.join("run", run_id)
absolute_path_to_run_dir = os.path.abspath(run_path)
print(f"Submitting a new run with ID: {run_id}")
# Update the global variables.
latest_run_id = run_id
latest_run_dir = absolute_path_to_run_dir
# Prepare a new run directory.
if os.path.exists(run_path):
raise Exception(f"The run path is not supposed to exist at this point. However the run path already exists: {run_path}")
os.makedirs(run_path)
# Create the initial plan file.
plan_file = PlanFile.create(plan_prompt)
plan_file.save(os.path.join(run_path, FilenameEnum.INITIAL_PLAN.value))
# Set the prompt in the environment so the pipeline can use it.
env = os.environ.copy()
env["RUN_ID"] = run_id
env["LLM_MODEL"] = llm_model
env["SPEED_VS_DETAIL"] = speedvsdetail
start_time = time.perf_counter()
# Launch the pipeline as a separate Python process.
command = [sys.executable, "-m", MODULE_PATH_PIPELINE]
print(f"Executing command: {' '.join(command)}")
# Spawn threads to print process output to the console.
if RELAY_PROCESS_OUTPUT:
active_proc = subprocess.Popen(
command,
cwd=".",
env=env,
stdout=None,
stderr=None
)
else:
active_proc = subprocess.Popen(
command,
cwd=".",
env=env,
stdout=subprocess.DEVNULL,
stderr=subprocess.DEVNULL
)
# Obtain process id
child_process_id = active_proc.pid
print(f"Process started. Process ID: {child_process_id}")
# Poll the output directory every second.
while True:
# Check if the process has ended.
if active_proc.poll() is not None:
break
# print("running...")
end_time = time.perf_counter()
duration = int(ceil(end_time - start_time))
# Check if a stop was requested.
if stop_event.is_set():
try:
active_proc.terminate()
except Exception as e:
print("Error terminating process:", e)
markdown_builder = MarkdownBuilder()
markdown_builder.status("Process terminated by user.")
markdown_builder.path_to_run_dir(absolute_path_to_run_dir)
markdown_builder.list_files(run_path)
yield markdown_builder.to_markdown()
break
last_update = ceil(time_since_last_modification(run_path))
markdown_builder = MarkdownBuilder()
markdown_builder.status(f"Working. {duration} seconds elapsed. Last output update was {last_update} seconds ago.")
markdown_builder.path_to_run_dir(absolute_path_to_run_dir)
markdown_builder.list_files(run_path)
yield markdown_builder.to_markdown()
# If the pipeline complete file is found, finish streaming.
if has_pipeline_complete_file(run_path):
break
time.sleep(1)
# Wait for the process to end (if it hasn't already) and clear our global.
returncode = 'NOT SET'
if active_proc is not None:
active_proc.wait()
returncode = active_proc.returncode
active_proc = None
# Process has completed.
end_time = time.perf_counter()
duration = int(ceil(end_time - start_time))
print(f"Process ended. returncode: {returncode}. Process ID: {child_process_id}. Duration: {duration} seconds.")
if has_pipeline_complete_file(run_path):
status_message = "Completed."
else:
status_message = "Stopped prematurely, the output may be incomplete."
# Final file listing update.
markdown_builder = MarkdownBuilder()
markdown_builder.status(f"{status_message} {duration} seconds elapsed.")
markdown_builder.path_to_run_dir(absolute_path_to_run_dir)
markdown_builder.list_files(run_path)
yield markdown_builder.to_markdown()
def stop_planner():
"""
Sets a global stop flag and, if there is an active process, attempts to terminate it.
Returns a status message.
"""
global active_proc, stop_event
stop_event.set()
if active_proc is not None:
try:
if active_proc.poll() is None: # Process is still running.
active_proc.terminate()
return "Stop signal sent. Process termination requested."
except Exception as e:
return f"Error terminating process: {e}"
else:
return "No active process to stop."
def open_output_dir():
"""
Opens the latest output directory in the native file explorer.
"""
if not latest_run_dir or not os.path.exists(latest_run_dir):
return "No output directory available."
try:
if sys.platform == "darwin": # macOS
subprocess.Popen(["open", latest_run_dir])
elif sys.platform == "win32": # Windows
subprocess.Popen(["explorer", latest_run_dir])
else: # Assume Linux or other unix-like OS
subprocess.Popen(["xdg-open", latest_run_dir])
return f"Opened the directory: {latest_run_dir}"
except Exception as e:
return f"Failed to open directory: {e}"
# Build the Gradio UI using Blocks.
with gr.Blocks(title="PlanExe") as demo_text2plan:
gr.Markdown("# PlanExe: crack open pandora’s box of ideas")
with gr.Tab("Main"):
with gr.Row():
with gr.Column(scale=2, min_width=300):
prompt_input = gr.Textbox(
label="Plan Description",
lines=5,
placeholder="Enter a description of your plan...",
value=gradio_default_example
)
with gr.Row():
submit_btn = gr.Button("Submit", variant='primary')
stop_btn = gr.Button("Stop")
retry_btn = gr.Button("Retry")
open_dir_btn = gr.Button("Open Output Dir")
output_markdown = gr.Markdown("Output will appear here...")
status_markdown = gr.Markdown("Status messages will appear here...")
with gr.Column(scale=1, min_width=300):
examples = gr.Examples(
examples=gradio_examples,
inputs=[prompt_input],
)
with gr.Tab("Settings"):
model_radio = gr.Radio(
available_model_names,
value=available_model_names[0],
label="Model",
interactive=True
)
speedvsdetail_items = [
("All details, but slow", SpeedVsDetailEnum.ALL_DETAILS_BUT_SLOW),
("Fast, but few details", SpeedVsDetailEnum.FAST_BUT_SKIP_DETAILS),
]
speedvsdetail_radio = gr.Radio(
speedvsdetail_items,
value=SpeedVsDetailEnum.ALL_DETAILS_BUT_SLOW,
label="Speed vs Detail",
interactive=True
)
with gr.Tab("Join the community"):
gr.Markdown("""
- [GitHub](https://github.com/neoneye/PlanExe) the source code.
- [Discord](https://neoneye.github.io/PlanExe-web/discord) join the community. Suggestions, feedback, and questions are welcome.
""")
# Start the planning assistant with streaming updates.
submit_btn.click(fn=run_planner, inputs=[submit_btn, prompt_input, model_radio, speedvsdetail_radio], outputs=output_markdown)
# The stop button simply calls stop_planner and displays its status message.
stop_btn.click(fn=stop_planner, outputs=status_markdown)
retry_btn.click(fn=run_planner, inputs=[retry_btn, prompt_input, model_radio, speedvsdetail_radio], outputs=output_markdown)
open_dir_btn.click(fn=open_output_dir, outputs=status_markdown)
if __name__ == "__main__":
# print("Environment variables Gradio:\n" + get_env_as_string() + "\n\n\n")
print("Press Ctrl+C to exit.")
demo_text2plan.launch()
|