aideml / aide /webui /app.py
Dixing (Dex) Xu
:sparkles: Add validation plot and score for webui (#28)
f3092ac unverified
raw
history blame
21.6 kB
import streamlit as st
import streamlit.components.v1 as components
from pathlib import Path
import tempfile
import shutil
import os
import json
from omegaconf import OmegaConf
from rich.console import Console
import sys
from dotenv import load_dotenv
import logging
from aide import Experiment
# Set up logging configuration
logging.basicConfig(
level=logging.INFO,
format="%(asctime)s - %(name)s - %(levelname)s - %(message)s",
handlers=[logging.StreamHandler(sys.stderr)],
)
logger = logging.getLogger("aide")
logger.setLevel(logging.INFO)
console = Console(file=sys.stderr)
class WebUI:
"""
WebUI encapsulates the Streamlit application logic for the AIDE Machine Learning Engineer Agent.
"""
def __init__(self):
"""
Initialize the WebUI with environment variables and session state.
"""
self.env_vars = self.load_env_variables()
self.project_root = Path(__file__).parent.parent.parent
self.config_session_state()
self.setup_page()
@staticmethod
def load_env_variables():
"""
Load API keys and environment variables from .env file.
Returns:
dict: Dictionary containing API keys.
"""
load_dotenv()
return {
"openai_key": os.getenv("OPENAI_API_KEY", ""),
"anthropic_key": os.getenv("ANTHROPIC_API_KEY", ""),
}
@staticmethod
def config_session_state():
"""
Configure default values for Streamlit session state.
"""
if "is_running" not in st.session_state:
st.session_state.is_running = False
if "current_step" not in st.session_state:
st.session_state.current_step = 0
if "total_steps" not in st.session_state:
st.session_state.total_steps = 0
if "progress" not in st.session_state:
st.session_state.progress = 0
if "results" not in st.session_state:
st.session_state.results = None
@staticmethod
def setup_page():
"""
Set up the Streamlit page configuration and load custom CSS.
"""
st.set_page_config(
page_title="AIDE: Machine Learning Engineer Agent",
layout="wide",
)
WebUI.load_css()
@staticmethod
def load_css():
"""
Load custom CSS styles from 'style.css' file.
"""
css_file = Path(__file__).parent / "style.css"
if css_file.exists():
with open(css_file) as f:
st.markdown(f"<style>{f.read()}</style>", unsafe_allow_html=True)
else:
st.warning(f"CSS file not found at: {css_file}")
def run(self):
"""
Run the main logic of the Streamlit application.
"""
self.render_sidebar()
input_col, results_col = st.columns([1, 3])
with input_col:
self.render_input_section(results_col)
with results_col:
self.render_results_section()
def render_sidebar(self):
"""
Render the sidebar with API key settings.
"""
with st.sidebar:
st.header("⚙️ Settings")
st.markdown(
"<p style='text-align: center;'>OpenAI API Key</p>",
unsafe_allow_html=True,
)
openai_key = st.text_input(
"OpenAI API Key",
value=self.env_vars["openai_key"],
type="password",
label_visibility="collapsed",
)
st.markdown(
"<p style='text-align: center;'>Anthropic API Key</p>",
unsafe_allow_html=True,
)
anthropic_key = st.text_input(
"Anthropic API Key",
value=self.env_vars["anthropic_key"],
type="password",
label_visibility="collapsed",
)
if st.button("Save API Keys", use_container_width=True):
st.session_state.openai_key = openai_key
st.session_state.anthropic_key = anthropic_key
st.success("API keys saved!")
def render_input_section(self, results_col):
"""
Render the input section of the application.
Args:
results_col (st.delta_generator.DeltaGenerator): The results column to pass to methods.
"""
st.header("Input")
uploaded_files = self.handle_file_upload()
goal_text, eval_text, num_steps = self.handle_user_inputs()
if st.button("Run AIDE", type="primary", use_container_width=True):
with st.spinner("AIDE is running..."):
results = self.run_aide(
uploaded_files, goal_text, eval_text, num_steps, results_col
)
st.session_state.results = results
def handle_file_upload(self):
"""
Handle file uploads and example file loading.
Returns:
list: List of uploaded or example files.
"""
# Only show file uploader if no example files are loaded
if not st.session_state.get("example_files"):
uploaded_files = st.file_uploader(
"Upload Data Files",
accept_multiple_files=True,
type=["csv", "txt", "json", "md"],
label_visibility="collapsed",
)
if uploaded_files:
st.session_state.pop(
"example_files", None
) # Remove example files if any
return uploaded_files
# Only show example button if no files are uploaded
if st.button(
"Load Example Experiment", type="primary", use_container_width=True
):
st.session_state.example_files = self.load_example_files()
if st.session_state.get("example_files"):
st.info("Example files loaded! Click 'Run AIDE' to proceed.")
with st.expander("View Loaded Files", expanded=False):
for file in st.session_state.example_files:
st.text(f"📄 {file['name']}")
return st.session_state.example_files
return [] # Return empty list if no files are uploaded or loaded
def handle_user_inputs(self):
"""
Handle goal, evaluation criteria, and number of steps inputs.
Returns:
tuple: Goal text, evaluation criteria text, and number of steps.
"""
goal_text = st.text_area(
"Goal",
value=st.session_state.get("goal", ""),
placeholder="Example: Predict the sales price for each house",
)
eval_text = st.text_area(
"Evaluation Criteria",
value=st.session_state.get("eval", ""),
placeholder="Example: Use the RMSE metric between the logarithm of the predicted and observed values.",
)
num_steps = st.slider(
"Number of Steps",
min_value=1,
max_value=20,
value=st.session_state.get("steps", 10),
)
return goal_text, eval_text, num_steps
@staticmethod
def load_example_files():
"""
Load example files from the 'example_tasks/house_prices' directory.
Returns:
list: List of example files with their paths.
"""
package_root = Path(__file__).parent.parent
example_dir = package_root / "example_tasks" / "house_prices"
if not example_dir.exists():
st.error(f"Example directory not found at: {example_dir}")
return []
example_files = []
for file_path in example_dir.glob("*"):
if file_path.suffix.lower() in [".csv", ".txt", ".json", ".md"]:
with tempfile.NamedTemporaryFile(
delete=False, suffix=file_path.suffix
) as tmp_file:
tmp_file.write(file_path.read_bytes())
example_files.append(
{"name": file_path.name, "path": tmp_file.name}
)
if not example_files:
st.warning("No example files found in the example directory")
st.session_state["goal"] = "Predict the sales price for each house"
st.session_state["eval"] = (
"Use the RMSE metric between the logarithm of the predicted and observed values."
)
return example_files
def run_aide(self, files, goal_text, eval_text, num_steps, results_col):
"""
Run the AIDE experiment with the provided inputs.
Args:
files (list): List of uploaded or example files.
goal_text (str): The goal of the experiment.
eval_text (str): The evaluation criteria.
num_steps (int): Number of steps to run.
results_col (st.delta_generator.DeltaGenerator): Results column for displaying progress.
Returns:
dict: Dictionary containing the results of the experiment.
"""
try:
self.initialize_run_state(num_steps)
self.set_api_keys()
input_dir = self.prepare_input_directory(files)
if not input_dir:
return None
experiment = self.initialize_experiment(input_dir, goal_text, eval_text)
placeholders = self.create_results_placeholders(results_col, experiment)
for step in range(num_steps):
st.session_state.current_step = step + 1
progress = (step + 1) / num_steps
self.update_results_placeholders(placeholders, progress)
experiment.run(steps=1)
self.clear_run_state(placeholders)
return self.collect_results(experiment)
except Exception as e:
st.session_state.is_running = False
console.print_exception()
st.error(f"Error occurred: {str(e)}")
return None
@staticmethod
def initialize_run_state(num_steps):
"""
Initialize the running state for the experiment.
Args:
num_steps (int): Total number of steps in the experiment.
"""
st.session_state.is_running = True
st.session_state.current_step = 0
st.session_state.total_steps = num_steps
st.session_state.progress = 0
@staticmethod
def set_api_keys():
"""
Set the API keys in the environment variables from the session state.
"""
if st.session_state.get("openai_key"):
os.environ["OPENAI_API_KEY"] = st.session_state.openai_key
if st.session_state.get("anthropic_key"):
os.environ["ANTHROPIC_API_KEY"] = st.session_state.anthropic_key
def prepare_input_directory(self, files):
"""
Prepare the input directory and handle uploaded files.
Args:
files (list): List of uploaded or example files.
Returns:
Path: The input directory path, or None if files are missing.
"""
input_dir = self.project_root / "input"
input_dir.mkdir(parents=True, exist_ok=True)
if files:
for file in files:
if isinstance(file, dict): # Example files
shutil.copy2(file["path"], input_dir / file["name"])
else: # Uploaded files
with open(input_dir / file.name, "wb") as f:
f.write(file.getbuffer())
else:
st.error("Please upload data files")
return None
return input_dir
@staticmethod
def initialize_experiment(input_dir, goal_text, eval_text):
"""
Initialize the AIDE Experiment.
Args:
input_dir (Path): Path to the input directory.
goal_text (str): The goal of the experiment.
eval_text (str): The evaluation criteria.
Returns:
Experiment: The initialized Experiment object.
"""
experiment = Experiment(data_dir=str(input_dir), goal=goal_text, eval=eval_text)
return experiment
@staticmethod
def create_results_placeholders(results_col, experiment):
"""
Create placeholders in the results column for dynamic content.
Args:
results_col (st.delta_generator.DeltaGenerator): The results column.
experiment (Experiment): The Experiment object.
Returns:
dict: Dictionary of placeholders.
"""
with results_col:
status_placeholder = st.empty()
step_placeholder = st.empty()
config_title_placeholder = st.empty()
config_placeholder = st.empty()
progress_placeholder = st.empty()
step_placeholder.markdown(
f"### 🔥 Running Step {st.session_state.current_step}/{st.session_state.total_steps}"
)
config_title_placeholder.markdown("### 📋 Configuration")
config_placeholder.code(OmegaConf.to_yaml(experiment.cfg), language="yaml")
progress_placeholder.progress(0)
placeholders = {
"status": status_placeholder,
"step": step_placeholder,
"config_title": config_title_placeholder,
"config": config_placeholder,
"progress": progress_placeholder,
}
return placeholders
@staticmethod
def update_results_placeholders(placeholders, progress):
"""
Update the placeholders with the current progress.
Args:
placeholders (dict): Dictionary of placeholders.
progress (float): Current progress value.
"""
placeholders["step"].markdown(
f"### 🔥 Running Step {st.session_state.current_step}/{st.session_state.total_steps}"
)
placeholders["progress"].progress(progress)
@staticmethod
def clear_run_state(placeholders):
"""
Clear the running state and placeholders after the experiment.
Args:
placeholders (dict): Dictionary of placeholders.
"""
st.session_state.is_running = False
placeholders["status"].empty()
placeholders["step"].empty()
placeholders["config_title"].empty()
placeholders["config"].empty()
placeholders["progress"].empty()
@staticmethod
def collect_results(experiment):
"""
Collect the results from the experiment.
Args:
experiment (Experiment): The Experiment object.
Returns:
dict: Dictionary containing the collected results.
"""
solution_path = experiment.cfg.log_dir / "best_solution.py"
if solution_path.exists():
solution = solution_path.read_text()
else:
solution = "No solution found"
journal_data = [
{
"step": node.step,
"code": str(node.code),
"metric": str(node.metric.value) if node.metric else None,
"is_buggy": node.is_buggy,
}
for node in experiment.journal.nodes
]
results = {
"solution": solution,
"config": OmegaConf.to_yaml(experiment.cfg),
"journal": json.dumps(journal_data, indent=2, default=str),
"tree_path": str(experiment.cfg.log_dir / "tree_plot.html"),
}
return results
def render_results_section(self):
"""
Render the results section with tabs for different outputs.
"""
st.header("Results")
if st.session_state.get("results"):
results = st.session_state.results
tabs = st.tabs(
[
"Tree Visualization",
"Best Solution",
"Config",
"Journal",
"Validation Plot",
]
)
with tabs[0]:
self.render_tree_visualization(results)
with tabs[1]:
self.render_best_solution(results)
with tabs[2]:
self.render_config(results)
with tabs[3]:
self.render_journal(results)
with tabs[4]:
# Display best score before the plot
best_metric = self.get_best_metric(results)
if best_metric is not None:
st.metric("Best Validation Score", f"{best_metric:.4f}")
self.render_validation_plot(results)
else:
st.info("No results to display. Please run an experiment.")
@staticmethod
def render_tree_visualization(results):
"""
Render the tree visualization from the experiment results.
Args:
results (dict): The results dictionary containing paths and data.
"""
if "tree_path" in results:
tree_path = Path(results["tree_path"])
logger.info(f"Loading tree visualization from: {tree_path}")
if tree_path.exists():
with open(tree_path, "r", encoding="utf-8") as f:
html_content = f.read()
components.html(html_content, height=600, scrolling=True)
else:
st.error(f"Tree visualization file not found at: {tree_path}")
logger.error(f"Tree file not found at: {tree_path}")
else:
st.info("No tree visualization available for this run.")
@staticmethod
def render_best_solution(results):
"""
Display the best solution code.
Args:
results (dict): The results dictionary containing the solution.
"""
if "solution" in results:
solution_code = results["solution"]
st.code(solution_code, language="python")
else:
st.info("No solution available.")
@staticmethod
def render_config(results):
"""
Display the configuration used in the experiment.
Args:
results (dict): The results dictionary containing the config.
"""
if "config" in results:
st.code(results["config"], language="yaml")
else:
st.info("No configuration available.")
@staticmethod
def render_journal(results):
"""
Display the experiment journal as JSON.
Args:
results (dict): The results dictionary containing the journal.
"""
if "journal" in results:
try:
journal_data = json.loads(results["journal"])
formatted_journal = json.dumps(journal_data, indent=2)
st.code(formatted_journal, language="json")
except json.JSONDecodeError:
st.code(results["journal"], language="json")
else:
st.info("No journal available.")
@staticmethod
def get_best_metric(results):
"""
Extract the best validation metric from results.
"""
try:
journal_data = json.loads(results["journal"])
metrics = []
for node in journal_data:
if node["metric"] is not None:
try:
# Convert string metric to float
metric_value = float(node["metric"])
metrics.append(metric_value)
except (ValueError, TypeError):
continue
return max(metrics) if metrics else None
except (json.JSONDecodeError, KeyError):
return None
@staticmethod
def render_validation_plot(results):
"""
Render the validation score plot.
"""
try:
journal_data = json.loads(results["journal"])
steps = []
metrics = []
for node in journal_data:
if node["metric"] is not None and node["metric"].lower() != "none":
try:
metric_value = float(node["metric"])
steps.append(node["step"])
metrics.append(metric_value)
except (ValueError, TypeError):
continue
if metrics:
import plotly.graph_objects as go
fig = go.Figure()
fig.add_trace(
go.Scatter(
x=steps,
y=metrics,
mode="lines+markers",
name="Validation Score",
line=dict(color="#F04370"),
marker=dict(color="#F04370"),
)
)
fig.update_layout(
title="Validation Score Progress",
xaxis_title="Step",
yaxis_title="Validation Score",
template="plotly_white",
hovermode="x unified",
plot_bgcolor="rgba(0,0,0,0)",
paper_bgcolor="rgba(0,0,0,0)",
)
st.plotly_chart(fig, use_container_width=True)
else:
st.info("No validation metrics available to plot.")
except (json.JSONDecodeError, KeyError):
st.error("Could not parse validation metrics data.")
if __name__ == "__main__":
app = WebUI()
app.run()