Dashboard Walkthrough

This comprehensive guide walks you through each step of the Rassket dashboard, mapping directly to the UI workflow. Follow along as you use the platform to understand what happens at each stage.

Step 1: Data Ingestion

What You See

  • Upload area with drag-and-drop interface
  • File information display (rows, columns, memory usage)
  • Domain selection interface
  • Preprocessing status indicator

What the System Does

  1. File Upload: Validates CSV format and size (max 50MB)
  2. Schema Detection: Analyzes column names, data types, and structure
  3. Quality Assessment: Detects missing values, duplicates, and data issues
  4. File Storage: Saves uploaded file for processing

Why This Step Matters

Data ingestion is the foundation of your ML pipeline. Proper validation ensures downstream processes work correctly. Rassket handles all validation automatically, catching common issues before they cause problems.

Pro Tip: Ensure your CSV has headers in the first row. Column names help Rassket understand your data structure and generate better explanations.
Upload interface with drag and drop

Upload your CSV file using drag-and-drop or click to browse

File information display

File information shows rows, columns, and memory usage

Upload complete with file details

After upload, file details are displayed

Step 2: Domain Selection

What You See

  • Two domain options: Energy and Research
  • Optional sub-domain selection
  • Preprocessing progress indicator
  • Success confirmation when preprocessing completes

What the System Does

  1. Domain Selection: You choose Energy or Research domain
  2. Sub-Domain Selection (Optional): Refine to specific use case
  3. Domain-Aware Preprocessing: Applies specialized feature engineering
  4. Feature Generation: Creates domain-specific features automatically

Why This Step Matters

Domain selection enables specialized feature engineering that improves model accuracy and generates domain-appropriate explanations. This is what makes Rassket different from generic AutoML tools.

See Domain Selection for detailed information about domains and sub-domains.

Domain selection interface

Choose between Energy and Research domains

Sub-domain selection

Optionally select a sub-domain for refined analysis

Preprocessing in progress

Domain-aware preprocessing applies specialized feature engineering

Step 3: Schema Understanding & Analysis

What You See

  • Target column selector (dropdown)
  • "Analyze Data" button
  • Problem type detection result
  • Recommended ML approaches (with selection checkboxes)
  • AI-powered insights (if OpenAI is configured)
  • Data characteristics summary

What the System Does

  1. Target Column Selection: You optionally select which column to predict
  2. Problem Type Detection: Automatically detects Regression, Binary Classification, or Multi-class Classification
  3. Model Recommendations: Suggests suitable ML approaches based on your data
  4. Data Analysis: Analyzes data characteristics (rows, columns, types, distributions)
  5. AI Insights: Generates domain-aware insights about your data (if enabled)

Why This Step Matters

Understanding your problem type ensures Rassket selects appropriate models and evaluation metrics. The analysis phase helps you understand your data before training begins.

Model Selection

You can:

  • Select Specific Models: Choose which models to train and compare
  • Select All: Train all recommended models
  • Deselect All: Let Rassket automatically choose the best model
Model Selection: If you don't select any models, Rassket will automatically select and train the best model for your problem type. Selecting multiple models enables comparison to find the best performer.
Data analysis interface

Select target column and analyze your data

Problem type detection

Automatic problem type detection and model recommendations

Model selection interface

Choose specific models to train or use automatic selection

Step 4: Feature Engineering

What Happens (Behind the Scenes)

Before model training, Rassket automatically performs feature engineering:

  • Missing Value Imputation: Handles missing data appropriately for your domain
  • Feature Creation: Generates domain-specific features
  • Encoding: Converts categorical variables to numeric
  • Scaling: Normalizes features for model compatibility
  • Feature Selection: Identifies and removes redundant features

Domain-Specific Engineering

Based on your domain selection:

  • Energy: Temporal features, lag variables, rolling statistics, peak indicators
  • Research: Statistical transformations, interaction terms, experimental features

Step 5: Model Training (AutoML)

What You See

  • Target column selector (required for training)
  • Problem type selector (Regression, Binary Classification, Multi-class Classification)
  • Selected models display (if you selected specific models)
  • "Train Model" button
  • Training progress indicator
  • Training results summary
  • Model comparison table (if multiple models trained)
  • Summary statistics table
  • AI explanation of results

What the System Does

  1. Data Splitting: Automatically splits data into train/validation/test sets
  2. Model Selection: Selects models based on your choices or automatically
  3. Hyperparameter Tuning: Uses Optuna to optimize hyperparameters for each model
  4. Model Training: Trains each selected model
  5. Model Comparison: Compares performance across models
  6. Best Model Selection: Identifies the best-performing model
  7. Comprehensive Evaluation: Generates detailed metrics and diagnostics

Why This Step Matters

This is where Rassket's AutoML engine shines. It automatically handles:

  • Model selection from a library of algorithms
  • Hyperparameter optimization (no manual tuning needed)
  • Cross-validation for robust evaluation
  • Model comparison to find the best performer
Training Time: Training can take several minutes, especially for large datasets or when training multiple models. Keep the page open during training. Complex models may take up to 30 minutes.

Training Results

After training completes, you'll see:

  • Model ID: Unique identifier for your trained model
  • Model Type: The algorithm used (XGBoost, LightGBM, CatBoost, etc.)
  • Status: Training completion status
  • Metrics: Performance metrics (R², RMSE, MAE for regression; Accuracy, Precision, Recall, F1 for classification)
  • Comparison Table: Side-by-side comparison if multiple models were trained
Training configuration

Configure target column and problem type for training

Training in progress

Watch real-time training progress

Training results summary

View training summary with model ID and metrics

Model comparison table

Compare multiple models side-by-side

Comprehensive metrics display

Detailed evaluation metrics and statistics

AI-powered analysis

AI-generated insights about model performance

Step 6: Evaluation & Metrics

What You See

  • Comprehensive metrics tables
  • Model comparison (if multiple models)
  • Summary statistics for all features
  • AI-powered analysis of results
  • Enhanced domain-aware analysis (if domain selected)

What the System Does

  1. Metric Calculation: Computes comprehensive evaluation metrics
  2. Model Diagnostics: Performs statistical diagnostics (for regression)
  3. Feature Importance: Calculates SHAP values for interpretability
  4. Domain Analysis: Generates domain-aware explanations (if domain provided)
  5. Reliability Assessment: Evaluates model reliability and potential issues

Evaluation Metrics

For Regression Problems

  • R² (R-squared): Proportion of variance explained
  • RMSE: Root Mean Squared Error
  • MAE: Mean Absolute Error
  • MSE: Mean Squared Error

For Classification Problems

  • Accuracy: Overall correctness
  • Precision: True positives / (True positives + False positives)
  • Recall: True positives / (True positives + False negatives)
  • F1 Score: Harmonic mean of precision and recall

Why This Step Matters

Comprehensive evaluation ensures you understand:

  • How well your model performs
  • Which features are most important
  • Potential issues or limitations
  • How to interpret results in your domain context

Step 7: Insights and Explanations

What You See

  • AI-powered insights about model performance
  • Enhanced domain-aware analysis (if domain selected)
  • Feature importance visualizations
  • Model reasoning panel
  • Real-world implications

What the System Does

  1. Performance Summary: Explains model performance in plain language
  2. Feature Insights: Identifies which features matter most and why
  3. Model Reliability: Assesses model confidence and potential issues
  4. Domain Context: Provides domain-specific interpretations
  5. Real-World Implications: Explains how to use predictions in decision-making

Types of Insights

Standard AI Analysis

Generic analysis of model performance, metrics, and feature importance. Available when OpenAI is configured.

Enhanced Domain-Aware Analysis

When a domain is selected, Rassket provides:

  • Performance Summary: Domain-contextualized performance assessment
  • Key Feature Insights: Domain-specific feature interpretations
  • Model Reliability: Statistical reliability assessment
  • Real-World Implications: How to use predictions in your domain

Why This Step Matters

Insights transform raw metrics into actionable intelligence. They help you:

  • Understand what your model learned
  • Identify important patterns in your data
  • Make informed decisions based on predictions
  • Communicate results to stakeholders

Step 8: Visualizations

What You See

  • Feature importance charts
  • Confusion matrices (for classification)
  • ROC curves (for classification)
  • Prediction distribution plots
  • Correlation heatmaps
  • Regression/classification metrics visualizations
  • Advanced diagnostics (optional)

What the System Does

  1. Automatic Visualization: Generates default visualizations after training
  2. Chart Generation: Creates interactive Plotly charts
  3. Advanced Diagnostics: Optional additional charts (residual diagnostics, calibration curves, etc.)
  4. Chart Regeneration: Allows regenerating charts with different options

Default Visualizations

  • Feature Importance: Shows which features contribute most to predictions
  • Confusion Matrix: Classification accuracy breakdown (classification only)
  • ROC Curve: Classification performance visualization (classification only)
  • Prediction Distribution: Distribution of predicted values
  • Metrics Charts: Visual representation of evaluation metrics
  • Correlation Heatmap: Feature correlation matrix

Advanced Diagnostics

Optional additional visualizations:

  • Residual Diagnostics: Q-Q plots, residuals vs fitted, scale-location (regression)
  • R² Indicators: Goodness-of-fit visualizations (regression)
  • Calibration Curve: Prediction probability calibration (classification)
  • Outcomes Summary: Pie chart of prediction outcomes

Why This Step Matters

Visualizations help you:

  • Understand model behavior visually
  • Identify potential issues (overfitting, bias, etc.)
  • Communicate results to non-technical stakeholders
  • Validate model assumptions
Feature importance visualization

Interactive feature importance charts

Model metrics visualization

Visual representation of model performance

Comprehensive results display

Complete results with metrics and visualizations

Advanced visualizations

Additional diagnostic charts and visualizations

Model reasoning panel

Detailed model reasoning and analysis

Export options

Export model packages and PDF reports

Final results summary

Complete workflow summary and next steps

Step 9: Decision-Ready Outputs

What You See

  • Export Model Package button
  • Export PDF Report button
  • Prediction interface (for making new predictions)

What the System Does

  1. Model Package Export: Creates a ZIP file containing:
    • Trained model files
    • Preprocessing pipeline
    • Feature engineering code
    • Model metadata
    • Usage instructions
  2. PDF Report Export: Generates a comprehensive report including:
    • Executive summary
    • Model performance metrics
    • Visualizations
    • Feature importance analysis
    • Domain-aware insights
    • Recommendations
  3. Prediction Interface: Allows making predictions on new data

Why This Step Matters

Export capabilities ensure you can:

  • Deploy models to production
  • Share results with stakeholders
  • Integrate predictions into your workflows
  • Maintain model documentation
Export Formats: Model packages are ready for deployment. PDF reports are suitable for presentations and documentation. Both include all necessary information for reproducing and using your models.

Workflow Summary

The complete Rassket workflow:

  1. Upload Data → CSV file with headers
  2. Select Domain → Energy or Research (with optional sub-domain)
  3. Analyze Data → Understand problem type and get recommendations
  4. Train Models → Automated model selection and hyperparameter tuning
  5. Evaluate Results → Comprehensive metrics and diagnostics
  6. Review Insights → Domain-aware explanations and analysis
  7. Explore Visualizations → Interactive charts and diagnostics
  8. Export Outputs → Model packages and reports

Next Steps

Now that you understand the workflow: