PitchHut logo
Descent Visualisers
Interactive tools to visualize machine learning algorithm behaviors.
Pitch

Descent Visualisers offers interactive labs that illuminate the behavior of various machine learning algorithms. From gradient descent to reinforcement learning, explore optimizers and environments in real-time with 3D visualizations. Customize your experience by adjusting hyperparameters and inputting your own loss surfaces.

Description

Descent Visualisers provides a suite of interactive labs designed to enhance understanding of various Machine Learning algorithms through engaging visualizations. This project focuses on two primary areas of study: Gradient Descent and Reinforcement Learning.

Gradient Descent Visualiser

The Gradient Descent Visualiser is a dynamic tool that allows users to observe how different gradient descent algorithms operate on complex loss surfaces in real time. Key features include:

  • Exploration of optimizers such as Vanilla SGD, Momentum, Adagrad, RMSProp, and Adam.
  • Real-time 3D visualizations facilitated by Plotly.
  • The ability to input custom loss surfaces by utilizing various mathematical functions.
  • Adjustable hyperparameters including learning rate, momentum, decay, and beta values.
  • Functionalities to play, pause, reset animations, and reinitialize points.

Preview Demonstrations

3D Loss SurfaceDifferent OptimizersCustom Surface
Loss SurfaceOptimizersCustom Surface

Reinforcement Learning Visualiser

The Reinforcement Learning Visualiser explores the behavior of various Reinforcement Learning algorithms such as Q-learning (QL), Deep Q-learning (DQL), and Proximal Policy Optimization (PPO) within custom simulated environments.

Environments & Features

Maze Environment

  • Utilizes tabular Q-Learning for training.
  • Adjustable hyperparameters to customize learning dynamics.
  • Editable configurations for unique environment setups.
  • Visualized training steps to observe learning progression.
  • Play mode for interacting with trained agents.

CartPole Environment

  • Trained using both DQL and PPO strategies.
  • Adjustable hyperparameters for optimal training experiences.
  • Configuration of environment parameters for tailored scenarios.
  • Visual aids to display training steps and agent behavior.
  • Play mode to observe the performance of trained agents.

Preview Demonstrations

Maze EnvironmentCartPole Environment
MazeCartPole

Live Demo

Access the live demo at: Descent Visualisers Live Demo

Engaging with these visualizations enables a deeper comprehension of algorithm behavior and optimizes the learning process in machine learning and reinforcement learning domains.

0 comments

No comments yet.

Sign in to be the first to comment.