Back to Blog
Iteration Process

Start From Scratch: The Local Setup

Voice Kanban project timeline showing local setup phase from October 16-19

October 16-18: Day one of development. Setting up HP AI Studio environment, installing dependencies, getting Whisper running. The 8-cell template methodology begins.

The Shared Folder Problem

Professor Bartlett's warning: "If we all open those notepads at the same time in the shared folder, we're gonna be live editing together. I don't even know what will happen. I honestly don't know how the system will handle 10 different people all simultaneously trying to run, execute kernels, restart kernels. Who knows?"

The solution: copy everything to local storage.

HP AI Studio has two folder types: shared (owned by the professor, disappears when he logs out) and local (your machine's storage, persistent). The first step of any project: copy template files from shared to local.

The BERT Week 3 Template

We started with Professor's BERT template. Not because BERT is smart—"I promise you this BERT is done. This is not a smart BERT"—but because it demonstrates the workflow.

The process:

  1. Log into HP AI Studio
  2. Navigate to Projects → BERT Week 3
  3. Copy template files from shared to local folder
  4. Open "run_workflow" notebook (the legible template)
  5. Hit play on the Nemo Framework workspace
  6. Watch it download dependencies

The 8-Cell Methodology

Professor Bartlett's structure: break complex processes into discrete cells. Each cell has one job. If something breaks, you know exactly which piece failed.

Cell 1-4: Execution & Configuration

Cell 1: Start execution

Cell 2-4: Configure GPU settings, verify CUDA drivers

Cell 5-7: Install & Import

Cell 5-6: Install required Python packages (PyTorch, transformers, Whisper)

Cell 7: Import libraries into notebook

Cell 8-10: Data Loading

Cell 8: Verify assets exist

Cell 9: Preprocess data

Cell 10: Load from pandas dataset

This structure isn't arbitrary. It's defensive programming. Each cell validates before the next one runs.

Understanding Cell Execution

HP AI Studio notebooks show execution state with numbers and asterisks:

  • Number (1, 2, 3...): Cell completed execution in that order
  • Asterisk (*): Cell is waiting to execute or currently running
  • Empty: Cell hasn't run yet

When you restart the notebook, all numbers disappear. Then cells execute sequentially, each asterisk becoming a number once complete.

The Localization Constraint

HP AI Studio machines are powerful: 64GB RAM, NVIDIA GPUs, capable of running Large V3 locally. But there's a tradeoff: everything is localized to these machines.

Professor's advice:

"Look at your schedule through the week and make time to come in here so that you can work. You can only do that if you're in this space. Really start focusing on that."

Voice Kanban development required physical presence. No remote work. No cloud alternatives. The GPUs were here, so I had to be here.

Installing Whisper Dependencies

The actual installation process for Whisper Large V3:

# Cell 5: Install PyTorch and dependencies
!pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118

# Cell 6: Install Whisper and audio libraries
!pip install openai-whisper
!pip install sounddevice soundfile

# Cell 7: Import libraries
import torch
import whisper
import sounddevice as sd
import soundfile as sf

PyTorch installation took 10+ minutes. Whisper download: another 5-7 minutes for the Large V3 model (1.5GB). Audio libraries: quick, under a minute.

The Requirements.txt Cheat Code

Professor showed us a shortcut: requirements.txt files. Instead of manually installing each package, create a text file listing everything:

torch==2.1.0
openai-whisper==20231117
sounddevice==0.4.6
soundfile==0.12.1
gradio==4.4.1
transformers==4.35.0

Then run one command:

!pip install -r requirements.txt

Python reads the file and installs everything automatically. "It's like a cheat code to make it download stuff in the background."

Configuring the GPU

HP AI Studio machines recognize CUDA automatically. But templates built from scratch need explicit GPU configuration:

# Verify CUDA availability
import torch
print(f"CUDA available: {torch.cuda.is_available()}")
print(f"CUDA device: {torch.cuda.get_device_name(0)}")

# Set device
device = "cuda" if torch.cuda.is_available() else "cpu"
model = whisper.load_model("large-v3").to(device)

This code checks if NVIDIA GPUs are available, identifies which one, and loads the Whisper model directly onto GPU memory for faster inference.

The Login Ritual

Every session started the same way:

  1. Sign out of everything from previous user
  2. Hard shutdown the machine
  3. Power on, log in with your credentials
  4. Open HP AI Studio
  5. Navigate to your project
  6. Verify local files still exist (they should)
  7. Launch workspace

The ritual mattered. HP AI Studio's complex login tied to personal accounts. If you forgot to sign out, the next person couldn't access their work.

Breaking Down Complexity

Professor's repeated advice:

"We work fast and iteratively. We fail quickly, so we can deduce how to get to that functional solution. Don't be afraid to mix and match skills. Take a cell, put it into Claude, ask where to insert the system prompt. Test one block at a time."

Voice Kanban wasn't one monolithic codebase. It was dozens of discrete cells, each testable independently. GPU config works? Green light. Whisper loads? Green light. Audio capture works? Green light.

Incremental validation. No 1000-line code dumps. One piece at a time.

What Actually Got Built

By October 18, the local setup included:

  • Environment: HP AI Studio with GPU access configured
  • Dependencies: PyTorch, Whisper, audio libraries installed
  • Model: Whisper Large V3 downloaded and loaded on GPU
  • Workflow: 8-cell template structure established
  • Testing: Sample audio transcription working

No Kanban integration yet. No GitHub API. No task routing algorithms. But the foundation was solid.

The Week Six Reality Check

Professor's red flag alarm: "We're in week six. I decided to look on your blogs and right now is the time. Right now. This is the moment where you need to start actually tracking and cataloging your work."

The expectation: screenshot everything. Reflect on every attempt. Document successes and failures. Practice-based research through creation, not just reading.

Ideate → Create → Reflect. Think of the idea. Attempt to make it. Reflect on what worked. Repeat in quick succession.

Why This Mattered

October 16-18 wasn't about building the full system. It was about establishing foundations that wouldn't need rebuilding.

GPU configuration: done once, never touched again. Dependencies: installed once, version-pinned, stable. 8-cell structure: established once, extended infinitely.

The local setup phase taught defensive development: test each piece before integrating. Validate before advancing. Document as you build.

Voice Kanban wouldn't succeed because of brilliant algorithms or novel AI techniques. It would succeed because the foundation was methodically built, one cell at a time.

Austen Maccherone | UX Designer & Applied AI | Portfolio