Skip to main content
J+IQ dashboard interface showing automated competitive report generation
Back to all work

Automating Analyst Workflows Across 3 Internal Products

Role: UX Designer Company: Jellyfish / BrandTech Group Period: 2022-2024

TL;DR

Three internal products, one common pattern: complex manual workflows holding back practitioners. J+IQ reduced competitive analysis from 72 hours to 15 minutes, saving 1,000+ hours in its first five months across 130 users in five cities. J+Bidding replaced a 5-hour manual API process with a 1-minute interface, supporting $800K+ in revenue. J+Carbon shipped a carbon footprint calculator for advertising campaigns in under two months, developed in partnership with Scope3.

Context

Jellyfish is a digital marketing agency within the BrandTech Group, operating a global network of advertising practitioners: account managers, data planners, media strategists, and paid media specialists working across dozens of countries. Their internal tool suite, the J+ platform, needed to automate the manual analytical work that consumed hours of practitioner time every day.

I worked on three products in this suite between 2022 and 2024. Each started from a different place but shared the same core problem: skilled practitioners spending their time on repetitive manual processes instead of the strategic work they were hired to do.

J+IQ

Context

J+IQ was an automated competitive intelligence tool that generated analysis reports comparing a brand against up to five competitors across digital footprint and Amazon market analysis dimensions. Before automation, building a single competitive report required analysts to manually gather data from SimilarWeb, PageSpeed, and other providers, build comparisons, format presentations, and deliver them to clients. The process took approximately 72 hours of analyst time per report.

The tool already existed when I joined, but it was not working well enough for broad rollout. The interface had usability issues that created friction across the workflow, and the product team needed validation before deploying to 130+ analysts across five global offices.

What I Designed

I redesigned the core workflows:

J+IQ dashboard showing analysis list with brand reports and generation progress tracker

Usability Validation

I designed and ran a multi-method usability study to validate the redesign. I recruited 7 users from 5 cities (Mexico City, Sao Paulo, Paris, Johannesburg, Brighton) across three experience levels: new users, single-use users, and frequent users. Each session combined a 30-minute exploration task with the standardized SUS questionnaire. I documented findings page by page, categorizing issues by type and providing specific improvement recommendations for every screen.

The tool achieved a SUS score of 83.9, rated "GOOD" and above the 68 industry average. Task completion hit 100%. Individual scores ranged from 75 to 97.5. The study also surfaced targeted fixes: the "Discover competitors" button was not immediately visible, URL copy-paste handling created friction, and status indicator dots were ambiguous. These were the kind of small issues that compound at scale across 130 users.

J+IQ competitor suggestion engine showing ranked similarity scores from SimilarWeb, organic, and paid data sources

J+IQ Results

J+Bidding

Context

J+Bidding was an ad spend optimization tool spanning two platforms: Meta and Google's DV360. Paid media practitioners needed to apply bid multipliers to fine-tune campaign targeting by demographics, geography, and other audience parameters.

Before this tool, applying bid multipliers on Meta required using the Graph Explorer API directly. This meant writing scripts, knowing technical targeting codes for countries, cities, and demographics, and having no visibility into which campaigns already had multipliers applied or how those multipliers affected spend. The process took approximately 5 hours per calculation, and most practitioners simply could not do it.

Research

I led end-to-end discovery across three countries (Spain, UK, US) with bilingual interview protocols in English and French. I designed distinct question sets for stakeholders (9 categories) and users (7 categories). From this I developed two personas representing the strategic director who needed portfolio-level visibility and the campaign manager who needed step-by-step guidance through a process she found intimidating.

I mapped the current five-stage workflow (Review, Initial Extraction, Modification and Implementation, Sending the Bid, Post-Implementation) with emotional tracking at each stage. Frustration peaked during the modification and implementation phase, the point where practitioners had to manually write API calls and guess at technical codes.

I also conducted shadow sessions observing actual bid management workflows and facilitated workshops that produced a categorized feature roadmap across four domains: bid management, dashboard/overview, simulation/prediction, and documentation/tooltips.

What I Designed

I built J+Bidding from scratch:

J+Bidding Bid Multiplier UI showing audience set configuration with demographic categories and Send to Meta action J+Bidding DV360 script setup interface showing auto-generate and manual creation paths

J+Bidding Results

J+Carbon

Context

J+Carbon was a carbon footprint calculator for digital advertising campaigns, developed in partnership with Scope3. The tool let practitioners measure the environmental impact of their media spend and generate sustainability reports.

What I Designed

This was a fast-turnaround project, shipped in under two months with minimal research. I designed:

The product also included a Sustainability in Action Slack channel to build an internal community around eco-friendly advertising practices.

J+Carbon landing page showing eco report generation and CSV upload features with platform integrations

Reflections

J+IQ and J+Bidding both ended up in the same place: 99%+ time reductions. But the research that got them there looked completely different. J+IQ needed validation of something that already existed. J+Bidding needed discovery for something that did not exist yet. The methodology I built on J+IQ directly informed how I approached J+Bidding eighteen months later.

J+Carbon was the fastest of the three. The problem was well scoped, the partnership with Scope3 provided the data layer, and the patterns I had already established on the other two products carried over. Under two months from start to launch.

Key Artifacts