Page cover

Clip web data to spreadsheet

If you've ever copied and pasted data from a website into a spreadsheet, this tutorial is for you. We're going to build a scraper that pulls info from any webpage and drops it straight into Google Sheets or Excel, automagically.

No code. Just a few clicks with a prompt and you're done.


What You'll Need

  • A free PixieBrix account

  • The Chrome extension installed

  • A Google Sheet or Excel file ready to go


Step 1: Set Up Your Spreadsheet

Before you do anything in PixieBrix, open Google Sheets (or Excel) and create a new spreadsheet. Add a header row at the top with the fields you want to collect. Think of these are what you are telling PixieBrix to look for on the page.

For example, if you're scraping LinkedIn profiles, your headers might look like:

| Name | Title | Company | URL |

If you're scraping Eventbrite for events:

| Event Name | Date | Location | URL |

The header names don't have to be fancy. Just make them clear and be ready to reference them when PixieBrix asks what you want to scrape.


Step 2: Go to the Page You Want to Scrape

Go to to the website you want to pull data from. PixieBrix works for scraping on most pages, including:

Once you're on the page, open the PixieBrix Page Editor. You can do this by hovering on the floating PixieBrix logo on the page and clicking the brick icon.


Step 3: Click "Clip Web Data to Spreadsheet"

Inside the Page Editor, you'll see a list of ready-to-use prompts, but use this one for your custom webscraper. Copy the following prompt:

... and click "Generate"

This kicks off a quick setup conversation where PixieBrix asks you a few questions:

  • What site are you scraping? (e.g., LinkedIn, Airbnb, Trustpilot, Eventbrite)

  • What info do you want to grab? (e.g., name, company, job title, URL, rating, price)

  • Where do you want to send it? Google Sheets or Excel

You'll be prompted to connect to your spreadsheet tool of choice. Just follow the steps, then select the spreadsheet you just created.


Step 4: Open the Mod and Test It

Once it's configured, click Open Mod. This opens your custom scraper so you can see what was built.

Click Try on <your site> and then right click on the LinkedIn page to see your new scraping menu.

PixieBrix will look at the current page, extract the data you asked for, and send it to your spreadsheet.

circle-info

Don't see the Try on <your site> modal? Just click the "Test" button in the top of the Page Editor and it will run on the connected page; no need to click the context menu.


Step 5: Check Your Spreadsheet

Flip over to your spreadsheet. You should see a new row with the data from that page filled in. If everything looks right, you're good to go.


Step 6: Tweak It If You Need To

Not quite right? No problem. Head back to the Page Editor and use the chat copilot on the left side to adjust things. You can describe what you want to change in plain language, like "also grab the email address" or "skip the name column."

You don't have to dig into this, but if you're curious or want to customize further, here's how the scraper is built:

The middle panel shows the steps in your workflow, called bricks. You'll see two main ones:

  1. Extract content from the page (this is the AI doing the scraping)

  2. Send to Google Sheets (or Excel, whichever you chose)

Click on any brick and the right panel shows you the settings for that step. You can adjust what fields you're pulling, change which spreadsheet it sends to, or add more actions.

For example, you could add a third brick to post the scraped data to Slack, or send yourself a notification. The workflow is yours to build on.


Step 7: Save the mod to use it again

Before you close the Page Editor, click the Save button in the top to create a username and save the mod so you can run it on any page without having to open the Page Editor.


That's It

You just built a web scraper. No code, no complicated setup, just a prompt and a few clicks. If you want to scrape more pages, just go to the next one and hit the button again. PixieBrix handles the rest.

Last updated

Was this helpful?