How To Use Python For SEO Redirect Mapping

How to use python for seo redirect mapping

In This Article...

TLDR; Summary

Python can automate SEO redirect mapping by comparing old and new website crawl data using fuzzy matching, dramatically reducing the time needed during site migrations.

This helps preserve rankings, prevent 404 errors and generate scalable redirect recommendations with far greater accuracy than manual mapping.

Learn how to do it in 8 easy steps.

Website migrations often lead to broken links and ranking losses if redirects are not mapped correctly. Python makes this process faster and more accurate by automating URL matching using page titles, headings and fuzzy matching.

Instead of manually mapping hundreds or thousands of URLs, you can use Python to compare old and new site crawl data and generate redirect recommendations in minutes. This is especially useful during SEO website migrations, where preserving rankings, traffic and user experience is critical.

Why Use Python for Redirect Mapping?

Manual redirect mapping becomes difficult at scale and increases the risk of:

  • Missed URLs
  • Incorrect redirects
  • Redirect chains
  • Lost SEO equity
  • 404 errors after launch

Python helps automate the process by:

  • Comparing old and new URLs
  • Matching pages using titles and headings
  • Scoring similarity between pages
  • Exporting redirect recommendations for review

For larger projects, this sits naturally alongside broader technical SEO work, including crawl optimisation, indexation checks and post-launch monitoring.

What You’ll Need

  • The latest version of Python
  • VS Code (or alternative code editor)
  • Screaming Frog (or alternative website crawler)

How To Install Python

Download the latest version of Python from:

Python.org

During installation:

  • Tick “Add Python to PATH”
  • Click Install Now

Once installed, restart your terminal or command prompt.

Check that Python is installed correctly

Open Terminal (Mac) or Command Prompt (Windows) and run:

Bash
python --version

You should see something similar to:

Terminal
Python 3.12.3

If that does not work on Windows, try:

Bash
py --version

Create a Project Folder

Create a new folder somewhere easy to access, for example:

File
seo-redirect-mapping

This will be the folder you will place your Screaming Frog crawls in.


Open the Folder in VS Code

Download:

Visual Studio Code

Then:

  • Open VS Code
  • Click File → Open Folder
  • Select your project folder

Create a new Python file called:

File
redirect-mapping.py

You are now ready to install the required libraries and run the script.

Gather Website Crawls

Before starting, gather crawl exports from both the old and new websites using a crawler such as Screaming Frog.

Download:

Screaming Frog

Your exports should include:

  • URL
  • Page title
  • H1
  • H2

Save both crawl exports as .xlsx files.


Preliminary Steps

Before running the steps below, open the integrated terminal in VS Code by clicking Terminal → New Terminal from the top menu.

This opens a terminal window directly inside your project folder. The pip install command in Step 1 is entered here.

For Steps 2 to 8, paste all of the code blocks into your redirect-mapping.py file in order, then run the script from the terminal using python redirect-mapping.py.

Before installing any libraries, it is recommended to set up a virtual environment. This keeps your project dependencies isolated and avoids conflicts with other Python projects on your machine.

Virtual environment — Mac

Bash
python -m venv venv
source venv/bin/activate

Virtual environment — Windows

Bash
python -m venv venv
venv\Scripts\activate

Once activated, you will see the environment name appear in your terminal. All libraries installed with pip will now be contained within this project folder. To deactivate the environment when you are finished, simply run deactivate as below:

Terminal
deactivate

Make sure the virtual environment is active in your terminal before moving on to Step 1. You will see (venv) at the start of your terminal prompt when it is activated. It should look a little something like this:

Terminal
(venv) your-machine-name:seo-redirect-mapping yourname$

Step-by-Step Guide

1. Install Required Libraries

Bash
pip install pandas polyfuzz rapidfuzz openpyxl

rapidfuzz is required separately for the matching model import to work correctly.


2. Import Libraries

Python
import pandas as pd
from polyfuzz import PolyFuzz
from polyfuzz.models import RapidFuzz

3. Load Crawl Data

Python
old_urls = pd.read_excel("old_site_crawl.xlsx")
new_urls = pd.read_excel("new_site_crawl.xlsx")

old_urls = old_urls.rename(columns={'Address': 'Old URL', 'Title 1': 'Old Title'})
new_urls = new_urls.rename(columns={'Address': 'New URL', 'Title 1': 'New Title'})

This assumes your Screaming Frog crawl exports use the default column names Address and Title 1, which are renamed in the code for clarity:

  • Title 1 → renamed to Old Title / New Title
  • Address → renamed to Old URL / New URL


4. Prepare Title Data

Python
old_titles_df = old_urls[['Old URL', 'Old Title']].dropna()
new_titles_df = new_urls[['New URL', 'New Title']].dropna()

old_titles = old_titles_df['Old Title'].tolist()
new_titles = new_titles_df['New Title'].tolist()

5. Run Fuzzy Matching

Python
matcher = PolyFuzz(RapidFuzz())
matcher.match(old_titles, new_titles)

results = matcher.get_matches()["RapidFuzz"]

The output includes:

  • Original title
  • Suggested matching title
  • Similarity score

6. Filter High-Confidence Matches

Python
filtered_results = results[results["Similarity"] >= 0.90]

You can lower this threshold for larger or messier migrations, but anything below 0.80 should usually be reviewed manually.


7. Map Matches Back to URLs

Python
redirect_map = filtered_results.merge(
    old_urls[['Old URL', 'Old Title']],
    left_on='From',
    right_on='Old Title'
)

redirect_map = redirect_map.merge(
    new_urls[['New URL', 'New Title']],
    left_on='To',
    right_on='New Title'
)

final_redirects = redirect_map[
    ['Old URL', 'New URL', 'Similarity']
].copy()

final_redirects.columns = [
    'Old URL',
    'New URL',
    'Confidence Score'
]

8. Export the Redirect Map

Python
final_redirects.to_excel(
    "redirect_mapping.xlsx",
    index=False
)

You will now have an Excel file containing:

  • Old URL
  • Suggested new URL
  • Match confidence score

Review lower-confidence matches manually before implementation.


Tips for Better Redirect Matching

To improve accuracy:

  • Match using H1s if titles are inconsistent
  • Remove duplicate titles before matching
  • Exclude paginated or parameter URLs
  • Manually override important pages
  • Review all low-confidence matches

For important migrations, it is also worth carrying out an SEO audit before launch so technical issues, content gaps and redirect risks are identified early.

Final Thoughts

Python dramatically simplifies redirect mapping during SEO migrations. By automating title and content matching, you can generate scalable redirect recommendations quickly while reducing manual errors.

The workflow above gives you a reliable foundation that works well for most migrations and can easily be expanded for larger, more complex websites.

Picture of Written By Nigel Adams

Written By Nigel Adams

Nigel is a freelance digital marketing consultant specialising in search marketing, SEO, and paid search (PPC). He helps businesses increase their visibility, attract high-quality traffic, and generate measurable growth through data-driven search marketing strategies.

About Me

Related Posts

Start Seeing Results

Ready To Start?

Get in touch for a free consultation.

Get In Touch

Please fill in your details below and I’ll be in touch. Alternatively, give me a call on 07930 533394

This site is protected by reCAPTCHA Enterprise and the Google Privacy Policy and Terms of Service apply.