python google nlp part 2
Estimated Read Time: 7 minute(s)
Common Topics: google, nlp, entities, data, flex

This is part 2 of a 2 part series. Please see Getting Started with Google NLP API Using Python first.

For Search Engines and SEO, Natural Language Processing (NLP) has been a revolution. NLP is simply the process and methodology for machines to understand human language. This is important for us to understand because machines are doing the bulk of page evaluation, not humans. While knowing at least some of the science behind NLP is interesting and beneficial, we now have the tools available to us to use NLP without needing a data science degree. By understanding how machines might understand our content, we can adjust for any misalignment or ambiguity. Let’s go!

In this intermediate tutorial part 2, using two web pages, I’ll show you how you can:

  • Compare entities and their salience between two web pages
  • Display missing entities between two pages

I highly recommend reading through the full Google NLP documentation for setting up the Google Cloud Platform, enabling the NLP API, and setting up authentication.

Requirements and Assumptions

  • Python 3 is installed and basic Python syntax understood
  • Access to a Linux installation (I recommend Ubuntu) or Google Colab
  • Google Cloud Platform account
  • NLP API Enabled
  • Credentials created (service account) and JSON file downloaded

Import Modules and Set Authentication

There are a number of modules we’ll need to install and import. If you are using Google Colab, these modules are preinstalled. If you are not, you will need to install the Google NLP module.

  • os – setting the environment variable for credentials
  • – Google’s NLP modules
  • pandas – for organizing data into a dataframes
  • fake_useragent – for generating a user agent when making a request
  • matplotlib – for the scatter plots

Out of those these 2 need to be installed. Google Colab has pandas installed but is outdated and we need the newest versions (as of this publish date)

!pip3 install fake_useragent

!pip3 install pandas==1.1.2

import os
from import language_v1
from import enums

from import language
from import types

import matplotlib.pyplot as plt
from matplotlib.pyplot import figure

from fake_useragent import UserAgent
import requests
import pandas as pd
import numpy as np

Next, we set our environment variable, which is a kind of system-wide variable that can be used across applications. It will contain the credentials JSON file for the API from Google Developer. Google requires it be in an environment variable. I am writing as if you are using Google Colab, which is the code block below (don’t forget to upload the file). To set the environment variable in Linux (I use Ubuntu) you can open ~/.profile and ~/.bashrc and add this line export GOOGLE_APPLICATION_CREDENTIALS="path_to_json_credentials_file". Change “path_to_json_credentials_file” as necessary. Keep this JSON file very safe.

os.environ['GOOGLE_APPLICATION_CREDENTIALS'] = "path_to_json_credentials_file"

Build NLP Function

Since we are using the same process to evaluate both pages we can create a function. This helps reduce redundant code. This function named processhtml() shown in the code below will:

  1. Create a new user agent for the request header
  2. Make the request to the web page and store the HTML content
  3. Initialize the Google NLP
  4. Communicate to Google that you are sending them HTML, rather than plain text
  5. Send the request to Google NLP
  6. Store the JSON response
  7. Convert the JSON into a python dictionary with the entities and salience scores (adjust rounding as needed)
  8. Convert the keys to lower case (for comparing)
  9. Return the new dictionary to the main script
def processhtml(url):

    ua = UserAgent() 
    headers = { 'User-Agent': } 
    res = requests.get(url,headers=headers) 
    html_page = res.text

    url_dict = {}

    client = language_v1.LanguageServiceClient()

    type_ = enums.Document.Type.HTML

    language = "en"
    document = {"content": html_page, "type": type_, "language": language}

    encoding_type = enums.EncodingType.UTF8

    response = client.analyze_entities(document, encoding_type=encoding_type)

    for entity in response.entities:
        url_dict[] = round(entity.salience,4)

    url_dict = {k.lower(): v for k, v in url_dict.items()}

    return url_dict

Process NLP Data and Calculate Salience Difference

Now that we have our function we can set the variables storing the web page URLs we want to compare and then send them to the function we just made.

url1 = "" 
url2 = "" 

url1_dict = processhtml(url1)
url2_dict = processhtml(url2)

We now have our NLP data for each URL. It’s time to compare the two entity lists and if there are matches, calculate the difference in salience if your competitor’s is higher. This code snippet:

  1. Create an empty datafame with 4 columns (Entity, URL1, URL2, Difference). URL1 and URL2 will contain the salience scores for each entity for that URL.
  2. Compare each entity between each list and if there is a match add the salience score for each in a variable
  3. If your competitor salience score for a keyword is greater than yours, record the difference (adjust the rounding as needed)
  4. Add the new comparison data for the entity to the dataframe
  5. Print out the dataframe after all entities are done being matched
df = pd.DataFrame([], columns=['Entity','URL1','URL2','Difference'])

for key in set(url1_dict) & set(url2_dict):
    url1_keywordnum = str(url1_dict.get(key,"n/a"))
    url2_keywordnum = str(url2_dict.get(key,"n/a"))
    if url2_keywordnum > url1_keywordnum:
        diff = str(round(float(url2_keywordnum) - float(url1_keywordnum),3))
        diff = "0"

    new_row = {'Keyword':key,'URL1':url1_keywordnum,'URL2':url2_keywordnum,'Difference':diff}
    df = df.append(new_row, ignore_index=True)

print(df.sort_values(by='Difference', ascending=False))

Example Output

This result tells us there are at least 9 entities found on both pages that are deemed by Google NLP more important (relative to the whole text) on the competitor page. These are keywords you may want to investigate and consider ways to communicate better on your page. I have rounded the salience scores to 3 decimal places, feel free to adjust to uncover finer differences.

NLP Entities

Find Difference in Named Entities

Next, it can be useful, especially for a competitor page that is outranking our page, to find the entities that exist on their page that is missing from your page. This snippet below:

  1. Use set() to compare entities between the two dictionaries. The entities found in the competitor list but not in your list, are left and stored in diff_lists.
  2. set() strips values from the dictionary, which is our salience score, so we need to add them back in
  3. Add the final_diff dictionary list we created to the dataframe
  4. Print the dataframe and sort by score descending
diff_lists = set(url2_dict) - set(url1_dict)

final_diff = {}

for k in diff_lists:
  for key,value in url2_dict.items():
    if k == key:

df = pd.DataFrame(final_diff.items(), columns=['Keyword','Score'])

print(df.head(25).sort_values(by='Score', ascending=False))

Example Output

This list shows the top 25 (adjust head() for more) entities by salience that appears on the competitor page that doesn’t appear on your page. This is useful to find entity opportunities where pages that outrank you are using but you are not.

NLP Entity Comparison


I hope you enjoyed this two-part series on getting started with NLP and how to compare entities between web pages. These scripts are foundations and can be extended to the limit of your imagination. Explore data blending with other sources and mine for further insights. Enjoy and as always, follow me on twitter and let me know what you think and how we’re using Google NLP!

Google NLP and Entities FAQ

How can I utilize Google NLP with Python to compare entities between two web pages?

Leverage the Google NLP API and Python scripts to extract entities from two web pages and compare the obtained entity lists for similarities or differences.

What Python libraries are commonly used for interacting with Google NLP to compare entities?

The primary Python library for interacting with the Google NLP API is google-cloud-language. Use this library to send requests, analyze web page entities, and compare the results.

What specific entities can be compared between two web pages using Google NLP and Python?

Google NLP can identify various entities, including people, organizations, locations, and more. Python scripts can extract and compare these entities between two web pages.

Are there any limitations or considerations when comparing entities using Google NLP and Python?

Consider the accuracy of entity extraction, potential false positives or negatives, and the need for pre-processing to handle variations in entity recognition when comparing entities between web pages.

Where can I find examples and documentation for using Google NLP to compare entities between web pages with Python?

Refer to the official Google Cloud documentation for the NLP API, which includes guides, reference documentation, and examples in Python. Additionally, explore online resources and tutorials for practical examples of comparing entities using Google NLP and Python.

Greg Bernhardt
Follow me

Leave a Reply