Advertisement
  1. SEJ
  2.  ⋅ 
  3. Tools

Doing More with Less: Automated, High-Quality Content Generation

Discover how technical SEOs can use the latest advances in Natural Language Understanding and Generation to create quality content at scale.

Hamet Batista - Doing More with Less - Automated, High-Quality Content Generation

How do you continue to deliver amazing results with limited time and resources?

Writing quality content that educates and persuades is still a surefire way to achieve your traffic and conversion goals.

But the process is an arduous, manual job that doesn’t scale.

Fortunately, the latest advances in Natural Language Understanding and Generation offer some promising and exciting results.

For his SEJ eSummit session, Hamlet Batista discussed what is possible right now using practical examples (and code) that technical SEO professionals can follow and adapt for their business.

Here’s a recap of his presentation.

Automated, High-Quality Content Generation

Autocomplete Suggestions

How many times have you encountered this?

You start typing on Gmail and Google automatically completes the whole part and it’s super accurate.

You know, it’s really fascinating, but at the same time, it can be really scary.

You might already be using AI technology in your work without you even realizing it.

Gmail autocomplete

If you’re using Google Docs’ Smart compose feature, Gmail, or even Microsoft Word and Outlook, you’re already leveraging this technology.

This is part of your day as a marketer when you’re communicating with clients.

The great thing is this technology is not only accessible to Google.

Check out the Write With Transformer website, start typing, and hit the tab key for full sentence ideas.

Batista demonstrated how after plugging in the title and a sentence from a recent SEJ article, the machine can start generating lines – you just need to hit the autocomplete command.

Write with Transformer

All of the highlighted text above was completely generated by a computer.

The cool thing about this is that the technology that makes this possible is freely available and accessible to anybody who wants to use it.

Intent-Based Searches

One of the shifts we’re seeing right now in SEO is the transition to intent-based searches.

As Mindy Weinstein puts it in her Search Engine Journal article, How to Go Deeper with Keyword Research:

“We are in the era where intent-based searches are more important to us than pure volume.”

“You should take the extra step to learn the questions customers are asking and how they describe their problems.”

“Go from keywords to questions”

This change brings about an opportunity for us when we’re writing content.

The Opportunity

Search engines are answering engines these days.

And one effective way to write original, and popular content is to answer your target audience’s most important questions.

Take a look at this example for the query “python for seo”.

The first result shows we can leverage content that answers questions, in this case using FAQ schema.

FAQ search snippets take more real estate in the SERPs.

python for seo

However, doing this manually for every piece of content you’re going to create can be expensive and time-consuming.

But what if we can automate it by leveraging AI and existing content assets?

Leveraging Existing Knowledge

Most established businesses already have valuable, proprietary knowledge bases that they have developed over time just by normal interactions with customers.

Many times these are not yet publicly available (support emails, chats, internal wikis).

Open Source AI + Proprietary Knowledge

Through a technique called “Transfer Learning”, we can produce original, quality content by combining proprietary knowledge bases and public deep learning models and datasets.

Transfer Learning

There are differences between traditional machine learning (ML) and deep learning, .

In traditional ML, you’re primarily doing classifications and leveraging existing knowledge to come up with the predictions.

Now with deep learning, you’re able to tap into common sense knowledge that has been built over time by big companies like Google, Facebook, Microsoft, and others.

During the session, Batista demonstrated how this can be done.

How to Automate Content Generation

Below are the steps to take when reviewing automated question and answer generation approaches.

  • Source popular questions using online tools.
  • Answer them using two NLG approaches:
    • A span search approach.
    • A “closed book” approach.
  • Add FAQ schema and validate using the SDTT.

Sourcing Popular Questions

Finding popular questions based on your keywords is not a big challenge since there are free tools you can use to do this.

Answer the Public

Simply type a keyword and you can get plenty of questions that users are asking.

Answer the Public

Question Analyzer by BuzzSumo

They aggregate information from forums and other places. You can also find more long-tail type of questions.

Question Analyzer by BuzzSumo

AlsoAsked.com

This tool scrapes the People Also Ask questions from Google.

AlsoAsked.com

Question & Answering System

The Algorithm

Papers With Codes is a great source of leading-edge research about Question Answering.

It allows you to freely tap into the latest research that is being published.

Academics and researchers post their research so they can get feedback from their peers.

They’re always challenging each other to come up with a better system.

What’s more interesting is that even people like us can access the code that we’re going to need to answer the questions.

For this task, we’re going to use T5, or Text-to-Text Transfer Transformer.

The Dataset

We also need the training data that the system is going to use to learn to answer questions.

The Stanford Question Answering Dataset 2.0 (SQuAD 2.0) is the most popular reading comprehension dataset.

SQuAD 2.0

Now that we have both the data set and the code, let’s talk about the two approaches we can use.

  • Open-book question answering: You know where the answer is.
  • Closed-book question answering: You don’t know where the answer is.

Approach #1: A Span Search Approach (Open Book)

With three simple lines of code, we can get the system to answer our questions.

This is something you can do in Google Colab.

Create a Colab notebook and type the following:

!pip install transformers

from transformers import pipeline
# Allocate a pipeline for question-answering
nlp = pipeline('question-answering')
nlp({
    'question': 'What is the name of the repository ?',
    'context': 'Pipeline have been included in the huggingface/transformers repository'
})

When you type the command – providing a question, as well as the context that you think has the answer to the question – you will see the system basically perform a search for the string that has the answer.

{'answer': 'huggingface/transformers',
'end': 59,
'score': 0.5135626548884602,
'start': 35}

The steps are simple:

  • Load the Transformers NLP library.
  • Allocate a Question Answering pipeline.
  • Provide the question and context (content/text most likely to include the answer).

So how are you going to get the context?

With a few lines of code.

!pip install requests-html

from requests_html import HTMLSession
session = HTMLSession()

url = "https://www.searchenginejournal.com/uncover-powerful-data-stories-phyton/328471/"

selector = "#post-328471 > div:nth-child(2) > div > div > div.sej-article-content.gototop-pos"

with session.get(url) as r:

post = r.html.find(selector, first=True)

text = post.text

Using the request HTML library, you can pull the URL – which is equivalent to navigating the browser to the URL – and providing a selector (which is the path of the element of the block of text on the page.)

I should simply make a call to pull the content and add it to the text – and that becomes my context.

In this instance, we’re going to ask a question that is included in an SEJ article.

That means we know where the answer is. We’re providing the article that has the answer.

But what if we don’t know what article contains the answer then we’re trying to ask?

Approach #2: Exploring the Limits of NLG with T5 & Turing-NLG (Closed Book)

Google’s T5 (11-billion parameter model) and Microsoft’s TuringNG (17-billion parameter model) are able to answer questions without providing any context.

They are so massive that they’re able to keep a memory of a lot of things when they were training.

The Google’s T5 team went head-to-head with the 11-billion parameter model in a pub trivia challenge and lost.

Let’s see how simple it is to train T5 to answer our own arbitrary questions.

In this example, one of the questions Batista asked is “Who is the best SEO in the world?”

T5 answering questionsT5 answering arbitrary questions.

The best SEO in the world, according to a model that was trained, by Google is SEOmoz.

SEOmoz - best SEO according to T5

How to Train, Fine-Tune & Leverage T5

Training T5

We are going to train the 3-billion parameter model using a free Google Colab TPU.

Here is the technical plan for using T5:

Copy the Colab Notebook to Your Google Drive

  • Change the runtime environment to Cloud TPU.

Change the Runtime Environment to Cloud TPU

Change the Runtime Environment to Cloud TPU

Create a Google Cloud Storage Bucket

  • Provide the bucket path to the notebook.

Provide the Bucket Path to the Notebook

  • Select the 3-billion parameters model.

Select the 3-billion Parameters Model

  • Run the remaining cells up to the prediction step.

Run the Remaining Cells up to the Prediction Step

And now you’ve got a model that can actually answer questions.

But how do we add your proprietary knowledge so that it can answer questions in your domain or industry from your website?

Adding New Proprietary Training Datasets

This is where we go into the fine-tuning step.

Just click on the Fine-tune option in the model.

Fine-tune

And there are some examples in the code of how to create new functionality and how to give new capabilities to the model.

Remember to:

  • Preprocess your proprietary knowledge base into a format that can work with T5.
  • Adapt the existing code for this purpose (Natural Questions, TriviaQA).

To learn the extract, transform, and load process for machine learning, read Batista’s Search Engine Journal article, A Practical Introduction to Machine Learning for SEO Professionals.

Adding FAQ Schema

This step is straight forward.

Simply go to the Google documentation for the FAQ: Mark up your FAQs with structured data.

Google Developers - FAQ markup

Add the JSON-LD structure for that.

JSON-LD FAQ

Do you want to do it automatically?

Batista also wrote an article about it: A Hands-On Introduction to Modern JavaScript for SEOs.

With JavaScript, you should be able to generate this JSON-LD.

Resources to Learn More:

Watch this Presentation

You can now watch Batista’s full presentation from SEJ eSummit on June 2.


Image Credits

Featured Image: Paulo Bobita
All screenshots taken by author, July 2020

Category SEO Tools
ADVERTISEMENT
SEJ STAFF Loren Baker Founder at Foundation Digital

Loren Baker is the Founder of SEJ, an Advisor at Alpha Brand Media and runs Foundation Digital, a digital marketing ...

Doing More with Less: Automated, High-Quality Content Generation

Subscribe To Our Newsletter.

Conquer your day with daily search marketing news.