Run in Google Colab | View source on GitHub |
This notebook demonstrates how to run inference on your custom framework using the ModelHandler class.
Named-entity recognition (NER) is one of the most common tasks for natural language processing (NLP). NLP locates named entities in unstructured text and classifies the entities using pre-defined labels, such as person name, organization, date, and so on.
This example illustrates how to use the popular spaCy
package to load a machine learning (ML) model and perform inference in an Apache Beam pipeline using the RunInference PTransform
.
For more information about the RunInference API, see About Beam ML in the Apache Beam documentation.
Install package dependencies
The RunInference library is available in Apache Beam versions 2.40 and later.
For this example, you need to install spaCy
and pandas
. A small NER model, en_core_web_sm
, is also installed, but you can use any valid spaCy
model.
# Uncomment the following lines to install the required packages.
# %pip install spacy pandas
# %pip install "apache-beam[gcp, dataframe, interactive]"
# !python -m spacy download en_core_web_sm
Learn about spaCy
To learn more about spaCy
, create a spaCy
language object in memory using spaCy
's trained models.
You can install these models as Python packages.
For more information, see spaCy's Models and Languages documentation.
import spacy
nlp = spacy.load("en_core_web_sm")
# Add text strings.
text_strings = [
"The New York Times is an American daily newspaper based in New York City with a worldwide readership.",
"It was founded in 1851 by Henry Jarvis Raymond and George Jones, and was initially published by Raymond, Jones & Company."
]
# Check which entities spaCy can recognize.
doc = nlp(text_strings[0])
for ent in doc.ents:
print(ent.text, ent.start_char, ent.end_char, ent.label_)
The New York Times 0 18 ORG American 25 33 NORP daily 34 39 DATE New York City 59 72 GPE
# Visualize the results.
from spacy import displacy
displacy.render(doc, style="ent")
# Visualize another example.
displacy.render(nlp(text_strings[1]), style="ent")
Create a model handler
This section demonstrates how to create your own ModelHandler
so that you can use spaCy
for inference.
import apache_beam as beam
from apache_beam.options.pipeline_options import PipelineOptions
import warnings
warnings.filterwarnings("ignore")
pipeline = beam.Pipeline()
# Print the results for verification.
with pipeline as p:
(p
| "CreateSentences" >> beam.Create(text_strings)
| beam.Map(print)
)
The New York Times is an American daily newspaper based in New York City with a worldwide readership. It was founded in 1851 by Henry Jarvis Raymond and George Jones, and was initially published by Raymond, Jones & Company.
# Define `SpacyModelHandler` to load the model and perform the inference.
from apache_beam.ml.inference.base import RunInference
from apache_beam.ml.inference.base import ModelHandler
from apache_beam.ml.inference.base import PredictionResult
from spacy import Language
from typing import Any
from typing import Dict
from typing import Iterable
from typing import Optional
from typing import Sequence
class SpacyModelHandler(ModelHandler[str,
PredictionResult,
Language]):
def __init__(
self,
model_name: str = "en_core_web_sm",
):
""" Implementation of the ModelHandler interface for spaCy using text as input.
Example Usage::
pcoll | RunInference(SpacyModelHandler())
Args:
model_name: The spaCy model name. Default is en_core_web_sm.
"""
self._model_name = model_name
self._env_vars = {}
def load_model(self) -> Language:
"""Loads and initializes a model for processing."""
return spacy.load(self._model_name)
def run_inference(
self,
batch: Sequence[str],
model: Language,
inference_args: Optional[Dict[str, Any]] = None
) -> Iterable[PredictionResult]:
"""Runs inferences on a batch of text strings.
Args:
batch: A sequence of examples as text strings.
model: A spaCy language model
inference_args: Any additional arguments for an inference.
Returns:
An Iterable of type PredictionResult.
"""
# Loop each text string, and use a tuple to store the inference results.
predictions = []
for one_text in batch:
doc = model(one_text)
predictions.append(
[(ent.text, ent.start_char, ent.end_char, ent.label_) for ent in doc.ents])
return [PredictionResult(x, y) for x, y in zip(batch, predictions)]
# Verify that the inference results are correct.
with pipeline as p:
(p
| "CreateSentences" >> beam.Create(text_strings)
| "RunInferenceSpacy" >> RunInference(SpacyModelHandler("en_core_web_sm"))
| beam.Map(print)
)
The New York Times is an American daily newspaper based in New York City with a worldwide readership. It was founded in 1851 by Henry Jarvis Raymond and George Jones, and was initially published by Raymond, Jones & Company. PredictionResult(example='The New York Times is an American daily newspaper based in New York City with a worldwide readership.', inference=[('The New York Times', 0, 18, 'ORG'), ('American', 25, 33, 'NORP'), ('daily', 34, 39, 'DATE'), ('New York City', 59, 72, 'GPE')]) PredictionResult(example='It was founded in 1851 by Henry Jarvis Raymond and George Jones, and was initially published by Raymond, Jones & Company.', inference=[('1851', 18, 22, 'DATE'), ('Henry Jarvis', 26, 38, 'PERSON'), ('Raymond', 39, 46, 'PERSON'), ('George Jones', 51, 63, 'PERSON'), ('Raymond, Jones & Company', 96, 120, 'ORG')])
Use KeyedModelHandler
to handle keyed data
If you have keyed data, use KeyedModelHandler
.
# You can use these text strings with keys to distinguish examples.
text_strings_with_keys = [
("example_0", "The New York Times is an American daily newspaper based in New York City with a worldwide readership."),
("example_1", "It was founded in 1851 by Henry Jarvis Raymond and George Jones, and was initially published by Raymond, Jones & Company.")
]
from apache_beam.runners.interactive.interactive_runner import InteractiveRunner
from apache_beam.ml.inference.base import KeyedModelHandler
from apache_beam.dataframe.convert import to_dataframe
pipeline = beam.Pipeline(InteractiveRunner())
keyed_spacy_model_handler = KeyedModelHandler(SpacyModelHandler("en_core_web_sm"))
# Verify that the inference results are correct.
with pipeline as p:
results = (p
| "CreateSentences" >> beam.Create(text_strings_with_keys)
| "RunInferenceSpacy" >> RunInference(keyed_spacy_model_handler)
# Generate a schema suitable for conversion to a dataframe using Map to Row objects.
| 'ToRows' >> beam.Map(lambda row: beam.Row(key=row[0], text=row[1][0], predictions=row[1][1]))
)
# Convert the results to a pandas dataframe.
import apache_beam.runners.interactive.interactive_beam as ib
beam_df = to_dataframe(results)
df = ib.collect(beam_df)
df