Home

Awesome

PROJECT NOT UNDER ACTIVE MANAGEMENT

This project will no longer be maintained by Intel.

Intel has ceased development and contributions including, but not limited to, maintenance, bug fixes, new releases, or updates, to this project.

Intel no longer accepts patches to this project.

If you have an ongoing need to use this project, are interested in independently developing it, or would like to maintain patches for the open source software community, please create your own fork of this project.

Contact: webadmin@linux.intel.com

Intelligent Indexing

Introduction

Many industries ingest massive volumes of complex documents and must utilize manual processes to both understand the contents of and route them to the relevant parties. AI-based Natural Language Processing (NLP) solutions for classifying documents can be one solution to automate this process, saving massive amounts of workforce, time, and cost while still maintaining human-level performance.

This example demonstrates one way of building an NLP pipeline for classifying documents to their respective topics and describe how we can leverage the Intel® oneAPI AI Analytics Toolkit (oneAPI) to accelerate the pipeline.

Solution Technical Overview

Methodology wise, the use case will train a Support Vector Classifiers (SVC) for multiclass classification which ingests a body of text and outputs the predicted topic of the document. At deployment, natural text is first mapped into Term Frequency-Inverse Document Frequency (TFIDF) vectors, which is then fed into our trained SVC to obtain predictions about the potential topic of the original text. SVC is a commonly and historically used algorithm for building powerful NLP classifiers using Machine Learning (ML) due to its ability to tackle the highly non-linear and complex relationships often found in text documents [1][2]. With recent advancements in NLP based solutions, it can be seen as a starting point before considering more advanced Deep Learning (DL) based NLP algorithms.

workflow

The savings gained from using Intel® oneAPI Data Analytics Library (oneDAL) can result in more efficient model training and inference, leading to more robust Artificial Intelligence (AI) powered systems.

oneDAL is used to achieve quick results even when the data for a model are huge. It provides the capability to reuse the code present in different languages so that the hardware utilization is optimized to provide these results.

The solution contained in this repo uses the following Intel® packages:

For more details, visit Intel® Distribution for Python*, Intel® Distribution of Modin* and Intel® Extension for Scikit-Learn*.

Solution Technical Details

In this section, we describe the data and how to replicate the results. The included code demonstrates a complete framework for:

  1. Setting up a virtual environment for Intel®-accelerated ML.
  2. Preprocessing data using Intel® Distribution of Modin* and NLTK*.
  3. Training an NLP model for text classification using Intel® Extension for Scikit-learn*.
  4. Predicting from the trained model on new data using Intel® Extension for Scikit-learn*.

Dataset

The dataset used for this demo is a set of ~200k news article with their respective topics obtained by mining the Huffington Post website originally obtained from https://www.kaggle.com/datasets/rmisra/news-category-dataset.

Please see this data set's applicable license for terms and conditions. Intel does not own the rights to this data set and does not confer any rights to it.

The included dataset is lightly preprocessed from above to split into train/test according to an 85:15 train test split. To download and setup this dataset for benchmarking, follow the instructions listed here.

Validated Hardware Details

There are workflow-specific hardware and software setup requirements to run this use case.

Recommended Hardware
CPU: Intel® 2nd Gen Xeon® Platinum 8280 CPU @ 2.70GHz or higher
RAM: 187 GB
Recommended Free Disk Space: 20 GB or more

Minimal Requirements

How it Works

To demonstrate the application of multi class document classification, using the News Category Dataset, we will build a model to predict the category of each news article based entirely on the headline, short_description, and URL of the given news article. In total, there are 42 unique categories which are described here.

Get Started

Start by defining an environment variable that will store the workspace path, this can be an existing directory or one to be created in further steps. This ENVVAR will be used for all the commands executed using absolute paths.

export WORKSPACE=$PWD/intelligent-indexing

Define DATA_DIR and OUTPUT_DIR as follows:

export DATA_DIR=$WORKSPACE/data
export OUTPUT_DIR=$WORKSPACE/output

Download the Workflow Repository

Create a working directory for the workflow and clone the Intelligent Indexing repository into your working directory.

mkdir -p $WORKSPACE && cd $WORKSPACE
git clone https://github.com/oneapi-src/intelligent-indexing.git $WORKSPACE

Set up conda

  1. Download the appropriate Miniconda Installer for linux.

    wget -q https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh
    
  2. In your terminal window, run.

    bash Miniconda3-latest-Linux-x86_64.sh
    
  3. Delete downloaded file.

    rm Miniconda3-latest-Linux-x86_64.sh
    

To learn more about conda installation, see the Conda Linux installation instructions.

Set Up Environment

Install and set the libmamba solver as default solver. Run the following commands:

conda install -n base conda-libmamba-solver -y
conda config --set solver libmamba

The $WORKSPACE/env/intel_env.yml file contains all dependencies to create the intel environment necessary for running the workflow.

Packages required in YAML fileVersion
python3.10
intelpython3_core2024.0.0
scikit-learn-intelex2024.0.1
modin-all0.24.1
nltk3.8.1
kaggle3.8.1

Execute next command to create the conda environment.

conda env create -f $WORKSPACE/env/intel_env.yml

Environment setup is required only once. This step does not cleanup the existing environment with the same name hence we need to make sure there is no conda environment with the same name. During this setup, intelligent_indexing_intel conda environment will be created with the dependencies listed in the YAML configuration.

Finally, activate intelligent_indexing_intel environment using the following command:

conda activate intelligent_indexing_intel

Download the Dataset

To setup the data for benchmarking, do the following:

  1. Configure your credentials and proxies.

  2. Download the data from https://www.kaggle.com/datasets/rmisra/news-category-dataset, save it to data directory and unzip it. This should produce a file called News_Category_Dataset_v3.json which we will need to split and save into the required files.

    cd $DATA_DIR
    kaggle datasets download -d rmisra/news-category-dataset
    unzip news-category-dataset.zip && rm news-category-dataset.zip
    

Please see this data set's applicable license for terms and conditions. Intel does not own the rights to this data set and does not confer any rights to it.

Supported Runtime Environment

You can execute the references pipelines using the following environments:

Run Using Bare Metal

Follow these instructions to set up and run this workflow on your own development system.

Set Up System Software

Our examples use the conda package and environment on your local computer. If you don't already have conda installed, go to Set up conda or see the Conda Linux installation instructions.

Run Workflow

To run the benchmarks with Intel® oneAPI technologies, the environment intelligent_indexing_intel should be activated using:

conda activate intelligent_indexing_intel
Setting up the data

The benchmarking scripts expects 2 files to be present in data/huffpost.

After downloading the data for benchmarking under these requirements, do the following:

All of the benchmarking can be run using the python script src/run_benchmarks.py.

The script reads and preprocesses the data, trains an SVC model, and predicts on unseen test data using the trained model, while also reporting on the execution time for these 3 steps.

Before running the script, we need to ensure that the appropriate conda environment is activated.

The run benchmark script takes the following arguments:

usage: run_benchmarks.py [-h] [-l LOGFILE] [-p] [-s SAVE_MODEL_DIR]

optional arguments:
  -h, --help            show this help message and exit
  -l LOGFILE, --logfile LOGFILE
                        log file to output benchmarking results to
  -p, --preprocessing_only
                        only perform preprocessing step
  -s SAVE_MODEL_DIR, --save_model_dir SAVE_MODEL_DIR
                        directory to save model to

To run with Intel® technologies, and log the performance to $OUTPUT_DIR/logs/intel.log, we would run (after creating the appropriate environment as above) from src directory:

cd $WORKSPACE/src
mkdir -p $OUTPUT_DIR/logs  # create logs dir in the OUTPUT_DIR dir if not present
python run_benchmarks.py -l $OUTPUT_DIR/logs/intel.log

Inspect the generated log to check Test Accuracy, Training Time, Inference Time and Total time data:

tail $OUTPUT_DIR/logs/intel.log

Clean Up Bare Metal

Follow these steps to restore your $WORKSPACE directory to an initial step. Please note that all downloaded dataset files, conda environment, and logs created by workflow will be deleted. Before executing next steps back up your important files.

conda deactivate
conda remove --name intelligent_indexing_intel --all -y
cd $DATA_DIR
rm -r huffpost News_Category_Dataset_v3.json
rm -r $OUTPUT_DIR/logs

Run Using Jupyter Notebook

You can directly access the Jupyter Notebook shared in this repo here.

  1. Follow the instructions described on Get Started to set required environment variables.

To launch Jupyter Notebook, execute the next commands:

  1. Execute Set Up Conda and Set Up environment steps.

  2. Activate Intel environment.

    conda activate intelligent_indexing_intel
    
  3. Install the IPython Kernel Package.

    conda install -c intel ipykernel -y
    
  4. Create a virtual environment and Install Jupyter Notebook.

    conda create -n jupyter_server -c intel nb_conda_kernels notebook -y
    
  5. Activate Jupyter Server environment.

    conda activate jupyter_server
    
  6. Change to working directory.

    cd $WORKSPACE
    
  7. Execute Jupyter command.

    jupyter notebook
    
    

Connect to Jupyter Notebook Server

Above command prints some information about the notebook server in your terminal, including the URL of the web application (by default, http://localhost:8888), for example:

To access the notebook, open this file in a browser: 
file:///path/to/jupyter/notebook/server/open.html
Or copy and paste one of these URLs: 
http://*********:8888/?token=***************************************** or 
http://127.0.0.1:8888/?token=*****************************************

Copy and paste one of the URLs into a web browser to open the Jupyter Notebook Dashboard.

Once in Jupyter, click on IntelligentIndexing.ipynb to get an interactive demo of the workflow.

Clean Up Jupyter Notebook

Clean Bare Metal Environment executing next commands:

conda activate base
conda remove --name intelligent_indexing_intel --all -y
conda remove --name jupyter_server --all -y
cd $DATA_DIR
rm -r huffpost News_Category_Dataset_v3.json
rm -r $OUTPUT_DIR/logs

Expected Output

Benchmark results are stored in the $OUTPUT_DIR/logs/intel.log file.

Check out the Test Accuracy, Training Time, Inference Time and Total time of the workflow. For example:

INFO:root:=======> Test Accuracy : 0.63
INFO:root:=======> Training Time : 229.324 secs
INFO:root:=======> Inference Time : 128.583 secs
INFO:root:=======> Total time : 374.083 secs

Summary and Next Steps

We break it down into the 3 primary tasks of this ML pipeline:

  1. Preprocessing data using Intel® Distribution of Modin* with the Ray* Backend.
  2. Training an NLP model for text classification using Intel® Extension for Scikit-learn*.
  3. Predicting from the trained model on new data using Intel® Extension for Scikit-learn*.

This exercise to categorize text data, can be used as a reference implementation across similar use cases with Intel AI optimizations enabled to accelerate the End-to-End (E2E) process.

Learn More

For more information about or to read about other relevant workflow examples, see these guides and software resources:

Support

If you have questions or issues about this use case, want help with troubleshooting, want to report a bug or submit enhancement requests, please submit a GitHub issue.

Appendix

References

<a id="joachims_1998">[1]</a> Joachims, Thorsten. "Text categorization with support vector machines: Learning with many relevant features." European conference on machine learning. Springer, Berlin, Heidelberg, 1998.

<a id="manning_2010">[2]</a> Manning, Christopher, Prabhakar Raghavan, and Hinrich Schütze. "Introduction to information retrieval. Chapter 15." Natural Language Engineering 16.1 (2010): 100-103.

*Other names and brands that may be claimed as the property of others. Trademarks.

Disclaimers

To the extent that any public or non-Intel datasets or models are referenced by or accessed using tools or code on this site those datasets or models are provided by the third party indicated as the content source. Intel does not create the content and does not warrant its accuracy or quality. By accessing the public content, or using materials trained on or with such content, you agree to the terms associated with that content and that your use complies with the applicable license.

Intel expressly disclaims the accuracy, adequacy, or completeness of any such public content, and is not liable for any errors, omissions, or defects in the content, or for any reliance on the content. Intel is not liable for any liability or damages relating to your use of public content.