Researchers from South Korea have used machine learning to develop an improved method for extracting actual content from web pages so that the ‘furniture’ of a web page – such as sidebars, footers and navigation headers, as well as advertisement blocks – disappears for the reader.
Though such functionality is either built into most popular web browsers, or else is easily available via extensions and plugins, these technologies rely on semantic formatting that may not be present in the web page, or which may have been deliberately compromised by the site owner in order to prevent the reader hiding the ‘full fat’ experience of the page.
Instead, the new method uses a grid-based system that iterates through the web page, evaluating how pertinent the content is to the core aim of the page.
Once a pertinent cell is identified, its relationship with nearby cells is also evaluated before being merged into the interpreted ‘core content’.
The central idea of the approach is to abandon code-based markup as an index of relevance (i.e. HTML tags that would normally denote the beginning of a paragraph, for instance, which can be replaced by alternate tags that will ‘fool’ screen readers and utilities such as Reader View), and deduce the content based solely on its visual appearance.
The approach, called Grid-Center-Expand (GCE), has been extended by the researchers into Deep Neural Network (DNN) models that exploit Google’s TabNet, an interpretative tabular learning architecture.
Get To the Point
The paper is titled Don’t read, just look: Main content extraction from web pages using visually apparent features, and comes from three researchers at Hanyang University, and one from the Institute of Convergence Technology, all located in Seoul.
Improved extraction of core web page content is potentially valuable not only for the casual end-user, but also for machine systems that are tasked with ingesting or indexing domain content for the purposes of Natural Language Processing (NLP), and other sectors in AI.
As it stands, if non-relevant content is included in such extraction processes, it may need to be manually filtered (or labeled), at great expense; worse, if the unwanted content is included with the core content, it could affect how the core content is interpreted, and the outcome of transformer and encoder/decoder systems that are relying on clean content.
An improved method, the researchers argue, is especially necessary because existing approaches often fail with non-English web pages.
Datasets and Training
The researchers compiled dataset material from English keywords in the GoogleTrends-2017 and GoogleTrends-2020 dataset, though they observe that, in terms of results, there were no practical differences between the two datasets.
Additionally, the authors gathered non-English keywords from South Korea, France, Japan, Russia, Indonesia and Saudi Arabia. Chinese keywords were added from a Baidu dataset, since Google Trends could not offer Chinese data.
Testing and Results
In testing the system, the authors found that it offer the same level of performance as recent DNN models, while providing better accommodation for a wider variety of languages.
For instance, the Boilernet architecture, while maintaining good performance in extracting pertinent content, adapts poorly to Chinese and Japanese datasets, while Web2Text, the authors find, has ‘relatively poor performance’ all round, with linguistic features that are not multilingual, and are unsuited for extracting central content from web pages.
Mozilla’s Readbility.js was found to achieve acceptable performance across multiple languages including English, even as a rule-based method. However the researchers found that its performance dropped notably on Japanese and French datasets, highlighting the limitations of trying to parse characteristics of a specific region entirely by rule-based approaches.
Meanwhile Google’s DOM Distiller, which blends heuristics and machine learning approaches, was found to perform well across the board.
The researchers conclude that ‘GCE does not need to keep up with the rapidly changing web environment because it relies on human nature—genuinely global and multilingual features’.
- Nanoantenna Enables Advanced Quantum Communication and Data Storage
- Human Brain Project Releases New Paper on Exascale Computing Power
- AI Researchers Estimate 97% Of EU Websites Fail GDPR Privacy Requirements- Especially User Profiling
- 5 Best Machine Learning & AI Books of All Time
- Earkick Raises $1M for Real-Time Mental Health Tracker