Skip to main content

Parse HTML Components

This page explains how to parse and extract information from a page (local or remote).

Parsing HTML and extract the relevant information is something we can use in many contexts: scan a page for a price change, extract a component, detect the broken links .. etc.

AppSeed, in particular, uses HTML parsing for two things:

  • Page structure detection
  • Component extraction

For newcomers, AppSeed uses automation tools to convert lifeless UI Kits into simple starters generated in many frameworks and patterns. For instance, this open-source design - Pixel Lite provided by Themesberg has been translated to Flask and Django using HTML parsing as the first phase of the translation process.

Required libraries and tools

The process

The flow explained in this article will execute a few simple steps:

  • Load the HTML content - this can be done from a local file or using a LIVE website
  • Analyze the page and extract XPATH expression for a component
  • Use Lxml library to extract the HTML
  • Format the component and save it on disk

Install libraries via PIP

$ pip install requests
$ pip install lxml
$ pip install beautifulsoup4

From this point, all the code is typed using a python console

$ python [ENTER]

Load the content from local file

>>> f = open('./app/templates/index.html','r')
>>> html_page =

Load content from remote HTML file (the LIVE sample)

>>> import requests
>>> page = requests.get('')
>>> html_page = page.content

At this point html_page variable contains the entire HTML content (string type) and we can use it in BS4 or Lxml to extract the components. To visualize the page structure we can use browser tools:

HTML Parser - Target Component Inspection.

The target component will be extracted using an XPATH expression provided by the browser:


To extract the component, this XPATH expression will be used in Lxml library to isolate the code.

>>> from lxml import html
>>> html_dom = html.fromstring( html_page )
>>> component = html_dom.xpath( '//*[@id="features"]' )

To extract the raw HTML from the component object we need to use tostring helper exposed by Lxml library:

>>> from lxml.etree import tostring
>>> component_html = tostring( component[0] )

The next step is to call Beautiful soup and prettify the HML for saving on disk

>>> from bs4 import BeautifulSoup as bs
>>> soup = bs( component_html )
>>> soup.prettify()

The component is fully extracted and parsable:

  <section class="section section-lg pb-0" id="features">
<div class="container">
<div class="row">


<div class="col-12 col-md-4">
<div class="icon-box text-center mb-5 mb-md-0">
<div class="icon icon-shape icon-lg bg-white shadow-lg border-light rounded-circle icon-secondary mb-3">
<span class="fas fa-box-open">
<h2 class="my-3 h5">
80 components
<p class="px-lg-4">
Beatifully crafted and creative components made with great care for each pixel



The rendered version:

HTML Parser - Extracted Component.