Search

Categories

    • categories-img Jacket, Women
    • categories-img Woolend Jacket
    • categories-img Western denim
    • categories-img Mini Dresss
    • categories-img Jacket, Women
    • categories-img Woolend Jacket
    • categories-img Western denim
    • categories-img Mini Dresss
    • categories-img Jacket, Women
    • categories-img Woolend Jacket
    • categories-img Western denim
    • categories-img Mini Dresss
    • categories-img Jacket, Women
    • categories-img Woolend Jacket
    • categories-img Western denim
    • categories-img Mini Dresss
    • categories-img Jacket, Women
    • categories-img Woolend Jacket
    • categories-img Western denim
    • categories-img Mini Dresss

Filter By Price

$
-
$

Dietary Needs

Top Rated Product

product-img product-img

Modern Chair

$165.00
product-img product-img

Plastic Chair

$165.00
product-img product-img

Design Rooms

$165.00

Brands

  • Wooden
  • Chair
  • Modern
  • Fabric
  • Shoulder
  • Winter
  • Accessories
  • Dress

Welcome and thank you for visiting us. For any query call us on 0799 626 359 or Email [email protected]

Offcanvas Menu Open

Shopping Cart

Africa largest book store

Sub Total:

Search for any Title

Getting Structured Data from the Internet : Running Web Crawlers/Scrapers on a Big Data Production Scale - 1st ed.

By: (Author) Jay M. Patel

Extended Catalogue

Ksh 9,900.00

Format: Paperback / Softback

ISBN-10: 1484265750

ISBN-13: 9781484265758

Edition: 1st ed.

Publisher: APress

Imprint: APress

Country of Manufacture: GB

Country of Publication: GB

Publication Date: Nov 13th, 2020

Print length: 397 Pages

Weight: 784 grams

Dimensions (height x width x thickness): 17.70 x 25.30 x 2.80 cms

Choose your Location

Shipping & Delivery

Door Delivery

Delivery fee

Delivery in 10 to 14 days

  • Description

  • Reviews

Utilize web scraping at scale to quickly get unlimited amounts of free data available on the web into a structured format. This book teaches you to use Python scripts to crawl through websites at scale and scrape data from HTML and JavaScript-enabled pages and convert it into structured data formats such as CSV, Excel, JSON, or load it into a SQL database of your choice. This book goes beyond the basics of web scraping and covers advanced topics such as natural language processing (NLP) and text analytics to extract names of people, places, email addresses, contact details, etc., from a page at production scale using distributed big data techniques on an Amazon Web Services (AWS)-based cloud infrastructure. It book covers developing a robust data processing and ingestion pipeline on the Common Crawl corpus, containing petabytes of data publicly available and a web crawl data set available on AWS's registry of open data. Getting Structured Data from the Internet also includes a step-by-step tutorial on deploying your own crawlers using a production web scraping framework (such as Scrapy) and dealing with real-world issues (such as breaking Captcha, proxy IP rotation, and more). Code used in the book is provided to help you understand the concepts in practice and write your own web crawler to power your business ideas. What You Will LearnUnderstand web scraping, its applications/uses, and how to avoid web scraping by hitting publicly available rest API endpoints to directly get dataDevelop a web scraper and crawler from scratch using lxml and BeautifulSoup library, and learn about scraping from JavaScript-enabled pages using SeleniumUse AWS-based cloud computing with EC2, S3, Athena, SQS, and SNS to analyze, extract, and store useful insights from crawled pagesUse SQL language on PostgreSQL running on Amazon Relational Database Service (RDS) and SQLite using SQLalchemyReview sci-kit learn, Gensim, and spaCy to perform NLP tasks on scraped web pages such as name entity recognition, topic clustering (Kmeans, Agglomerative Clustering), topic modeling (LDA, NMF, LSI), topic classification (naive Bayes, Gradient Boosting Classifier) and text similarity (cosine distance-based nearest neighbors)Handle web archival file formats and explore Common Crawl open data on AWSIllustrate practical applications for web crawl data by building a similar website tool and a technology profiler similar to builtwith.comWrite scripts to create a backlinks database on a web scale similar to Ahrefs.com, Moz.com, Majestic.com, etc., for search engine optimization (SEO), competitor research, and determining website domain authority and rankingUse web crawl data to build a news sentiment analysis system or alternative financial analysis covering stock market trading signalsWrite a production-ready crawlerin Python using Scrapy framework and deal with practical workarounds for Captchas, IP rotation, and moreWho This Book Is ForPrimary audience: data analysts and scientists with little to no exposure to real-world data processing challenges, secondary: experienced software developers doing web-heavy data processing who need a primer, tertiary: business owners and startup founders who need to know more about implementation to better direct their technical team

Utilize web scraping at scale to quickly get unlimited amounts of free data available on the web into a structured format. This book teaches you to use Python scripts to crawl through websites at scale and scrape data from HTML and JavaScript-enabled pages and convert it into structured data formats such as CSV, Excel, JSON, or load it into a SQL database of your choice.

This book goes beyond the basics of web scraping and covers advanced topics such as natural language processing (NLP) and text analytics to extract names of people, places, email addresses, contact details, etc., from a page at production scale using distributed big data techniques on an Amazon Web Services (AWS)-based cloud infrastructure. It book covers developing a robust data processing and ingestion pipeline on the Common Crawl corpus, containing petabytes of data publicly available and a web crawl data set available on AWS''s registry of open data.

Getting Structured Data from the Internet also includes a step-by-step tutorial on deploying your own crawlers using a production web scraping framework (such as Scrapy) and dealing with real-world issues (such as breaking Captcha, proxy IP rotation, and more). Code used in the book is provided to help you understand the concepts in practice and write your own web crawler to power your business ideas.


What You Will Learn

  • Understand web scraping, its applications/uses, and how to avoid web scraping by hitting publicly available rest API endpoints to directly get data
  • Develop a web scraper and crawler from scratch using lxml and BeautifulSoup library, and learn about scraping from JavaScript-enabled pages using Selenium
  • Use AWS-based cloud computing with EC2, S3, Athena, SQS, and SNS to analyze, extract, and store useful insights from crawled pages
  • Use SQL language on PostgreSQL running on Amazon Relational Database Service (RDS) and SQLite using SQLalchemy
  • Review sci-kit learn, Gensim, and spaCy to perform NLP tasks on scraped web pages such as name entity recognition, topic clustering (Kmeans, Agglomerative Clustering), topic modeling (LDA, NMF, LSI), topic classification (naive Bayes, Gradient Boosting Classifier) and text similarity (cosine distance-based nearest neighbors)
  • Handle web archival file formats and explore Common Crawl open data on AWS
  • Illustrate practical applications for web crawl data by building a similar website tool and a technology profiler similar to builtwith.com
  • Write scripts to create a backlinks database on a web scale similar to Ahrefs.com, Moz.com, Majestic.com, etc., for search engine optimization (SEO), competitor research, and determining website domain authority and ranking
  • Use web crawl data to build a news sentiment analysis system or alternative financial analysis covering stock market trading signals
  • Write a production-ready crawler in Python using Scrapy framework and deal with practical workarounds for Captchas, IP rotation, and more


Who This Book Is For

Primary audience: data analysts and scientists with little to no exposure to real-world data processing challenges, secondary: experienced software developers doing web-heavy data processing who need a primer, tertiary: business owners and startup founders who need to know more about implementation to better direct their technical team


Get Getting Structured Data from the Internet by at the best price and quality guranteed only at Werezi Africa largest book ecommerce store. The book was published by APress and it has pages. Enjoy Shopping Best Offers & Deals on books Online from Werezi - Receive at your doorstep - Fast Delivery - Secure mode of Payment

Customer Reviews

Based on 0 reviews

Mind, Body, & Spirit