Saturday, 31 January 2015

I Don’t Need No Stinking API: Web Scraping For Fun and Profit

If you’ve ever needed to pull data from a third party website, chances are you started by checking to see if they had an official API. But did you know that there’s a source of structured data that virtually every website on the internet supports automatically, by default?
scraper toolThat’s right, we’re talking about pulling our data straight out of HTML — otherwise known as web scraping. Here’s why web scraping is awesome:

Any content that can be viewed on a webpage can be scraped. Period.

If a website provides a way for a visitor’s browser to download content and render that content in a structured way, then almost by definition, that content can be accessed programmatically. In this article, I’ll show you how.

Over the past few years, I’ve scraped dozens of websites — from music blogs and fashion retailers to the USPTO and undocumented JSON endpoints I found by inspecting network traffic in my browser.

There are some tricks that site owners will use to thwart this type of access — which we’ll dive into later — but they almost all have simple work-arounds.

Why You Should Scrape

But first we’ll start with some great reasons why you should consider web scraping first, before you start looking for APIs or RSS feeds or other, more traditional forms of structured data.

Websites are More Important Than APIs

The biggest one is that site owners generally care way more about maintaining their public-facing visitor website than they do about their structured data feeds.

We’ve seen it very publicly with Twitter clamping down on their developer ecosystem, and I’ve seen it multiple times in my projects where APIs change or feeds move without warning.

Sometimes it’s deliberate, but most of the time these sorts of problems happen because no one at the organization really cares or maintains the structured data. If it goes offline or gets horribly mangled, no one really notices.

Whereas if the website goes down or is having issues, that’s a more of an in-your-face, drop-everything-until-this-is-fixed kind of problem, and gets dealt with quickly.

No Rate-Limiting

Another thing to think about is that the concept of rate-limiting is virtually non-existent for public websites.

Aside from the occasional captchas on sign up pages, most businesses generally don’t build a lot of defenses against automated access. I’ve scraped a single site for over 4 hours at a time and not seen any issues.

Unless you’re making concurrent requests, you probably won’t be viewed as a DDOS attack, you’ll just show up as a super-avid visitor in the logs, in case anyone’s looking.

Anonymous Access

There are also fewer ways for the website’s administrators to track your behavior, which can be useful if you want gather data more privately.

With APIs, you often have to register to get a key and then send along that key with every request. But with simple HTTP requests, you’re basically anonymous besides your IP address and cookies, which can be easily spoofed.

The Data’s Already in Your Face

Web scraping is also universally available, as I mentioned earlier. You don’t have to wait for a site to open up an API or even contact anyone at the organization. Just spend some time browsing the site until you find the data you need and figure out some basic access patterns — which we’ll talk about next.

Let’s Get to Scraping

So you’ve decided you want to dive in and start grabbing data like a true hacker. Awesome.

Just like reading API docs, it takes a bit of work up front to figure out how the data is structured and how you can access it. Unlike APIs however, there’s really no documentation so you have to be a little clever about it.

I’ll share some of the tips I’ve learned along the way.

Fetching the Data

So the first thing you’re going to need to do is fetch the data. You’ll need to start by finding your “endpoints” — the URL or URLs that return the data you need.

If you know you need your information organized in a certain way — or only need a specific subset of it — you can browse through the site using their navigation. Pay attention to the URLs and how they change as you click between sections and drill down into sub-sections.

The other option for getting started is to go straight to the site’s search functionality. Try typing in a few different terms and again, pay attention to the URL and how it changes depending on what you search for. You’ll probably see a GET parameter like q= that always changes based on you search term.

Try removing other unnecessary GET parameters from the URL, until you’re left with only the ones you need to load your data. Make sure that there’s always a beginning ? to start the query string and a & between each key/value pair.

Dealing with Pagination

At this point, you should be starting to see the data you want access to, but there’s usually some sort of pagination issue keeping you from seeing all of it at once. Most regular APIs do this as well, to keep single requests from slamming the database.

Usually, clicking to page 2 adds some sort of offset= parameter to the URL, which is usually either the page number or else the number of items displayed on the page. Try changing this to some really high number and see what response you get when you “fall off the end” of the data.

With this information, you can now iterate over every page of results, incrementing the offset parameter as necessary, until you hit that “end of data” condition.

The other thing you can try doing is changing the “Display X Per Page” which most pagination UIs now have. Again, look for a new GET parameter to be appended to the URL which indicates how many items are on the page.

Try setting this to some arbitrarily large number to see if the server will return all the information you need in a single request. Sometimes there’ll be some limits enforced server-side that you can’t get around by tampering with this, but it’s still worth a shot since it can cut down on the number of pages you must paginate through to get all the data you need.

AJAX Isn’t That Bad!

Sometimes people see web pages with URL fragments # and AJAX content loading and think a site can’t be scraped. On the contrary! If a site is using AJAX to load the data, that probably makes it even easier to pull the information you need.

The AJAX response is probably coming back in some nicely-structured way (probably JSON!) in order to be rendered on the page with Javscript.

All you have to do is pull up the network tab in Web Inspector or Firebug and look through the XHR requests for the ones that seem to be pulling in your data.

Once you find it, you can leave the crufty HTML behind and focus instead on this endpoint, which is essentially an undocumented API.

(Un)structured Data?

Now that you’ve figured out how to get the data you need from the server, the somewhat tricky part is getting the data you need out of the page’s markup.

Use CSS Hooks

In my experience, this is usually straightforward since most web designers litter the markup with tons of classes and ids to provide hooks for their CSS.

You can piggyback on these to jump to the parts of the markup that contain the data you need.

Just right click on a section of information you need and pull up the Web Inspector or Firebug to look at it. Zoom up and down through the DOM tree until you find the outermost <div> around the item you want.

This <div> should be the outer wrapper around a single item you want access to. It probably has some class attribute which you can use to easily pull out all of the other wrapper elements on the page. You can then iterate over these just as you would iterate over the items returned by an API response.

A note here though: the DOM tree that is presented by the inspector isn’t always the same as the DOM tree represented by the HTML sent back by the website. It’s possible that the DOM you see in the inspector has been modified by Javascript — or sometime even the browser, if it’s in quirks mode.

Once you find the right node in the DOM tree, you should always view the source of the page (“right click” > “View Source”) to make sure the elements you need are actually showing up in the raw HTML.

This issue has caused me a number of head-scratchers.

Get a Good HTML Parsing Library

It is probably a horrible idea to try parsing the HTML of the page as a long string (although there are times I’ve needed to fall back on that). Spend some time doing research for a good HTML parsing library in your language of choice.

Most of the code I write is in Python, and I love BeautifulSoup for its error handling and super-simple API. I also love its motto:

    You didn’t write that awful page. You’re just trying to get some data out of it. Beautiful Soup is here to help. :)

You’re going to have a bad time if you try to use an XML parser since most websites out there don’t actually validate as properly formed XML (sorry XHTML!) and will give you a ton of errors.

A good library will read in the HTML that you pull in using some HTTP library (hat tip to the Requests library if you’re writing Python) and turn it into an object that you can traverse and iterate over to your heart’s content, similar to a JSON object.

Some Traps To Know About

I should mention that some websites explicitly prohibit the use of automated scraping, so it’s a good idea to read your target site’s Terms of Use to see if you’re going to make anyone upset by scraping.

For two-thirds of the website I’ve scraped, the above steps are all you need. Just fire off a request to your “endpoint” and parse the returned data.

But sometimes, you’ll find that the response you get when scraping isn’t what you saw when you visited the site yourself.

When In Doubt, Spoof Headers

Some websites require that your User Agent string is set to something they allow, or you need to set certain cookies or other headers in order to get a proper response.

Depending on the HTTP library you’re using to make requests, this is usually pretty straightforward. I just browse the site in my web browser and then grab all of the headers that my browser is automatically sending. Then I put those in a dictionary and send them along with my request.

Note that this might mean grabbing some login or other session cookie, which might identify you and make your scraping less anonymous. It’s up to you how serious of a risk that is.

Content Behind A Login

Sometimes you might need to create an account and login to access the information you need. If you have a good HTTP library that handles logins and automatically sending session cookies (did I mention how awesome Requests is?), then you just need your scraper login before it gets to work.

Note that this obviously makes you totally non-anonymous to the third party website so all of your scraping behavior is probably pretty easy to trace back to you if anyone on their side cared to look.

Rate Limiting

I’ve never actually run into this issue myself, although I did have to plan for it one time. I was using a web service that had a strict rate limit that I knew I’d exceed fairly quickly.

Since the third party service conducted rate-limiting based on IP address (stated in their docs), my solution was to put the code that hit their service into some client-side Javascript, and then send the results back to my server from each of the clients.

This way, the requests would appear to come from thousands of different places, since each client would presumably have their own unique IP address, and none of them would individually be going over the rate limit.

Depending on your application, this could work for you.

Poorly Formed Markup

Sadly, this is the one condition that there really is no cure for. If the markup doesn’t come close to validating, then the site is not only keeping you out, but also serving a degraded browsing experience to all of their visitors.

It’s worth digging into your HTML parsing library to see if there’s any setting for error tolerance. Sometimes this can help.

If not, you can always try falling back on treating the entire HTML document as a long string and do all of your parsing as string splitting or — God forbid — a giant regex.



Well there’s 2000 words to get you started on web scraping. Hopefully I’ve convinced you that it’s actually a legitimate way of collecting data.

It’s a real hacker challenge to read through some HTML soup and look for patterns and structure in the markup in order to pull out the data you need. It usually doesn’t take much longer than reading some API docs and getting up to speed with a client. Plus it’s way more fun!

Source: https://blog.hartleybrody.com/web-scraping/

Monday, 26 January 2015

Living Your Dream Life When Others Are Scraping Their Windshields

One Internet marketer once remarked that he knew he was successful when he was awakened one early winter morning by the sound of his neighbors scraping the ice off of their windshields so that they could drive to work. As he burrowed deeper under his covers, he grinned to himself and realized that while he wasn't sipping cocktails aboard his own yacht yet, he had reached an enviable measure of success. If you're ready to begin living your dream life, then you're ready to become an Internet marketer.

What does living your dream life mean to you? Does it mean being able to be your own boss and not having to adhere to another's schedule? Does living your dream life mean taking vacations to exotic locations? The beauty of Internet marketing is that you can realize your goals and live the life you choose. You can do all the things you've dreamed of doing and go to the places you've dreamed of seeing. You answer to no one but yourself, although most Internet marketers learn that they still have a boss even after they quit their jobs. That new boss, actually bosses, are your customers and you'll have to do a great job in order to get them and keep them.

Most Internet marketers, the ones who truly have an entrepreneurial spirit, are naturals when it comes to pleasing customers. They want to earn their business, they care about their needs and their likes and dislikes and will go out on a limb to satisfy their customers. While customer service is severely lacking in many of today's businesses, without it, you'll be dead in the water before you even get started. If living your dream life is important to you, plan now to offer the kind of customer service that no one else can rival.

Imagine living your dream life for the rest of your life. Imagine being able to take your family to the kinds of places you've only seen in movies and read about in books. Imagine having money in the bank to tide you over when times are tough and having the peace of mind of knowing that your home is not at risk. You can have all of this and more if you are willing to work hard. The rewards for Internet marketers aren't instant, but they are sweet and they can be yours.

Source: http://ezinearticles.com/?Living-Your-Dream-Life-When-Others-Are-Scraping-Their-Windshields&id=2983939

Wednesday, 21 January 2015

What is Blog Content Scraping?

Blog content scraping is an act usually performed with scripts that extract content from numerous sources and pulls it into one site. It is so easy now that anyone can install a WordPress site, put a free or commercial theme, and install a few plugins that will go and scrape content from selected blogs, so it can be published on their site.

Why are they Stealing my Content?

Some of our users have asked us why are they stealing my content? The simple answer is because you are AWESOME. The truth is that these content scrapers have ulterior motives. Below are just few reasons why someone would scrape your content:

•    Affiliate commission – There are some dirty affiliate marketers out there that just wants to exploit the system to make few extra bucks. They will use your content and other’s content to bring traffic to their site through search engine. These sites are usually targeted towards a specific niche, so they have related products that they are promoting.

•    Lead Generation – Often we see lawyers and realtors doing this. They want to seem like industry leaders in their small communities. They do not have the bandwidth to produce quality content, so they go out and scrape content from other sources. Sometimes, they are not even aware of this because they are paying some scumbag $30/month to add content and help them get better SEO. We have encountered quite a few of these in the past.

•    Advertising Revenue – Some folks just want to create a “hub” of knowledge. A one-stop-shop for users in a specific niche. If I had a penny for every time someone has done this with our content, then we would have a few hundred pennies. Often we notice that our site content is being scraped. The scraper always replies, I was doing this for the good of the community. Except the site is plastered with ads.

These are just a few reasons why someone would steal your content.

Source:http://www.wpbeginner.com/beginners-guide/beginners-guide-to-preventing-blog-content-scraping-in-wordpress/

Tuesday, 6 January 2015

Data Mining's Importance in Today's Corporate Industry

A large amount of information is collected normally in business, government departments and research & development organizations. They are typically stored in large information warehouses or bases. For data mining tasks suitable data has to be extracted, linked, cleaned and integrated with external sources. In other words, it is the retrieval of useful information from large masses of information, which is also presented in an analyzed form for specific decision-making.

Data mining is the automated analysis of large information sets to find patterns and trends that might otherwise go undiscovered. It is largely used in several applications such as understanding consumer research marketing, product analysis, demand and supply analysis, telecommunications and so on. Data Mining is based on mathematical algorithm and analytical skills to drive the desired results from the huge database collection.

It can be technically defined as the automated mining of hidden information from large databases for predictive analysis. Web mining requires the use of mathematical algorithms and statistical techniques integrated with software tools.

Data mining includes a number of different technical approaches, such as:

    Clustering
    Data Summarization
    Learning Classification Rules
    Finding Dependency Networks
    Analyzing Changes
    Detecting Anomalies

The software enables users to analyze large databases to provide solutions to business decision problems. Data mining is a technology and not a business solution like statistics. Thus the data mining software provides an idea about the customers that would be intrigued by the new product.

It is available in various forms like text, web, audio & video data mining, pictorial data mining, relational databases, and social networks. Data mining is thus also known as Knowledge Discovery in Databases since it involves searching for implicit information in large databases. The main kinds of data mining software are: clustering and segmentation software, statistical analysis software, text analysis, mining and information retrieval software and visualization software.

Data Mining therefore has arrived on the scene at the very appropriate time, helping these enterprises to achieve a number of complex tasks that would have taken up ages but for the advent of this marvelous new technology.

Our web research provides detailed information on data mining, business intelligence data mining, web data mining, online data research, web research services. We will closely work with you; we guarantee clear, focused and relevant information that meets your specifications.

Source:http://ezinearticles.com/?Data-Minings-Importance-in-Todays-Corporate-Industry&id=2057401

Thursday, 1 January 2015

Choosing Hand-Scraped Flooring for Your Home

Are you looking for a unique appearance for your floor? If you have considered adding hardwood to your home, opt for hand-scraped flooring. With a distressed appearance, hand-scraped flooring is offered by many manufacturers, from those as all-encompassing as Columbia and Mullican to specialty brands like BR 111 and Homerwood. Regardless of the brand, however, exotic and domestic species, solid and engineered flooring, and unfinished and prefinished varieties are all available hand scraped. In fact, hand scraped lock flooring is even available.

As a trend for the moment, hand-scraped flooring gives a space a rustic and aged appearance. Nevertheless, before you decide on hand-scraped hardwood, determine if you will want this look long-term in your home. For those that do, the floor may need to be refinished. Instead of sanding and adding a stain, the wood should be scraped with a Floor Abrader in order to remove the finish and keep the contours. After this, a new finish can be applied.

Sanding wears away at the unique appearance of hand-scraped flooring, but some may want to do away with it entirely years later, especially if the look goes out of style. But, be careful upfront in this regard. If you see yourself wanting to get rid of hand-scraped flooring many years down the line, add solid hardwood that can be sanded down or opt for an engineered variety that can easily be replaced. Engineered hand-scraped flooring is often distressed down to the veneers below the wear layer, and sanding it away will reveal the lower plys.

At the same time, realize that not all hand-scraped flooring is the same, and several methods are used to create aged or rustic looks. These include:

• Aging. Time worn aged and antique hand-scraped flooring are aged over time, and the resulting appearance is further accented with dark-colored staining, highlighting, and contouring.

• Physically altering the wood. Hardwood can be wire brushed to accent the grain, hand sculpted, or, for the roughest appearance available, hand hewn and rough sawn.

• French bleed. Beveled edges are deeper and highlighted with darker stain.

• Custom unfinished. Once unfinished hardwood is installed, a professional custom distresses it through bleaching, pickeling, beating with chains, or fastening with antique nails. A finish is applied once the wood is distressed.

Source:http://www.articlesbase.com/home-improvement-articles/choosing-hand-scraped-flooring-for-your-home-5383626.html