Skip to content

Latest commit

 

History

History
32 lines (21 loc) · 1.25 KB

README.md

File metadata and controls

32 lines (21 loc) · 1.25 KB

Set of scripts to download content from NYTimes and Foxnews for vocabulary analysis.

First you will need to get the site content, I used wget to make this happen

Here is an explanations of the wget options I used:

  • -l 2 says follow 2 levels deep
  • -I 2012 is specific to NYTimes - it says to only include stuff in the 2012 directory which are articles otherwise we get a bunch of random website stuff
  • -w 2 says to wait for 2 seconds between each retrieval

To run:

wget -r -l 3 -I 2012 -w 2 http://www.nytimes.com/most-popular
wget -r -l 3 -w 2 http://www.foxnews.com

You will now have a www.foxnews.com and www.nytimes.com directories.

execute iterate_nytimes.sh and iterate_fox.sh to process all the HTML content and turn it into straight text, which is much easier to process. we use 'lynx' to do this, a text-based web browser, so you may need to yum install lynx

you need to direct output

$ ./iterate_fox.sh > FOX_OUT

word_check.lua (requires lua) will read the words in and count individual words. it discounts 'words' that have non-alpha characters in it. the output will be in csv format. the word followed by a comma followed by the number of occurrences. you can import this into a spreadsheet.

$ ./word_check.lua < FOX_OUT > FOX_OUT.csv