diff --git a/README.md b/README.md index 94583eb..700056f 100644 --- a/README.md +++ b/README.md @@ -11,13 +11,13 @@ I hope some of you find this useful. Make sure all these python modules are intalled: + BeautifulSoup -+ urllib2 -+ urlparse ++ urllib3+ ++ urlparse3 example: -```bash -sudo pip install BeautifulSoup +```bash +sudo pip install -r requirements.txt ``` ### Usage @@ -30,7 +30,9 @@ To get all documents: Files that exist on disk will not be re-downloaded (so by default only new sections/files are downloaded). To override this default and force re-download of files that exist on disk, use +```bash ./getAWSdocs.py --force +``` Thats it! diff --git a/getAWSdocs.py b/getAWSdocs.py index 7d1dd0c..0a601b0 100755 --- a/getAWSdocs.py +++ b/getAWSdocs.py @@ -9,7 +9,7 @@ # Build a list of the amazon service sections def get_services(): - html_page = urllib2.urlopen("http://aws.amazon.com/documentation/") + html_page = urllib.urlopen("http://aws.amazon.com/documentation/") # Parse the HTML page soup = BeautifulSoup(html_page) urls = [] diff --git a/requirements.txt b/requirements.txt index ea7a949..cbf3faf 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1 +1,3 @@ -BeautifulSoup==3.2.1 +BeautifulSoup +urllib3 +urlparse3