diff --git a/keywords.txt b/keywords.txt
index 825349a..c6a1b17 100644
--- a/keywords.txt
+++ b/keywords.txt
@@ -16,7 +16,6 @@ Tkinter
scala
einstein
feynman
-deep learning
multithreading
parallel computing
apache framework
diff --git a/runtime.txt b/runtime.txt
index fc81e75..cfa5aa5 100644
--- a/runtime.txt
+++ b/runtime.txt
@@ -1 +1 @@
-python-3.6.2
\ No newline at end of file
+python-3.6.2
diff --git a/useful_twitter.py b/useful_twitter.py
index 6d44e00..9cff43a 100644
--- a/useful_twitter.py
+++ b/useful_twitter.py
@@ -2,6 +2,8 @@
Script defining my Twitter bot, using sixohsix's Python wrapper for the
Twitter API.
"""
+# Employ machine learning - follow only those people who follow back, and unfollow only
+# those people who don't unfollow back!
# Instead of searching tweets and then doing actions on them, why not try
# streaming interesting tweets in realtime and then performing actions on them?
@@ -27,6 +29,8 @@
flags=re.IGNORECASE) #Copyright (c) 2013-2016 Molly White
#Above offensive compilation is not my stuff
+news_block_expr = re.compile(r'(?s).*?(.*?).*?')
+latest_expr = re.compile(r'(?s)
')
try:
oauth = OAuth(
@@ -131,17 +135,22 @@ def print_tweet(tweet):
hashtags.append(h["text"])
print(hashtags)
-def find_news():
+def find_news(): # I'm adventuring with regular expressions for parsing!
nyTech = requests.get('https://nytimes.com/section/technology')
- latest_patt = r'(?s)'
- latest = re.search(latest_patt, nyTech.text)
- news = re.findall(r'(?s)(.*?)', latest.group(1))
- news = [item.strip() for item in list(set(news))]
- for i in range(len(news)):
- item = news[i]
- if item.startswith('Daily Report: '):
- news[i] = item[14:]
- tv = requests.get('https://theverge.com', headers={'User-Agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_11_5) AppleWebKit/537.36 (KHTML, like Gecko) Cafari/537.36'})
+
+ latest = latest_expr.search(nyTech.text)
+ #news = re.findall(r'(?s)(.*?)', latest.group(1))
+ news_blocks = news_block_expr.findall(latest.group(1))
+ #news_links = re.findall(r'(?s)', latest.group(1))
+ #news = [item.strip() for item in list(set(news))]
+ news = []
+ for i in range(len(news_blocks)):
+ item = news_blocks[i][1].strip() + ' ' + news_blocks[i][0]
+ if item[1].startswith('Daily Report: '):
+ item = item[14:]
+ news.append(item)
+
+ '''tv = requests.get('https://theverge.com', headers={'User-Agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_11_5) AppleWebKit/537.36 (KHTML, like Gecko) Cafari/537.36'})
feed_patt = r'(?s)(.*?)