2019-06-20 16:16:20 +02:00
|
|
|
import httpclient, asyncdispatch, htmlparser, times
|
2019-07-03 11:46:03 +02:00
|
|
|
import sequtils, strutils, json, xmltree, uri
|
2019-06-20 16:16:20 +02:00
|
|
|
|
2019-07-03 11:46:03 +02:00
|
|
|
import types, parser, parserutils, formatters, search
|
2019-06-20 16:16:20 +02:00
|
|
|
|
2019-06-21 03:51:14 +02:00
|
|
|
const
|
|
|
|
agent = "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/71.0.3578.98 Safari/537.36"
|
2019-06-29 14:11:23 +02:00
|
|
|
lang = "en-US,en;q=0.9"
|
2019-06-24 05:14:14 +02:00
|
|
|
auth = "Bearer AAAAAAAAAAAAAAAAAAAAAPYXBAAAAAAACLXUNDekMxqa8h%2F40K4moUkGsoc%3DTYfbDKbT3jJPCEVnMYqilB28NHfOPqkca3qaAxGfsyKCs0wRbw"
|
2019-06-29 14:11:23 +02:00
|
|
|
cardAccept = "text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3"
|
2019-07-03 11:46:03 +02:00
|
|
|
jsonAccept = "application/json, text/javascript, */*; q=0.01"
|
2019-06-24 05:14:14 +02:00
|
|
|
|
|
|
|
base = parseUri("https://twitter.com/")
|
|
|
|
apiBase = parseUri("https://api.twitter.com/1.1/")
|
|
|
|
|
2019-07-03 11:46:03 +02:00
|
|
|
timelineUrl = "i/profiles/show/$1/timeline/tweets"
|
|
|
|
timelineSearchUrl = "i/search/timeline"
|
2019-07-04 04:18:32 +02:00
|
|
|
timelineMediaUrl = "i/profiles/show/$1/media_timeline"
|
2019-06-21 03:51:14 +02:00
|
|
|
profilePopupUrl = "i/profiles/popup"
|
|
|
|
profileIntentUrl = "intent/user"
|
2019-06-27 20:13:46 +02:00
|
|
|
tweetUrl = "status"
|
2019-06-24 05:14:14 +02:00
|
|
|
videoUrl = "videos/tweet/config/$1.json"
|
|
|
|
tokenUrl = "guest/activate.json"
|
2019-06-29 14:11:23 +02:00
|
|
|
cardUrl = "i/cards/tfw/v1/$1"
|
|
|
|
pollUrl = cardUrl & "?cardname=poll2choice_text_only&lang=en"
|
2019-06-21 03:51:14 +02:00
|
|
|
|
2019-06-24 05:29:47 +02:00
|
|
|
var
|
2019-06-25 15:09:13 +02:00
|
|
|
guestToken = ""
|
|
|
|
tokenUses = 0
|
|
|
|
tokenMaxUses = 230
|
2019-06-24 05:29:47 +02:00
|
|
|
tokenUpdated: Time
|
2019-06-25 15:09:13 +02:00
|
|
|
tokenLifetime = initDuration(minutes=20)
|
2019-06-24 05:29:47 +02:00
|
|
|
|
|
|
|
template newClient() {.dirty.} =
|
2019-06-21 03:51:14 +02:00
|
|
|
var client = newAsyncHttpClient()
|
|
|
|
defer: client.close()
|
|
|
|
client.headers = headers
|
2019-06-20 16:16:20 +02:00
|
|
|
|
2019-06-24 05:29:47 +02:00
|
|
|
proc fetchHtml(url: Uri; headers: HttpHeaders; jsonKey = ""): Future[XmlNode] {.async.} =
|
|
|
|
newClient()
|
|
|
|
|
2019-06-21 02:15:46 +02:00
|
|
|
var resp = ""
|
|
|
|
try:
|
|
|
|
resp = await client.getContent($url)
|
|
|
|
except:
|
|
|
|
return nil
|
|
|
|
|
|
|
|
if jsonKey.len > 0:
|
|
|
|
let json = parseJson(resp)[jsonKey].str
|
|
|
|
return parseHtml(json)
|
|
|
|
else:
|
|
|
|
return parseHtml(resp)
|
|
|
|
|
2019-06-24 05:14:14 +02:00
|
|
|
proc fetchJson(url: Uri; headers: HttpHeaders): Future[JsonNode] {.async.} =
|
2019-06-24 05:29:47 +02:00
|
|
|
newClient()
|
2019-06-24 05:14:14 +02:00
|
|
|
|
|
|
|
var resp = ""
|
|
|
|
try:
|
|
|
|
resp = await client.getContent($url)
|
2019-06-25 13:18:44 +02:00
|
|
|
result = parseJson(resp)
|
2019-06-24 05:14:14 +02:00
|
|
|
except:
|
|
|
|
return nil
|
|
|
|
|
2019-06-25 15:09:13 +02:00
|
|
|
proc getGuestToken(force=false): Future[string] {.async.} =
|
|
|
|
if getTime() - tokenUpdated < tokenLifetime and
|
|
|
|
not force and tokenUses < tokenMaxUses:
|
|
|
|
return guestToken
|
2019-06-21 02:15:46 +02:00
|
|
|
|
2019-06-24 05:29:47 +02:00
|
|
|
tokenUpdated = getTime()
|
2019-06-25 15:09:13 +02:00
|
|
|
tokenUses = 0
|
2019-06-20 16:16:20 +02:00
|
|
|
|
2019-06-24 05:14:14 +02:00
|
|
|
let headers = newHttpHeaders({
|
2019-07-03 11:46:03 +02:00
|
|
|
"Accept": jsonAccept,
|
2019-06-24 05:14:14 +02:00
|
|
|
"Referer": $base,
|
|
|
|
"User-Agent": agent,
|
|
|
|
"Authorization": auth
|
|
|
|
})
|
|
|
|
|
2019-06-24 05:29:47 +02:00
|
|
|
newClient()
|
2019-06-24 05:14:14 +02:00
|
|
|
|
|
|
|
let
|
2019-06-29 14:11:23 +02:00
|
|
|
url = apiBase / tokenUrl
|
2019-06-24 05:14:14 +02:00
|
|
|
json = parseJson(await client.postContent($url))
|
|
|
|
|
|
|
|
result = json["guest_token"].to(string)
|
2019-06-25 15:09:13 +02:00
|
|
|
guestToken = result
|
2019-06-24 05:14:14 +02:00
|
|
|
|
|
|
|
proc getVideo*(tweet: Tweet; token: string) {.async.} =
|
2019-06-25 02:38:18 +02:00
|
|
|
if tweet.video.isNone(): return
|
2019-06-24 05:29:47 +02:00
|
|
|
|
2019-06-24 05:14:14 +02:00
|
|
|
let headers = newHttpHeaders({
|
2019-07-03 11:46:03 +02:00
|
|
|
"Accept": jsonAccept,
|
2019-07-01 23:14:36 +02:00
|
|
|
"Referer": $(base / getLink(tweet)),
|
2019-06-24 05:14:14 +02:00
|
|
|
"User-Agent": agent,
|
|
|
|
"Authorization": auth,
|
|
|
|
"x-guest-token": token
|
|
|
|
})
|
|
|
|
|
2019-07-03 07:15:52 +02:00
|
|
|
let url = apiBase / (videoUrl % tweet.id)
|
|
|
|
let json = await fetchJson(url, headers)
|
2019-06-24 05:14:14 +02:00
|
|
|
|
2019-06-27 21:07:29 +02:00
|
|
|
if json == nil:
|
2019-06-25 15:09:13 +02:00
|
|
|
if getTime() - tokenUpdated > initDuration(seconds=1):
|
|
|
|
tokenUpdated = getTime()
|
2019-07-03 07:15:52 +02:00
|
|
|
discard await getGuestToken(force=true)
|
2019-06-25 15:09:13 +02:00
|
|
|
await getVideo(tweet, guestToken)
|
|
|
|
return
|
|
|
|
|
2019-06-24 05:14:14 +02:00
|
|
|
tweet.video = some(parseVideo(json))
|
2019-06-25 15:09:13 +02:00
|
|
|
tokenUses.inc
|
2019-06-24 05:14:14 +02:00
|
|
|
|
2019-07-01 03:13:12 +02:00
|
|
|
proc getVideos*(thread: Thread; token="") {.async.} =
|
2019-07-03 07:18:19 +02:00
|
|
|
if thread == nil: return
|
2019-06-24 05:14:14 +02:00
|
|
|
|
2019-07-03 07:15:52 +02:00
|
|
|
var gToken = token
|
2019-06-25 15:09:13 +02:00
|
|
|
if gToken.len == 0:
|
|
|
|
gToken = await getGuestToken()
|
|
|
|
|
2019-07-03 07:15:52 +02:00
|
|
|
var videoFuts: seq[Future[void]]
|
2019-07-01 03:13:12 +02:00
|
|
|
for tweet in thread.tweets.filterIt(it.video.isSome):
|
2019-07-03 07:15:52 +02:00
|
|
|
videoFuts.add getVideo(tweet, gToken)
|
2019-06-24 05:14:14 +02:00
|
|
|
|
|
|
|
await all(videoFuts)
|
|
|
|
|
|
|
|
proc getConversationVideos*(convo: Conversation) {.async.} =
|
|
|
|
var token = await getGuestToken()
|
|
|
|
var futs: seq[Future[void]]
|
|
|
|
|
|
|
|
futs.add getVideo(convo.tweet, token)
|
2019-06-25 15:09:13 +02:00
|
|
|
futs.add convo.replies.mapIt(getVideos(it, token))
|
2019-07-03 07:15:52 +02:00
|
|
|
futs.add getVideos(convo.before, token)
|
|
|
|
futs.add getVideos(convo.after, token)
|
2019-06-24 05:14:14 +02:00
|
|
|
|
|
|
|
await all(futs)
|
|
|
|
|
2019-06-29 14:11:23 +02:00
|
|
|
proc getPoll*(tweet: Tweet) {.async.} =
|
|
|
|
if tweet.poll.isNone(): return
|
|
|
|
|
|
|
|
let headers = newHttpHeaders({
|
|
|
|
"Accept": cardAccept,
|
2019-07-01 23:14:36 +02:00
|
|
|
"Referer": $(base / getLink(tweet)),
|
2019-06-29 14:11:23 +02:00
|
|
|
"User-Agent": agent,
|
|
|
|
"Authority": "twitter.com",
|
|
|
|
"Accept-Language": lang,
|
|
|
|
})
|
|
|
|
|
|
|
|
let url = base / (pollUrl % tweet.id)
|
|
|
|
let html = await fetchHtml(url, headers)
|
|
|
|
if html == nil: return
|
2019-06-24 05:29:47 +02:00
|
|
|
|
2019-06-29 14:11:23 +02:00
|
|
|
tweet.poll = some(parsePoll(html))
|
|
|
|
|
2019-07-01 03:13:12 +02:00
|
|
|
proc getPolls*(thread: Thread) {.async.} =
|
2019-07-03 07:18:19 +02:00
|
|
|
if thread == nil: return
|
2019-07-01 03:13:12 +02:00
|
|
|
var polls = thread.tweets.filterIt(it.poll.isSome)
|
2019-06-29 14:11:23 +02:00
|
|
|
await all(polls.map(getPoll))
|
|
|
|
|
|
|
|
proc getConversationPolls*(convo: Conversation) {.async.} =
|
|
|
|
var futs: seq[Future[void]]
|
|
|
|
futs.add getPoll(convo.tweet)
|
|
|
|
futs.add getPolls(convo.before)
|
|
|
|
futs.add getPolls(convo.after)
|
|
|
|
futs.add convo.replies.map(getPolls)
|
|
|
|
await all(futs)
|
|
|
|
|
2019-07-11 19:22:23 +02:00
|
|
|
proc getCard*(tweet: Tweet) {.async.} =
|
|
|
|
if tweet.card.isNone(): return
|
|
|
|
|
|
|
|
let headers = newHttpHeaders({
|
|
|
|
"Accept": cardAccept,
|
|
|
|
"Referer": $(base / getLink(tweet)),
|
|
|
|
"User-Agent": agent,
|
|
|
|
"Authority": "twitter.com",
|
|
|
|
"Accept-Language": lang,
|
|
|
|
})
|
|
|
|
|
2019-07-15 03:44:33 +02:00
|
|
|
let query = get(tweet.card).query.replace("sensitive=true", "sensitive=false")
|
|
|
|
let html = await fetchHtml(base / query, headers)
|
2019-07-11 19:22:23 +02:00
|
|
|
if html == nil: return
|
|
|
|
|
|
|
|
parseCard(get(tweet.card), html)
|
|
|
|
|
|
|
|
proc getCards*(thread: Thread) {.async.} =
|
|
|
|
if thread == nil: return
|
|
|
|
var cards = thread.tweets.filterIt(it.card.isSome)
|
|
|
|
await all(cards.map(getCard))
|
|
|
|
|
|
|
|
proc getConversationCards*(convo: Conversation) {.async.} =
|
|
|
|
var futs: seq[Future[void]]
|
|
|
|
futs.add getCard(convo.tweet)
|
|
|
|
futs.add getCards(convo.before)
|
|
|
|
futs.add getCards(convo.after)
|
|
|
|
futs.add convo.replies.map(getCards)
|
|
|
|
await all(futs)
|
|
|
|
|
2019-07-04 04:18:32 +02:00
|
|
|
proc getPhotoRail*(username: string): Future[seq[GalleryPhoto]] {.async.} =
|
|
|
|
let headers = newHttpHeaders({
|
|
|
|
"Accept": jsonAccept,
|
|
|
|
"Referer": $(base / username),
|
|
|
|
"User-Agent": agent,
|
|
|
|
"X-Requested-With": "XMLHttpRequest"
|
|
|
|
})
|
|
|
|
|
|
|
|
let params = {
|
|
|
|
"for_photo_rail": "true",
|
|
|
|
"oldest_unread_id": "0"
|
|
|
|
}
|
|
|
|
|
|
|
|
let url = base / (timelineMediaUrl % username) ? params
|
|
|
|
let html = await fetchHtml(url, headers, jsonKey="items_html")
|
|
|
|
|
|
|
|
result = parsePhotoRail(html)
|
|
|
|
|
2019-06-29 14:11:23 +02:00
|
|
|
proc getProfileFallback(username: string; headers: HttpHeaders): Future[Profile] {.async.} =
|
|
|
|
let url = base / profileIntentUrl ? {"screen_name": username}
|
|
|
|
let html = await fetchHtml(url, headers)
|
2019-06-27 21:07:29 +02:00
|
|
|
if html == nil: return Profile()
|
2019-06-25 00:55:41 +02:00
|
|
|
|
2019-06-24 05:29:47 +02:00
|
|
|
result = parseIntentProfile(html)
|
|
|
|
|
|
|
|
proc getProfile*(username: string): Future[Profile] {.async.} =
|
|
|
|
let headers = newHttpHeaders({
|
|
|
|
"Accept": "text/html,application/xhtml+xml,application/xml;q=0.9",
|
|
|
|
"Referer": $(base / username),
|
|
|
|
"User-Agent": agent,
|
|
|
|
"X-Twitter-Active-User": "yes",
|
|
|
|
"X-Requested-With": "XMLHttpRequest",
|
2019-06-29 14:11:23 +02:00
|
|
|
"Accept-Language": lang
|
2019-06-24 05:29:47 +02:00
|
|
|
})
|
|
|
|
|
|
|
|
let
|
|
|
|
params = {
|
|
|
|
"screen_name": username,
|
|
|
|
"wants_hovercard": "true",
|
|
|
|
"_": $(epochTime().int)
|
|
|
|
}
|
|
|
|
url = base / profilePopupUrl ? params
|
|
|
|
html = await fetchHtml(url, headers, jsonKey="html")
|
|
|
|
|
2019-06-27 21:07:29 +02:00
|
|
|
if html == nil: return Profile()
|
2019-06-25 00:55:41 +02:00
|
|
|
|
2019-06-27 21:07:29 +02:00
|
|
|
if html.select(".ProfileCard-sensitiveWarningContainer") != nil:
|
2019-06-24 05:29:47 +02:00
|
|
|
return await getProfileFallback(username, headers)
|
|
|
|
|
|
|
|
result = parsePopupProfile(html)
|
|
|
|
|
2019-07-03 11:46:03 +02:00
|
|
|
proc getTweet*(username, id: string): Future[Conversation] {.async.} =
|
2019-06-21 03:51:14 +02:00
|
|
|
let headers = newHttpHeaders({
|
2019-07-03 11:46:03 +02:00
|
|
|
"Accept": jsonAccept,
|
|
|
|
"Referer": $base,
|
2019-06-20 16:16:20 +02:00
|
|
|
"User-Agent": agent,
|
|
|
|
"X-Twitter-Active-User": "yes",
|
|
|
|
"X-Requested-With": "XMLHttpRequest",
|
2019-07-03 11:46:03 +02:00
|
|
|
"Accept-Language": lang,
|
|
|
|
"pragma": "no-cache",
|
|
|
|
"x-previous-page-name": "profile"
|
2019-06-20 16:16:20 +02:00
|
|
|
})
|
|
|
|
|
2019-07-03 11:46:03 +02:00
|
|
|
let
|
|
|
|
url = base / username / tweetUrl / id
|
|
|
|
html = await fetchHtml(url, headers)
|
|
|
|
|
|
|
|
if html == nil: return
|
|
|
|
|
|
|
|
result = parseConversation(html)
|
|
|
|
|
2019-07-11 19:22:23 +02:00
|
|
|
let
|
|
|
|
vidsFut = getConversationVideos(result)
|
|
|
|
pollFut = getConversationPolls(result)
|
|
|
|
cardFut = getConversationCards(result)
|
|
|
|
|
|
|
|
await all(vidsFut, pollFut, cardFut)
|
2019-06-20 16:16:20 +02:00
|
|
|
|
2019-07-11 00:42:31 +02:00
|
|
|
proc finishTimeline(json: JsonNode; query: Option[Query]; after: string): Future[Timeline] {.async.} =
|
2019-06-27 21:07:29 +02:00
|
|
|
if json == nil: return Timeline()
|
2019-06-20 16:16:20 +02:00
|
|
|
|
2019-06-25 07:36:36 +02:00
|
|
|
result = Timeline(
|
|
|
|
hasMore: json["has_more_items"].to(bool),
|
2019-06-25 13:07:49 +02:00
|
|
|
maxId: json.getOrDefault("max_position").getStr(""),
|
2019-07-03 11:46:03 +02:00
|
|
|
minId: json.getOrDefault("min_position").getStr("").cleanPos(),
|
2019-07-11 00:42:31 +02:00
|
|
|
query: query,
|
|
|
|
beginning: after.len == 0
|
2019-06-25 07:36:36 +02:00
|
|
|
)
|
|
|
|
|
2019-06-27 21:48:59 +02:00
|
|
|
if json["new_latent_count"].to(int) == 0: return
|
|
|
|
if not json.hasKey("items_html"): return
|
2019-06-25 07:36:36 +02:00
|
|
|
|
2019-07-01 03:13:12 +02:00
|
|
|
let
|
|
|
|
html = parseHtml(json["items_html"].to(string))
|
|
|
|
thread = parseThread(html)
|
|
|
|
vidsFut = getVideos(thread)
|
|
|
|
pollFut = getPolls(thread)
|
2019-07-11 19:22:23 +02:00
|
|
|
cardFut = getCards(thread)
|
2019-06-29 14:11:23 +02:00
|
|
|
|
2019-07-11 19:22:23 +02:00
|
|
|
await all(vidsFut, pollFut, cardFut)
|
2019-07-01 03:13:12 +02:00
|
|
|
result.tweets = thread.tweets
|
2019-06-20 16:16:20 +02:00
|
|
|
|
2019-07-03 11:46:03 +02:00
|
|
|
proc getTimeline*(username, after: string): Future[Timeline] {.async.} =
|
2019-06-21 03:51:14 +02:00
|
|
|
let headers = newHttpHeaders({
|
2019-07-03 11:46:03 +02:00
|
|
|
"Accept": jsonAccept,
|
|
|
|
"Referer": $(base / username),
|
2019-06-20 16:16:20 +02:00
|
|
|
"User-Agent": agent,
|
|
|
|
"X-Twitter-Active-User": "yes",
|
|
|
|
"X-Requested-With": "XMLHttpRequest",
|
2019-07-03 11:46:03 +02:00
|
|
|
"Accept-Language": lang
|
2019-06-20 16:16:20 +02:00
|
|
|
})
|
|
|
|
|
2019-07-03 11:46:03 +02:00
|
|
|
var params = toSeq({
|
|
|
|
"include_available_features": "1",
|
|
|
|
"include_entities": "1",
|
|
|
|
"include_new_items_bar": "false",
|
|
|
|
"reset_error_state": "false"
|
|
|
|
})
|
2019-06-20 16:16:20 +02:00
|
|
|
|
2019-07-03 11:46:03 +02:00
|
|
|
if after.len > 0:
|
|
|
|
params.add {"max_position": after}
|
2019-06-25 13:07:49 +02:00
|
|
|
|
2019-07-03 11:46:03 +02:00
|
|
|
let json = await fetchJson(base / (timelineUrl % username) ? params, headers)
|
2019-07-11 00:42:31 +02:00
|
|
|
result = await finishTimeline(json, none(Query), after)
|
2019-06-29 14:11:23 +02:00
|
|
|
|
2019-07-03 11:46:03 +02:00
|
|
|
proc getTimelineSearch*(username, after: string; query: Query): Future[Timeline] {.async.} =
|
2019-07-04 14:54:15 +02:00
|
|
|
let queryParam = genQueryParam(query)
|
|
|
|
let queryEncoded = encodeUrl(queryParam, usePlus=false)
|
|
|
|
|
2019-07-03 11:46:03 +02:00
|
|
|
let headers = newHttpHeaders({
|
|
|
|
"Accept": jsonAccept,
|
2019-07-04 14:54:15 +02:00
|
|
|
"Referer": $(base / ("search?f=tweets&vertical=default&q=$1&src=typd" % queryEncoded)),
|
2019-07-03 11:46:03 +02:00
|
|
|
"User-Agent": agent,
|
|
|
|
"X-Requested-With": "XMLHttpRequest",
|
|
|
|
"Authority": "twitter.com",
|
|
|
|
"Accept-Language": lang
|
|
|
|
})
|
|
|
|
|
|
|
|
let params = {
|
|
|
|
"f": "tweets",
|
|
|
|
"vertical": "default",
|
2019-07-04 14:54:15 +02:00
|
|
|
"q": queryParam,
|
2019-07-03 11:46:03 +02:00
|
|
|
"src": "typd",
|
|
|
|
"include_available_features": "1",
|
|
|
|
"include_entities": "1",
|
|
|
|
"max_position": if after.len > 0: genPos(after) else: "0",
|
|
|
|
"reset_error_state": "false"
|
|
|
|
}
|
|
|
|
|
|
|
|
let json = await fetchJson(base / timelineSearchUrl ? params, headers)
|
2019-07-11 00:42:31 +02:00
|
|
|
result = await finishTimeline(json, some(query), after)
|