invidious-experimenting/src/helpers.cr

645 lines
17 KiB
Crystal
Raw Normal View History

macro add_mapping(mapping)
def initialize({{*mapping.keys.map { |id| "@#{id}".id }}})
end
def to_a
return [{{*mapping.keys.map { |id| "@#{id}".id }}}]
end
DB.mapping({{mapping}})
end
macro templated(filename)
render "src/views/#{{{filename}}}.ecr", "src/views/layout.ecr"
end
2018-03-10 00:12:23 +05:30
class Config
YAML.mapping({
2018-05-02 05:21:16 +05:30
crawl_threads: Int32,
channel_threads: Int32,
2018-04-28 21:20:02 +05:30
video_threads: Int32,
db: NamedTuple(
2018-03-10 00:12:23 +05:30
user: String,
password: String,
host: String,
port: Int32,
dbname: String,
),
dl_api_key: String?,
2018-03-10 00:12:23 +05:30
})
end
2018-04-16 09:26:58 +05:30
class FilteredCompressHandler < Kemal::Handler
exclude ["/videoplayback"]
def call(env)
return call_next env if exclude_match? env
{% if flag?(:without_zlib) %}
call_next env
{% else %}
request_headers = env.request.headers
if request_headers.includes_word?("Accept-Encoding", "gzip")
env.response.headers["Content-Encoding"] = "gzip"
env.response.output = Gzip::Writer.new(env.response.output, sync_close: true)
elsif request_headers.includes_word?("Accept-Encoding", "deflate")
env.response.headers["Content-Encoding"] = "deflate"
env.response.output = Flate::Writer.new(env.response.output, sync_close: true)
end
call_next env
{% end %}
end
end
2018-01-28 07:39:27 +05:30
class Video
module HTTPParamConverter
def self.from_rs(rs)
HTTP::Params.parse(rs.read(String))
end
end
add_mapping({
2018-01-28 07:39:27 +05:30
id: String,
info: {
type: HTTP::Params,
default: HTTP::Params.parse(""),
converter: Video::HTTPParamConverter,
},
updated: Time,
title: String,
views: Int64,
likes: Int32,
dislikes: Int32,
wilson_score: Float64,
2018-02-03 09:14:10 +05:30
published: Time,
2018-02-27 06:28:45 +05:30
description: String,
2018-03-17 06:15:37 +05:30
language: String?,
2018-01-28 07:39:27 +05:30
})
end
2018-03-25 09:08:35 +05:30
class InvidiousChannel
module XMLConverter
def self.from_rs(rs)
XML.parse_html(rs.read(String))
end
end
add_mapping({
id: String,
2018-03-25 09:08:35 +05:30
author: String,
updated: Time,
})
end
class ChannelVideo
add_mapping({
id: String,
title: String,
published: Time,
updated: Time,
ucid: String,
author: String,
2018-03-25 09:08:35 +05:30
})
end
2018-03-30 08:11:05 +05:30
class User
add_mapping({
2018-04-01 05:39:27 +05:30
id: String,
updated: Time,
notifications: Array(String),
subscriptions: Array(String),
email: String,
2018-03-30 08:11:05 +05:30
})
end
2018-04-22 04:34:01 +05:30
class RedditThing
2018-03-04 02:36:14 +05:30
JSON.mapping({
2018-04-22 04:34:01 +05:30
kind: String,
data: RedditComment | RedditLink | RedditMore | RedditListing,
2018-03-04 02:36:14 +05:30
})
end
2018-04-22 04:34:01 +05:30
class RedditComment
2018-03-04 02:36:14 +05:30
JSON.mapping({
2018-04-22 04:34:01 +05:30
author: String,
body_html: String,
replies: RedditThing | String,
score: Int32,
depth: Int32,
2018-03-04 02:36:14 +05:30
})
end
2018-04-22 04:34:01 +05:30
class RedditLink
2018-03-04 02:36:14 +05:30
JSON.mapping({
2018-04-22 04:34:01 +05:30
author: String,
score: Int32,
2018-03-04 02:36:14 +05:30
subreddit: String,
num_comments: Int32,
2018-04-22 04:34:01 +05:30
id: String,
2018-03-04 02:36:14 +05:30
permalink: String,
title: String,
})
end
2018-04-22 04:34:01 +05:30
class RedditMore
JSON.mapping({
children: Array(String),
count: Int32,
depth: Int32,
})
end
class RedditListing
JSON.mapping({
children: Array(RedditThing),
modhash: String,
})
end
2018-01-21 05:49:12 +05:30
# See http://www.evanmiller.org/how-not-to-sort-by-average-rating.html
def ci_lower_bound(pos, n)
if n == 0
2018-01-28 07:39:27 +05:30
return 0.0
2018-01-21 05:49:12 +05:30
end
# z value here represents a confidence level of 0.95
z = 1.96
phat = 1.0*pos/n
return (phat + z*z/(2*n) - z * Math.sqrt((phat*(1 - phat) + z*z/(4*n))/n))/(1 + z*z/n)
end
def elapsed_text(elapsed)
millis = elapsed.total_milliseconds
return "#{millis.round(2)}ms" if millis >= 1
"#{(millis * 1000).round(2)}µs"
end
2018-01-28 07:39:27 +05:30
def fetch_video(id, client)
2018-03-04 20:24:19 +05:30
info = client.get("/get_video_info?video_id=#{id}&el=detailpage&ps=default&eurl=&gl=US&hl=en").body
2018-03-09 22:25:14 +05:30
html = client.get("/watch?v=#{id}&bpctr=#{Time.new.epoch + 2000}").body
2018-01-21 05:49:12 +05:30
html = XML.parse_html(html)
2018-01-28 07:39:27 +05:30
info = HTTP::Params.parse(info)
2018-01-21 05:49:12 +05:30
if info["reason"]?
2018-02-03 09:34:34 +05:30
info = client.get("/get_video_info?video_id=#{id}&ps=default&eurl=&gl=US&hl=en").body
info = HTTP::Params.parse(info)
if info["reason"]?
raise info["reason"]
end
2018-01-21 05:49:12 +05:30
end
2018-01-28 07:39:27 +05:30
title = info["title"]
views = info["view_count"].to_i64
2018-01-21 05:49:12 +05:30
2018-01-28 07:39:27 +05:30
likes = html.xpath_node(%q(//button[@title="I like this"]/span))
2018-03-15 04:36:21 +05:30
likes = likes.try &.content.delete(",").try &.to_i
likes ||= 0
2018-01-28 07:39:27 +05:30
dislikes = html.xpath_node(%q(//button[@title="I dislike this"]/span))
2018-03-15 04:36:21 +05:30
dislikes = dislikes.try &.content.delete(",").try &.to_i
dislikes ||= 0
2018-01-28 07:39:27 +05:30
2018-02-27 06:28:45 +05:30
description = html.xpath_node(%q(//p[@id="eow-description"]))
description = description ? description.to_xml : ""
2018-01-28 07:39:27 +05:30
wilson_score = ci_lower_bound(likes, likes + dislikes)
2018-02-05 07:12:13 +05:30
published = html.xpath_node(%q(//strong[contains(@class,"watch-time-text")]))
2018-02-03 09:14:10 +05:30
if published
published = published.content
2018-02-05 07:12:13 +05:30
else
raise "Could not find date published"
end
published = published.lchop("Published ")
published = published.lchop("Started streaming ")
2018-03-17 09:04:26 +05:30
published = published.lchop("Streamed live ")
published = published.lchop("Uploaded ")
2018-02-05 07:12:13 +05:30
published = published.lchop("on ")
published = published.lchop("Scheduled for ")
2018-02-06 05:27:03 +05:30
if !published.includes?("ago")
published = Time.parse(published, "%b %-d, %Y")
else
# Time matches format "20 hours ago", "40 minutes ago"...
2018-02-05 07:12:13 +05:30
delta = published.split(" ")[0].to_i
case published
when .includes? "minute"
published = Time.now - delta.minutes
when .includes? "hour"
published = Time.now - delta.hours
2018-02-06 05:27:03 +05:30
else
2018-02-05 07:12:13 +05:30
raise "Could not parse #{published}"
2018-02-06 05:27:03 +05:30
end
2018-02-03 09:14:10 +05:30
end
2018-03-17 06:15:37 +05:30
video = Video.new(id, info, Time.now, title, views, likes, dislikes, wilson_score, published, description, nil)
2018-01-21 05:49:12 +05:30
return video
end
2018-01-28 07:39:27 +05:30
def get_video(id, client, db, refresh = true)
if db.query_one?("SELECT EXISTS (SELECT true FROM videos WHERE id = $1)", id, as: Bool)
video = db.query_one("SELECT * FROM videos WHERE id = $1", id, as: Video)
2018-01-21 05:49:12 +05:30
2018-01-28 07:39:27 +05:30
# If record was last updated over an hour ago, refresh (expire param in response lasts for 6 hours)
2018-03-30 05:51:44 +05:30
if refresh && Time.now - video.updated > 1.hour
begin
2018-03-31 20:21:44 +05:30
video = fetch_video(id, client)
2018-03-30 05:33:00 +05:30
video_array = video.to_a
args = arg_array(video_array[1..-1], 2)
2018-03-30 05:33:00 +05:30
db.exec("UPDATE videos SET (info,updated,title,views,likes,dislikes,wilson_score,published,description,language)\
= (#{args}) WHERE id = $1", video_array)
rescue ex
db.exec("DELETE FROM videos * WHERE id = $1", id)
2018-03-31 20:21:44 +05:30
end
2018-01-21 05:49:12 +05:30
end
else
2018-01-28 07:39:27 +05:30
video = fetch_video(id, client)
2018-03-04 20:24:19 +05:30
args = arg_array(video.to_a)
db.exec("INSERT INTO videos VALUES (#{args})", video.to_a)
2018-01-21 05:49:12 +05:30
end
return video
end
2018-01-22 05:19:27 +05:30
2018-01-28 07:39:27 +05:30
def search(query, client)
2018-03-04 20:24:19 +05:30
html = client.get("https://www.youtube.com/results?q=#{query}&sp=EgIQAVAU").body
2018-01-22 05:19:27 +05:30
html = XML.parse_html(html)
html.xpath_nodes(%q(//ol[@class="item-section"]/li)).each do |item|
root = item.xpath_node(%q(div[contains(@class,"yt-lockup-video")]/div))
if root
link = root.xpath_node(%q(div[contains(@class,"yt-lockup-thumbnail")]/a/@href))
if link
yield link.content.split("=")[1]
end
end
end
end
def splice(a, b)
c = a[0]
a[0] = a[b % a.size]
a[b % a.size] = c
return a
end
def decrypt_signature(a)
a = a.split("")
2018-05-06 07:19:46 +05:30
a.delete_at(0..1)
2018-04-14 05:47:53 +05:30
a.reverse!
2018-04-24 23:31:43 +05:30
a.delete_at(0..1)
2018-04-14 05:47:53 +05:30
a.reverse!
2018-05-06 07:19:46 +05:30
a.delete_at(0..2)
a = splice(a, 60)
return a.join("")
end
2018-04-28 19:52:06 +05:30
def rank_videos(db, n, filter, url)
2018-02-06 05:26:40 +05:30
top = [] of {Float64, String}
2018-03-17 10:27:31 +05:30
db.query("SELECT id, wilson_score, published FROM videos WHERE views > 5000 ORDER BY published DESC LIMIT 1000") do |rs|
2018-02-06 05:26:40 +05:30
rs.each do
id = rs.read(String)
wilson_score = rs.read(Float64)
published = rs.read(Time)
# Exponential decay, older videos tend to rank lower
2018-02-10 21:36:37 +05:30
temperature = wilson_score * Math.exp(-0.000005*((Time.now - published).total_minutes))
2018-02-06 05:26:40 +05:30
top << {temperature, id}
end
end
top.sort!
# Make hottest come first
top.reverse!
top = top.map { |a, b| b }
2018-03-17 06:06:49 +05:30
if filter
language_list = [] of String
top.each do |id|
if language_list.size == n
break
else
2018-04-28 19:52:06 +05:30
client = make_client(url)
2018-03-19 23:05:35 +05:30
begin
video = get_video(id, client, db)
rescue ex
next
end
2018-03-17 06:15:37 +05:30
if video.language
language = video.language
else
description = XML.parse(video.description)
content = [video.title, description.content].join(" ")
2018-03-19 23:05:35 +05:30
content = content[0, 10000]
2018-03-17 06:06:49 +05:30
2018-03-17 06:15:37 +05:30
results = DetectLanguage.detect(content)
language = results[0].language
2018-03-17 06:06:49 +05:30
2018-03-17 06:15:37 +05:30
db.exec("UPDATE videos SET language = $1 WHERE id = $2", language, id)
end
if language == "en"
2018-03-17 06:06:49 +05:30
language_list << id
end
end
end
return language_list
else
return top[0..n - 1]
end
2018-02-06 05:26:40 +05:30
end
2018-02-06 06:37:49 +05:30
2018-03-05 09:55:03 +05:30
def make_client(url)
context = OpenSSL::SSL::Context::Client.new
context.add_options(
OpenSSL::SSL::Options::ALL |
OpenSSL::SSL::Options::NO_SSL_V2 |
OpenSSL::SSL::Options::NO_SSL_V3
)
2018-02-27 06:29:02 +05:30
client = HTTP::Client.new(url, context)
client.read_timeout = 10.seconds
client.connect_timeout = 10.seconds
2018-02-06 06:37:49 +05:30
return client
end
2018-03-04 02:36:14 +05:30
def get_reddit_comments(id, client, headers)
query = "(url:3D#{id}%20OR%20url:#{id})%20(site:youtube.com%20OR%20site:youtu.be)"
search_results = client.get("/search.json?q=#{query}", headers)
2018-03-09 22:25:14 +05:30
if search_results.status_code == 200
2018-04-22 04:34:01 +05:30
search_results = RedditThing.from_json(search_results.body)
thread = search_results.data.as(RedditListing).children.sort_by { |child| child.data.as(RedditLink).score }[-1]
thread = thread.data.as(RedditLink)
2018-03-09 22:25:14 +05:30
2018-04-22 04:34:01 +05:30
result = client.get("/r/#{thread.subreddit}/comments/#{thread.id}?limit=100&sort=top", headers).body
result = Array(RedditThing).from_json(result)
elsif search_results.status_code == 302
2018-04-22 04:34:01 +05:30
result = client.get(search_results.headers["Location"], headers).body
result = Array(RedditThing).from_json(result)
2018-04-22 04:34:01 +05:30
thread = result[0].data.as(RedditListing).children[0].data.as(RedditLink)
else
raise "Got error code #{search_results.status_code}"
2018-03-04 06:40:25 +05:30
end
2018-04-22 04:34:01 +05:30
comments = result[1].data.as(RedditListing).children
2018-03-04 06:40:25 +05:30
return comments, thread
2018-03-04 02:36:14 +05:30
end
def template_comments(root)
html = ""
root.each do |child|
2018-04-22 04:34:01 +05:30
if child.data.is_a?(RedditComment)
child = child.data.as(RedditComment)
author = child.author
score = child.score
body_html = HTML.unescape(child.body_html)
2018-03-04 02:36:14 +05:30
replies_html = ""
2018-04-22 04:34:01 +05:30
if child.replies.is_a?(RedditThing)
replies = child.replies.as(RedditThing)
replies_html = template_comments(replies.data.as(RedditListing).children)
2018-03-04 02:36:14 +05:30
end
content = <<-END_HTML
<p>
<a href="javascript:void(0)" onclick="toggle(this)">[ - ]</a> #{score} <b>#{author}</b>
2018-03-04 02:36:14 +05:30
</p>
<div>
#{body_html}
2018-03-04 02:36:14 +05:30
#{replies_html}
</div>
2018-03-04 02:36:14 +05:30
END_HTML
2018-04-22 04:34:01 +05:30
if child.depth > 0
2018-03-04 02:36:14 +05:30
html += <<-END_HTML
<div class="pure-g">
<div class="pure-u-1-24">
</div>
2018-03-04 02:36:14 +05:30
<div class="pure-u-23-24">
#{content}
</div>
</div>
END_HTML
else
html += <<-END_HTML
<div class="pure-g">
<div class="pure-u-1">
#{content}
</div>
</div>
END_HTML
end
end
end
return html
end
def number_with_separator(number)
number.to_s.reverse.gsub(/(\d{3})(?=\d)/, "\\1,").reverse
end
2018-03-04 20:24:19 +05:30
2018-03-30 05:33:00 +05:30
def arg_array(array, start = 1)
2018-04-01 20:16:13 +05:30
if array.size == 0
args = "NULL"
else
args = [] of String
(start..array.size + start - 1).each { |i| args << "($#{i})" }
args = args.join(",")
end
2018-03-04 20:24:19 +05:30
return args
end
2018-03-07 09:30:35 +05:30
def add_alt_links(html)
alt_links = [] of {Int32, String}
# This is painful but is likely the only way to accomplish this in Crystal,
# as Crystigiri and others are not able to insert XML Nodes into a document.
# The goal here is to use as little regex as possible
html.scan(/<a[^>]*>([^<]+)<\/a>/) do |match|
anchor = XML.parse_html(match[0])
anchor = anchor.xpath_node("//a").not_nil!
url = URI.parse(HTML.unescape(anchor["href"]))
if ["www.youtube.com", "m.youtube.com"].includes?(url.host) && url.path == "/watch" || url.path == "/redirect"
2018-03-07 09:30:35 +05:30
alt_link = <<-END_HTML
<a href="#{url.full_path}">
2018-03-07 09:30:35 +05:30
<i class="fa fa-link" aria-hidden="true"></i>
</a>
END_HTML
2018-03-10 01:36:35 +05:30
elsif url.host == "youtu.be"
alt_link = <<-END_HTML
<a href="/watch?v=#{url.full_path.lchop("/")}">
<i class="fa fa-link" aria-hidden="true"></i>
</a>
END_HTML
else
alt_link = ""
2018-03-07 09:30:35 +05:30
end
2018-03-10 01:36:35 +05:30
alt_links << {match.end.not_nil!, alt_link}
2018-03-07 09:30:35 +05:30
end
alt_links.reverse!
alt_links.each do |position, alt_link|
html = html.insert(position, alt_link)
end
return html
end
def fill_links(html, scheme, host)
html = XML.parse_html(html)
html.xpath_nodes("//a").each do |match|
url = URI.parse(match["href"])
2018-03-26 08:51:24 +05:30
# Reddit links don't have host
if !url.host && !match["href"].starts_with?("javascript")
2018-03-07 09:30:35 +05:30
url.scheme = scheme
url.host = host
match["href"] = url
end
end
html = html.to_xml
end
2018-03-16 22:10:29 +05:30
def login_req(login_form, f_req)
data = {
"pstMsg" => "1",
"checkConnection" => "youtube",
"checkedDomains" => "youtube",
"hl" => "en",
"deviceinfo" => %q([null,null,null,[],null,"US",null,null,[],"GlifWebSignIn",null,[null,null,[]]]),
"f.req" => f_req,
"flowName" => "GlifWebSignIn",
"flowEntry" => "ServiceLogin",
}
2018-04-28 19:57:05 +05:30
data = login_form.merge(data)
2018-03-16 22:10:29 +05:30
return HTTP::Params.encode(data)
end
2018-03-25 09:08:35 +05:30
def get_channel(id, client, db)
if db.query_one?("SELECT EXISTS (SELECT true FROM channels WHERE id = $1)", id, as: Bool)
channel = db.query_one("SELECT * FROM channels WHERE id = $1", id, as: InvidiousChannel)
2018-04-12 05:06:36 +05:30
if Time.now - channel.updated > 1.minute
channel = fetch_channel(id, client, db)
channel_array = channel.to_a
2018-03-26 08:48:29 +05:30
args = arg_array(channel_array)
db.exec("INSERT INTO channels VALUES (#{args}) \
2018-03-30 05:20:24 +05:30
ON CONFLICT (id) DO UPDATE SET updated = $3", channel_array)
2018-03-25 09:08:35 +05:30
end
else
channel = fetch_channel(id, client, db)
2018-03-25 09:08:35 +05:30
args = arg_array(channel.to_a)
db.exec("INSERT INTO channels VALUES (#{args})", channel.to_a)
end
return channel
end
def fetch_channel(id, client, db)
2018-03-25 09:08:35 +05:30
rss = client.get("/feeds/videos.xml?channel_id=#{id}").body
rss = XML.parse_html(rss)
2018-04-01 03:03:01 +05:30
db.exec("DELETE FROM channel_videos * WHERE ucid = $1", id)
rss.xpath_nodes("//feed/entry").each do |entry|
video_id = entry.xpath_node("videoid").not_nil!.content
title = entry.xpath_node("title").not_nil!.content
published = Time.parse(entry.xpath_node("published").not_nil!.content, "%FT%X%z")
updated = Time.parse(entry.xpath_node("updated").not_nil!.content, "%FT%X%z")
author = entry.xpath_node("author/name").not_nil!.content
ucid = entry.xpath_node("channelid").not_nil!.content
video = ChannelVideo.new(video_id, title, published, updated, ucid, author)
2018-03-31 20:21:44 +05:30
video_array = video.to_a
args = arg_array(video_array)
2018-04-01 05:39:27 +05:30
db.exec("UPDATE users SET notifications = notifications || $1 \
WHERE updated < $2 AND $3 = ANY(subscriptions) AND $1 <> ALL(notifications)", video_id, published, ucid)
# TODO: Update record on conflict
2018-03-31 20:21:44 +05:30
db.exec("INSERT INTO channel_videos VALUES (#{args})\
ON CONFLICT (id) DO NOTHING", video_array)
end
author = rss.xpath_node("//feed/author/name").try &.content
author ||= ""
2018-03-25 09:08:35 +05:30
channel = InvidiousChannel.new(id, author, Time.now)
2018-03-25 09:08:35 +05:30
return channel
end
2018-03-30 08:11:05 +05:30
def get_user(sid, client, headers, db)
if db.query_one?("SELECT EXISTS (SELECT true FROM users WHERE id = $1)", sid, as: Bool)
user = db.query_one("SELECT * FROM users WHERE id = $1", sid, as: User)
if Time.now - user.updated > 1.minutes
user = fetch_user(sid, client, headers)
user_array = user.to_a
args = arg_array(user_array)
2018-03-31 20:21:44 +05:30
2018-03-30 08:11:05 +05:30
db.exec("INSERT INTO users VALUES (#{args}) \
2018-04-04 03:38:44 +05:30
ON CONFLICT (email) DO UPDATE SET id = $1, updated = $2, notifications = $3, subscriptions = $4", user_array)
2018-03-30 08:11:05 +05:30
end
else
user = fetch_user(sid, client, headers)
2018-03-31 21:00:17 +05:30
user_array = user.to_a
2018-03-30 08:11:05 +05:30
args = arg_array(user.to_a)
2018-03-31 21:00:17 +05:30
db.exec("INSERT INTO users VALUES (#{args}) \
ON CONFLICT (email) DO UPDATE SET id = $1, updated = $2, subscriptions = $4", user_array)
2018-03-30 08:11:05 +05:30
end
return user
end
def fetch_user(sid, client, headers)
2018-04-29 20:10:33 +05:30
feed = client.get("/subscription_manager?disable_polymer=1", headers)
feed = XML.parse_html(feed.body)
2018-03-30 08:11:05 +05:30
channels = [] of String
2018-05-04 07:07:17 +05:30
feed.xpath_nodes(%q(//ul[@id="guide-channels"]/li/a)).each do |channel|
if !["Popular on YouTube", "Music", "Sports", "Gaming"].includes? channel["title"]
channel_id = channel["href"].lstrip("/channel/")
2018-05-04 07:07:17 +05:30
get_channel(channel_id, client, PG_DB)
2018-03-31 21:00:17 +05:30
2018-05-04 07:07:17 +05:30
channels << channel_id
end
2018-03-31 21:00:17 +05:30
end
email = feed.xpath_node(%q(//a[@class="yt-masthead-picker-header yt-masthead-picker-active-account"]))
if email
2018-04-10 09:45:01 +05:30
email = email.content.strip
2018-03-31 21:00:17 +05:30
else
email = ""
2018-03-30 08:11:05 +05:30
end
2018-04-01 05:39:27 +05:30
user = User.new(sid, Time.now, [] of String, channels, email)
2018-03-30 08:11:05 +05:30
return user
end