forked from midou/invidious
Remove 'crawl_threads', fix sleep in fibers
This commit is contained in:
parent
a3b767bb13
commit
591a6b330a
@ -1,5 +1,3 @@
|
|||||||
video_threads: 0
|
|
||||||
crawl_threads: 0
|
|
||||||
channel_threads: 1
|
channel_threads: 1
|
||||||
feed_threads: 1
|
feed_threads: 1
|
||||||
db:
|
db:
|
||||||
|
@ -36,14 +36,6 @@ logger = Invidious::LogHandler.new
|
|||||||
|
|
||||||
Kemal.config.extra_options do |parser|
|
Kemal.config.extra_options do |parser|
|
||||||
parser.banner = "Usage: invidious [arguments]"
|
parser.banner = "Usage: invidious [arguments]"
|
||||||
parser.on("-t THREADS", "--crawl-threads=THREADS", "Number of threads for crawling YouTube (default: #{config.crawl_threads})") do |number|
|
|
||||||
begin
|
|
||||||
config.crawl_threads = number.to_i
|
|
||||||
rescue ex
|
|
||||||
puts "THREADS must be integer"
|
|
||||||
exit
|
|
||||||
end
|
|
||||||
end
|
|
||||||
parser.on("-c THREADS", "--channel-threads=THREADS", "Number of threads for refreshing channels (default: #{config.channel_threads})") do |number|
|
parser.on("-c THREADS", "--channel-threads=THREADS", "Number of threads for refreshing channels (default: #{config.channel_threads})") do |number|
|
||||||
begin
|
begin
|
||||||
config.channel_threads = number.to_i
|
config.channel_threads = number.to_i
|
||||||
@ -60,14 +52,6 @@ Kemal.config.extra_options do |parser|
|
|||||||
exit
|
exit
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
parser.on("-v THREADS", "--video-threads=THREADS", "Number of threads for refreshing videos (default: #{config.video_threads})") do |number|
|
|
||||||
begin
|
|
||||||
config.video_threads = number.to_i
|
|
||||||
rescue ex
|
|
||||||
puts "THREADS must be integer"
|
|
||||||
exit
|
|
||||||
end
|
|
||||||
end
|
|
||||||
parser.on("-o OUTPUT", "--output=OUTPUT", "Redirect output (default: STDOUT)") do |output|
|
parser.on("-o OUTPUT", "--output=OUTPUT", "Redirect output (default: STDOUT)") do |output|
|
||||||
FileUtils.mkdir_p(File.dirname(output))
|
FileUtils.mkdir_p(File.dirname(output))
|
||||||
logger = Invidious::LogHandler.new(File.open(output, mode: "a"))
|
logger = Invidious::LogHandler.new(File.open(output, mode: "a"))
|
||||||
@ -108,24 +92,12 @@ LOCALES = {
|
|||||||
"ru" => load_locale("ru"),
|
"ru" => load_locale("ru"),
|
||||||
}
|
}
|
||||||
|
|
||||||
config.crawl_threads.times do
|
|
||||||
spawn do
|
|
||||||
crawl_videos(PG_DB, logger)
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
refresh_channels(PG_DB, logger, config.channel_threads, config.full_refresh)
|
refresh_channels(PG_DB, logger, config.channel_threads, config.full_refresh)
|
||||||
|
|
||||||
refresh_feeds(PG_DB, logger, config.feed_threads)
|
refresh_feeds(PG_DB, logger, config.feed_threads)
|
||||||
|
|
||||||
subscribe_to_feeds(PG_DB, logger, HMAC_KEY, config)
|
subscribe_to_feeds(PG_DB, logger, HMAC_KEY, config)
|
||||||
|
|
||||||
config.video_threads.times do |i|
|
|
||||||
spawn do
|
|
||||||
refresh_videos(PG_DB, logger)
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
statistics = {
|
statistics = {
|
||||||
"error" => "Statistics are not availabile.",
|
"error" => "Statistics are not availabile.",
|
||||||
}
|
}
|
||||||
@ -154,7 +126,6 @@ if config.statistics_enabled
|
|||||||
}
|
}
|
||||||
|
|
||||||
sleep 1.minute
|
sleep 1.minute
|
||||||
Fiber.yield
|
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
@ -164,8 +135,7 @@ if config.top_enabled
|
|||||||
spawn do
|
spawn do
|
||||||
pull_top_videos(config, PG_DB) do |videos|
|
pull_top_videos(config, PG_DB) do |videos|
|
||||||
top_videos = videos
|
top_videos = videos
|
||||||
sleep 1.minutes
|
sleep 1.minute
|
||||||
Fiber.yield
|
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
@ -174,8 +144,7 @@ popular_videos = [] of ChannelVideo
|
|||||||
spawn do
|
spawn do
|
||||||
pull_popular_videos(PG_DB) do |videos|
|
pull_popular_videos(PG_DB) do |videos|
|
||||||
popular_videos = videos
|
popular_videos = videos
|
||||||
sleep 1.minutes
|
sleep 1.minute
|
||||||
Fiber.yield
|
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
@ -183,8 +152,6 @@ decrypt_function = [] of {name: String, value: Int32}
|
|||||||
spawn do
|
spawn do
|
||||||
update_decrypt_function do |function|
|
update_decrypt_function do |function|
|
||||||
decrypt_function = function
|
decrypt_function = function
|
||||||
sleep 1.minutes
|
|
||||||
Fiber.yield
|
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
@ -4284,6 +4251,7 @@ get "/videoplayback" do |env|
|
|||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
|
# We need this so the below route works as expected
|
||||||
get "/ggpht*" do |env|
|
get "/ggpht*" do |env|
|
||||||
end
|
end
|
||||||
|
|
||||||
|
@ -1,7 +1,5 @@
|
|||||||
class Config
|
class Config
|
||||||
YAML.mapping({
|
YAML.mapping({
|
||||||
video_threads: Int32, # Number of threads to use for updating videos in cache (mostly non-functional)
|
|
||||||
crawl_threads: Int32, # Number of threads to use for finding new videos from YouTube (used to populate "top" page)
|
|
||||||
channel_threads: Int32, # Number of threads to use for crawling videos from channels (for updating subscriptions)
|
channel_threads: Int32, # Number of threads to use for crawling videos from channels (for updating subscriptions)
|
||||||
feed_threads: Int32, # Number of threads to use for updating feeds
|
feed_threads: Int32, # Number of threads to use for updating feeds
|
||||||
db: NamedTuple( # Database configuration
|
db: NamedTuple( # Database configuration
|
||||||
|
@ -1,51 +1,3 @@
|
|||||||
def crawl_videos(db, logger)
|
|
||||||
ids = Deque(String).new
|
|
||||||
random = Random.new
|
|
||||||
|
|
||||||
search(random.base64(3)).as(Tuple)[1].each do |video|
|
|
||||||
if video.is_a?(SearchVideo)
|
|
||||||
ids << video.id
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
loop do
|
|
||||||
if ids.empty?
|
|
||||||
search(random.base64(3)).as(Tuple)[1].each do |video|
|
|
||||||
if video.is_a?(SearchVideo)
|
|
||||||
ids << video.id
|
|
||||||
end
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
begin
|
|
||||||
id = ids[0]
|
|
||||||
video = get_video(id, db)
|
|
||||||
rescue ex
|
|
||||||
logger.write("#{id} : #{ex.message}\n")
|
|
||||||
next
|
|
||||||
ensure
|
|
||||||
ids.delete(id)
|
|
||||||
end
|
|
||||||
|
|
||||||
rvs = [] of Hash(String, String)
|
|
||||||
video.info["rvs"]?.try &.split(",").each do |rv|
|
|
||||||
rvs << HTTP::Params.parse(rv).to_h
|
|
||||||
end
|
|
||||||
|
|
||||||
rvs.each do |rv|
|
|
||||||
if rv.has_key?("id") && !db.query_one?("SELECT EXISTS (SELECT true FROM videos WHERE id = $1)", rv["id"], as: Bool)
|
|
||||||
ids.delete(id)
|
|
||||||
ids << rv["id"]
|
|
||||||
if ids.size == 150
|
|
||||||
ids.shift
|
|
||||||
end
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
Fiber.yield
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
def refresh_channels(db, logger, max_threads = 1, full_refresh = false)
|
def refresh_channels(db, logger, max_threads = 1, full_refresh = false)
|
||||||
max_channel = Channel(Int32).new
|
max_channel = Channel(Int32).new
|
||||||
|
|
||||||
@ -82,30 +34,14 @@ def refresh_channels(db, logger, max_threads = 1, full_refresh = false)
|
|||||||
end
|
end
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
|
sleep 1.minute
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
max_channel.send(max_threads)
|
max_channel.send(max_threads)
|
||||||
end
|
end
|
||||||
|
|
||||||
def refresh_videos(db, logger)
|
|
||||||
loop do
|
|
||||||
db.query("SELECT id FROM videos ORDER BY updated") do |rs|
|
|
||||||
rs.each do
|
|
||||||
begin
|
|
||||||
id = rs.read(String)
|
|
||||||
video = get_video(id, db)
|
|
||||||
rescue ex
|
|
||||||
logger.write("#{id} : #{ex.message}\n")
|
|
||||||
next
|
|
||||||
end
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
Fiber.yield
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
def refresh_feeds(db, logger, max_threads = 1)
|
def refresh_feeds(db, logger, max_threads = 1)
|
||||||
max_channel = Channel(Int32).new
|
max_channel = Channel(Int32).new
|
||||||
|
|
||||||
@ -158,6 +94,8 @@ def refresh_feeds(db, logger, max_threads = 1)
|
|||||||
end
|
end
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
|
sleep 1.minute
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
@ -180,7 +118,6 @@ def subscribe_to_feeds(db, logger, key, config)
|
|||||||
end
|
end
|
||||||
|
|
||||||
sleep 1.minute
|
sleep 1.minute
|
||||||
Fiber.yield
|
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
@ -211,7 +148,7 @@ def pull_top_videos(config, db)
|
|||||||
end
|
end
|
||||||
|
|
||||||
yield videos
|
yield videos
|
||||||
Fiber.yield
|
sleep 1.minute
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
@ -226,7 +163,7 @@ def pull_popular_videos(db)
|
|||||||
ORDER BY ucid, published DESC", subscriptions, as: ChannelVideo).sort_by { |video| video.published }.reverse
|
ORDER BY ucid, published DESC", subscriptions, as: ChannelVideo).sort_by { |video| video.published }.reverse
|
||||||
|
|
||||||
yield videos
|
yield videos
|
||||||
Fiber.yield
|
sleep 1.minute
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
@ -239,6 +176,7 @@ def update_decrypt_function
|
|||||||
end
|
end
|
||||||
|
|
||||||
yield decrypt_function
|
yield decrypt_function
|
||||||
|
sleep 1.minute
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
@ -250,7 +188,8 @@ def find_working_proxies(regions)
|
|||||||
# proxies = filter_proxies(proxies)
|
# proxies = filter_proxies(proxies)
|
||||||
|
|
||||||
yield region, proxies
|
yield region, proxies
|
||||||
Fiber.yield
|
end
|
||||||
end
|
|
||||||
|
sleep 1.minute
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
Loading…
Reference in New Issue
Block a user