Minor refactor

This commit is contained in:
Omar Roth 2019-08-27 08:08:26 -05:00
parent 7d3ecd2297
commit 26107bd6c3
No known key found for this signature in database
GPG key ID: B8254FB7EC3D37F2
3 changed files with 86 additions and 131 deletions

View file

@ -148,7 +148,6 @@ statistics = {
} }
if config.statistics_enabled if config.statistics_enabled
spawn do spawn do
loop do
statistics = { statistics = {
"version" => "2.0", "version" => "2.0",
"software" => SOFTWARE, "software" => SOFTWARE,
@ -162,12 +161,19 @@ if config.statistics_enabled
}, },
"metadata" => { "metadata" => {
"updatedAt" => Time.utc.to_unix, "updatedAt" => Time.utc.to_unix,
"lastChannelRefreshedAt" => PG_DB.query_one?("SELECT updated FROM channels ORDER BY updated DESC LIMIT 1", as: Time).try &.to_unix || 0, "lastChannelRefreshedAt" => PG_DB.query_one?("SELECT updated FROM channels ORDER BY updated DESC LIMIT 1", as: Time).try &.to_unix || 0_i64,
}, },
} }
loop do
sleep 1.minute sleep 1.minute
Fiber.yield Fiber.yield
statistics["usage"].as(Hash)["users"].as(Hash)["total"] = PG_DB.query_one("SELECT count(*) FROM users", as: Int64)
statistics["usage"].as(Hash)["users"].as(Hash)["activeHalfyear"] = PG_DB.query_one("SELECT count(*) FROM users WHERE CURRENT_TIMESTAMP - updated < '6 months'", as: Int64)
statistics["usage"].as(Hash)["users"].as(Hash)["activeMonth"] = PG_DB.query_one("SELECT count(*) FROM users WHERE CURRENT_TIMESTAMP - updated < '1 month'", as: Int64)
statistics["metadata"].as(Hash(String, Int64))["updatedAt"] = Time.utc.to_unix
statistics["metadata"].as(Hash(String, Int64))["lastChannelRefreshedAt"] = PG_DB.query_one?("SELECT updated FROM channels ORDER BY updated DESC LIMIT 1", as: Time).try &.to_unix || 0_i64
end end
end end
end end
@ -3223,7 +3229,8 @@ get "/api/v1/storyboards/:id" do |env|
storyboard = storyboard[0] storyboard = storyboard[0]
end end
webvtt = <<-END_VTT String.build do |str|
str << <<-END_VTT
WEBVTT WEBVTT
@ -3238,7 +3245,7 @@ get "/api/v1/storyboards/:id" do |env|
storyboard[:storyboard_height].times do |j| storyboard[:storyboard_height].times do |j|
storyboard[:storyboard_width].times do |k| storyboard[:storyboard_width].times do |k|
webvtt += <<-END_CUE str << <<-END_CUE
#{start_time}.000 --> #{end_time}.000 #{start_time}.000 --> #{end_time}.000
#{url}#xywh=#{storyboard[:width] * k},#{storyboard[:height] * j},#{storyboard[:width]},#{storyboard[:height]} #{url}#xywh=#{storyboard[:width] * k},#{storyboard[:height] * j},#{storyboard[:width]},#{storyboard[:height]}
@ -3250,8 +3257,7 @@ get "/api/v1/storyboards/:id" do |env|
end end
end end
end end
end
webvtt
end end
get "/api/v1/captions/:id" do |env| get "/api/v1/captions/:id" do |env|
@ -3321,7 +3327,7 @@ get "/api/v1/captions/:id" do |env|
caption = caption[0] caption = caption[0]
end end
url = caption.baseUrl + "&tlang=#{tlang}" url = "#{caption.baseUrl}&tlang=#{tlang}"
# Auto-generated captions often have cues that aren't aligned properly with the video, # Auto-generated captions often have cues that aren't aligned properly with the video,
# as well as some other markup that makes it cumbersome, so we try to fix that here # as well as some other markup that makes it cumbersome, so we try to fix that here
@ -3329,7 +3335,8 @@ get "/api/v1/captions/:id" do |env|
caption_xml = client.get(url).body caption_xml = client.get(url).body
caption_xml = XML.parse(caption_xml) caption_xml = XML.parse(caption_xml)
webvtt = <<-END_VTT webvtt = String.build do |str|
str << <<-END_VTT
WEBVTT WEBVTT
Kind: captions Kind: captions
Language: #{tlang || caption.languageCode} Language: #{tlang || caption.languageCode}
@ -3359,16 +3366,16 @@ get "/api/v1/captions/:id" do |env|
text = "<v #{md["name"]}>#{md["text"]}</v>" text = "<v #{md["name"]}>#{md["text"]}</v>"
end end
webvtt += <<-END_CUE str << <<-END_CUE
#{start_time} --> #{end_time} #{start_time} --> #{end_time}
#{text} #{text}
END_CUE END_CUE
end end
end
else else
url += "&format=vtt" webvtt = client.get("#{url}&format=vtt").body
webvtt = client.get(url).body
end end
if title = env.params.query["title"]? if title = env.params.query["title"]?
@ -4833,43 +4840,6 @@ get "/videoplayback" do |env|
end end
end end
response = HTTP::Client::Response.new(403)
5.times do
begin
client = make_client(URI.parse(host), region)
response = client.head(url, headers)
break
rescue Socket::Addrinfo::Error
if !mns.empty?
mn = mns.pop
end
fvip = "3"
host = "https://r#{fvip}---#{mn}.googlevideo.com"
rescue ex
end
end
if response.headers["Location"]?
url = URI.parse(response.headers["Location"])
host = url.host
env.response.headers["Access-Control-Allow-Origin"] = "*"
url = url.full_path
url += "&host=#{host}"
if region
url += "&region=#{region}"
end
next env.redirect url
end
if response.status_code >= 400
env.response.status_code = response.status_code
next
end
if url.includes? "&file=seg.ts" if url.includes? "&file=seg.ts"
if CONFIG.disabled?("livestreams") if CONFIG.disabled?("livestreams")
env.response.status_code = 403 env.response.status_code = 403
@ -4957,11 +4927,7 @@ get "/videoplayback" do |env|
if location = response.headers["Location"]? if location = response.headers["Location"]?
location = URI.parse(location) location = URI.parse(location)
location = "#{location.full_path}&host=#{location.host}" location = "#{location.full_path}&host=#{location.host}#{region ? "&region=#{region}" : ""}"
if region
location += "&region=#{region}"
end
env.redirect location env.redirect location
break break

View file

@ -159,10 +159,9 @@ class APIHandler < Kemal::Handler
call_next env call_next env
env.response.output.rewind env.response.output.rewind
response = env.response.output.gets_to_end
if env.response.headers["Content-Type"]?.try &.== "application/json" if env.response.headers.includes_word?("Content-Type", "application/json")
response = JSON.parse(response) response = JSON.parse(env.response.output)
if fields_text = env.params.query["fields"]? if fields_text = env.params.query["fields"]?
begin begin
@ -178,6 +177,8 @@ class APIHandler < Kemal::Handler
else else
response = response.to_json response = response.to_json
end end
else
response = env.response.output.gets_to_end
end end
rescue ex rescue ex
ensure ensure

View file

@ -295,8 +295,7 @@ def get_subscription_feed(db, user, max_results = 40, page = 1)
args = arg_array(notifications) args = arg_array(notifications)
notifications = db.query_all("SELECT * FROM channel_videos WHERE id IN (#{args}) notifications = db.query_all("SELECT * FROM channel_videos WHERE id IN (#{args}) ORDER BY published DESC", notifications, as: ChannelVideo)
ORDER BY published DESC", notifications, as: ChannelVideo)
videos = [] of ChannelVideo videos = [] of ChannelVideo
notifications.sort_by! { |video| video.published }.reverse! notifications.sort_by! { |video| video.published }.reverse!
@ -322,14 +321,11 @@ def get_subscription_feed(db, user, max_results = 40, page = 1)
else else
values = "VALUES #{user.watched.map { |id| %(('#{id}')) }.join(",")}" values = "VALUES #{user.watched.map { |id| %(('#{id}')) }.join(",")}"
end end
videos = PG_DB.query_all("SELECT DISTINCT ON (ucid) * FROM #{view_name} WHERE \ videos = PG_DB.query_all("SELECT DISTINCT ON (ucid) * FROM #{view_name} WHERE NOT id = ANY (#{values}) ORDER BY ucid, published DESC", as: ChannelVideo)
NOT id = ANY (#{values}) \
ORDER BY ucid, published DESC", as: ChannelVideo)
else else
# Show latest video from each channel # Show latest video from each channel
videos = PG_DB.query_all("SELECT DISTINCT ON (ucid) * FROM #{view_name} \ videos = PG_DB.query_all("SELECT DISTINCT ON (ucid) * FROM #{view_name} ORDER BY ucid, published DESC", as: ChannelVideo)
ORDER BY ucid, published DESC", as: ChannelVideo)
end end
videos.sort_by! { |video| video.published }.reverse! videos.sort_by! { |video| video.published }.reverse!
@ -342,14 +338,11 @@ def get_subscription_feed(db, user, max_results = 40, page = 1)
else else
values = "VALUES #{user.watched.map { |id| %(('#{id}')) }.join(",")}" values = "VALUES #{user.watched.map { |id| %(('#{id}')) }.join(",")}"
end end
videos = PG_DB.query_all("SELECT * FROM #{view_name} WHERE \ videos = PG_DB.query_all("SELECT * FROM #{view_name} WHERE NOT id = ANY (#{values}) ORDER BY published DESC LIMIT $1 OFFSET $2", limit, offset, as: ChannelVideo)
NOT id = ANY (#{values}) \
ORDER BY published DESC LIMIT $1 OFFSET $2", limit, offset, as: ChannelVideo)
else else
# Sort subscriptions as normal # Sort subscriptions as normal
videos = PG_DB.query_all("SELECT * FROM #{view_name} \ videos = PG_DB.query_all("SELECT * FROM #{view_name} ORDER BY published DESC LIMIT $1 OFFSET $2", limit, offset, as: ChannelVideo)
ORDER BY published DESC LIMIT $1 OFFSET $2", limit, offset, as: ChannelVideo)
end end
end end
@ -366,16 +359,11 @@ def get_subscription_feed(db, user, max_results = 40, page = 1)
videos.sort_by! { |video| video.author }.reverse! videos.sort_by! { |video| video.author }.reverse!
end end
notifications = PG_DB.query_one("SELECT notifications FROM users WHERE email = $1", user.email, notifications = PG_DB.query_one("SELECT notifications FROM users WHERE email = $1", user.email, as: Array(String))
as: Array(String))
notifications = videos.select { |v| notifications.includes? v.id } notifications = videos.select { |v| notifications.includes? v.id }
videos = videos - notifications videos = videos - notifications
end end
if !limit
videos = videos[0..max_results]
end
return videos, notifications return videos, notifications
end end