fixes ig
This commit is contained in:
parent
9bc0183800
commit
9666af1eb6
496
1
Normal file
496
1
Normal file
@ -0,0 +1,496 @@
|
||||
const std = @import("std");
|
||||
const json = std.json;
|
||||
const mem = std.mem;
|
||||
const http = std.http;
|
||||
const ws = @import("ws");
|
||||
const builtin = @import("builtin");
|
||||
const HttpClient = @import("tls12");
|
||||
const net = std.net;
|
||||
const crypto = std.crypto;
|
||||
const tls = std.crypto.tls;
|
||||
//const TlsClient = @import("tls12").TlsClient;
|
||||
//const Certificate = @import("tls12").Certificate;
|
||||
// todo use this to read compressed messages
|
||||
const zlib = @import("zlib");
|
||||
|
||||
const Opcode = enum(u4) {
|
||||
Dispatch = 0,
|
||||
Heartbeat = 1,
|
||||
Identify = 2,
|
||||
PresenceUpdate = 3,
|
||||
VoiceStateUpdate = 4,
|
||||
Resume = 6,
|
||||
Reconnect = 7,
|
||||
RequestGuildMember = 8,
|
||||
InvalidSession = 9,
|
||||
Hello = 10,
|
||||
HeartbeatACK = 11,
|
||||
};
|
||||
|
||||
const ShardSocketCloseCodes = enum(u16) {
|
||||
Shutdown = 3000,
|
||||
ZombiedConnection = 3010,
|
||||
};
|
||||
|
||||
const BASE_URL = "https://discord.com/api/v10";
|
||||
|
||||
pub const Intents = packed struct {
|
||||
guilds: bool = false,
|
||||
guild_members: bool = false,
|
||||
guild_bans: bool = false,
|
||||
guild_emojis: bool = false,
|
||||
guild_integrations: bool = false,
|
||||
guild_webhooks: bool = false,
|
||||
guild_invites: bool = false,
|
||||
guild_voice_states: bool = false,
|
||||
|
||||
guild_presences: bool = false,
|
||||
guild_messages: bool = false,
|
||||
guild_message_reactions: bool = false,
|
||||
guild_message_typing: bool = false,
|
||||
direct_messages: bool = false,
|
||||
direct_message_reactions: bool = false,
|
||||
direct_message_typing: bool = false,
|
||||
message_content: bool = false,
|
||||
|
||||
guild_scheduled_events: bool = false,
|
||||
_pad: u3 = 0,
|
||||
auto_moderation_configuration: bool = false,
|
||||
auto_moderation_execution: bool = false,
|
||||
_pad2: u2 = 0,
|
||||
|
||||
_pad3: u8 = 0,
|
||||
|
||||
pub fn toRaw(self: Intents) u32 {
|
||||
return @as(u32, @bitCast(self));
|
||||
}
|
||||
|
||||
pub fn fromRaw(raw: u32) Intents {
|
||||
return @as(Intents, @bitCast(raw));
|
||||
}
|
||||
|
||||
pub fn jsonStringify(self: Intents, options: std.json.StringifyOptions, writer: anytype) !void {
|
||||
_ = options;
|
||||
try writer.print("{}", .{self.toRaw()});
|
||||
}
|
||||
};
|
||||
|
||||
pub fn main() !void {
|
||||
const TOKEN = "Bot MTI5ODgzOTgzMDY3OTEzMDE4OA.GNojts.iyblGKK0xTWU57QCG5n3hr2Be1whyylTGr44P0";
|
||||
var gpa = std.heap.GeneralPurposeAllocator(.{}){};
|
||||
defer if (gpa.deinit() == .leak) {
|
||||
std.log.warn("Has leaked\n", .{});
|
||||
};
|
||||
const alloc = gpa.allocator();
|
||||
|
||||
var handler = try Handler.init(alloc, .{ .token = TOKEN, .intents = Intents.fromRaw(513) });
|
||||
//errdefer handler.deinit();
|
||||
|
||||
// try handler.readMessage();
|
||||
const thread = try std.Thread.spawn(.{}, Handler.readMessage, .{&handler});
|
||||
thread.detach();
|
||||
//try handler.identify();
|
||||
}
|
||||
|
||||
const HeartbeatHandler = struct {
|
||||
gateway: *Handler,
|
||||
heartbeatInterval: u64,
|
||||
lastHeartbeatAck: bool,
|
||||
/// useful for calculating ping
|
||||
lastBeat: u64 = 0,
|
||||
mutex: std.Thread.Mutex = .{},
|
||||
thread: ?std.Thread,
|
||||
|
||||
pub fn init(gateway: *Handler, interval: u64) !HeartbeatHandler {
|
||||
var gateway_mut = gateway.*;
|
||||
return .{
|
||||
.gateway = &gateway_mut,
|
||||
.heartbeatInterval = interval,
|
||||
.lastHeartbeatAck = false,
|
||||
.lastBeat = 0,
|
||||
.thread = null,
|
||||
};
|
||||
}
|
||||
|
||||
pub fn deinit(self: *HeartbeatHandler) void {
|
||||
_ = self;
|
||||
}
|
||||
|
||||
pub fn run(self: *HeartbeatHandler) !void {
|
||||
while (true) {
|
||||
std.time.sleep(self.heartbeatInterval);
|
||||
try self.gateway.heartbeat(false);
|
||||
}
|
||||
}
|
||||
|
||||
pub fn loop(self: *HeartbeatHandler) !void {
|
||||
//_ = self;
|
||||
std.debug.print("start loop\n", .{});
|
||||
var self_mut = self.*;
|
||||
self.thread = std.Thread.spawn(.{}, HeartbeatHandler.run, .{&self_mut}) catch null;
|
||||
}
|
||||
|
||||
pub fn readAck(self: *HeartbeatHandler) bool {
|
||||
self.mutex.lock();
|
||||
defer self.mutex.lock();
|
||||
|
||||
return self.lastHeartbeatAck;
|
||||
}
|
||||
|
||||
pub fn updateAck(self: *HeartbeatHandler, ack: bool) void {
|
||||
self.mutex.lock();
|
||||
defer self.mutex.lock();
|
||||
|
||||
self.lastHeartbeatAck = ack;
|
||||
}
|
||||
};
|
||||
|
||||
const FetchReq = struct {
|
||||
allocator: mem.Allocator,
|
||||
token: []const u8,
|
||||
client: HttpClient,
|
||||
body: std.ArrayList(u8),
|
||||
|
||||
pub fn init(allocator: mem.Allocator, token: []const u8) FetchReq {
|
||||
const client = HttpClient{ .allocator = allocator };
|
||||
return FetchReq{
|
||||
.allocator = allocator,
|
||||
.client = client,
|
||||
.body = std.ArrayList(u8).init(allocator),
|
||||
.token = token,
|
||||
};
|
||||
}
|
||||
|
||||
pub fn deinit(self: *FetchReq) void {
|
||||
self.client.deinit();
|
||||
self.body.deinit();
|
||||
}
|
||||
|
||||
pub fn makeRequest(self: *FetchReq, method: http.Method, path: []const u8, body: ?[]const u8) !HttpClient.FetchResult {
|
||||
var fetch_options = HttpClient.FetchOptions{
|
||||
.location = HttpClient.FetchOptions.Location{
|
||||
.url = path,
|
||||
},
|
||||
.extra_headers = &[_]http.Header{
|
||||
http.Header{ .name = "Accept", .value = "application/json" },
|
||||
http.Header{ .name = "Content-Type", .value = "application/json" },
|
||||
http.Header{ .name = "Authorization", .value = self.token },
|
||||
},
|
||||
.method = method,
|
||||
.response_storage = .{ .dynamic = &self.body },
|
||||
};
|
||||
|
||||
if (body != null) {
|
||||
fetch_options.payload = body;
|
||||
}
|
||||
|
||||
const res = try self.client.fetch(fetch_options);
|
||||
return res;
|
||||
}
|
||||
};
|
||||
|
||||
pub const Handler = struct {
|
||||
///
|
||||
/// https://discord.com/developers/docs/topics/gateway#get-gateway
|
||||
///
|
||||
const GatewayInfo = struct {
|
||||
/// The WSS URL that can be used for connecting to the gateway
|
||||
url: []const u8,
|
||||
};
|
||||
|
||||
///
|
||||
/// https://discord.com/developers/docs/events/gateway#session-start-limit-object
|
||||
///
|
||||
const GatewaySessionStartLimit = struct {
|
||||
/// Total number of session starts the current user is allowed
|
||||
total: u32,
|
||||
/// Remaining number of session starts the current user is allowed
|
||||
remaining: u32,
|
||||
/// Number of milliseconds after which the limit resets
|
||||
reset_after: u32,
|
||||
/// Number of identify requests allowed per 5 seconds
|
||||
max_concurrency: u32,
|
||||
};
|
||||
|
||||
///
|
||||
/// https://discord.com/developers/docs/topics/gateway#get-gateway-bot
|
||||
///
|
||||
const GatewayBotInfo = struct {
|
||||
url: []const u8,
|
||||
///
|
||||
/// The recommended number of shards to use when connecting
|
||||
///
|
||||
/// See https://discord.com/developers/docs/topics/gateway#sharding
|
||||
///
|
||||
shards: u32,
|
||||
///
|
||||
/// Information on the current session start limit
|
||||
///
|
||||
/// See https://discord.com/developers/docs/topics/gateway#session-start-limit-object
|
||||
///
|
||||
session_start_limit: ?GatewaySessionStartLimit,
|
||||
};
|
||||
|
||||
const IdentifyProperties = struct {
|
||||
///
|
||||
/// Operating system the shard runs on.
|
||||
///
|
||||
os: []const u8,
|
||||
///
|
||||
/// The "browser" where this shard is running on.
|
||||
///
|
||||
browser: []const u8,
|
||||
///
|
||||
/// The device on which the shard is running.
|
||||
///
|
||||
device: []const u8,
|
||||
};
|
||||
|
||||
const _default_properties = IdentifyProperties{
|
||||
.os = @tagName(builtin.os.tag),
|
||||
.browser = "seyfert",
|
||||
.device = "seyfert",
|
||||
};
|
||||
|
||||
client: ws.Client,
|
||||
token: []const u8,
|
||||
intents: Intents,
|
||||
session_id: ?[]const u8,
|
||||
sequence: ?u16,
|
||||
heartbeater: ?HeartbeatHandler,
|
||||
allocator: mem.Allocator,
|
||||
resume_gateway_url: ?[]const u8 = null,
|
||||
info: GatewayBotInfo,
|
||||
|
||||
pub fn resume_conn(self: *Handler, out: anytype) !void {
|
||||
const data = .{ .op = @intFromEnum(Opcode.Resume), .d = .{
|
||||
.token = self.token,
|
||||
.session_id = self.session_id,
|
||||
.seq = self.sequence,
|
||||
} };
|
||||
|
||||
try json.stringify(data, .{}, out);
|
||||
|
||||
try self.client.write(&out);
|
||||
}
|
||||
|
||||
inline fn gateway_url(self: ?*Handler) []const u8 {
|
||||
// wtf is this?
|
||||
if (self) |s| {
|
||||
return s.resume_gateway_url orelse s.info.url;
|
||||
}
|
||||
|
||||
return "wss://gateway.discord.gg";
|
||||
}
|
||||
|
||||
// identifies in order to connect to Discord and get the online status, this shall be done on hello perhaps
|
||||
fn identify(self: *Handler) !void {
|
||||
std.debug.print("identifying now...\n", .{});
|
||||
const data = .{
|
||||
.op = @intFromEnum(Opcode.Identify),
|
||||
.d = .{
|
||||
//.compress = false,
|
||||
.intents = self.intents.toRaw(),
|
||||
.properties = Handler._default_properties,
|
||||
.token = self.token,
|
||||
},
|
||||
};
|
||||
|
||||
var buf: [1000]u8 = undefined;
|
||||
var fba = std.heap.FixedBufferAllocator.init(&buf);
|
||||
var string = std.ArrayList(u8).init(fba.allocator());
|
||||
try std.json.stringify(data, .{}, string.writer());
|
||||
|
||||
std.debug.print("{s}\n", .{string.items});
|
||||
// try posting our shitty data
|
||||
try self.client.write(string.items);
|
||||
}
|
||||
|
||||
// asks /gateway/bot initializes both the ws client and the http client
|
||||
pub fn init(allocator: mem.Allocator, args: struct { token: []const u8, intents: Intents }) !Handler {
|
||||
var req = FetchReq.init(allocator, args.token);
|
||||
defer req.deinit();
|
||||
|
||||
const res = try req.makeRequest(.GET, BASE_URL ++ "/gateway/bot", null);
|
||||
const body = try req.body.toOwnedSlice();
|
||||
defer allocator.free(body);
|
||||
|
||||
// check status idk
|
||||
if (res.status != http.Status.ok) {
|
||||
@panic("we are cooked");
|
||||
}
|
||||
|
||||
const parsed = try json.parseFromSlice(GatewayBotInfo, allocator, body, .{});
|
||||
defer parsed.deinit();
|
||||
|
||||
return .{
|
||||
.allocator = allocator,
|
||||
.token = args.token,
|
||||
.intents = args.intents,
|
||||
// maybe there is a better way to do this
|
||||
.client = try Handler._connect_ws(allocator, parsed.value.url["wss://".len..]),
|
||||
.session_id = undefined,
|
||||
.sequence = undefined,
|
||||
.heartbeater = null,
|
||||
.info = parsed.value,
|
||||
};
|
||||
}
|
||||
|
||||
inline fn _connect_ws(allocator: mem.Allocator, url: []const u8) !ws.Client {
|
||||
var conn = try ws.Client.init(allocator, .{
|
||||
.tls = true, // important: zig.http doesn't support this, type shit
|
||||
.port = 443,
|
||||
.host = url,
|
||||
//.ca_bundle = @import("tls12").Certificate.Bundle{},
|
||||
});
|
||||
|
||||
try conn.handshake("/?v=10&encoding=json", .{
|
||||
.timeout_ms = 1000,
|
||||
.headers = "host: gateway.discord.gg",
|
||||
});
|
||||
|
||||
return conn;
|
||||
}
|
||||
|
||||
pub fn deinit(self: *Handler) void {
|
||||
defer self.client.deinit();
|
||||
if (self.heartbeater) |*hb| {
|
||||
hb.deinit();
|
||||
}
|
||||
}
|
||||
|
||||
// listens for messages
|
||||
pub fn readMessage(self: *Handler) !void {
|
||||
try self.client.readTimeout(std.time.ms_per_s * 1);
|
||||
while (true) {
|
||||
const msg = (try self.client.read()) orelse {
|
||||
// no message after our 1 second
|
||||
std.debug.print(".", .{});
|
||||
continue;
|
||||
};
|
||||
|
||||
// must be called once you're done processing the request
|
||||
defer self.client.done(msg);
|
||||
|
||||
std.debug.print("received: {s}\n", .{msg.data});
|
||||
|
||||
const DiscordData = struct {
|
||||
s: ?u16, //well figure it out
|
||||
op: Opcode,
|
||||
d: json.Value, // needs parsing
|
||||
t: ?[]const u8,
|
||||
};
|
||||
|
||||
const raw = try json.parseFromSlice(DiscordData, self.allocator, msg.data, .{});
|
||||
defer raw.deinit();
|
||||
|
||||
const payload = raw.value;
|
||||
|
||||
if (payload.op == Opcode.Dispatch) {
|
||||
self.sequence = @as(?u16, payload.s);
|
||||
}
|
||||
|
||||
switch (payload.op) {
|
||||
Opcode.Dispatch => {},
|
||||
Opcode.Hello => {
|
||||
const HelloPayload = struct { heartbeat_interval: u32, _trace: [][]const u8 };
|
||||
const parsed = try json.parseFromValue(HelloPayload, self.allocator, payload.d, .{});
|
||||
const helloPayload = parsed.value;
|
||||
|
||||
// PARSE NEW URL IN READY
|
||||
|
||||
if (self.heartbeater == null) {
|
||||
var self_mut = self.*;
|
||||
|
||||
self.heartbeater = try HeartbeatHandler.init(
|
||||
// we cooking
|
||||
&self_mut, helloPayload.heartbeat_interval);
|
||||
}
|
||||
|
||||
var heartbeater = self.heartbeater.?;
|
||||
|
||||
heartbeater.heartbeatInterval = helloPayload.heartbeat_interval;
|
||||
|
||||
try self.heartbeat(false);
|
||||
try heartbeater.loop();
|
||||
try self.identify();
|
||||
}, // heartbeat_interval
|
||||
Opcode.HeartbeatACK => {
|
||||
if (self.heartbeater) |*hb| {
|
||||
hb.lastHeartbeatAck = true;
|
||||
}
|
||||
}, // keep this shit alive otherwise kill it
|
||||
Opcode.Reconnect => {},
|
||||
Opcode.Resume => {
|
||||
const WithSequence = struct {
|
||||
token: []const u8,
|
||||
session_id: []const u8,
|
||||
seq: ?u16,
|
||||
};
|
||||
const parsed = try json.parseFromValue(WithSequence, self.allocator, payload.d, .{});
|
||||
const payload_new = parsed.value;
|
||||
|
||||
self.sequence = @as(?u16, payload_new.seq);
|
||||
self.session_id = payload_new.session_id;
|
||||
},
|
||||
else => {
|
||||
std.debug.print("Unhandled {} -- {s}", .{ payload.op, "none" });
|
||||
},
|
||||
}
|
||||
//try client.write(message.data);
|
||||
}
|
||||
}
|
||||
|
||||
pub fn heartbeat(self: *Handler, requested: bool) !void {
|
||||
var heartbeater = self.heartbeater.?;
|
||||
//std.time.sleep(heartbeater.heartbeatInterval);
|
||||
|
||||
if (!requested) {
|
||||
if (!heartbeater.readAck()) {
|
||||
try self.close(ShardSocketCloseCodes.ZombiedConnection, "Zombied connection");
|
||||
heartbeater.deinit();
|
||||
return;
|
||||
}
|
||||
heartbeater.updateAck(false);
|
||||
}
|
||||
const data = .{ .op = @intFromEnum(Opcode.Heartbeat), .d = self.sequence };
|
||||
|
||||
heartbeater.mutex.lock();
|
||||
defer heartbeater.mutex.unlock();
|
||||
|
||||
var buf: [1000]u8 = undefined;
|
||||
var fba = std.heap.FixedBufferAllocator.init(&buf);
|
||||
var string = std.ArrayList(u8).init(fba.allocator());
|
||||
try std.json.stringify(data, .{}, string.writer());
|
||||
|
||||
// try posting our shitty data
|
||||
std.debug.print("sending heartbeat rn\n", .{});
|
||||
try self.client.write(string.items);
|
||||
}
|
||||
|
||||
pub inline fn reconnect(self: *Handler) !void {
|
||||
_ = self;
|
||||
//try self.disconnect();
|
||||
//_ = try self.connect();
|
||||
}
|
||||
|
||||
pub fn connect(self: *Handler) !Handler {
|
||||
std.time.sleep(std.time.ms_per_s * 5);
|
||||
self.client = try Handler._connect_ws(self.allocator, self.gateway_url());
|
||||
|
||||
return self.*;
|
||||
}
|
||||
|
||||
pub fn disconnect(self: *Handler) !void {
|
||||
try self.close(ShardSocketCloseCodes.Shutdown, "Shard down request");
|
||||
}
|
||||
|
||||
pub fn close(self: *Handler, code: ShardSocketCloseCodes, reason: []const u8) !void {
|
||||
std.debug.print("cooked closing ws conn...\n", .{});
|
||||
// Implement reconnection logic here
|
||||
try self.client.close(.{
|
||||
.code = @intFromEnum(code), //u16
|
||||
.reason = reason, //[]const u8
|
||||
});
|
||||
}
|
||||
};
|
179
src/main.zig
179
src/main.zig
@ -12,6 +12,7 @@ const tls = std.crypto.tls;
|
||||
//const Certificate = @import("tls12").Certificate;
|
||||
// todo use this to read compressed messages
|
||||
const zlib = @import("zlib");
|
||||
const Scheduler = @import("scheduler.zig").Scheduler;
|
||||
|
||||
const Opcode = enum(u4) {
|
||||
Dispatch = 0,
|
||||
@ -84,48 +85,11 @@ pub fn main() !void {
|
||||
const alloc = gpa.allocator();
|
||||
|
||||
var handler = try Handler.init(alloc, .{ .token = TOKEN, .intents = Intents.fromRaw(513) });
|
||||
//errdefer handler.deinit();
|
||||
errdefer handler.deinit();
|
||||
|
||||
try handler.readMessage();
|
||||
//try handler.identify();
|
||||
}
|
||||
|
||||
const HeartbeatHandler = struct {
|
||||
gateway: *Handler,
|
||||
heartbeatInterval: u64,
|
||||
lastHeartbeatAck: bool,
|
||||
/// useful for calculating ping
|
||||
lastBeat: u64 = 0,
|
||||
|
||||
pub fn init(gateway: *Handler, interval: u64) !HeartbeatHandler {
|
||||
var gateway_mut = gateway.*;
|
||||
return .{
|
||||
.gateway = &gateway_mut,
|
||||
.heartbeatInterval = interval,
|
||||
.lastHeartbeatAck = false,
|
||||
.lastBeat = 0,
|
||||
};
|
||||
}
|
||||
|
||||
pub fn deinit(self: *HeartbeatHandler) void {
|
||||
_ = self;
|
||||
}
|
||||
|
||||
pub fn run(self: *HeartbeatHandler) !void {
|
||||
while (true) {
|
||||
std.time.sleep(self.heartbeatInterval);
|
||||
try self.gateway.heartbeat(false);
|
||||
}
|
||||
}
|
||||
|
||||
pub fn loop(self: *HeartbeatHandler) !void {
|
||||
_ = self;
|
||||
std.debug.print("start loop\n", .{});
|
||||
//var self_mut = self.*;
|
||||
//self.thread = try std.Thread.spawn(.{}, HeartbeatHandler.run, .{&self_mut});
|
||||
}
|
||||
};
|
||||
|
||||
const FetchReq = struct {
|
||||
allocator: mem.Allocator,
|
||||
token: []const u8,
|
||||
@ -233,26 +197,79 @@ pub const Handler = struct {
|
||||
.device = "seyfert",
|
||||
};
|
||||
|
||||
const Heart = struct {
|
||||
heartbeatInterval: u32,
|
||||
ack: bool,
|
||||
/// useful for calculating ping
|
||||
lastBeat: u64,
|
||||
};
|
||||
|
||||
const HeartTask = union(enum) {
|
||||
heartbeat,
|
||||
ack,
|
||||
|
||||
pub fn run(task: HeartTask, ctx: *Handler, next_at: i64) void {
|
||||
std.debug.print("task successful\n", .{});
|
||||
_ = next_at;
|
||||
switch (task) {
|
||||
.heartbeat => {
|
||||
var ctx_mut = ctx.*;
|
||||
HeartTask.sendhb(&ctx_mut) catch {
|
||||
@panic("we are fucking cooked didn't send heartbeat fuckfuckfuck");
|
||||
};
|
||||
},
|
||||
.ack => {
|
||||
var ctx_mut = ctx.*;
|
||||
ctx.heart.ack = true;
|
||||
HeartTask.sendhb(&ctx_mut) catch {
|
||||
@panic("we are fucking cooked didn't send heartbeat fuckfuckfuck");
|
||||
};
|
||||
std.debug.print("<< ♥\n", .{});
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
fn sendhb(ctx: *Handler) !void {
|
||||
std.debug.print(">> ♥\n", .{});
|
||||
try ctx.heartbeat(false);
|
||||
}
|
||||
};
|
||||
|
||||
client: ws.Client,
|
||||
token: []const u8,
|
||||
intents: Intents,
|
||||
session_id: ?[]const u8,
|
||||
sequence: ?u16,
|
||||
heartbeater: ?HeartbeatHandler,
|
||||
heart: Heart = .{ .heartbeatInterval = 45000, .ack = false, .lastBeat = 0 },
|
||||
allocator: mem.Allocator,
|
||||
resume_gateway_url: ?[]const u8 = null,
|
||||
info: GatewayBotInfo,
|
||||
heart_task: Scheduler(HeartTask, *Handler),
|
||||
mutex: std.Thread.Mutex = .{},
|
||||
|
||||
pub fn resume_conn(self: *Handler, out: anytype) !void {
|
||||
inline fn jitter() i1 {
|
||||
return 0;
|
||||
}
|
||||
|
||||
pub inline fn resumable(self: *Handler) bool {
|
||||
return self.resume_gateway_url != null and
|
||||
self.session_id != null and
|
||||
self.sequence != null;
|
||||
}
|
||||
|
||||
pub fn resume_(self: *Handler) !void {
|
||||
const data = .{ .op = @intFromEnum(Opcode.Resume), .d = .{
|
||||
.token = self.token,
|
||||
.session_id = self.session_id,
|
||||
.seq = self.sequence,
|
||||
} };
|
||||
|
||||
try json.stringify(data, .{}, out);
|
||||
var buf: [1000]u8 = undefined;
|
||||
var fba = std.heap.FixedBufferAllocator.init(&buf);
|
||||
var string = std.ArrayList(u8).init(fba.allocator());
|
||||
try std.json.stringify(data, .{}, string.writer());
|
||||
|
||||
try self.client.write(&out);
|
||||
try self.client.write(string.items);
|
||||
}
|
||||
|
||||
inline fn gateway_url(self: ?*Handler) []const u8 {
|
||||
@ -312,8 +329,9 @@ pub const Handler = struct {
|
||||
.client = try Handler._connect_ws(allocator, parsed.value.url["wss://".len..]),
|
||||
.session_id = undefined,
|
||||
.sequence = undefined,
|
||||
.heartbeater = null,
|
||||
.heart = undefined,
|
||||
.info = parsed.value,
|
||||
.heart_task = Scheduler(HeartTask, *Handler).init(allocator),
|
||||
};
|
||||
}
|
||||
|
||||
@ -334,10 +352,9 @@ pub const Handler = struct {
|
||||
}
|
||||
|
||||
pub fn deinit(self: *Handler) void {
|
||||
defer self.heart_task.deinit();
|
||||
defer self.client.deinit();
|
||||
if (self.heartbeater) |*hb| {
|
||||
hb.deinit();
|
||||
}
|
||||
@panic("killing the whole bot");
|
||||
}
|
||||
|
||||
// listens for messages
|
||||
@ -349,7 +366,6 @@ pub const Handler = struct {
|
||||
std.debug.print(".", .{});
|
||||
continue;
|
||||
};
|
||||
|
||||
// must be called once you're done processing the request
|
||||
defer self.client.done(msg);
|
||||
|
||||
@ -362,7 +378,10 @@ pub const Handler = struct {
|
||||
t: ?[]const u8,
|
||||
};
|
||||
|
||||
const raw = try json.parseFromSlice(DiscordData, self.allocator, msg.data, .{});
|
||||
const raw = json.parseFromSlice(DiscordData, self.allocator, msg.data, .{}) catch {
|
||||
std.debug.print("unexpected payload, continuing: {s}\n", .{msg.data});
|
||||
continue;
|
||||
};
|
||||
defer raw.deinit();
|
||||
|
||||
const payload = raw.value;
|
||||
@ -376,32 +395,43 @@ pub const Handler = struct {
|
||||
Opcode.Hello => {
|
||||
const HelloPayload = struct { heartbeat_interval: u32, _trace: [][]const u8 };
|
||||
const parsed = try json.parseFromValue(HelloPayload, self.allocator, payload.d, .{});
|
||||
defer parsed.deinit();
|
||||
const helloPayload = parsed.value;
|
||||
|
||||
// PARSE NEW URL IN READY
|
||||
|
||||
if (self.heartbeater == null) {
|
||||
var self_mut = self.*;
|
||||
|
||||
self.heartbeater = try HeartbeatHandler.init(
|
||||
// we cooking
|
||||
&self_mut, helloPayload.heartbeat_interval);
|
||||
}
|
||||
|
||||
var heartbeater = self.heartbeater.?;
|
||||
|
||||
heartbeater.heartbeatInterval = helloPayload.heartbeat_interval;
|
||||
self.heart = Heart{
|
||||
// TODO: fix bug
|
||||
.heartbeatInterval = helloPayload.heartbeat_interval,
|
||||
.ack = false,
|
||||
.lastBeat = 0,
|
||||
};
|
||||
|
||||
try self.heartbeat(false);
|
||||
try heartbeater.loop();
|
||||
try self.identify();
|
||||
}, // heartbeat_interval
|
||||
Opcode.HeartbeatACK => {
|
||||
if (self.heartbeater) |*hb| {
|
||||
hb.lastHeartbeatAck = true;
|
||||
std.debug.print("starting heart beater. seconds:{d}...\n", .{self.heart.heartbeatInterval});
|
||||
var self_mut = self.*;
|
||||
try self.heart_task.start(&self_mut);
|
||||
errdefer self.heart_task.deinit();
|
||||
try self.heart_task.scheduleIn(.heartbeat, @as(i64, @intCast(self.heart.heartbeatInterval)));
|
||||
if (self.resumable()) {
|
||||
try self.resume_();
|
||||
return;
|
||||
} else {
|
||||
try self.identify();
|
||||
}
|
||||
},
|
||||
Opcode.HeartbeatACK => {
|
||||
// perhaps lock the thread w a mutex?
|
||||
self.heart.ack = true;
|
||||
std.debug.print("got heartbeat\n", .{});
|
||||
}, // keep this shit alive otherwise kill it
|
||||
Opcode.Reconnect => {},
|
||||
Opcode.Heartbeat => {
|
||||
try self.heartbeat(true);
|
||||
},
|
||||
Opcode.Reconnect => {
|
||||
std.debug.print("reconnecting\n", .{});
|
||||
try self.reconnect();
|
||||
},
|
||||
Opcode.Resume => {
|
||||
const WithSequence = struct {
|
||||
token: []const u8,
|
||||
@ -414,25 +444,23 @@ pub const Handler = struct {
|
||||
self.sequence = @as(?u16, payload_new.seq);
|
||||
self.session_id = payload_new.session_id;
|
||||
},
|
||||
Opcode.InvalidSession => {},
|
||||
else => {
|
||||
std.debug.print("Unhandled {} -- {s}", .{ payload.op, "none" });
|
||||
},
|
||||
}
|
||||
//try client.write(message.data);
|
||||
}
|
||||
}
|
||||
|
||||
pub fn heartbeat(self: *Handler, requested: bool) !void {
|
||||
var heartbeater = self.heartbeater.?;
|
||||
//std.time.sleep(heartbeater.heartbeatInterval);
|
||||
|
||||
if (!requested) {
|
||||
if (!heartbeater.lastHeartbeatAck) {
|
||||
if (!self.heart.ack) {
|
||||
//try self.close(ShardSocketCloseCodes.ZombiedConnection, "Zombied connection");
|
||||
heartbeater.deinit();
|
||||
return;
|
||||
}
|
||||
heartbeater.lastHeartbeatAck = false;
|
||||
self.heart.ack = false;
|
||||
}
|
||||
const data = .{ .op = @intFromEnum(Opcode.Heartbeat), .d = self.sequence };
|
||||
|
||||
@ -447,13 +475,12 @@ pub const Handler = struct {
|
||||
}
|
||||
|
||||
pub inline fn reconnect(self: *Handler) !void {
|
||||
_ = self;
|
||||
//try self.disconnect();
|
||||
//_ = try self.connect();
|
||||
try self.disconnect();
|
||||
_ = try self.connect();
|
||||
}
|
||||
|
||||
pub fn connect(self: *Handler) !Handler {
|
||||
std.time.sleep(std.time.ms_per_s * 5);
|
||||
//std.time.sleep(std.time.ms_per_s * 5);
|
||||
self.client = try Handler._connect_ws(self.allocator, self.gateway_url());
|
||||
|
||||
return self.*;
|
||||
@ -464,6 +491,8 @@ pub const Handler = struct {
|
||||
}
|
||||
|
||||
pub fn close(self: *Handler, code: ShardSocketCloseCodes, reason: []const u8) !void {
|
||||
self.mutex.lock();
|
||||
defer self.mutex.unlock();
|
||||
std.debug.print("cooked closing ws conn...\n", .{});
|
||||
// Implement reconnection logic here
|
||||
try self.client.close(.{
|
||||
|
156
src/scheduler.zig
Normal file
156
src/scheduler.zig
Normal file
@ -0,0 +1,156 @@
|
||||
const std = @import("std");
|
||||
|
||||
const Thread = std.Thread;
|
||||
const Allocator = std.mem.Allocator;
|
||||
|
||||
fn Job(comptime T: type) type {
|
||||
return struct {
|
||||
at: i64,
|
||||
task: T,
|
||||
};
|
||||
}
|
||||
|
||||
pub fn Scheduler(comptime T: type, comptime C: type) type {
|
||||
return struct {
|
||||
queue: Q,
|
||||
running: bool,
|
||||
mutex: Thread.Mutex,
|
||||
cond: Thread.Condition,
|
||||
thread: ?Thread,
|
||||
|
||||
const Q = std.PriorityQueue(Job(T), void, compare);
|
||||
|
||||
fn compare(_: void, a: Job(T), b: Job(T)) std.math.Order {
|
||||
return std.math.order(a.at, b.at);
|
||||
}
|
||||
|
||||
const Self = @This();
|
||||
|
||||
pub fn init(allocator: Allocator) Self {
|
||||
return .{
|
||||
.cond = .{},
|
||||
.mutex = .{},
|
||||
.thread = null,
|
||||
.running = false,
|
||||
.queue = Q.init(allocator, {}),
|
||||
};
|
||||
}
|
||||
|
||||
pub fn deinit(self: *Self) void {
|
||||
self.stop();
|
||||
self.queue.deinit();
|
||||
}
|
||||
|
||||
pub fn start(self: *Self, ctx: C) !void {
|
||||
{
|
||||
self.mutex.lock();
|
||||
defer self.mutex.unlock();
|
||||
if (self.running == true) {
|
||||
return error.AlreadyRunning;
|
||||
}
|
||||
self.running = true;
|
||||
}
|
||||
self.thread = try Thread.spawn(.{}, Self.run, .{ self, ctx });
|
||||
}
|
||||
|
||||
pub fn stop(self: *Self) void {
|
||||
{
|
||||
self.mutex.lock();
|
||||
defer self.mutex.unlock();
|
||||
if (self.running == false) {
|
||||
return;
|
||||
}
|
||||
self.running = false;
|
||||
}
|
||||
|
||||
self.cond.signal();
|
||||
self.thread.?.join();
|
||||
}
|
||||
|
||||
pub fn scheduleIn(self: *Self, task: T, ms: i64) !void {
|
||||
return self.schedule(task, std.time.milliTimestamp() + ms);
|
||||
}
|
||||
|
||||
pub fn schedule(self: *Self, task: T, at: i64) !void {
|
||||
const job: Job(T) = .{
|
||||
.at = at,
|
||||
.task = task,
|
||||
};
|
||||
|
||||
var reschedule = false;
|
||||
{
|
||||
self.mutex.lock();
|
||||
defer self.mutex.unlock();
|
||||
|
||||
if (self.queue.peek()) |*next| {
|
||||
if (at < next.at) {
|
||||
reschedule = true;
|
||||
}
|
||||
} else {
|
||||
reschedule = true;
|
||||
}
|
||||
try self.queue.add(job);
|
||||
}
|
||||
|
||||
if (reschedule) {
|
||||
// Our new job is scheduled before our previous earlier job
|
||||
// (or we had no previous jobs)
|
||||
// We need to reset our schedule
|
||||
self.cond.signal();
|
||||
}
|
||||
}
|
||||
|
||||
// this is running in a separate thread, started by start()
|
||||
fn run(self: *Self, ctx: C) void {
|
||||
self.mutex.lock();
|
||||
|
||||
while (true) {
|
||||
const ms_until_next = self.processPending(ctx);
|
||||
|
||||
// mutex is locked when returning for processPending
|
||||
|
||||
if (self.running == false) {
|
||||
self.mutex.unlock();
|
||||
return;
|
||||
}
|
||||
|
||||
if (ms_until_next) |timeout| {
|
||||
const ns = @as(u64, @intCast(timeout * std.time.ns_per_ms));
|
||||
self.cond.timedWait(&self.mutex, ns) catch |err| {
|
||||
std.debug.assert(err == error.Timeout);
|
||||
// on success or error, cond locks mutex, which is what we want
|
||||
};
|
||||
} else {
|
||||
self.cond.wait(&self.mutex);
|
||||
}
|
||||
// if we woke up, it's because a new job was added with a more recent
|
||||
// scheduled time. This new job MAY not be ready to run yet, and
|
||||
// it's even possible for our cond variable to wake up randomly (as per
|
||||
// the docs), but processPending is defensive and will check this for us.
|
||||
}
|
||||
}
|
||||
|
||||
// we enter this function with mutex locked
|
||||
// and we exit this function with the mutex locked
|
||||
// importantly, we don't lock the mutex will process the task
|
||||
fn processPending(self: *Self, ctx: C) ?i64 {
|
||||
while (true) {
|
||||
const next = self.queue.peek() orelse {
|
||||
// yes, we must return this function with a locked mutex
|
||||
return null;
|
||||
};
|
||||
const seconds_until_next = next.at - std.time.milliTimestamp();
|
||||
if (seconds_until_next > 0) {
|
||||
// this job isn't ready, yes, the mutex should remain locked!
|
||||
return seconds_until_next;
|
||||
}
|
||||
|
||||
// delete the peeked job from the queue, because we're going to process it
|
||||
const job = self.queue.remove();
|
||||
self.mutex.unlock();
|
||||
defer self.mutex.lock();
|
||||
job.task.run(ctx, next.at);
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
Loading…
x
Reference in New Issue
Block a user