1
0

Compare commits

...

2 Commits

View File

@@ -136,12 +136,9 @@ async def connect(vm_obj: dict):
} }
ws_url = vm_obj["ws_url"] ws_url = vm_obj["ws_url"]
log_directory = getattr(config, "log_directory", "./logs") log_directory = getattr(config, "log_directory", "./logs")
# Ensure the log directory exists # Create VM-specific log directory
os.makedirs(log_directory, exist_ok=True) vm_log_directory = os.path.join(log_directory, log_label)
log_file_path = os.path.join(log_directory, f"{log_label}.json") os.makedirs(vm_log_directory, exist_ok=True)
if not os.path.exists(log_file_path):
with open(log_file_path, "w") as log_file:
log_file.write("{}")
origin = Origin(vm_obj.get("origin_override", get_origin_from_ws_url(ws_url))) origin = Origin(vm_obj.get("origin_override", get_origin_from_ws_url(ws_url)))
@@ -240,37 +237,30 @@ async def connect(vm_obj: dict):
utc_day = utc_now.strftime("%Y-%m-%d") utc_day = utc_now.strftime("%Y-%m-%d")
timestamp = utc_now.isoformat() timestamp = utc_now.isoformat()
with open(log_file_path, "r+") as log_file: # Get daily log file path
try: daily_log_path = os.path.join(vm_log_directory, f"{utc_day}.json")
log_data = json.load(log_file)
except json.JSONDecodeError: # Load existing log data or create new
log_data = {} if os.path.exists(daily_log_path):
with open(daily_log_path, "r") as log_file:
try:
log_data = json.load(log_file)
except json.JSONDecodeError:
log_data = []
else:
log_data = []
if utc_day not in log_data: log_data.append(
log_data[utc_day] = [] {
"type": "chat",
# for i in range(0, len(backlog), 2): "timestamp": timestamp,
# backlog_user = backlog[i] "username": user,
# backlog_message = backlog[i + 1] "message": message,
# if not any(entry["message"] == backlog_message and entry["username"] == backlog_user for entry in log_data[utc_day]): }
# log.info(f"[{vm_name} - {backlog_user} (backlog)]: {backlog_message}") )
# log_data[utc_day].append({
# "timestamp": timestamp, with open(daily_log_path, "w") as log_file:
# "username": backlog_user,
# "message": backlog_message
# })
log_data[utc_day].append(
{
"type": "chat",
"timestamp": timestamp,
"username": user,
"message": message,
}
)
log_file.seek(0)
json.dump(log_data, log_file, indent=4) json.dump(log_data, log_file, indent=4)
log_file.truncate()
if config.commands["enabled"] and message.startswith( if config.commands["enabled"] and message.startswith(
config.commands["prefix"] config.commands["prefix"]
@@ -424,27 +414,30 @@ async def connect(vm_obj: dict):
utc_day = utc_now.strftime("%Y-%m-%d") utc_day = utc_now.strftime("%Y-%m-%d")
timestamp = utc_now.isoformat() timestamp = utc_now.isoformat()
with open(log_file_path, "r+") as log_file: # Get daily log file path
try: daily_log_path = os.path.join(vm_log_directory, f"{utc_day}.json")
log_data = json.load(log_file)
except json.JSONDecodeError: # Load existing log data or create new
log_data = {} if os.path.exists(daily_log_path):
with open(daily_log_path, "r") as log_file:
try:
log_data = json.load(log_file)
except json.JSONDecodeError:
log_data = []
else:
log_data = []
if utc_day not in log_data: log_data.append(
log_data[utc_day] = [] {
"type": "turn",
log_data[utc_day].append( "timestamp": timestamp,
{ "active_turn_user": current_turn,
"type": "turn", "queue": queue,
"timestamp": timestamp, }
"active_turn_user": current_turn, )
"queue": queue,
} with open(daily_log_path, "w") as log_file:
)
log_file.seek(0)
json.dump(log_data, log_file, indent=4) json.dump(log_data, log_file, indent=4)
log_file.truncate()
case ["remuser", count, *list]: case ["remuser", count, *list]:
for i in range(int(count)): for i in range(int(count)):