Skip Navigation
InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)LA
Posts
17
Comments
5
Joined
2 yr. ago
Announcements (lazysoci.al) @lazysoci.al
lazyadmin @lazysoci.al

Alternative Lemmy Frontends; Other changes

cross-posted from: https://lazysoci.al/post/24004627

Along with the upgrade to 0.19.10 today I have made the following additional changes

Lazy Meta @lazysoci.al
lazyadmin @lazysoci.al

Alternative Lemmy Frontends; Other changes

Along with the upgrade to 0.19.10 today I have made the following additional changes

Announcements (lazysoci.al) @lazysoci.al
lazyadmin @lazysoci.al

lazysoci.al Maintenance 2025-04-02

When this post is 1 hour old, I shall be updating lazysoci.al to Lemmy verison 0.9.10

Downtime is expected to be ~10 minutes or less

Announcements (lazysoci.al) @lazysoci.al
lazyadmin @lazysoci.al

lazysoci.al Update 2025-03-1

This evening I have updated Lemmy from 0.19.8 to 0.19.9

There has been an increasing number of new users. Many of these have turned out to be dishonest trolls and subsequently banned. One of the traits of the internet I guess.

For those that remain, warm welcome :)

Announcements (lazysoci.al) @lazysoci.al
lazyadmin @lazysoci.al

lazysoci.al Maintenance 2025-01-20

I shall be moving the pict-rs data from local filesystem to local S3 this evening.

The site shall be offline for an extended period during this migration. There is ~450G in just over 3 million files currently hosted on the pict-rs instance.

Lemmy and all its containers are on an NVMe backed zfs pool. With pict-rs fast approaching 500G I've been exploring alternatives. So I introduced an S3 service using spinning rust.

I expect the migration to take several hours so will start this just before I head off for bed between 2300-0000 GMT.

Update 2025-01-21 03:09 GMT

Downtime 3h 17m -- Migration completed. Images seem to be working, both new and old. Site is catching up on missed posts.

Announcements (lazysoci.al) @lazysoci.al
lazyadmin @lazysoci.al

lazysoci.al Upcoming Maintenance 2025-01-02

The new server is prepped and ready to take to the datacentre. I've created a vxlan (mikrotik eoip) to the office so have been transferring services to the new server when I had time over the last few days.

I just performed the transfer of Lemmy to the new server which was an unannounced 2h30m downtime (sorry!). It was a good backup restore test though.

To complete the project, I shall be visiting the datacentre on Thursday, 2025-01-02 to swap the old server for the new. lazysoci.al will be offline on 2025-01-02 from 0830 to 1200 GMT

Apologies for the inconvenience.

Update: server swap completed without issue.

lazysoci.al Update 2024-12-26

  • Site is back online. pict-rs performs its purge/delete requests as part of a job queue which will increase load on the server for a while.

     undefined
        
    pictrs=# select count(*) from job_queue;
     count
    --------
     957546
    (1 row)
    
      

    A few jobs for it to complete :)

  • Announcements (lazysoci.al) @lazysoci.al
    lazyadmin @lazysoci.al

    lazysoci.al Update 2024-12-26

    When this post is 2 hours old (2100 UTC) I shall be performing maintenance on the pict-rs container and updating lazysoci.al to Lemmy 0.19.8.

    During the maintenance period, the site shall be offline.

    The Lemmy update should be straight forward, however I also need to purge old cached images in pict-rs which could take a few hours.

    Edit: Site back online

    Announcements (lazysoci.al) @lazysoci.al
    lazyadmin @lazysoci.al

    lazysoci.al Update 2024-12-18

    Holidays are coming, and with that, spare time!

    I shall be updating the instance over the next couple of weeks to 0.19.8. I'll post an update some hours before the upgrade.

    Over the holidays I shall be decommissioning the physical server that this Lemmy instance is hosted on. The new server is being built and will be a Dell PowerEdge R640 running all U.2 NVMe drives. Hopefully, this will make the server the fastest it can possibly be.

    Announcements (lazysoci.al) @lazysoci.al
    lazyadmin @lazysoci.al

    lazysoci.al Outage 2024-11-18

    Our lemmy instance stopped processing new activity sometime on Monday 2024-11-18 morning.

    The root cause remains unknown. Services were online. Database was responsive.

    Lemmy server logs were showing the incoming ActivityPub requests, no errors, but no response was being returned to the sender. The system was restarted on 2024-11-19 and processing of requests resumed.

    Luckily, the protocol allows for some caching of requests across all servers, so after 30 minutes of heavy load, our server had mostly caught up.

    I was away on Monday, and I did notice the issue, but I initially thought it was a problem with my mobile app (recently moved to Boost). I normally view Lemmy sorted by "Top - Last Twelve Hours" and on Tuesday this returned zero results, which prompted a closer look.

    I have added additional monitoring to the system, checking for the age of the latest post. I shall now receive an alert if a new post has not been received for 15 minutes. This may result in some false positive

    Announcements (lazysoci.al) @lazysoci.al
    lazyadmin @lazysoci.al

    lazysoci.al Outage 2024-10-11

    Our hosting provider had an outage which lasted ~30 minutes.

    I raised a support case as soon as I got the alert as I couldn't access the systems via their link, but the 4G backup confirmed servers were online.

    Announcements (lazysoci.al) @lazysoci.al
    lazyadmin @lazysoci.al

    lazysoci.al Outage 2024-08-28

    lazysoci.al was offline whilst the cluster was being updated.

    Outage was 38 minutes. 11:00 to 11:38 BST

    This was expected to be <15 minutes. The extended outage was due to an issue bringing up a docker container which was a pre-requisite for the load balancer.

    Announcements (lazysoci.al) @lazysoci.al
    lazyadmin @lazysoci.al

    lazysoci.al Update 2024-07-13

    I shall be migrating pict-rs to postgres, now.

    The pict-rs service shall need to be offline which shall effect thumbnail generation and the ability to post images.

    Should be completed within 4 hours of this post.

    Edit: Migration has completed at 1515 UK time.

    lazysoci.al Update 2024-06-27

  • Update has been completed. A few months ago I foolishly updated to a beta version, and this introduced database changes that are not part of the standard migration path. So of course, the standard migration scripts that get run on startup failed, and I'm left with a broken instance.

    Luckily some others had the same issue and this was discussed at https://github.com/LemmyNet/lemmy/issues/4641#issuecomment-2071033771

    I had two functions of the same name when the migration script was only expecting the one. So I removed the one that didn't match the expected function body, and the migraton worked.

    I also wanted to migrate pict-rs from the disk-based SLED database to its own PostGres instance, but at the rate it was going, it was going to take 4 hours. The server its running on is full NVMe, enterprise hardware (Dell Poweredge) but its getting on a bit (I'll be purchasing a newer server for the datacenter later this year).

    I shall retry the pict-rs migration in the next few days but it will mean an extended downtime.

  • Announcements (lazysoci.al) @lazysoci.al
    lazyadmin @lazysoci.al

    lazysoci.al Update 2024-06-27

    Registration requirements

    I've had to ban multiple vile accounts this morning. It seems this instance has found itself on the radar of trolls.

    To that end, sign-ups now require users to fill in a questionnaire prior to joining. I always thought it was lame, and it won't really prevent a troll account from joining, but it would slow them down, and likely cause them to find another easier-to-join instance.

    Upgrade

    I shall be upgrading the instance to 0.19.4 0.19.5 this afternoon on Friday morning, so expect a little downtime

    I'll unpin this notice once the update is complete.

    Edit: Unfortunately work got in the way of me performing the update today, so postponing this to first thing tomorrow morning.

    Cached images

    We use the standard group of services for Lemmy, including the pict-rs image/thumbnail cache. This image cache grew to 700G recently and continues to grow as Lemmy grows. Therefore some effort has been made to keep it under control.

    Cached i

    Announcements (lazysoci.al) @lazysoci.al
    lazyadmin @lazysoci.al

    Outage: 2024-01-14 - Upgraded to 0.19.2

    Upgraded to 0.19.2

    Upgrade itself went OK. Also updated pict-rs to 0.5.1 which performed a metadata update.

    I tested this upgrade last Thursday 2024-01-11 but rolled back when outgoing federation didn't work.

    Found a fix from matrix.org (thanks to [email protected] and [email protected]) by updating the "updated" column on the instance table.

    UPDATE instance SET updated = now() WHERE updated > now() - interval '14 days';

    Total downtime was 12 minutes.

    Announcements (lazysoci.al) @lazysoci.al
    lazyadmin @lazysoci.al

    Outage: 2023-12-17

    Upgraded to 0.19.0

    Looks good so far. Took a snapshot prior to upgrade, but initial monitoring shows no issues.

    Total downtime was 42 minutes.

    Announcements (lazysoci.al) @lazysoci.al
    lazyadmin @lazysoci.al

    Outage: 2023-11-11

    What was meant to be a quick blip ended up being over an hour. Migrating the reverse proxy that sits in front of the Lemmy server failed as dockerhub is having an outage.

    Announcements (lazysoci.al) @lazysoci.al
    lazyadmin @lazysoci.al

    Outage; 2023-09-10

    Overview

    lazysoci.al was offline for 3h 15m today following a database corruption. Server is now back online, federated data is flowing again.

    Details

    I moved the server to its own dedicated host this morning, for both the performance and security (dedicated vlan) impact. Should have been a simple case of moving the virtual disk with the Lemmy data to the new VM and spinning up the new docker image.

    The docker logs didn't show any initial issues, however writing to the database gave errors of ERROR: relation "approvals" does not exist for every UPDATE query.

    After some troubleshooting, I finally thought the database was corrupted, so I started a restore from last nights backup. This took approx. 2h 30m to restore.

    Post-restore, the same issue. I then performed an update to the latest beta, and the issue is now resolved.

    This has highlighted one problem. I use proxmox-backup-server and proxmox-virtual-environment. You can't easily restore a single disk from a VM into a ZFS vo