Announcements

228 readers
1 users here now

Announcements about system updates or other things related to this instance.

founded 1 year ago
MODERATORS
1
 
 

Hi everyone,

We just had some unexpected downtime due to a disk quickly getting full over night... :/

We will adjust the monitoring to start alerting us earlier for disk space running out so we can actually do something about it.

I took the chance to install the latest OS patches and updates now though since the server was anyway down.

Hope you didnt notice the downtime too much!

2
29
submitted 6 months ago* (last edited 6 months ago) by [email protected] to c/[email protected]
 
 

Hi all,

We were getting several reports of people linking to csam content in the signal groups community. We tried to contact the moderator but no response in several days.

Its difficult to moderate this kind of content. Moderators of a lemmy community cannot be expected to visit each posted group around the clock to make sure it doesnt have csam content. Signal groups can change their content at any time.

So sorry to say, it had to be removed. If you were using it, I hope you find some other way to find new signal groups.

3
 
 

We decided to double the server memory and add more cpu, since we were close to running out sometimes.

So if you notice higher performance, you know what the reason is now. :)

Short post, but just wanted to mention it.

4
 
 

We just deployed the latest version of Lemmy 0.19.2 where the github code includes a possible fix for the outgoing federation issues we have been having.

But lets see before we celebrate. Help us test if outgoing federation seems to work now by making comments, posts and upvotes and see if they appear on other instances.

Of course if the other instances are on Lemmy 0.19.0 or 0.19.1, they could have issues with outgoing federation still until they update.

Release notes for 0.19.2: https://join-lemmy.org/news/2024-01-10_-_Lemmy_Release_v0.19.2_-_More_Federation_Fixes

5
 
 

Please try to comment and post things now, and see if they federate again.

Hopefully you see your activity instantly federated. I have tried making comments both to instances running Lemmy 0.19.1 and 0.18.5 and they all federate like they should.

Hope you have the same experience! 🥳

6
24
submitted 10 months ago* (last edited 10 months ago) by [email protected] to c/[email protected]
 
 

As you may have read in other threads, this version of Lemmy (0.19.1) seems to have bugs in outgoing federation on some instances.

As a temporary fix, we have added a scheduled restart of Lemmy every hour. It only takes a few seconds to restart, and the big advantage is that your comments and posts are only delayed up to 1 hour before they federate to other instances. You probably wont notice the restart even.

This will be in effect until a bug fix arrives from Lemmy developers, probably after new years sometime.

Thanks for reading and merry x-mas to everyone. :)

7
 
 

Today we spent some time preparing for the big upgrade, by kicking the Lemmy version up to 0.18.5, merging in the latest changes from lemmy-ansible git repository and cleaning up some disk space on the instance.

So tomorrow at 02.00 am Oregon time we will do the update to Lemmy 0.19.

This is 11.00 am CET for people in Europe.

Hopefully all goes well and we come out the other side with a nice new 0.19 version. Its supposed to take about 30 minutes of downtime if there are no issues to solve.

Wish us luck :)


IMPORTANT: You probably need to log out and log in again to be able to post anything since they reworked authentication in this release.


8
 
 

Hi,

We are planning to install Lemmy 0.19 soon, hopefully in the coming week or so. It's a huge release with many new features, and I personally really like that it allows you users to block other instances if you want to.

You can read about all the new features in the link above.

More info coming in a few days about planned downtime and so on. :)

9
 
 

We just had another unscheduled downtime due to a linux kernel bug.

Yesterday we noticed some issues with the server - mainly that we couldnt stop some docker containers. As you know, we are running some extra user web interfaces for Lemmy and we noticed that they started acting weirdly and had one cpu running at 100% constantly.

I wanted to restart those containers but I couldnt stop them. Found some posts online that this is a bug in the Ubuntu linux kernel: https://github.com/moby/moby/issues/43094.

Our stop attempts caused the docker platform to behave weirdly and it started to affect the main Lemmy software, so we did a reboot of the server and installed the latest updates.

We are very sorry for this unscheduled downtime. :/ Did you guys notice weirdness with Lemmy in the last 8 hours or so?

10
 
 

Hi guys!

This weekend we will move lemmy.today over to using object storage for images. We will be serving images from Amazon S3 in the Oregon region (western USA).

The way lemmy software is designed right now, it caches every image federated from other instances. So even if we are small instance, we still have to store a lot of federated images locally on our disk. This leads to disk space running out quickly and we previously had to delete images because of this.

When we delete images, it removes not only those cached images but also user profile icons and banners, as well as community icons and banners. This is why we have some missing images under Communities right now, and also why users have lost their profile pics.

Its been very embarrassing to have to do this, and now we will move to object storage to prevent this from happening in the future. Its much cheaper compared to ordinary disk space and gives better performance for users, so its a win-win. We just need to do a one-time migration over to it.

**Estimated downtime hours: **

Oregon time: Sunday 3 am - 6 am

CET: Sunday 12 pm to 15 pm.

If you have any questions, you know what to do. :)


EDIT: Looks like it went well and images are now served from S3 instead of filling up our disks. :) The url to the images still looks like they are served by the instance, but thats by design appearently. In the background, they are fetched from S3.

Please fill free to re-upload any banners, user avatars or community pictures you had in place before that may have been broken by the disk cleaning before.

  • When you do, you have to create a new picture with a new name for Lemmy to actually replace the image. Otherwise it wont work - ive tried myself. :)

11
 
 

Hi everyone,

As part of cleaning old cached images when the disk went full, it seems also images like your profile picture and banners (if you had those), got deleted.

If you dont mind, would you upload those again? And when you do, you cant upload the same picture. I made attempts myself to upload the same picture, but it needs to be a new picture (not even a rename of the pic works).

Next time I will make a DB query to figure out what pics are local and which are not, and delete only remote ones (cached images from other instances). There is a column in the DB for that, so just need to export a list of remote images and then delete only those.

Despite these growing pains, I think lemmy is still pretty awesome, and there will be tools to make these sort of issues go away in the future. I hear they are working on something for next version already so we will see.

Anyway, enjoy the weekend and once again, sorry for the mess around this issue.

12
 
 

Hi all,

The disk on the instance ran out of space today, due to the way Lemmy software caches images from all other instances. That cache had filled up about 60 GB's of disk, despite us being a small instance with very little local activity.

I had to delete the last 10 days of cached images again, and I plan to delete quite a lot of older cached images as well. The mobile apps seems to not be affected by this (they have a local image cache I believe), but on the web site, this leads to missing thumbnail images.

They are working on a fix for this in the Lemmy software so the disks dont fill up so enormously with cached thumbnails, and as soon as its out, we will install it here.

Hope you guys didnt get too annoyed or sad by the instance being unavailable for a while.

13
6
submitted 1 year ago* (last edited 1 year ago) by [email protected] to c/[email protected]
 
 

Some bots posted lots of illegal pictures in the https://lemmy.world/c/lemmyshitpost community, and because of federation, those pictures have spread to all instances, including this one.

The lemmy software doesnt have good moderation tools for abuse like this, and the quickest way to get rid of them was to delete all cached images for the last couple of days.

You may see some thumbnail images missing in the web interface, but I personally dont see any missing images in my mobile app. I guess it has its own thumbnail cache.

14
9
submitted 1 year ago* (last edited 1 year ago) by [email protected] to c/[email protected]
 
 

These ones have been added:

People like them on lemmy.world so didnt want them to missing here. :)

15
 
 

I noticed that the web interface sometimes didnt show all pictures when doing a full reload of the front page. This has been fixed now. It was related to some custom security settings I added last week, and I didnt notice the problem since i use mobile apps myself. But for everyone who uses the web interface a lot, this must have been annoying and has been fixed.

16
5
Lemmy themes! (lemmy.today)
submitted 1 year ago* (last edited 1 year ago) by [email protected] to c/[email protected]
 
 

I added some themes for people who use the web interface. Some are pretty nice I think:

Modern Light:

Hanubeki Cold

Hanubeki Mint Alt Lt

And others.

How to use

  • After you switch to a theme and save your settings, its really important to reload your browser cache, otherwise the theme will look wonky.

  • Do this by holding shift and clicking the Reload current page button in your browser. (or press Shift-Control-R if you are on firefox)

17
4
submitted 1 year ago* (last edited 1 year ago) by [email protected] to c/[email protected]
 
 

New version of Lemmy just got released by the devs. :)

I plan to wait a few days before I upgrade the instance, just to make sure there isn't any weirdness being discovered. It's a small bug fix release so nothing major.

Link to the announcement: https://lemmy.ml/post/3021118

18
3
submitted 1 year ago* (last edited 1 year ago) by [email protected] to c/[email protected]
 
 

Hi guys,

Lemmy.today was just updated to 0.18.3 that came out yesterday.

Major changes

This version brings major optimizations to the database queries, which significantly reduces CPU usage. There is also a change to the way federation activities are stored, which reduces database size by around 80%. Special thanks to @phiresky for their work on DB optimizations.

The federation code now includes a check for dead instances which is used when sending activities. This helps to reduce the amount of outgoing POST requests, and also reduce server load.

In terms of security, Lemmy now performs HTML sanitization on all messages which are submitted through the API or received via federation. Together with the tightened content-security-policy from 0.18.2, cross-site scripting attacks are now much more difficult.

Other than that, there are numerous bug fixes and minor enhancements.