←back to thread

597 points classichasclass | 5 comments | | HN request time: 1.053s | source
Show context
lwansbrough ◴[] No.45010657[source]
We solved a lot of our problems by blocking all Chinese ASNs. Admittedly, not the friendliest solution, but there were so many issues originating from Chinese clients that it was easier to just ban the entire country.

It's not like we can capitalize on commerce in China anyway, so I think it's a fairly pragmatic approach.

replies(6): >>45010748 #>>45010787 #>>45010871 #>>45011590 #>>45011656 #>>45011732 #
sugarpimpdorsey ◴[] No.45010787[source]
There's some weird ones you'd never think of that originate an inordinate amount of bad traffic. Like Seychelles. A tiny little island nation in the middle of the ocean inhabited by... bots apparently? Cyprus is another one.

Re: China, their cloud services seem to stretch to Singapore and beyond. I had to blacklist all of Alibaba Cloud and Tencent and the ASNs stretched well beyond PRC borders.

replies(5): >>45010898 #>>45010946 #>>45011282 #>>45011573 #>>45014393 #
johnisgood[dead post] ◴[] No.45011573[source]
[flagged]
pabs3 ◴[] No.45011797[source]
Which site is it?
replies(1): >>45011806 #
johnisgood ◴[] No.45011806[source]
My own shitty personal website that is so uninteresting that I do not even wish to disclose here. Hence my lack of understanding of the down-votes for me doing what works for my OWN shitty website, well, server.

In fact, I bet it would choke on a small amount of traffic from here considering it has a shitty vCPU with 512 MB RAM.

replies(2): >>45012578 #>>45022105 #
1. pabs3 ◴[] No.45022105[source]
Personal sites are definitely interesting, way more interesting than most of the rest of the web.

I was thinking I would put your site into archive.org, using ArchiveBot, with reasonable crawl delay, so that it is preserved if your hardware dies. Ask on the ArchiveTeam IRC if you want that to happen.

https://chat.hackint.org/?join=%23archiveteam-bs

replies(1): >>45024534 #
2. johnisgood ◴[] No.45024534[source]
It is a public git repository for the most part, that is the essence of my website, not really much writings besides READMEs, comments in code and commits.
replies(1): >>45034397 #
3. pabs3 ◴[] No.45034397[source]
A public git repository is even more interesting, for both ArchiveTeam Codearchiver, and Software Heritage. The latter offers an interface for saving code automatically.

https://wiki.archiveteam.org/index.php/Codearchiver https://wiki.archiveteam.org/index.php/Software_Heritage https://archive.softwareheritage.org/save/

replies(1): >>45039703 #
4. johnisgood ◴[] No.45039703{3}[source]
After initial save, do they perform automatic git pulls? What happens if there are potential conflicts? I wonder how it all works behind the surface. I know I ran into issues with "git pull --all" before, for example. Or what if it is public software that is not mine? I saved some git repositories (should I do .tar.gz too for the same project? Does it know anything about versions?).
replies(1): >>45046848 #
5. pabs3 ◴[] No.45046848{4}[source]
On Software Heritage: for forges (GitLab, cgit etc), every couple of months SWH lists all repos, pulls new/updated ones. I think if you save an individual repo, it gets pulled later too, but I'm not sure of the schedule. They have custom tooling (open source) for doing the importing of repos, tarballs and other things. They deduplicate on the backend, so if you cloned some repos then the files/commits that are shared between them are saved once. They import the git tags (and other refs) too.

ArchiveTeam Codearchiver is quite a bit different, it does one-shot archiving of repos into VCS-native export formats, like git bundles. There is some deduplication based on commit hashes I think.