I like sysadmin, scripting, manga and football.

  • 1 Post
  • 23 Comments
Joined 2 years ago
cake
Cake day: June 13th, 2023

help-circle


  • The easiest way by far is downloading an existing dump from kiwix

    Per example wikipedia_en_all_nopic_2024-06.zim is only 54GB since it only contains text. Then via docker you could use this compose file where you have your .zim files in the wikis volume:

    services:
      kiwix:
        image: ghcr.io/kiwix/kiwix-serve
        container_name: kiwix_app
        command: '*'
        ports:
          - '8080:8080'
        volumes:
          - "/wikis:/data"
        restart: always
    

    Theorically you can actually one of the wikipedia database dumps with mediawiki but I don’t known of any easy plug and play guide









  • Clean all the cache downloads of Arch Linux Packages

    pacman -Scc

    Remove unused docker networks and images

    docker system prune --all

    Cleanup untracked git files that might be in .gitignore such as build and out directories (beware of losing data, use “n” instead of “f” for a dry run)

    git clean -xdf

    Do an aggresive pruning of objects in git (MIGHT BE VERY SLOW)

    git gc --aggressive --prune=now

    Remove old journal logs, keeping last seven days

    journalctl --vacuum-time 7days

    Remove pip cache

    pip cache purge





  • 4-2-1-1 for me I guess 🫣 or 4-2-2?

    Two copies at home, synced daily, one of them in an external drive that I like to refer as the emergency grab and run copy lol

    One at a family member synced weekly and manually every time I visit.

    All of those three copies are always within a 10 kilometer radius in a valley overseen by a volcano so…

    One partial copy of the so-critical-would-cry-if-Iost data is synced every few days to a backblaze bucket.