Canadian software engineer living in Europe.

  • 5 Posts
  • 170 Comments
Joined 2 years ago
cake
Cake day: June 7th, 2023

help-circle


  • I use them quite heavily in combination with Cookie Autodelete. I then create a separate profile for each surveillance capitalist service I work with. So for example, here’s my list of containers:

    A screenshot of my container list

    Every time I visit one of these sites, Firefox opens them in the respective container, and the cookies they create are isolated to that container. When I’m in the LinkedIn container, Cookie AutoDelete nukes every cookie that isn’t from LinkedIn (including Google, GitHub, etc.). When I’m not in any container, all cookies are deleted everywhere.

    Basically it’s a nice way to leverage Cookie Autodelete without having to whitelist Big Tech for all my browsing.


  • I don’t think there’s an official “way”, but here’s mine (which I love):

    On start-up I open all the apps I usually use, one per designated workspace:

    1. Slack/Teams/Mattermost, whatever my work requires.
    2. Thunderbird
    3. Kitty
    4. PyCharm/RustRover, whatever the job requires
    5. Firefox

    Workspaces 6-9 are left empty, ready for whatever app I need in the moment, but only ever one app per workspace.

    With this setup, I’ve mapped Ctrl+Fx to each workspace, so Ctrl+F4 takes me to PyCharm where I write the code, and Ctrl+F5 followed by another F5 takes me to Firefox and reloads the page. Ctrl+F3 is always the terminal, etc., so you quickly start building these shortcuts to mean Fwhatever is $APP_NAME.

    I almost never use the mouse, unless what I’m doing is necessarily mouse-driven: browsing or drawing charts etc. Everything else is keyboard-driven.



  • I have a few interesting ones.

    Download a video:

    alias yt="yt-dlp -o '%(title)s-%(id)s.%(ext)s' "
    

    Execute the previous command as root:

    alias please='sudo $(fc -n -l -1)'
    

    Delete all the Docker things. I do this surprisingly often:

    alias docker-nuke="docker system prune --all --volumes --force"
    

    This is a handy one for detecting a hard link

    function is-hardlink {
      count=$(stat -c %h -- "${1}")
      if [ "${count}" -gt 1 ]; then
        echo "Yes.  There are ${count} links to this file."
      else
        echo "Nope.  This file is unique."
      fi
    }
    

    I run this one pretty much every day. Regardless of the distro I’m using, it Updates All The Things:

    function up {
      if [[ $(command -v yay) ]]; then
        yay -Syu --noconfirm
        yay -Yc --noconfirm
      elif [[ $(command -v apt) ]]; then
        sudo apt update
        sudo apt upgrade -y
        sudo apt autoremove -y
      fi
      flatpak update --assumeyes
      flatpak remove --unused --assumeyes
    }
    

    I maintain an aliases file in GitLab with all the stuff I have in my environment if anyone is curious.


  • I have much the same:

    • Files on the network with NFS
    • Kodi on an old laptop under the TV so we can watch said files.
    • Syncthing on our phones and laptops to pull films from there onto that file server.

    The only difference is that I’m using a Synology 'cause I have 15TB and don’t know how to do RAID myself, let alone how to do it with an old laptop. I can’t really recommend a Synology though. It’s got too many useless add-ons and simple tools like rsync never work properly with it.







  • Lowering the barrier to entry by moving from a technology few use (mercurial) to something popular (git) makes sense. Requiring participation on a proprietary platform owned by Microsoft instead of an open one like Codeberg or GitLab is just lazy. If someone wants to contribute to Firefox, asking them to create an account is a small ask, and I’d argue that if they’re unwilling to do even that, then their participation in the community is likely to be far from useful.




  • Granted, sudo isn’t in coreutils, but it’s sufficiently standard that I’d argue that the licence is very relevant to the wider Linux community.

    Anyway, I answered this at length the last time this subject came up here, but the TL;DR is that private companies (like Canonical, who owns Ubuntu) love the MIT license because it allows them to take the code and make proprietary versions of it without having to release the source code. Consider the implications of a sudo binary that’s Built For Ubuntu™ with closed-source proprietary hooks into Canonical’s cloud auth provider. It’s death by a thousand MIT-licensed cuts to our once Free operating system.





  • Daniel Quinn@lemmy.catoLinux@lemmy.mlThis looks cool but can it game?
    link
    fedilink
    English
    arrow-up
    20
    arrow-down
    1
    ·
    2 months ago

    The Ampre Altra runs from 32 to 128 cores (dear gods that’s beautiful), but with that architecture, and the company’s stated purpose, it makes more sense in a computer meant to be used as a server rather than a desktop gaming rig. You’d use a chip like that in a Kubernetes cluster for example.

    Combined with an Nvidia card, a brand notorious for being a Pain In The Ass in Linuxland, I’m going to go out on a limb here and suggest that the intended purpose of a box like this is a server for AI/ML-based services.