• 0 Posts
  • 100 Comments
Joined 11 months ago
cake
Cake day: May 6th, 2025

help-circle






  • extract () {
      if [ $# -ne 1 ]
      then
        echo "Error: No file specified."
        return 1
      fi
        if [ -f "$1" ] ; then
            case "$1" in
                *.tar.bz2) tar xvjf "$1"   ;;
                *.tar.gz)  tar xvzf "$1"   ;;
                *.tar.xz)  tar xvf "$1"    ;;
                *.tar.zst) tar axvf "$1"   ;;
                *.xz)      xz -kd "$1"     ;;
                *.bz2)     bunzip2 "$1"    ;;
                *.gz)      gunzip "$1"     ;;
                *.tar)     tar xvf "$1"    ;;
                *.tbz2)    tar xvjf "$1"   ;;
                *.tgz)     tar xvzf "$1"   ;;
                *.lzma)    unlzma "$1"     ;;
                *.rar)     unrar x "$1"    ;;
                *.zip)     unzip "$1"      ;;
                *.Z)       uncompress "$1" ;;
                *.7z)      7z x "$1"       ;;
                *.exe)     cabextract "$1" ;;
                *.deb)     ar x "$1"       ;;
                *.jar)     jar xf "$1"     ;;
                *)         echo "'$1' cannot be extracted via extract" ;;
            esac
        else
            echo "'$1' is not a valid file"
        fi
    }
    









  • People publish things for all sorts of reasons, sure, but the main one is that they think the information has value. Whether some corporation profits from it is completely irrelevant, the data doesn’t care who benefits. The author doesn’t matter either, what matters is the information itself. Licenses just try to put fences around what should be free, and most of the time they only get in the way. People can follow them or ignore them, and life goes on. The scene understands that better than anyone else. It’s the purest form of the web: information shared for its own sake, without permission, limits, or fake moral theater.





  • I completely disagree. The road to hell is paved with good intentions. What he proposes is just another flavor of technofascism, control disguised as ethics. These so-called humane scraping barriers will end up blocking humans, not machines. We have seen this before with CAPTCHAs, reCAPTCHAs, and all those “human verification” gimmicks, always bypassed by bots and always annoying to real people. Personally, I have nothing against crawlers and bots; they should do their shady jobs. Trying to wall yourself off is just meh. And those fancy licenses or “ethical use” terms won’t change anything either. The web is not the United States, and nobody really cares about someone’s imaginary social contracts. Maybe it’s time to accept a simple fact: once something goes on the web, it becomes public territory, and no one can still pretend to control the flow of information.