John Colagioia

Hi, I work on a variety of things, most of which I talk about more on my blog than on social media. Here, you’ll probably find me talking mostly talking about Free Culture works and sometimes technology.

  • 0 Posts
  • 5 Comments
Joined 2 years ago
cake
Cake day: July 23rd, 2023

help-circle
  • I’ve been using different versions of SearX for a long while (sometimes on my server, sometimes through a provider like Disroot) as my standard search engine, since I’ve never had great luck with the big names, and it’s decent, but between upstream provider quota limits, and just the fact that it relies on corporate search APIs at all, sometimes the quality craters.

    While I haven’t had the energy to run YaCy on my own, and public instances tend to not have a long life, I don’t have nearly as much experience with it, but when I have gotten to try it out, the search itself looked great, but generally didn’t have as broad or current an index. Long-term, though, it (and its protocol) is probably going to be the way to go, if only because a company can’t randomly tank it like they can with the meta-search systems or their own interfaces.

    Looking at Presearch for the first time now, the search results look almost surprisingly good if poorly sorted, but the fact that I now know orders of magnitude more about their finances and their cryptocurrency token than what and how the thing actually searches makes me worry a bit about its future.


  • I believe that YouTube supports RSS. I haven’t used it in years, but gPodder allowed subscribing to channels.

    Ah, yeah. From this post:

    • Go to the YouTube channel page.
    • Click more for the About box.
    • Scroll down to click Share channel. Choose Copy channel ID.
    • Get the feed from https://www.youtube.com/feeds/videos.xml?channel_id plus that channel ID from the previous step.

    From there, something (like a podcast client) needs to grab the video.

    Otherwise, I’ve been using Tartube to download to my media server, which is not great but fine, except for needing to delete the lock file when it (or the computer) crashes, and the fact that the media server hasn’t the foggiest idea of how to organize the “episodes.”




  • I keep saying “no” to this sort of thing, for a variety of reasons.

    1. “You can use this code for anything you want as long as you don’t work in a field that I don’t like” is pretty much the opposite of the spirit of the GPL.
    2. The enormous companies slurping up all content available on the Internet do not care about copyright. The GPL already forbids adapting and redistributing code without licensing under the GPL, and they’re not doing that. So another clause that says “hey, if you’re training an AI, leave me out” is wasted text that nobody is going to read.
    3. Making “AI” an issue instead of “big corporate abuse” means that academics and hobbyists can’t legally train a language model on your code, even if they would otherwise comply with the license.
    4. The FSF has never cared about anything unless Stallman personally cared about it on his personal computer, and they’ve recently proven that he matters to them more than the community, so we probably shouldn’t ever expect a new GPL.
    5. The GPL has so many problems (because it’s been based on one person’s personal focuses) that they don’t care about or isolate in random silos (like the AGPL, as if the web is still a fringe thing) that AI barely seems relevant.

    I mean, I get it. The language-model people are exhausting, and their disinterest in copyright law is unpleasant. But asking an organization that doesn’t care to add restrictions to a license that the companies don’t read isn’t going to solve the problem.