• 0 Posts
  • 58 Comments
Joined 1 year ago
cake
Cake day: June 6th, 2023

help-circle
rss

  • Right, now get a borderline computer-illiterate person to connect to your network, ensure their firewall isn’t misconfigured to block all incoming traffic (with TeamViewer, this configuration would still work because the device just connects to the TV server) and open and set up a completely separate screen sharing program.

    I know none of these steps are difficult if you have any idea what you’re doing, but I’ve met plenty of people who would most likely need assistance going through the motions. Funnily enough, the best way to do it remotely would probably be to get them to install TeamViewer to then set this up for them remotely.

    By the way, as far as networking goes, Tailscale does the same thing TeamViewer does, just for a VPN instead of a screen sharing application - it will try to do all the NAT punchthrough techniques and IPv6 connection and fall back on tunneling through relay servers if all else fails. It’s not any more of a direct connection than TV.


  • So I did look more into it, and apparently the open firmware is technically compatible with PCIe cards using this chip, but doesn’t provide any advantages over just wiping the firmware and letting the chip default to its built-in fallback firmware, and so the maintainer doesn’t see any value in explicitly supporting it.

    Now the question is whether you consider the proprietary fallback firmware to be acceptable to run - this might sound weird, but for example FSF has explicitly made exceptions for devices with built-in firmware to be able to qualify for the Respects Your Freedom certification, so if your view aligns with theirs, you might consider this to be completely OK. If not, the free firmware appears to have a similar feature set, you’ll just have to jump through more hoops.

    Also do note that both the fallback firmware and the free firmware are missing many features of the proprietary firmware, so make sure to check it’s not missing anything you need (wake on LAN, Jumbo frames and PXE boot seem like the most notable missing features to me).

    More info on support for various PCIe cards


  • Convenience (after you install it, all you have to do is enter the code and you’re connected, no other setup required), familiarity (it’s the default name people will think of or find if they want remote access - that alone means they can get away with pushing their users slightly more) and - IMHO most importantly - connectivity: if two computers can connect to the TeamViewer servers, they will be able to connect to each other.

    That’s huge in the world of broken Internet where peer to peer networking feels like rocket science - pretty much every consumer device will be sitting behind a NAT, which means “just connecting” is not possible. You can set up port forwarding (either manually or automatically using UPnP, which is its own bag of problems), or you can use IPv6 (which appears to be currently available to roughly 40% users globally; to use it, both sides need to have functional IPv6), or you can try various NAT traversal techniques (which only work with certain kinds of NAT and always require a coordinating server to pull off - this is one of the functions provided by TeamViewer servers). Oh, and if you’re behind CGNAT (a kind of NAT used by internet providers; apparently it’s moderately common), then neither port forwarding or NAT traversal are possible. So if both sides are behind CGNAT and at least one doesn’t have IPv6, establishing a direct link is impossible.

    With a relay server (like TeamViewer provides), you don’t have to worry about being unable to connect - it will try to get you a direct link, but if that fails, it will just act as a tunnel and pass the data between both devices.

    Sure, you can self host all this, but that takes time and effort to do right. And if your ISP happens to use CGNAT, that means renting a VPS because you can’t host it at home. With TeamViewer, you’re paying for someone else to worry about all that (and pay for the servers that coordinate NAT traversal and relay data, and their internet bandwidth, neither of which is free).


  • It’s less about the computer and more about the card itself - Talos II and Blackbird both use the BCM5719 chip for their integrated NICs. Basically, you’re flashing part of the motherboard with this firmware. A PCIe card built around the same chip might connect the interfaces in a different way, and firmware doesn’t generally have a way of poking around to find out how everything’s set up from the hardware side of things - it needs to just know this, and that’s why there are separate firmware builds for different hardware.

    If you flash one of these files to that card, it might just so happen to work perfectly, but it most likely won’t. You would need to figure out how it’s wired up and modify the firmware with that knowledge. And then you could use the modified open firmware with that specific card model on any computer that supports the proprietary firmware, because IIUC this is meant to be functionally identical.

    So in short, no, you cannot currently use this open firmware on any computer other than Talos II and Blackbird, but for slightly different reason than you might think.



  • If it is an Arch-based distro (sorry, I don’t recognize the package manager), then this might just be the recent Wine update that made it 700 MB smaller (which would mean the rest of your system grew 300 MB)

    I made a post here about it: this one

    Btw, is there a way to link to a post in a way that resolves on everyone’s separate instance instead of hard coding it to my instance?


  • Also, some programs, such as many terminal emulators, can cache you PW so you don’t have to enter it multiple times.

    Terminal emulators don’t (or at least shouldn’t) do any such thing. sudo itself is responsible for letting you do privilege escalation without password for some time after successfully passing once - whenever you run it and successfully authenticate, it saves your user id, current time and a session identifier (each open shell gets a unique identifier) into a file. Then, when you attempt to do anything, it will check this file to see if you’ve if you’ve authenticated within the last few minutes in this terminal, and only ask for a password if you haven’t.

    For more info, see man sudoers_timestamp



  • CS:GO got a controversial update and got renamed. Old versions are still available under CS2, you just can’t use Valve’s servers anymore. Playing old versions on private servers is possible. But OK, I give you half a point for this one - you can’t play matchmaking with old smoke physics anymore (but then again, it’s not like it’s the first CS:GO update to change the gameplay in a fundamental way).

    Moving on, Artifact. It’s in my library, ready to be played - Valve definitely didn’t “make me lose Artifact” like you claimed. The community is dead, but there are still 40 people playing right now according to SteamDB and servers are up. One point down for easily verifiable lie.

    And finally, Team Fortress 1. I assume you don’t mean the Valve’s game called Team Fortress Classic, because that one is still available for purchase on the Steam Store and oscillates between 40 and 100 active players at any time. So that leaves us with Team Fortress, a mod for Quake. But that one is available from ModDB without any problems, so… What’s the issue supposed to be, exactly? No points, because I have no idea if there’s more to your claim.

    Hint: blatantly lying about some points heavily undermines the other points you make. So at least try to be subtle.


  • It should not be controlled by a company that is known to make you lose your games.

    Are you referring to the fact that Valve promotes digital game distribution (which is a very fair view), or are you talking about some incident where Valve removed games from people’s libraries? Because if it’s the second one, then I would really like to hear about it.


  • Are you sure you didn’t set DNS directly on some/all of your devices? Because in that case they won’t care about the router settings and will use whatever you set them to.

    Also as the other commenter said, DNS changes might not propagate to other devices on the network until the next time they connect - a reboot or unplugging the cable from your computer for a few seconds is a dirty but pretty OS agnostic way to do that.


  • No worries, nothing wrong with not knowing everything about every random subject. I would like to apologize for being overly harsh, I assumed that people in c/opensource would have general knowledge of this definition, but that assumption was clearly bad. So again, sorry.

    I assume the term is misused often.

    Yes, companies sometimes do that. Open source is marketable as a guarantee that you won’t fully lose access to a piece of software, and there aren’t any real consequences of misusing it. But there’s also a scheme called dual licensing where the software is available under two licenses - one license is open source but annoying for commercial use, and the other is a “normal” proprietary license under which businesses can buy the code. This is fine (as long as the provider has copyright to all the code being dual licensed) and is pretty common and makes the software open source.


  • But “open source” doesn’t even mean that you can reproduce it or use it for free.

    You’re thinking of source-available licenses. Open source has a clear and widely accepted definition that requires a certain level of freedom. You’re free to ignore this definition, but you can’t expect the rest of the world to do the same.

    To be clear, open source allows for only providing access to paying customers, but those paying customers are then free to use and distribute their copies without any further payment. Then it wouldn’t be open source anymore.


  • I mean, it’s called “LaTeX by example”, so there’s a pretty good chance it’s written in LaTeX, which you do indeed compile to get the PDF or whatever output you want.

    Also, just having access to the source doesn’t make it open source - that requires more freedoms. For example, here’s GitLab Enterprise Edition source code, fully functional and ready to be used. And also officially described as the proprietary edition of GitLab by the GitLab company itself. Why? Because its license pretty much boils down to “you can use this only for testing and development, unless you have paid for it”.



  • Not a fair comparison IMHO - Ethernet is designed to be a connection between two or more otherwise independent peers (on L2), while USB’s goal was to allow connecting simple peripheral devices to computers. There was never meant to be a situation where it’s unclear which side is the Host.

    Also note that the bridging “cable” is literally just two USB devices, one for each computer (although they are both on the same chip, so there’s that), with some internal link to pass the data.


  • You could put your .desktop files in a separate directory and just symlink them to ~/.local/share/applications. If you want to have all your aliases together, you could have a directory like Aliases or whatever and then have Aliases/desktop for .desktop launchers, Aliases/bin for scripts or binaries (and have it in PATH), Aliases/bashrc for your bash aliases (and just put source .../Aliases/bashrc in ~/.bashrc), etc.

    Of course everyone has their own opinions on how to organize stuff, but this is IMHO pretty clean for what you probably want to do.

    Edit: and to quickly (re)create the symlinks, you could use a bash one-liner like for f in .../Aliases/desktop/*; do ln -s "$f" "~/.local/share/applications/$(basename "$f")"; done - put it into your bashrc as a function and remember to run it whenever you create a new .desktop launcher, and you should be golden.


  • I don’t think that’s a similar situation - the Linux kernel lost some functionality there, but in this case Ext2 filesystems are still fully supported by the Ext4 driver, so there’s no difference in “hardware” support.

    The separate Ext2 driver was being kept for embedded devices with extreme memory or storage limitations where saving some kilobytes by not having all the new Ext3/4 features was useful, but when you can afford the extra memory, there’s no reason not to just use the Ext4 driver for all Ext2/3/4 filesystems.


  • So, if I understand this correctly, open source means free beer, just not if you sell the end product.

    Yes, once you give the beer to someone, you can’t require any further payments no matter what they do with it. Free software philosophy says users are free to use the software however they wish and for whatever purpose they wish without any barriers (like having to pay for commercial use).

    its all a scam for free work for corpos then. Very disappointing.

    I’m sorry you feel that way, and it’s becoming a not-so-rare sentiment lately (or at least I’ve started noticing it more), but I don’t agree. Look at (A)GPL and how many companies are doing their best to avoid such code - like when Google made their own C library for Android and even stated that its main goal was to avoid copyleft licenses. I’ve also seen plenty of people say that GPL code is pretty much useless for their work due to their company’s policies forbidding its use.

    I also think that revenue-based loyalties screw over small companies the most - sure, you get the donations from the massive companies that can work with 1% of their revenue gone while also keeping it free for non-commercial users, but in my view you also help those same massive corporations by making the software less viable for their smaller competitors who don’t have the economies of scale on their side, and for whom that 1% might legitimately break the bank.

    And to be clear, I don’t mean any of my arguments as some kind of “gotcha! Look, I’m right and you’re wrong”, I just thought I might share my reasoning for why I don’t think your statement is fair.