• 1 Post
  • 681 Comments
Joined 3 years ago
cake
Cake day: July 5th, 2023

help-circle

  • I think it’s worth being clear about the scope of the rating. iFixit has always been about repairability defined by parts availability, and its ratings consider software restrictions only to the point where it interferes with the user experience when replacing parts to restore things to the original performance.

    Customizability (in software or otherwise) isn’t part of the score. Durability/longevity isn’t part of the score, either. Those are things that I want, too, but I can recognize those are outside the scope of what iFixit advocates for.

    I do have some concerns about the partnerships creating a conflict of interest, but sometimes that feedback loop is helpful for improving the product, where the maintainer of a standard also has a consulting business in helping others meet that standard. Ideally there’s a wall between the two sides (advisors versus raters), but the mere fact that one company might do both things isn’t that big of a deal in itself.




  • Yeah. AI is really good at generating footage and stills for these infrared night vision types of surveillance cameras, in large part because they don’t need to be high resolution and are in black and white.

    Because of the way they’re trained on cats and deer have eyes that reflect light back at the camera, these generated images also tend to give that effect even to animals (like humans) that wouldn’t have as pronounced of that effect.


  • Unintuitive locked down garbage that can’t do anything a PC can’t do for half the price.

    From the user perspective, enterprise managed Windows is locked down, too, and somehow less reliable.

    Most of the software engineers I know in FAANG and similar tier companies use Macbooks to program. Poke around a coffee shop in the bay area during a weekday and look around.

    And personally, I switched to Mac about 15 years ago mainly because dependency management and the shell made more sense to me coming from Linux. Windows has always been trash, and most other non-Apple OEMs make the actual physical laptop experience worse (hinges, behavior on closing the lid, trackpad behavior and size, power management, display quality in both brightness and pixel density, webcam/audio behavior).





  • You can reason from a few principles:

    • At its core, the math functions being optimized by these AI tools and their specialized hardware is that they can perform inference and pattern recognition at huge scales across enormous data sets.
    • Inferring a rule set for pattern also allows generation of new data that fits that pattern.
    • Some portion of human cognitive work falls within the general framework of finding patterns or finding new data that fits an old pattern.

    So when people start making claims about things with clear, objective definitions (a win condition in chess, the fastest route to take through a maze, a highest lossless compression algorithm for real world text), it’s reasonable to believe that the current AI infrastructure can lead to breakthroughs on that front. So image recognition, voice recognition, and things like that were largely solved a decade ago. Text generation with clear and simple definitions of good or bad (simple summaries, basic code that accomplishes a clearly defined goal) is what LLMs have been doing well.

    On things that have much more fuzzy or even internally inconsistent definitions, the AI world gets much more controversial.

    But I happen to believe that finding and exploiting bugs or security vulnerabilities falls more into the well defined problem with well defined successes and failures. So I take it seriously when people claim that AI tools are helpful for developing certain exploits.


  • but isn’t the memory on the Neo on the same die as the processor?

    Not actually on the same die, but in the same package, stacked on top using TSMC’s Integrated Fan-Out Package on Package (InFO-PoP).

    So the memory still needs to be sourced from memory manufacturers, sent to TSMC, and then have TSMC package it all together in a single package. It’s unclear whether they had locked up this supply at pre-AI prices, though. The underlying A18 Pro chip/package was annoinced and launched about 18 months ago, so if they had the manufacturing pipeline set up for that they might have kept the contractual rights to continue buying memory at the old prices.




  • On the flip side, I’m a former sysadmin and I only stick around for 5 years because I had the educational credentials to move onto another field (and then another field). I’m glad I did the IT thing in my 20’s, and still like to tinker with homelab stuff 20 years later, but in the end it was a stepping stone towards something else (that does require formal schooling). The degree is a tool that can be used to control on a few more things in your life, in the hopes that you can go where you want to end up.


  • No, it’s not volunteering, at least not anymore.

    Subpoena is legal Latin for “under penalty,” because noncompliance with a subpoena carries a penalty.

    Originally, it was an information request from the feds, and Reddit refused. Then they escalated to getting a grand jury subpoena (which means they got a bunch of normal citizens to agree that the information was relevant to a criminal investigation), so now noncompliance carries a penalty.

    Reddit notified the users, who hired their own lawyers, who are resisting the subpoena and will litigate it to where they need a judge to decide whether Reddit will have to turn the information over.

    That’s the process for these things, and we’re a couple steps in already.





  • LLM companies have argued they should get to ignore all copyright, and now that one of their code leaked, suddenly they care greatly about copyright.

    Anthropic itself has argued that digitizing and using the digitized copies to train models is fair use, so long as:

    • They don’t redistribute the physical copies they bought
    • They don’t allow an end user to retrieve the contents of any one specific work at the user interface (if you ask Claude to spit out the entire text of a copyrighted work used to train it, it is designed to resist copying too much out of a single work)

    So they don’t argue that copyright doesn’t count, exactly. They argue that copyright doesn’t prevent model training from ingesting an entire copyrighted work, as long as it’s done with so many other copyrighted works that any given original isn’t a huge contributor to the model or its outputs.

    There’s tension in their positions, but not so much that it would totally fall apart.