• 0 Posts
  • 142 Comments
Joined 3 years ago
Cake day: July 5th, 2023


  • You can reason from a few principles:

    • At its core, the math functions being optimized by these AI tools and their specialized hardware is that they can perform inference and pattern recognition at huge scales across enormous data sets.
    • Inferring a rule set for pattern also allows generation of new data that fits that pattern.
    • Some portion of human cognitive work falls within the general framework of finding patterns or finding new data that fits an old pattern.

    So when people start making claims about things with clear, objective definitions (a win condition in chess, the fastest route to take through a maze, a highest lossless compression algorithm for real world text), it’s reasonable to believe that the current AI infrastructure can lead to breakthroughs on that front. So image recognition, voice recognition, and things like that were largely solved a decade ago. Text generation with clear and simple definitions of good or bad (simple summaries, basic code that accomplishes a clearly defined goal) is what LLMs have been doing well.

    On things that have much more fuzzy or even internally inconsistent definitions, the AI world gets much more controversial.

    But I happen to believe that finding and exploiting bugs or security vulnerabilities falls more into the well defined problem with well defined successes and failures. So I take it seriously when people claim that AI tools are helpful for developing certain exploits.


  • but isn’t the memory on the Neo on the same die as the processor?

    Not actually on the same die, but in the same package, stacked on top using TSMC’s Integrated Fan-Out Package on Package (InFO-PoP).

    So the memory still needs to be sourced from memory manufacturers, sent to TSMC, and then have TSMC package it all together in a single package. It’s unclear whether they had locked up this supply at pre-AI prices, though. The underlying A18 Pro chip/package was annoinced and launched about 18 months ago, so if they had the manufacturing pipeline set up for that they might have kept the contractual rights to continue buying memory at the old prices.



  • No, it’s not volunteering, at least not anymore.

    Subpoena is legal Latin for “under penalty,” because noncompliance with a subpoena carries a penalty.

    Originally, it was an information request from the feds, and Reddit refused. Then they escalated to getting a grand jury subpoena (which means they got a bunch of normal citizens to agree that the information was relevant to a criminal investigation), so now noncompliance carries a penalty.

    Reddit notified the users, who hired their own lawyers, who are resisting the subpoena and will litigate it to where they need a judge to decide whether Reddit will have to turn the information over.

    That’s the process for these things, and we’re a couple steps in already.


  • According to your POV here, companies can claim whatever and it’s my job now to figure out if they are lying or to what extent.

    No, the actual claims here, that describe specific bugs in specific software, can be evaluated. Even without whipping out a test environment to try to reproduce the results with your own proof of concept, you can read the text and evaluate whether the claims make sense on their face.

    a broken clock is never right, reality momentarily aligns with it, which is a completely different thing

    And that’s why the substance of a statement matters. I don’t believe in the supernatural, so if someone says “I’m a psychic and the missing girl on the news is in a shed near the water,” that doesn’t register with me at all. But if that person says “I’m a psychic and the missing girl is in a shed at 1234 Main Street” that raises eyebrows because it is easily falsifiable. And if the person says “I’m a psychic and the missing girl is in a shed, so I looked and found her and reported it to the cops, and here’s a cryptographic hash of my description of how I found her, which I’ll publish once the cops confirm she’s safe” that’s gonna be a much more serious statement. Even if I don’t believe that the person actually is a psychic, I can pay attention to how the whole thing played out because the person claims serious non-psychic validation of the results, and the results themselves are important entirely externally from the claim of whether psychics have powers.

    This is a story about several cybersecurity vulnerabilities, some of which sound medium or high severity in very commonly used software. That’s important in itself, outside of AI mattering at all. And if they claim to have the receipts in a falsifiable way, that’s the kind of thing that shows a high degree of confidence in the genuineness of what was found.

    I don’t give a shit about AI and I’m generally a skeptic of the future of any of these AI companies. But if someone uses AI tools to discover something new in the subjects that I do care about, like cybersecurity, then I’ll pay attention to the results and what they publish in that field.


  • This is really a corporate problem of their own making and their responsibility to fix. They have lied so much, I do not owe then a single iota of trust.

    The statements can stand for themselves, evaluated on the merits of the claims, regardless of authorship. That’s how these things should work. Someone who has a great history of finding vulnerabilities still has to stand by each exploit/proof of concept they write, on its own merits. On the flip side, the corollary to the adage that a broken clock is still right twice a day is that you can’t just say “oh the broken clock said this so I can ignore it.”

    Do you really think any of them would post something like “yeah, we found a vulnerability but it’s basically a typo that could not be seriously exploited”?

    The blog post literally describes exactly that, for ffmpeg. And several of the other described vulnerabilities sound like they’re in that category of “here’s a bug but we didn’t find an exploit.”

    Simply refusing to engage with these big claims just because of the source is an irresponsible way to approach cybersecurity.

    even if the whole scenario is real, it may not have the intervention of Ai they are claiming

    …who cares? If it’s a real bug and a real PR addressing the bug, why does authorship or methodology matter?

    It’s just the ad hominem fallacy (or the close relative, appeal to authority). Let the actual substance stand and fall on its merits. Read the described vulnerabilities and exploits and decide whether you think those need to be patched and how critical/severe the bugs/vulnerabilities are.

    And maybe your priorities are different from mine, but the core of the claim (we found some vulnerabilities) trigger a responsibility to address them (confirm and patch). I don’t care about marketing or corporate interests or whatever in those circumstances, I’m just focused on fixing problems that have been found.


  • Yes I understand, but I’m also putting the direct claims right there, not filtered through Anthropic’s PR or an article from the IT industry press interpreting those PR statements.

    These are real CVEs that have actually been submitted to the code maintainers for both FOSS and closed source software that is foundational to the computing world. Some of them are published in this post. And many more are simply described with a hash of the full writeup indicating that they have it written out and are waiting for the patches to be applied. I’m especially interested in the Virtual Machine Monitor and the exploits for jumping out of browser sandboxes for “all major browsers.”

    Some of the published CVEs in the blog post seem pretty serious, especially the FreeBSD remote root access for devices running NFS. The OpenBSD one is a critical DOS vector, and the FFMPEG one is just a bug that doesn’t seem to actually expose the software to any practical exploits but should still be patched.

    But they’ve staked it out with their public disclosure of the hashes and a description of a few of the problems. These are big bold claims that are provided in a format that will be easily falsifiable in due time. And treating it as just marketing fluff ignores the shades of gray that actually apply to corporate claims.





  • What if license and copyright was washed by using an LLM to translate Claude into another language?

    The law doesn’t allow you to launder copyright like that. That’s just a derivative work, which can be restricted by the copyright holder in the original. As an example, in fictional writing, distinct characters are copyrighted, and using an LLM to generate new works using those copyrighted characters would still be a derivative work that the original copyright owner would have the right to deny distribution.

    So if you have a copyrighted codebase and you try to implement that codebase using some kind of transformation of that code, that’d still be a derivative work and infringe the original copyright.

    Now if you have some kind of clean room implementation where you can show that it was written without copying the original code itself, only working to implement its functionality through documentation/reverse engineering how the code worked, you’d be able to escape out of calling it a derivative work and could distribute it without the original copyright holder’s permission (Compaq did this with the IBM BIOS to make unauthorized/unlicensed PC clones, and Google did this with the Java API to make Android without a license from Sun/Oracle and won at the Supreme Court).

    Claude can’t be copyrighted because it’s a product of an LLM.

    No, because Claude’s code is still created by humans with the assistance of non-human tools. There’s a spectrum from spelling correction and tab completion in IDEs all the way to full vibe coding with a prompt describing the raw functionality (where the prompt is so uncreative that it isn’t itself copyrightable). Anthropic has never claimed that there was no human in the loop, or that the prompts it uses are so uncreative and purely functional so that the outputs aren’t copyrightable.


  • Unless it can be paper thin this does not look better than magnetic tape.

    As the article explains, the whole purpose here is to be able to store data on a medium that can endure harsh conditions, including heat, moisture, radiation, and physical abrasion. The company’s website claims the medium can retain data for 5000 years without power, and is water and fire resistant.

    I reckon you could scratch it pretty easily.

    The underlying ceramic film is already used for protecting tools like drill bits and saw blades from physical damage, which is why it was chosen for this project. They already found one of the most durable materials in the world, and asked whether they could store data using that already-durable material.



  • The targeted court cases are to argue that the previously passed legislation already covers these particular facts.

    If the legislature passes a law that says “making false statements to another in order to obtain something of value is fraud,” you can expect litigation about the actual contours of what is or isn’t fraud.

    Same with legislation against driving at an unsafe speed, causing a nuisance to your neighbors, discriminating against employees on the basis of sex, etc. Court cases decide the edge cases.

    If the legislature passes a law banning gambling outside of licensed institutions, and banning gambling for minors, you can expect litigation about what actually is or isn’t gambling.



  • It might be possible to use separate accounts related to separate interests

    That’s what people should do. And the natural consequence is that there is code switching, where people subtly use different jargon and references and writing style when talking to different audiences.

    Nobody is gonna correlate my shitposts or joke comments to my work email, because the way I write in a professional environment is totally different from the way I write with my friends and family, or in casual contexts organized around different interests. Even between different friends, family, or colleagues, I have a sense of my audience, and my tone/style differs significantly for different people.

    So at that point, if I have a Linux/technology account and a separate account for the sports I like and a separate account for the local things happening in my city, who’s going to be able to link them by their very different textual styles?



  • Because each sensor broadcasts a fixed unique ID, the same car can be recognized repeatedly without reading a license plate. This makes TPMS-based tracking cheaper, harder to detect, and more difficult to avoid than camera-based surveillance, and therefore a stronger privacy threat.

    This seems like a real stretch.

    Cameras and automated license plate recognition are absurdly cheap at this point. And cameras have much greater range and reliability than whatever wireless signal interception this is, which the researchers have said is effective up to 50 meters.

    Meanwhile, from the office where I sit (which happens to be more than 50 meters above street level), I can see a highway and read the license plates of all the cars maybe 100-300m away. Plug in a cheap phone as a simple webcam and I can probably log all the license plates that drive by, maybe even correlate that to makes and models of vehicles for redundancy.

    And who’s going to detect that I’ve got a cell phone camera pointed out of my office window, or that I’m running that type of image recognition on the phone?