• 0 Posts
  • 33 Comments
Joined 2 years ago
cake
Cake day: July 14th, 2023

help-circle
  • Why should we know this?

    Not watching that video for a number of reasons, namely that ten seconds in they hadn’t said anything of substance, their first claim was incorrect (Amazon does not prohibit use of gen ai in books, nor do they require its use be disclosed to the public, no matter how much you might wish it did), and there was nothing in the description of substance, which in instances like this generally means the video will largely be devoid of substance.

    What books is the Math Sorcerer selling? Are they the ones on Amazon linked from their page? Are they selling all of those or just promoting most of them?

    Why do we think they were generated with AI?

    When you say “generated with AI,” what do you mean?

    • Generated entirely with AI, without even editing? Then why do they have so many 5 star reviews?
    • Generated with AI and then heavily edited?
    • Written partly by hand with some pieces written by unedited GenAI?
    • Written partly by hand with some pieces written by edited GenAI?
    • AI was used for ideation?
    • AI was used during editing? E.g., Grammarly?
    • GenAI was used during editing?E.g., “ChatGPT, review this chapter and give me any feedback. If sections need rewritten go ahead and take a first pass.”
    • AI might have been used, but we don’t know for sure, and the issue is that some passages just “read like AI?”

    And what’s the result? Are the books misleading in some way? That’s the most legitimate actual concern I can think of (I’m sure the people screaming that AI isn’t fair use would disagree, but if that’s the concern, settle it in court).


  • Look up “LLM quantization.” The idea is that each parameter is a number; by default they use 16 bits of precision, but if you scale them into smaller sizes, you use less space and have less precision, but you still have the same parameters. There’s not much quality loss going from 16 bits to 8, but it gets more noticeable as you get lower and lower. (That said, there’s are ternary bit models being trained from scratch that use 1.58 bits per parameter and are allegedly just as good as fp16 models of the same parameter count.)

    If you’re using a 4-bit quantization, then you need about half that number in VRAM. Q4_K_M is better than Q4, but also a bit larger. Ollama generally defaults to Q4_K_M. If you can handle a higher quantization, Q6_K is generally best. If you can’t quite fit it, Q5_K_M is generally better than any other option, followed by Q5_K_S.

    For example, Llama3.3 70B, which has 70.6 billion parameters, has the following sizes for some of its quantizations:

    • q4_K_M (the default): 43 GB
    • fp16: 141 GB
    • q8: 75 GB
    • q6_K: 58 GB
    • q5_k_m: 50 GB
    • q4: 40 GB
    • q3_K_M: 34 GB
    • q2_K: 26 GB

    This is why I run a lot of Q4_K_M 70B models on two 3090s.

    Generally speaking, there’s not a perceptible quality drop going to Q6_K from 8 bit quantization (though I have heard this is less true with MoE models). Below Q6, there’s a bit of a drop between it and 5 and then 4, but the model’s still decent. Below 4-bit quantizations you can generally get better results from a smaller parameter model at a higher quantization.

    TheBloke on Huggingface has a lot of GGUF quantization repos, and most, if not all of them, have a blurb about the different quantization types and which are recommended. When Ollama.com doesn’t have a model I want, I’m generally able to find one there.


  • I recommend a used 3090, as that has 24 GB of VRAM and generally can be found for $800ish or less (at least when I last checked, in February). It’s much cheaper than a 4090 and while admittedly more expensive than the inexpensive 24GB Nvidia Tesla card (the P40?) it also has much better performance and CUDA support.

    I have dual 3090s so my performance won’t translate directly to what a single GPU would get, but it’s pretty easy to find stats on 3090 performance.






  • stuck with the GPL forever

    If you accept a patch and don’t have the ability to relicense it, you can remove it and re-license the new codebase. You can even re-implement changes made by the patch in many cases, whether those changes are bug fixes or new features.

    If you re-implement the change, you do need to ensure this is done in a way that doesn’t cause it to become a derivative work, but it’s much easier if you have copyright to 99% of a work already and only need to re-implement 1% or so. If you’ve received substantial community contributions and the community is opposed to relicensing, it will be much harder to do so.

    A clean room implementation - where the person rewriting the code doesn’t look at the original code, and is only given a description of the functionality - which can include a detailed description of the algorithm - is the most defensible way to perform such a rewrite and relicense, but it’s not the only option.

    You should generally consult an attorney when relicensing and shouldn’t just do it casually. But a single patch certainly doesn’t mean you’re locked in forever.









  • Further, “Whether another user actually downloaded the content that Meta made available” through torrenting “is irrelevant,” the authors alleged. “Meta ‘reproduced’ the works as soon as it made them available to other peers.”

    Is there existing case law for what making something “available” means? If I say “Alright, I’ll send you this book if you want, just ask,” have I made it available? What if, when someone asks, I don’t actually send them anything?

    I’m thinking outside of contexts of piracy and torrenting, to be clear - like if a software license requires you to make any changed versions available to anyone who uses the software. Can you say it’s available if your distribution platform is configured to prevent downloads?

    If not, then why would it be any different when torrenting?

    Meta ‘reproduced’ the works as soon as it made them available to other peers.

    The argument that a copyrighted work has been reproduced when “made available,” when “made available” has such a low bar is also perplexing. If I post an ad on Craigslist for the sale of the Mona Lisa, have I reproduced it?

    What if it was for a car?

    I’m selling a brand new 2026 Alfa Romeo 4E, DM me your offers. I’ve now “reproduced” a car - come at me, MPAA.



  • It’s okay, the author of the article didn’t actually read (or understand) the Copyright Office’s recommendations. They are:

    Based on an analysis of copyright law and policy, informed by the many thoughtful comments in response to our NOI, the Office makes the following conclusions and recommendations:

    • Questions of copyrightability and AI can be resolved pursuant to existing law, without the need for legislative change.
    • The use of AI tools to assist rather than stand in for human creativity does not affect the availability of copyright protection for the output.
    • Copyright protects the original expression in a work created by a human author, even if the work also includes AI-generated material.
    • Copyright does not extend to purely AI-generated material, or material where there is insufficient human control over the expressive elements.
    • Whether human contributions to AI-generated outputs are sufficient to constitute authorship must be analyzed on a case-by-case basis.
    • Based on the functioning of current generally available technology, prompts do not alone provide sufficient control.
    • Human authors are entitled to copyright in their works of authorship that are perceptible in AI-generated outputs, as well as the creative selection, coordination, or arrangement of material in the outputs, or creative modifications of the outputs.
    • The case has not been made for additional copyright or sui generis protection for AI- generated content.

    Pretty much everything the article’s author stated is contradicted by the above.



  • From the feature comparison at https://github.com/meichthys/foss_note_apps only two FOSS apps support handwriting: Joplin (with a plugin) which gets a subjective 6/10 score, and TriliumNext, which gets a subjective 2/10 score. I personally dislike Joplin but many people love it, so I recommend giving it a shot. EDIT: I installed Joplin using the APK from the site and both the handwriting and Excalidraw plugins were “not available on mobile,” so I have to rescind my recommendation. On my iOS device, the plugins didn’t even show up in the search.

    I think TriliumNext is great, but the mobile experience is still lacking (though they are tracking several issues to improve here). There’s no dedicated mobile app but they at least have a PWA. It also needs to be self-hosted, but doing so is straightforward if you’re already using Docker. The handwriting is done via a built-in Excalidraw integration.

    Here are some options not captured in that list:

    Obsidian is not open source, but also has an Excalidraw plugin. I’ve not used it yet but I’ve seen multiple discussions saying that it’s very well done and has additional functionality on top of base Excalidraw. There’s also an open source (MIT) plugin for Obsidian that adds support for handwritten notes. I only use Obsidian on my work computer and haven’t used it either, though I plan to install the Excalidraw plugin Monday.

    StylusLabs Write is FOSS (AGPL 3.0), multiplatform, and has a free Android apk available. Note that the Google Play version has had updates suspended. I just learned about it and don’t know how it otherwise measures up. I’m planning to check it out, though.

    You can use any note app that has Excalidraw support, so long as you don’t need your handwritten text to be OCRed. That means that the following are all options: