• 0 Posts
  • 22 Comments
Joined 9 months ago
cake
Cake day: February 4th, 2024

help-circle
  • You can install Plex on your mobile device and toggle the “share media from this device” setting. Otherwise, a steam deck would have everything an RPI has plus a GPU and a touch screen. Since there are two radios (2 and 5Ghz) on the device, you should be able to set it up as a bridge device, but I’ve not tried this personally.



  • Yeah. I’m thinking more along the lines of research and open models than anything to do with OpenAI. Fair use, above all else, generally requires that the derivative work not threaten the economic viability of the original and that’s categorically untrue of ChatGPT/Copilot which are marketed and sold as products meant to replace human workers.

    The clean room development analogy is definitely an analogy I can get behind, but raises further questions since LLMs are multi stage. Technically, only the tokenization stage will “see” the source code, which is a bit like a “clean room” from the perspective of subsequent stages. When does something stop being just a list of technical requirements and veer into infringement? I’m not sure that line is so clear.

    I don’t think the generative copyright thing is so straightforward since the model requires a human agent to generate the input even if the output is deterministic. I know, for example, Microsoft’s Image Generator says that the images fall under creative Commons, which is distinct from public domain given that some rights are withheld. Maybe that won’t hold up in court forever, but Microsoft’s lawyers seem to think it’s a bit more nuanced than “this output can’t be copyrighted”. If it’s not subject to copyright, then what product are they selling? Maybe the court agrees that LLMs and monkeys are the same, but I’m skeptical that that will happen considering how much money these tech companies have poured into it and how much the United States seems to bend over backwards to accommodate tech monopolies and their human rights violations.

    Again, I think it’s clear that commerical entities using their market position to eliminate the need for artists and writers is clearly against the spirit of copyright and intellectual property, but I also think there are genuinely interesting questions when it comes to models that are themselves open source or non-commercial.


  • For example, if I ask it to produce python code for addition, which GPL’d library is it drawing from?

    I think it’s clear that the fair use doctrine no longer applies when OpenAI turns it into a commercial code assistant, but then it gets a bit trickier when used for research or education purposes, right?

    I’m not trying to be obtuse-- I’m an AI researcher who is highly skeptical of AI. I just think the imperfect compression that neural networks use to “store” data is a bit less clear than copy/pasting code wholesale.

    would you agree that somebody reading source code and then reimplenting it (assuming no reverse engineering or proprietary source code) would not violate the GPL?

    If so, then the argument that these models infringe on right holders seems to hinge on the verbatim argument that their exact work was used without attribution/license requirements. This surely happens sometimes, but is not, in general, a thing these models are capable of since they’re using loss-y compression to “learn” the model parameters. As an additional point, it would be straightforward to then comply with DMCA requests using any number of published “forced forgetting” methods.

    Then, that raises a further question.

    If I as an academic researcher wanted to make a model that writes code using GPL’d training data, would I be in compliance if I listed the training data and licensed my resulting model under the GPL?

    I work for a university and hate big tech as much as anyone on Lemmy. I am just not entirely sure GPL makes sense here. GPL 3 was written because GPL 2 had loopholes that Microsoft exploited and I suspect their lawyers are pretty informed on the topic.


  • I hate big tech too, but I’m not really sure how the GPL or MIT licenses (for example) would apply. LLMs don’t really memorize stuff like a database would and there are certain (academic/research) domains that would almost certainly fall under fair use. LLMs aren’t really capable of storing the entire training set, though I admit there are almost certainly edge cases where stuff is taken verbatim.

    I’m not advocating for OpenAI by any means, but I’m genuinely skeptical that most copyleft licenses have any stake in this. There’s no static linking or source code distribution happening. Many basic algorithms don’t follow under copyright, and, in practice, stack overflow code is copy/pasted all the time without that being released under any special license.

    If your code is on GitHub, it really doesn’t matter what license you provide in the repository – you’ve already agreed to allowing any user to “fork” it for any reason whatsoever.



  • and my point was explaining that that work has likely been done because the paper I linked was 20 years old and they talk about the deep connection between “similarity” and “compresses well”. I bet if you read the paper, you’d see exactly why I chose to share it-- particularly the equations that define NID and NCD.

    The difference between “seeing how well similar images compress” and figuring out “which of these images are similar” is the quantized, classficiation step which is trivial compared to doing the distance comparison across all samples with all other samples. My point was that this distance measure (using compressors to measure similarity) has been published for at least 20 years and that you should probably google “normalized compression distance” before spending any time implementing stuff, since it’s very much been done before.


  • I think there’s probably a difference between an intro to computer science course and the PhD level papers that discuss the ability of machines to learn and decide, but my experience in this is limited to my PhD in the topic.

    And, no, textbooks are often not peer reviewed in the same way and generally written by graduate students. They have mistakes in them all the time. Or grand statements taken out of context. Or are simplified explanations because introducing the nuances of PAC-learnability to somebody who doesn’t understand a “for” loop is probably not very productive.

    I came here to share some interesting material from my PhD research topic and you’re calling me an asshole. It sounds like you did not have a wonderful day and I’m sorry for that.

    Did you try learning about how computers learn things and make decisions? It’s pretty neat


  • You seem very upset, so I hate to inform you that neither one of those are peer reviewed sources and that they are simplifying things.

    “Learning” is definitely something a machine can do and then they can use that experience to coordinate actions based on data that is inaccesible to the programmer. If that’s not “making a decision”, then we aren’t speaking the same language. Call it what you want and argue with the entire published field or AI, I guess. That’s certainly an option, but generally I find it useful for words to mean things without getting too pedantic.


  • Yeah. I understand. But first you have to cluster your images so you know which ones are similar and can then do the deduplication. This would be a powerful way to do that. It’s just expensive compared to other clustering algorithms.

    My point in linking the paper is that “the probe” you suggested is a 20 year old metric that is well understood. Using normalized compression distance as a measure of Kolmogorov Complexity is what the linked paper is about. You don’t need to spend time showing similar images will compress more than dissimilar ones. The compression length is itself a measure of similarity.




  • Agree to disagree. Something makes a decision about how to classify the images and it’s certainly not the person writing 10 lines of code. I’d be interested in having a good faith discussion, but repeating a personal opinion isn’t really that. I suspect this is more of a metaphysics argument than anything and I don’t really care to spend more time on it.

    I hope you have a wonderful day, even if we disagree.







  • We tried that too. That was the USSR. It famously didn’t work out so well either, despite the superior working hours, vacation time, and education level when compared to the west. Even with well intentioned people acting on the best of intentions, there tends to be unpredictable side effects.

    How Amsterdam accidentally created a violence and crime ridden ghetto by building consolidated public housing:

    https://youtu.be/sJsu7Tv-fRY?si=1xAsX2Ipvu6L6ybN

    Same idea, but Chicago:

    https://youtu.be/_CogQmmBL9k?si=pSuz9dvf6G4vC8Qk

    Yugoslavia:

    https://en.m.wikipedia.org/wiki/Western_City_Gate

    Granted, that’s not always the case. The “Kruschevkas” (commie blocks) from the middle period of the USSR are still in use across the former USSR and seem to be working well enough, especially since they were intended to be demolished in the 80s. But there wasn’t really income disparity like we see the west, so it’s hard to compare directly. We do know that over the long term, people got demotivated by shortages of luxury goods, tvs, chocolate, etc and this lead to a vicious cycle of shortages, being corrupt to circumvent the shortages, and demotivation because even if you worked really hard and saved your money, there wasn’t really anything to buy.

    By the time they opened the borders, it was already doomed because, unfortunately, the average society citizen liked chocolate more than they liked the ideal of not relying on slavery and child labor to make exotic candy.


  • I agree with the platitude, but subsidizing interest isn’t giving money to poor people-- it’s cutting a check to a bank, using tax money that rich people avoid paying. If you want to increase the supply of cheap houses, you have to build cheap houses. If you want to lower the price of housing, artificially injecting a bunch of money into a market is going to raise demand and do nothing for the supply. Nevermind that in the US in particular, the problems exacerbated by a massive demographic shift towards dense urban centers that were not bombed out and rebuilt in the mid 20th century and therefore do not have housing stock appropriate for the lifestyles of today.

    You also ignored the comment that subsidized interest is already an existing thing in the US and that the last time they tried giving houses to poor people that they could not afford to maintain, it crashed the global economy and was the single greatest wealth transfer towards hedge funds in history. NYC’s newest policy (after moving away from the structural failure of consolidated public housing) is that all new construction must have a certain number of low income, middle income, and high income units. That’s an actual supply side solution and the ratios and income thresholds are determined by the already existing demographics of the neighborhood as a way to fight both gentrification and white flight.

    I’d be happy to have an actual policy discussion rooted in facts, but you can’t wave your hand over a stack of tax revenue and solve structural problems about housing stock, centuries of segregation, and the horrors of capitalism simply by wishing it away. There’s a lot of actual work to be done, and some of that might mean replacing historical housing districts with high density, mixed income developments. But that won’t ever happen, because the generations before us relied on home ownership as an investment vehicle for retirement and have spent decades campaigning against stuff like public transit, low income housing, and the existence of homeless people anywhere near them.

    Feel free to diagree, but I’d rather give my money to a homeless person directly than subsidize the actuarial risk of a billion dollar investment bank via the coervice use of state violence if I decided not to pay my taxes.

    Like, if you can’t come up with 3% of the value of a house (the down payment standard for subsidized homeowners in the US) how are you supposed to repair a roof or replace a major appliance? More government grants? From what revenue?

    if the proposal is to disposses all the property owned by billionaires and use that to fund a reimagining of our cities such that they’re built for people and not cars, I’m game. If your idea is to raise taxes on working people to subsidize other working people while the banks, the construction firms, and every material vendor takes a profit, then fuck that.

    Most of Sweden’s housing is owned by local governments (kommun in Swedish), but it’s also a 5 year wait list to get anything, impossible for immigrants/students to find anything because they don’t qualify for the queue immediately, and tied to a specific place. So, if you get a better job somewhere else, you’re still fucked because you would have needed to get on their queue 5 years ago or pay the inflated “market” rent because the supply of non-state-owned housing is so low. Denmark and the Netherlands have the same problem, but perhaps that’s more understandable considering their absolute size and population density.

    As a side point, are you aware that the government has relied on deficit spending for more than a decade and that every new dollar spent is a dollar + interest that must be paid back to an investment fund?

    For example, every time you ride the subway in NYC, you’re paying something like $.30 to Chase bank to service debt the MTA used to renovate the system in the 70s.