Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

The next major leap in LLMs (in the next year) is probably going to be the prompt context size. Right now we have 2k, 4k, 8k ... but OpenAI also has a 32k model that they're not really giving access to unfortunately.

The 8k model is nice but it's GPT4 so it's slow.

I think the thing that you're missing is that zero shot learning is VERY hard but anything > GPT3 is actually pretty good once you give it some real world examples.

I think prompt engineering is going to be here for a while just because, on a lot of task, examples are needed.

Doesn't mean it needs to be a herculean effort of course. Just that you need to come up with some concrete examples.

This is going to be ESPECIALLY true with Open Source LLMs that aren't anywhere near as sophisticated as GPT4.

In fact, I think there's a huge opportunity to use GPT4 to train the prompts of smaller models, come up with more examples, and help improve their precision/recall without massive prompt engineering efforts.



>> The next major leap in LLMs (in the next year) is probably going to be the prompt context size. Right now we have 2k, 4k, 8k ... but OpenAI also has a 32k model that they're not really giving access to unfortunately.

Saw this article today about a different approach that opens up orders of magnitude larger contexts

https://hazyresearch.stanford.edu/blog/2023-03-07-hyena


You can’t commercially use anything you train off OpenAI outputs.


This is just ToS violation which will just result in loss of access to OpenAI. There is nothing they can do to stop you from commercially competing, given there is no copyright law precedence


You can as long as the resulting model does not compete with OpenAI.


It can be argued that if you build a model using their outputs such that you can then stop using their API, your model is effectively competing with their’s. Let’s just say that if you’re a startup or SMB, you do not want to be the one dragged to court to iron out whether this holds or not.


Can you elaborate?


They're probably talking about the TOS a user would've had to agree to when using their services. It's actually a lot more permissive then I expected

> Restrictions. You may not (i) use the Services in a way that infringes, misappropriates or violates any person’s rights; (ii) reverse assemble, reverse compile, decompile, translate or otherwise attempt to discover the source code or underlying components of models, algorithms, and systems of the Services (except to the extent such restrictions are contrary to applicable law); (iii) use output from the Services to develop models that compete with OpenAI;


> use output from the Services to develop models that compete with OpenAI

It can be argued that if you build a model using their outputs such that you can then stop using their API, your model is effectively competing with their’s.

Let’s just say that if you’re a startup or SMB, you do not want to be the one dragged to court to iron out whether this holds or not.


Their API TOS basically forbid it. Simple as that.


Someone who acquires these outputs who has never consented to their ToS is not bound by their ToS.


Sure, but the ways of acquiring those outputs legally have vampiric licensing that bind you to those ToS, since the re-licenser is bound by the original ToS.

It's like distributing GPL code in a nonfree application. Even if you didn't "consent to [the original author's] ToS," you are still going to be bound to it via the redistributors license.


There’s no license. OpenAI is not an author of their models’ outputs, and they know it.

OpenAI can’t just start suing random people on the street without any legal basis. That’s how lawyers become not-lawyers.

There’s only a (somewhat dubiously enforceable) ToS contract between OpenAI and the user of OpenAI’s website. This is probably bullshit too - what legitimate interest does OpenAI have in a model output that doesn’t even belong to them, but it’s less obviously bullshit.


> Even if you didn't "consent to [the original author's] ToS," you are still going to be bound to it via the redistributors license.

In the context of the GPL, are there real examples of judgements which bind defendants to a license they never saw or knew anything about, because of the errant actions of an intermediary?


It doesn’t matter.

It gives OpenAI a legal basis to launch a law suit if they want to.

Would it succeed? Is it right? Do they care? Eh.

…but, if I as some random reddit user say I might sue you for making a LLM you for training on data that may or may not have my posts in it, you can probably safely ignore me.

If you go and build a massive LLM using high quality data that couldn’t possibly come from anywhere other than openai, and they have a log of all the content that api key XXX generated; they both know and have a legal basis for litigation.

There’s a difference, even if you’re a third party (not the owner of the api key) or don’t care.

(And I’m not saying they would, or even they would win; but it’s sufficient cause for them to be able to make a case if they want to)


Just being able to make a case doesn't mean they will consider the legal fees and resulting judgment to be valuable enough to their business, nor that the suit will even make it into the courts resulting in a final judgment.

A lot of behavior that rides this line is rationalized via a careful cost-benefit analysis.


Sure, I'm just saying that in that cost-benefit analysis the 'risk of case failing and getting nothing from it' is significantly lower; it's your call as a consumer to do your due diligence and decide:

"I don't think they'll be bothered chasing after me"

vs.

"If it came to it I think the court will rule that they don't have a case after we play the pay-the-lawyers-all-the-money game"

vs.

"how screwed am I if they do, I lose and I have clearly, blatantly and provably violated their terms of service"

^

...because, and this is the point I'm making. There is no question; it is very very obvious if you do this, and it's not very difficult for them to prove it.

All they need is to slap you with a discovery order, look at your training data, and compare it to their output logs.


That means they can refuse you service. That's all. You can use or share the output


No you are not, the outputs of an LLM are afforded no intellectual property rights protections.

The internet is increasingly covered with OpenAI's outputs and they do not have grounds for a claim to prevent people from using them.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: