Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

So the public access one isn't private and secure?



The concern is that ChatGPT is training on your chats (by default, you can opt out but you lose chat history last I checked).

So in general enterprises cannot allow internal users to paste private code into ChatGPT, for example.


As an example of this. I found that GPT4 wouldn't agree with me that C(A) = C(AA^T) until I explained the proof. A few weeks later it would agree in new chats and would explain using the same proof I did presented the same way.


I’ve found that the behavior of ChatGPT can vary widely from session to session. The recent information about GPT4 being a “mixture of experts” might also be relevant.

Do we know that it wouldn’t have varied in its answer by just as much, if you had tried in a new session at the same time?


There is randomness even at t=0, there was another HN submission about that


I tested it several times, new chats never got this right at first. I tried at least 6 times. I was experimenting and found that GPT4 couldn't be fooled by faulty proofs. Only a valid proof could change its mind.

Now it seems to know this mathematical property from first prompt though.


This is kinda creepy. But at the same time, how do they do that? I thought the training of these models stopped in September 2021/2022. So how do they do these incremental trainings?


All the public and (leaked) private statements I have seen state that this is not happening. As siblings noted, MoE probably explains this variance.

AIUI they are using current chat data for training GPT-5, not re-finetuning the existing models.


The exact phrase they previously used on the homepage was "Limited knowledge of world and events after 2021" - so maybe as a finetune?


but doesn’t finetuning result in forgetting previous knowledge? it seems that finetuning is most usable to train “structures” not new knowledge. am i missing something?


Kind of implies that OpenAI are lying and using customer input to train their models


Unless you have an NDA with Open AI, you are giving them whatever you put in that prompt.


Also, at some point some users ended up with other users’ chat history [0]. So they’ve proven to be a bit weak on that side.

[0]: https://www.theverge.com/2023/3/21/23649806/chatgpt-chat-his...


> However, ChatGPT risks exposing confidential intellectual property.

I don't remember seeing this disclaimer on the ChatGPT website, gee maybe OpenAI should add this so folks stop using it.


If you use ChatGPT through the app or website they can use the data for training, unless you turn it off. https://help.openai.com/en/articles/5722486-how-your-data-is...


Providing my data for training doesn't imply that it risks being exposed.

If you understand what happens on a technical level, it might be possible, but OpenAI has never said this was a risk by using their product.


Absolutely. For example it doesn't say that OpenAI employees can't look at everything you write.


It's pretty clear in the FAQ to be fair.


The comment you are responding to is sarcastic


I believe it’s implying the free ChatGPT collects data and this one doesn’t.


I thought sama said they don’t use data going through the api for training. Guess we can’t trust that statement


That is correct, they do not use the data going through the API for training, but they do use the data from the web and mobile interfaces (unless you explicitly turn it off).


“We don’t water down your beer”.

Oh nice!

“But that is lager”


Another thing is that using ChatGPT for European companies might be in violation with GDPR – Azure OpenAI Services are available on European servers.


No

Edit: yes


I just love this comment.




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: