Here is what GPT-J 6B know about purple CIA files:
the CIA purple files are about aliens as well as government cover-up of UFO facts. But there are more files in that huge storehouse that I need to see and read about.
They all involve US government cover-up of UFO facts, and the secret of what is being covered up.
And that’s the key.
What kind of secret would the CIA have about an UFO?
And the same goes for the NSA, the FBI and the army.
And the FBI?
And the military?
What would any of them do to hide such a secret?
Looks like GPT-J 6B is more sensitive, worried and sincere. :)
Really, there is nothing common with AI at all. Еxaggerating, it is just a program that runs over giant database and try to find most relevant continuation for the previous phrase.
As for GPT-3 model, sooner or later it will inevitable leak out, so you will have a chance to play with it.
I thought you had replied in that philosopher AI thread.
No. I replied only here.
What do you got on this philosopher AI
Philosopher AI is not a philosopher and not an AI. It is interface to simple neural network that is pre-teached to predict next word in a sentence. It is just a text syntesizer. It is trained on many texts and result of that training named GPT-3. OpenAI (not open and not AI) sell this closed pre-trained model for money using modern hype about AI.
What's this GPT-J 6B?
There are another Generative Pre-trained Transformer (GPT) models, open and free, f.e. GPT-2 and GPT-J you could easily play with. GPT-J 6B is larger than GPT-3. GPT-J 6B is a GPT-J transform model that has 6 billion trainable parameters. https://huggingface.co/EleutherAI/gpt-j-6B
I'd like to use something where it wouldn't say "nonsense" all the time.
You have to spend some time to get something not "nonsense" from any GPT. Examples that you saw about Philosopher AI (GPT-3) is just a little part selected from a huge pile of nonsense answers. Marketing, you know...
Coefficients in neural network. Hard to tell what they adjust exactly, but you are free to fine tune them and see what happens. There are tooltips if you hover cursor over the field, but they are not very informative, except recommended ranges for the value.
"seed" is a number used to "seed" neural network. If set to 0 it will use random seed for every pass. You will always get different completion texts. If you set it to some numeric value, it will use that value every time and you will always get same text for same phrase. If you want to play with other settings, set seed to some value, to view how settings will change output, otherwise you will get different texts each time and will not notice changes coming from changing other settings.
Where it seemed to have access to government 3 letter agency databases.
It does not have access to anything, really.
GPT just try to continue text in the way it trained. That's all. It's like somebody read all texts on Earth and figure out the most probable words sequences that should follow some already existing sequence. Nothing more, nothing less. There are no any texts at all, just, say, a list of all words and neural network coefficients that select next words sequence using previous sequence as input.
There are no any magic or even some complex things. It's just a simple pre-trained neural network that sets output according to input as trained.
You want a site that an idiot can type stuff in, and doesn't have to do code.
Here it is - https://bellard.org/textsynth/
Just type in anything you want and press "Complete text". That's all. If you want go further, you could play with parameters Fabrice made changeable. You will get different results with different settings.
is it going to be the same databases as that one philosopher AI was using?
It depends are OpenAI lied about source or not. If they tell a truth, both GPT-2 and GPT-J trained on the same data. It's hard to check, really.
The thing was really fishy because it would say "sensitive information"
Really it is hard to tell anything about that information. AFAIK, "The Pile" which was used to train GPTs is a 800Gb of different texts, including SciFi and other things. So, when GPT tries to complete text, it could easily generate a sequence that is trained from SciFi or Conspiracy books, or any other text that get into "The Pile", including possible confidential documents or their parts from books.
You can download "The Pile" and try to check it by yourself, but it is >800Gb, so it could take time. :)
It's too late, but cheers!
Here is what GPT-J 6B know about purple CIA files:
Looks like GPT-J 6B is more sensitive, worried and sincere. :)
Really, there is nothing common with AI at all. Еxaggerating, it is just a program that runs over giant database and try to find most relevant continuation for the previous phrase.
As for GPT-3 model, sooner or later it will inevitable leak out, so you will have a chance to play with it.
No. I replied only here.
Philosopher AI is not a philosopher and not an AI. It is interface to simple neural network that is pre-teached to predict next word in a sentence. It is just a text syntesizer. It is trained on many texts and result of that training named GPT-3. OpenAI (not open and not AI) sell this closed pre-trained model for money using modern hype about AI.
There are another Generative Pre-trained Transformer (GPT) models, open and free, f.e. GPT-2 and GPT-J you could easily play with. GPT-J 6B is larger than GPT-3. GPT-J 6B is a GPT-J transform model that has 6 billion trainable parameters. https://huggingface.co/EleutherAI/gpt-j-6B
You are not limited in length here - https://bellard.org/textsynth/
You have to spend some time to get something not "nonsense" from any GPT. Examples that you saw about Philosopher AI (GPT-3) is just a little part selected from a huge pile of nonsense answers. Marketing, you know...
Coefficients in neural network. Hard to tell what they adjust exactly, but you are free to fine tune them and see what happens. There are tooltips if you hover cursor over the field, but they are not very informative, except recommended ranges for the value.
"seed" is a number used to "seed" neural network. If set to 0 it will use random seed for every pass. You will always get different completion texts. If you set it to some numeric value, it will use that value every time and you will always get same text for same phrase. If you want to play with other settings, set seed to some value, to view how settings will change output, otherwise you will get different texts each time and will not notice changes coming from changing other settings.
It does not have access to anything, really. GPT just try to continue text in the way it trained. That's all. It's like somebody read all texts on Earth and figure out the most probable words sequences that should follow some already existing sequence. Nothing more, nothing less. There are no any texts at all, just, say, a list of all words and neural network coefficients that select next words sequence using previous sequence as input.
There are no any magic or even some complex things. It's just a simple pre-trained neural network that sets output according to input as trained.
Here it is - https://bellard.org/textsynth/ Just type in anything you want and press "Complete text". That's all. If you want go further, you could play with parameters Fabrice made changeable. You will get different results with different settings.
It depends are OpenAI lied about source or not. If they tell a truth, both GPT-2 and GPT-J trained on the same data. It's hard to check, really.
Really it is hard to tell anything about that information. AFAIK, "The Pile" which was used to train GPTs is a 800Gb of different texts, including SciFi and other things. So, when GPT tries to complete text, it could easily generate a sequence that is trained from SciFi or Conspiracy books, or any other text that get into "The Pile", including possible confidential documents or their parts from books. You can download "The Pile" and try to check it by yourself, but it is >800Gb, so it could take time. :)