They don’t actually understand what you’re asking for so they aren’t going to go do the task. They’ll give whatever answer seems plausible based on what everyone else in their training data has said. So you might get a random string that looks like it could be a SHA-512 hash with 12 leading zeros, but I’d be surprised if it actually is one.
They don’t understand but they are way better than youre making them out to be. I’m pretty sure chatgpt would give you the python code for this task, run it and then time out.
They don’t actually understand what you’re asking for so they aren’t going to go do the task. They’ll give whatever answer seems plausible based on what everyone else in their training data has said. So you might get a random string that looks like it could be a SHA-512 hash with 12 leading zeros, but I’d be surprised if it actually is one.
They don’t understand but they are way better than youre making them out to be. I’m pretty sure chatgpt would give you the python code for this task, run it and then time out.
it would only give you python code if you asked for it