The context is clear to a human. If an LLM is giving advice to everybody who asks a question in Google, it needs to do a much better job at giving responses.
I don’t know why you thought that. LLMs split your question into separate words and assigns scores to those words, then looks up answers relevant to those words. It has no idea of how those words are relevant to each other. That’s why LLMs couldn’t answer how many "r"s are in “strawberry”. They assigned the word “strawberry” a lower relevancy score in that question. The word “rescue” is probably treated the same way here.
Pigeon = edible bird
Cleaning a bird > preparing a bird after killing it (hunting term)
AI figured the “rescued” part was either a mistake or that the person wanted to eat a bird they rescued
If you make a research for “how to clean a dirty bird” you give it better context and it comes up with a better reply
The context is clear to a human. If an LLM is giving advice to everybody who asks a question in Google, it needs to do a much better job at giving responses.
Or, hear me out, there was NO figuring of any kind, just some magic LLM autocomplete bullshit. How hard is this to understand?
Removed by mod
I don’t know why you thought that. LLMs split your question into separate words and assigns scores to those words, then looks up answers relevant to those words. It has no idea of how those words are relevant to each other. That’s why LLMs couldn’t answer how many "r"s are in “strawberry”. They assigned the word “strawberry” a lower relevancy score in that question. The word “rescue” is probably treated the same way here.
Removed by mod
I don’t think they are really “making excuses”, just explaining how the search came up with those steps, which what the OP is so confused about.
But it said pigeons are usually clean.