Is not the answer here the same? Build the trust in the AI systems. Benchmarks help and are a start. Consistent results over time are another measure. We'll learn over the next few years which models, or which companies developing model can be trusted for certain sets of tasks. If these exceed alternatives, and have consistency over a margin I can reason about, I'll make use of them, same as any other tool.
I can ask the exact same question of an LLM multiple times and get different answers with the same degree of confidence. Hard to trust that, and also hard to fix.
Which wouldn’t be so problematic if people didn’t just turn off their brains when interacting with them.
Either way, everything you’re suggesting are possibilities for the future, which may or may not pan out. The bad comparisons to humans are happening today.