I spent all day prompting GPT-4 and I am afraid to tell you it still has the hallucination problem and the "inventing false libraries" problem. I prompted it for knowledge of an obscure book series and it just made up characters. I promoted it for knowledge from the field I work in and it produced subtly incorrect results.
It has no way of knowing if what it says is true, has no indication of its confidence in any topic, and all the text it produces has a distinctive boilerplate feel.
The ONLY thing it can do well in my experience is bugfixing in code, and generating boilerplate/common code. You NEED an experienced dev to look over the code it makes, and the scientific facts it conjures up are often subtly wrong. It is better than GPT-3 at making convincing sentences, but not correct ones.
It ISN'T over, we still have a couple more years I think before it can replace PhD level knowledge. Nothing to worry about yet.