Tried 4.7 on a few of my regular workloads. The quality ceiling is definitely higher than 4.6 when it actually engages — but that's the problem. "Adaptive thinking" seems to actively avoid thinking on tasks where I'd expect it to reason carefully, and I end up getting flat, fast answers where I wanted depth.
Turning off adaptive thinking and bumping effort to high gets me closer to what I want, but at that point the token cost becomes hard to justify vs. just using a smaller model with explicit CoT. Feels like Anthropic is solving a cost optimization problem and calling it a feature.