Gemini 2.5: Our most intelligent models are getting even better

  • The new 2.5 Pro (05-06) definitely does not have any sort of meaningful 1 million context window, as many users have pointed out. It does not even remember to generate its reasoning block at 50k+ tokens.

    Their new pro model seemed to just trade off fluid intelligence and creativity for performance on closed-end coding tasks (and hence benchmarks), which unfortunately seems to be a general pattern for LLM development now.

  • I wish Google would provide a WebRTC endpoint for their Live mode like Open AI does for their Realtime models [1]. Makes it so much easier to deploy without needing something like LiveKit or Pipecat.

    1. https://platform.openai.com/docs/guides/realtime#connect-wit...

  • I think it’s pretty strange how time and time again I see the scores for other models go up, but when I actually use them it sucks, and then I go back to Claude.

    It’s also nice Claude just doesn’t update until they have actual improvements to show.

  • I strongly dislike the “updating of versions” whenever possible. Versions are rarely better in all ways, makes things harder. Just make it version 2.6.

  • I love Gemini.

    I just wish they’d give powerful options for getting files out of it.

    I’m so sick of cutting and pasting.

    It would be nice to git push and pull into AI Studio chats or SFTP.

  • Why don’t companies publish hashes of emitted answers so that we, eg teachers, could verify if the AI produced this result?