It was paid for through code assist enterprise and had all the flags enabled for the "preview" models. Still the only way to get gemini 3+ was to open and close the application 5 to 10 times and sometimes you would get 3 for a bit and then get dumped back to 2.5 and no matter what you do it would not use 3.
I tossed it after spending like 3 hours total messing around the google cloud console and trying a bunch of shit from the github issues. The other offerings don't waste my time (or waste less of it anyway). If they want me to beta test their shit they shouldn't charge for it.
I would like one for the vram but I am sure they will be unobtainable after the initial stock sells out as I assume they were produced before the RAM prices went up.
It definitely speeds the effects of dementia and similar because your brain insists on filling in what you didn’t hear and it tends to be wildly negative, at least in my two experiences of having gone through it.
I mentioned this could be a possible falsification of the idea. It's also possible there are multiple causes and the modality I mentioned is a cause for some. I'm not sure. There are definitely cases where isolation contributes to cognitive decline.
Exactly. My mom lost her job because of early onset. She was very social, read tons of books, etc…. Now, I’m happy she at least still knows who am, but she can’t put a sentence together.
I feel like big / old companies thrive on process and are bogged down in bureaucracy.
Sure there is a process to get a library approved, and that abstraction makes you feel better but for the guy who's job it is to approve they are not going to spend an entire day reviewing a lib. The abstraction hides what is essentially a "LGTM" its just that takes a week for someone to check it off their outlook todos.
They are caching internal LLM state, which is in the 10s of GB for each session. It's called a KV cache (because the internal state that is cached are the K and V matrices) and it is fundamental to how LLM inference works; it's not some Anthropic-specific design decision. See my other comment for more detail and a reference.
Everyone is hung up making their shit "scalable" like its a systems design interview at google in 2010.
Now a days you get a box with 600+ cores and 4TB of RAM. That is going to cover a very very large percentage of most enterprises.
reply