Lua.
Don’t call the ambulance, it’s too late for me
Lua.
Don’t call the ambulance, it’s too late for me
You are totally correct, but I feel like pointing out that a surprising number of games use the 4k texture nomenclature in a totally illogical way; they label it 4k because it’s meant to look good on a 4k screen, not because the texture itself is at that resolution (or any loosely related resolution).
Which is itself really annoying. But I guess less savvy crowd might not actually understand what ‘real’ 4k textures even refer to?
We got warnings of this in my area, but we just barely missed the danger zone I guess. All we ended up with was a few days of steady rain. Seems it got a lot worse elsewhere…
I’m not so sure we’re missing that much personally, I think it’s more just sheer scale, as well as the complexity of the input and output connections (I guess unlike machine learning networks, living things tend to have a much more ‘fuzzy’ sense of inputs and outputs). And of course sheer computational speed; our virtual networks are basically at a standstill compared to the paralellism of a real brain.
Just my thoughts though!
To me, what is surprising is that people refuse to see the similarity between how our brains work and how neural networks work. I mean, it’s in the name. We are fundamentally the same, just on different scales. I belive we work exactly like that, but with way more inputs and outputs and way deeper network, but fundamental principles i think are the same.
Same for me. As long as opening Relay brings me to reddit, it’s hard to stop using it. But once that stops, or becomes ad ridden or whatever, there’s no way in hell I will install the official reddit app or anything like that, and I hate using a browser on mobile so not doing that either… So yeah. That’ll be it for me. So far Beehaw/lemmy is shaping up to replace it though.
My partner is in her mid 30s and is a game designer. Suffice to say she likes games, board games / pen & paper, video games…
I can’t give an authorative answer (not my domain), but I think there are two ways these types of things are done.
First is just observing the page or service as an external entity; basically requesting a page or hitting an endpoint, and just tracking whether you get a response (if not, it must be down), or for measuring load level in a very naive way, track the response time. This is easy in the sense that you need no special access to the target. But it’s also limited in its accuracy.
Second way, like what your github example is doing, is having access to special api endpoints for (or direct access to) performance metrics. Since the github status page is literally ran by Github, they obviously have easy access to any metric they could want. They probably (certainly) run services whose entire job is to produce reliable data for their status page.
The minute details of each of these options is pretty open ended; many ways to do it.
Just my 5¢ as a non-web developer.