I don’t want to sound jaded, but I have a feeling that eventually someone will use it to run Electron apps in Flatpak on Linux compiled to WASM running on Chrome, all because portability is hard and Qt’s licensing is unpleasant.
still hoping that rust will bring us one native gui framework to rule them all that will offload at least some of the burden from electron or whatever else js developers use 🤷♀️
It won’t. 95% of applications would be just fine with garbage collection, and the extra complexity of Rust’s memory model will make it so that current Javascript developers never use it. Go is more likely to win (but honestly, just a more lightweight JS runtime with a more lightweight UI framework would be fine by me)
a native gui framework doesn’t necessarily mean that one will have to write the application in rust: even today, when rust gui libs are in their very infancy, some already provide bindings to other languages (js included)
cross-platform native gui frameworks are super hard, but having a language that’s more accessible than c++, more modern and safe, will hopefully bring this closer to reality
I don’t think most electron developers are there because of Qt’s license. Qt and C++ are a pain in the ass compared to JS and modern web frameworks.
Lasers, graphene - those will be expensive logic gates. Why not use other kinds of light?
In few years our power super gaming PC seems a Abacus to us. The first personal Quandum computer is in the market since last year, for $5000 (2 Qbits, not really usefull yet), but remeber the developement since 15-20 years with the current exponential advances. Current cheap smartphones are high end PC 15 years ago.
There were good reasons to think that silicon based electronics will have exponential advances (basically scaling down components makes them exponentially faster by default). Do we have similar physical properties of quantum computing tech? Perhaps silicon was special and now we are at the end of it with nothing that has this exponential property anymore to replace it.
Yes, scaling down silicon based electronics gaining in speed, but only until certain limits which have silicon to scaling down the components, currently un atomic scale. Because of this are investigated other materials, like syntetic diamonds, which soports much more heat and with this a major processing speed. Quantum computing ofrfers much more speed, because it isn’t limited to 1 OR 0, it can process simultaneous 1 AND O AND everything in between, until now the difficult is that it need very low temperatures and the lack of stability with minimum interferences. But technology is advancing and cooling by magnetic means makes it possible to dispense with nitrogen/helium cooling and allows its application in desktop PCs, although at the moment only with three Qbits, suitable for both research and education use for complex calculations. Well, with the first PCs the same thing happened.
I’m gonna go full ludite here but I don’t see the point of having faster computer.
Servers. There’s going to be more people, more people that have access to technology and internet and a lot of services are online.
Not only that, but look at how big data centers, how much cooling they need (They need a LOT of water), it could make it more efficient.
But to be fair this article writes very little about something that is just a proof of concept.
For personal computing, sure. That’s not full ludite, we’ve basically reached a point where most things a person does on their computer can be done well with a $500 laptop or phone.
For servers, media rendering, hash cracking, prime searching, machine-learning training, video and image enhancement, medical simulations and other applications, we still love more power.
And it’s not just having more powerful hardware available. It can also mean producing less hardware in the first place. Or in the case of a phone, maybe more and more people can just hook it up to a USB-C dock when they need the form factor of a laptop, but otherwise carry around phone.
Most of ML industrial application are just fluff, I don’t need my bank, my social network and my governement to profile me nor predict my behavior… And entertainment industry can tell beautiful stories with or without computers.
The only legit application imo is medical research which has positive outcome on people lives but then I’m sure that if the gafam repurposed their massive tracking infrastructure we’re good for a while.
Those applications you listed are fluff, but there are a significant amount of others (‘most’ or not, I don’t know) that aren’t fluff. Medical research, other engineering research, accessibility tools. I appreciate GAN content upscaling and de-noising for older art, but that’s a more personal case.
I do think it’s unfortunate that we live in a system that encourages powerful tools to be absolutely wasted on fluff.
Couldn’t have put it better than your last sentence.
Have a nice day :)
deleted by creator
How about solving protein folding problems to develop new medical treatments, or running chemical simulations to come up with more climate-friendly industrial processes?
Honestly? I don’t think tech’s gonna save us from climate changes.
I agree with the medical treatment point.
Breaking encryption and stuffs?!