So many skibidi vibes here man, it was all vibes rizzing up this place and until I came and unvibed everything. The vibes are so joever vro I was just mogged. 😔
So many skibidi vibes here man, it was all vibes rizzing up this place and until I came and unvibed everything. The vibes are so joever vro I was just mogged. 😔
Do you even need quantum mechanics to make that argument, then? You’re basically saying weak emergence is evidence of being in a simulation because you can approximate nature much simpler when “zoomed out.” It seems like even if we did not have quantum mechanics you could still make that argument.
Why would an optimization make things more complicated? The point of optimizations in any simulation is to simplify the complexity of the computation. The entire reason why there is a multi-billionaire industry to research quantum computers is because they are exponentially more difficult to simulate than classical physics, so they are not practical to simulate on a classical computer. Seems weird to me that a simulator would “optimize” things by making them enormously more complex.
Yeah, the jacket is very different as well if you look at the front chest area. While people do say maybe he just changed his clothes, the problem is if he also changed his backpack, he couldn’t have just put the clothes in the backpack, meaning he would’ve had to have left them somewhere and there would’ve been a trail that probably would’ve been found by now. It doesn’t really add up for them to be the same person.
Complex numbers are just a way of representing an additional degree of freedom in an equation. You have to represent complex numbers not on a number line but on the complex plane, so each complex number is associated with two numbers. That means if you create a function that requires two inputs and two outputs, you could “compress” that function into a single input and output by using complex numbers.
Complex numbers are used all throughout classical mechanics. Waves are two-dimensional objects because they both have an amplitude and a wavelength. Classical wave dynamics thus very often use complex numbers because you can capture the properties of waves more concisely. An example of this is the Fourier transform. If you look up the function, it looks very scary, it has an integral and Euler’s number raised to the negative power of the imaginary number multiplied by pi. However, if you’ve worked with complex numbers a lot, you’d immediately recognize that raising Euler’s number to pi times the imaginary number is just how you represent rotations on the complex plane.
Despite how scary the Fourier transform looks, literally all it is actually doing is wrapping a wave around a circle. 3Blue1Brown has a good video on his channel of how to visualize the Fourier transform. The Fourier transform, again, isn’t inherently anything quantum mechanical, we use it all the time in classical mechanics, for example, if you ever used an old dial-up model and wondered why it made those weird noises, it was encoding data as sound wave by representing them as different harmonic waves that it would then add together, producing that sound. The Fourier transform could then be used by the modem at the other end to break the sound back apart into those harmonic waves and then decode it back into data.
In quantum mechanics, properties of systems always have an additional kind of “orientation” to them. When particles interact, if their orientations are aligned, the outcome of the interaction is deterministic. If they are misaligned, then it introduces randomness. For example, an electron’s spin state can either be up or down. However, its spin state also has a particular orientation to it, so you can only measure it “correctly” by having the orientation of the measuring device aligned with the electron. If they are misaligned, you introduce randomness. These orientations often are associated with physical rotations, for example, with the electron spins state, you measure it with something known as a Stern-Gerlach apparatus, and to measure the electron on a different orientation you have to physically rotate the whole apparatus.
Because the probability of measuring certain things directly relates to the relative orientation between your measuring device and the particle, it would be nice if we had a way to represent both the relative orientation and the probability at the same time. And, of course, you guessed it, we do. It turns out you can achieve this simply by representing your probability amplitudes (the % chance of something occurring) as complex numbers. This means in quantum mechanics, for example, an event can have a -70.7i% chance of occurring.
While that sounds weird at first, you quickly realize that the only reason we represent it this way is because it directly connects the relative orientation between the systems interacting and the probabilities of certain outcomes. You see, you can convert quantum probabilities to classical just by computing the distance from 0% on the complex plane and squaring it, which in the case of -70.7i% would give you 50%, which tells you this just means it is basically a fair coin flip. However, you can also compute from this number the relative orientation of the two measuring devices, which in this case you would find it to be rotated 90 degrees. Hence, because both values can be computed from the same number, if you rotate the measuring device it must necessarily alter the probabilities of different outcomes.
You technically don’t need to ever use complex numbers. You could, for example, take the Schrodinger equation and just break it up into two separate equations for the real and imaginary part, and have them both act on real numbers. Indeed, if you actually build a quantum computer simulator in a classical computer, most programming languages don’t include complex numbers, so all your algorithms have to break the complex numbers into two real numbers. It’s just when you are writing down these equations, they can get very messy this way. Complex numbers are just far more concise to represent additional degrees of freedom without needing additional equations/functions.
Yep. Technically you could in principle use Grover’s algorithm to speed up cracking a symmetrical cipher, but the size typically used for the keys is too large so even though it’d technically be faster it still not be possible in practice. Even with asymmetrical ciphers we already have replacements that are quantum safe, although most companies have not implemented them yet.
Honestly, the random number generation on quantum computers is practically useless. Speeds will not get anywhere near as close to a pseudorandom number generator, and there are very simple ones you can implement that are blazing fast, far faster than any quantum computer will spit out, and produce numbers that are widely considered in the industry to be cryptographically secure. You can use AES for example as a PRNG and most modern CPUs like x86 processor have hardware-level AES implementation. This is why modern computers allow you to encrypt your drive, because you can have like a file that is a terabyte big that is encrypted but your CPU can decrypt it as fast as it takes for the window to pop up after you double-click it.
While PRNG does require an entropy pool, the entropy pool does not need to be large, you can spit out terabytes of cryptographically secure pseudorandom numbers on a fraction of a kilobyte of entropy data, and again, most modern CPUs actually include instructions to grab this entropy data, such as Intel’s CPUs have an RDSEED instruction which let you grab thermal noise from the CPU. In order to avoid someone discovering a potential exploit, most modern OSes will mix into this pool other sources as well, like fluctuations in fan voltage.
Indeed, used to with Linux, you had a separate way to read random numbers directly from the entropy pool and another way to read pseudorandom numbers, those being /dev/random and /dev/urandom. If you read from the entropy pool, if it ran out, the program would freeze until it could collect more, so some old Linux programs you would see the program freeze until you did things like move your mouse around.
But you don’t see this anymore because generating enormous amounts of cryptographysically secure random nubmers is so easy with modern algorithms that modern Linux just collects a little bit of entropy at boot and it uses that to generate all pseudorandom numbers after, and just got rid of needing to read it directly, both /dev/random and /dev/urandom now just internally in the OS have the same behavior. Any time your PC needs a random number it just pulls from the pseudorandom number generator that was configured at boot, and you have just from the short window of collecting entropy data at boot the ability to generate sufficient pseudorandom numbers basically forever, and these are the numbers used for any cryptographic application you may choose to run.
The point of all this is to just say random number generation is genuinely a solved problem, people don’t get just how easy it is to basically produce practically infinite cryptographically secure pseudorandom numbers. While on paper quantum computers are “more secure” because their random numbers would be truly random, in practice you literally would never notice a difference. If you gave two PhD mathematicians or statisticians the same message, one encrypted using a quantum random number generator and one encrypted with a PRNG like AES or ChaCha20, and asked them to decipher them, they would not be able to decipher either. In fact, I doubt they would even be able to identify which one was even encoded using the quantum random number generator. A string of random numbers looks just as “random” to any random number test suite whether or not it came from a QRNG or a high-quality PRNG (usually called CSPRNG).
I do think at least on paper quantum computers could be a big deal if the engineering challenge can ever be overcome, but quantum cryptography such as “the quantum internet” are largely a scam. All the cryptographic aspects of quantum computers are practically the same, if not worse, than traditional cryptography, with only theoretical benefits that are technically there on paper but nobody would ever notice in practice.
the study that found the universe is not locally real. Things only happen once they are observed
This is only true if you operate under a very specific and strict criterion of “realism” known as metaphysical realism. Einstein put forward a criterion of what he thought this philosophy implied for a physical theory, and his criterion is sometimes called scientific realism.
Metaphysical realism is a very complex philosophy. One of its premises is that there exists an “absolute” reality where all objects are made up of properties that are independent of perspective. Everything we perceive is wholly dependent upon perspective, so metaphysical realism claims that what we perceive is not “true” reality but sort of an illusion created by the brain. “True” reality is then treated as the absolute spacetime filled with particles captured in the mathematics of Newton’s theory.
The reason it relies on this premise is because by assigning objects perspective invariant properties, then they can continue to exist even if no other object is interacting with them, or, more specifically, they continue to exist even if “no one is looking at them.” For example, if you fire a cannonball from point A to point B, and you only observe it leaving point A and arriving at point B, Newtonian mechanics allows you to “track” its path between these two points even if you did not observe it.
The problem is that you cannot do this in quantum mechanics. If you fire a photon from point A to point B, the theory simply disallows you from unambiguously filling in the “gaps” between the two points. People then declare that “realism is dead,” but this is a bit misleading because this is really only a problem for metaphysical/scientific realism. There are many other kinds of realism in literature.
For example, the philosopher Jocelyn Benoist’s contextual realism argues that the exact opposite. The mathematical theory is not “true reality” but is instead a description of reality. A description of reality is not the same as reality. Would a description of the Eiffel Tower substitute actually seeing it in reality? Of course not, they’re not the same. Contextual realism instead argues that what is real is not the mathematical description but is precisely what we perceive. The reason we perceive reality in a way that depends upon perspective is because reality is just relative (or “contextual”). There is no “absolute” reality but only a contextual reality and that contextual reality we perceive directly as it really is.
Thus for contextual realism, there is no issue with the fact that we cannot “track” things unambiguously, because it has no attachment to treating particles as if they persist as autonomous entities. It is perfectly fine with just treating it as if the particle hops from point A to point B according to some predictable laws and relative to the context in which the observer occupies. That is just how objective reality works. Observation isn’t important, and indeed, not even measurement, because whatever you observe in the experimental setting is just what reality is like in that context. The only thing that “arises” is your identification.
Why did physicists start using the word “real” and “realism”? It’s a philosophical term, not a physical one, and it leads to a lot of confusion. “Local” has a clear physical meaning, “realism” gets confusing. I have seen some papers that use “realism” in a way that has a clear physical definition, such as one I came across defined it in terms of a hidden variable theory. Yet, I also saw a paper coauthored by the great Anton Zeilinger that speaks of “local realism,” but very explicitly uses “realism” with its philosophical meaning, that there is an objective reality independent of the observer, which to me it is absurd to pretend that physics in any way calls this into account.
If you read John Bell’s original paper “On the Einstein Podolsky Rosen Paradox,” he never once use the term “realism.” The only time I have seen “real” used at all in this early discourse is in the original EPR paper, but this was merely a “criterion” (meaning a minimum but not sufficient condition) for what would constitute a theory that is a complete description of reality. Einstein/Podolsky/Rosen in no way presented this as a definition of “reality” or a kind of “realism.”
Indeed, even using the term “realism” on its own is ambiguous, as there are many kinds of “realisms” in the literature. The phrase “local realism” on its own is bound to lead to confusion, and it does, because I pointed out, even in the published literature physicists do not always use “realism” consistently. If you are going to talk about “realism,” you need to preface it to be clear what kind of realism you are specifically talking about.
If the reason physicists started to talk about “realism” is because they specifically are referring to something that includes the EPR criterion, then they should call it “EPR realism” or something like that. Just saying “realism” is so absurdly ridiculous it is almost as if they are intentionally trying to cause confusion. I don’t really blame anyone who gets confused on this because like I said if you even read the literature there is not even consistent usage in the peer-reviewed papers.
The phrase “observer-dependence” is also very popular in the published literature. So, while I am not disagreeing with you that “observation” is just an interaction, this is actually a rather uncommon position known as relational quantum mechanics.
You don’t have to be sorry, that was stupid of me to write that.
Because the same functionality would be available as a cloud service (like AI now). This reduces costs and the need to carry liquid nitrogen around.
Okay, you are just misrepresenting my argument at this point.
Why are you isolating a single algorithm? There are tons of them that speed up various aspects of linear algebra and not just that single one, and many improvements to these algorithms since they were first introduced, there are a lot more in the literature than just in the popular consciousness.
The point is not that it will speed up every major calculation, but these are calculations that could be made use of, and there will likely even be more similar algorithms discovered if quantum computers are more commonplace. There is a whole branch of research called quantum machine learning that is centered solely around figuring out how to make use of these algorithms to provide performance benefits for machine learning algorithms.
If they would offer speed benefits, then why wouldn’t you want to have the chip that offers the speed benefits in your phone? Of course, in practical terms, we likely will not have this due to the difficulty and expense of quantum chips, and the fact they currently have to be cooled below to near zero degrees Kelvin. But your argument suggests that if somehow consumers could have access to technology in their phone that would offer performance benefits to their software that they wouldn’t want it.
That just makes no sense to me. The issue is not that quantum computers could not offer performance benefits in theory. The issue is more about whether or not the theory can be implemented in practical engineering terms, as well as a cost-to-performance ratio. The engineering would have to be good enough to both bring the price down and make the performance benefits high enough to make it worth it.
It is the same with GPUs. A GPU can only speed up certain problems, and it would thus be even more inefficient to try and force every calculation through the GPU. You have libraries that only call the GPU when it is needed for certain calculations. This ends up offering major performance benefits and if the price of the GPU is low enough and the performance benefits high enough to match what the consumers want, they will buy it. We also have separate AI chips now as well which are making their way into some phones. While there’s no reason at the current moment to believe we will see quantum technology shrunk small and cheap enough to show up in consumer phones, if hypothetically that was the case, I don’t see why consumers wouldn’t want it.
I am sure clever software developers would figure out how to make use of them if they were available like that. They likely will not be available like that any time in the near future, if ever, but assuming they are, there would probably be a lot of interesting use cases for them that have not even been thought of yet. They will likely remain something largely used by businesses but in my view it will be mostly because of practical concerns. The benefits of them won’t outweigh the cost anytime soon.
Uh… one of those algorithms in your list is literally for speeding up linear algebra. Do you think just because it sounds technical it’s “businessy”? All modern technology is technical, that’s what technology is. It would be like someone saying, “GPUs would be useless to regular people because all they mainly do is speed up matrix multiplication. Who cares about that except for businesses?” Many of these algorithms here offer potential speedup for linear algebra operations. That is the basis of both graphics and AI. One of those algorithms is even for machine learning in that list. There are various algorithms for potentially speeding up matrix multiplication in the linear. It’s huge for regular consumers… assuming the technology could ever progress to come to regular consumers.
A person who would state they fully understand quantum mechanics is the last person i would trust to have any understanding of it.
I find this sentiment can lead to devolving into quantum woo and mysticism. If you think anyone trying to tell you quantum mechanics can be made sense of rationally must be wrong, then you implicitly are suggesting that quantum mechanics is something that cannot be made sense of, and thus it logically follows that people who are speaking in a way that does not make sense and have no expertise in the subject so they do not even claim to make sense are the more reliable sources.
It’s really a sentiment I am not a fan of. When we encounter difficult problems that seem mysterious to us, we should treat the mystery as an opportunity to learn. It is very enjoyable, in my view, to read all the different views people put forward to try and make sense of quantum mechanics, to understand it, and then to contemplate on what they have to offer. To me, the joy of a mystery is not to revel in the mystery, but to search for solutions for it, and I will say the academic literature is filled with pretty good accounts of QM these days. It’s been around for a century, a lot of ideas are very developed.
I also would not take the game Outer Wilds that seriously. It plays into the myth that quantum effects depend upon whether or not you are “looking,” which is simply not the case and largely a myth. You end up with very bizarre and misleading results from this, for example, in the part where you land on the quantum moon and have to look at the picture of it for it to not disappear because your vision is obscured by fog. This makes no sense in light of real physics because the fog is still part of the moon and your ship is still interacting with the fog, so there is no reason it should hop to somewhere else.
Now quantum science isn’t exactly philosophy, ive always been interested in philosophy but its by studying quantum mechanics, inspired by that game that i learned about the mechanic of emerging properties. I think on a video about the dual slit experiment.
The double-slit experiment is a great example of something often misunderstood as somehow evidence observation plays some fundamental role in quantum mechanics. Yes, if you observe the path the two particles take through the slits, the interference pattern disappears. Yet, you can also trivially prove in a few line of calculation that if the particle interacts with a single other particle when it passes through the two slits then it would also lead to a destruction of the interference effects.
You model this by computing what is called a density matrix for both the particle going through the two slits and the particle it interacts with, and then you do what is called a partial trace whereby you “trace out” the particle it interacts with giving you a reduced density matrix of only the particle that passes through the two slits, and you find as a result of interacting with another particle its coherence terms would reduce to zero, i.e. it would decohere and thus lose the ability to interfere with itself.
If a single particle interaction can do this, then it is not surprising it interacting with a whole measuring device can do this. It has nothing to do with humans looking at it.
At that point i did not yet know that emergence was already a known topic in philosophy just quantum science, because i still tried to avoid external influences but it really was the breakthrough I needed and i have gained many new insights from this knowledge since.
Eh, you should be reading books and papers in the literature if you are serious about this topic. I agree that a lot of philosophy out there is bad so sometimes external influences can be negative, but the solution to that shouldn’t be to entirely avoid reading anything at all, but to dig through the trash to find the hidden gems.
My views when it comes to philosophy are pretty fringe as most academics believe the human brain can transcend reality and I reject this notion, and I find most philosophy falls right into place if you reject this notion. However, because my views are a bit fringe, I do find most philosophical literature out there unhelpful, but I don’t entirely not engage with it. I have found plenty of philosophers and physicists who have significantly helped develop my views, such as Jocelyn Benoist, Carlo Rovelli, Francois-Igor Pris, and Alexander Bogdanov.
This is why many philosophers came to criticize metaphysical logic in the 1800s, viewing it as dealing with absolutes when reality does not actually exist in absolutes, stating that we need some other logical system which could deal with the “fuzziness” of reality more accurately. That was the origin of the notion of dialectical logic from philosophers like Hegel and Engels, which caught on with some popularity in the east but then was mostly forgotten in the west outside of some fringe sections of academia. Even long prior to Bell’s theorem, the physicist Dmitry Blokhintsev, who adhered to this dialectical materialist mode of thought, wrote a whole book on quantum mechanics where the first part he discusses the need to abandon the false illusion of the rigidity and concreteness of reality and shows how this is an illusion even in the classical sciences where everything has uncertainty, all predictions eventually break down, nothing is never possible to actually fully separate something from its environment. These kinds of views heavily influenced the contemporary physicist Carlo Rovelli as well.
This is accurate, yes. The cat in the box is conscious presumably, in my opinion of cats at least, but still can be “not an observer” from the POV of the scientist observing the experiment from outside the box.
“Consciousness” is not relevant here at all. You can write down the wave function of a system relative to a rock if you wanted, in a comparable way as writing down the velocity of a train from the “point of view” of a rock. It is coordinate. It has nothing to do with “consciousness.” The cat would perceive a definite state of the system from its reference frame, but the person outside the box would not until they interact with it.
QM is about quite a lot more than coordinate systems
Obviously QM is not just coordinate systems. The coordinate nature of quantum mechanics, the relative nature of it, is merely a property of the theory and not the whole theory. But the rest of the theory does not have any relevance to “consciousness.”
and in my opinion will make it look weird in retrospect once physics expands to a more coherent whole
The theory is fully coherent and internally consistent. It amazes me how many people choose to deny QM and always want to rush to change it. Your philosophy should be guided by the physical sciences, not the other way around. People see QM going against their basic intuitions and their first thought is it must be incomplete and needs to have additional complexity added to it to make it fit their intuitions, rather than just questioning that maybe their basic intuitions are wrong.
Your other comment was to a Wikipedia page which if you clicked the link on your own source it would’ve told you that the scientific consensus on that topic is that what you’re presenting is a misinterpretation.
A simple search on YouTube could’ve also brought up several videos explaining this to you.
Edit: Placing my response here as an edit since I don’t care to continue this conversation so I don’t want to notify.
Yes, that was what I said. Er, well… QM, as I understand it, doesn’t have to do anything with shifting coordinate systems per se (and in fact is still incompatible with relativity). They’re just sort of similar in that they both have to define some point of view and make everything else in the model relative to it. I’m still not sure why you brought coordinate systems into it.
A point of view is just a colloquial term to refer to a coordinate system. They are not coordinate in the exact same way but they are both coordinate.
My point was that communication of state to the observer in the system, or not, causes a difference in the outcome. And that from the general intuitions that drive almost all of the rest of physics, that’s weird and sort of should be impossible.
No, it doesn’t not, and you’re never demonstrated that.
Sure. How is it when combined with macro-scale intuition about the way natural laws work, or with general relativity?
We have never observed quantum effects on the scale where gravitational effects would also be observable, so such a theory, if we proposed one, would not be based on empirical evidence.
This is very, very very much not what I am doing. What did I say that gave you the impression I was adding anything to it?
You literally said in your own words we need to take additional things into account we currently are not. You’re now just doing a 180 and pretending you did not say what literally anyone can scroll up and see that you said.
I am not talking about anything about retrocausality here, except maybe accidentally.
Then you don’t understand the experiment since the only reason it is considered interesting is because if you interpret it in certain ways it seems to imply retrocausality. Literally no one has ever treated it as anything more than that. You are just making up your own wild implications from the experiment.
I was emphasizing the second paragraph; “wave behavior can be restored by erasing or otherwise making permanently unavailable the ‘which path’ information.”
The behavior of the system physically changes when it undergoes a physical interaction. How surprising!
There is “observer-dependence” in quantum mechanics in a comparable way that there is observer-dependence in general relativity. It has nothing to do with some “fundamental role of consciousness” but comes from the fact that reality itself depends on how you look at it, it is reference frame dependent. The “observer” is just a chosen coordinate system in which to describe other things. I know, you probably got this from Kastrup too, right? Idealists have been getting desperate and resorting to quantum woo, pretending that something that changes based on coordinate system proves fundamental consciousnesses.
Kastrup is entirely unconvincing because he pretends the only two schools of philosophy in the whole universe are his specific idealism and metaphysical realism which he falsely calls the latter “materialism.” He thus never feels the need to ever address anything outside of a critique of a single Laymen understanding of materialism which is more popular in western countries than eastern countries, ignoring the actual wealth of philosophical literature.
Anyone who actually reads books on philosophy would inevitably find Kastrup to be incredibly unconvincing as he, by focusing primarily on a single school, never justifies many of his premises. He begins from the very beginning talking about “conscious experience” and whatnot when, if you’re not a metaphysical realist, that is what you are supposed to be arguing in the first place. Unless you’re already a dualist or metaphysical realist, if you are pretty much any other philosophical school like contextual realist, dialectical materialist, empiriomonist, etc, you probably already view reality as inherently observable, and thus perception is just reality from a particular point-of-view. It then becomes invalid to add qualifiers to it like “conscious experience” or “subjective experience” as reality itself cannot had qualifiers.
I mean, the whole notion of “subjective experience” goes back to Nagel who was a metaphysical realist through-and-through and wrote a whole paper defending that notion, “What is it like to be a Bat?”, and this is what Kastrup assumes his audience already agrees with from the get-go. He never addresses any of the criticisms of metaphysical realism but pretends like they don’t exist and he is the unique sole critic of it and constantly calls metaphysical realism “materialism” as if they’re the same philosophy at all. He then builds all of his arguments off of this premise.
You should look into contextual realism. You might find it interesting. It is a philosophical school from the philosopher Jocelyn Benoist that basically argues that the best way to solve most of the major philosophical problems and paradoxes (i.e. mind-body problem) is to presume the natural world is context variant all the way down, i.e. there simply is no reality independent of specifying some sort of context under which it is described (kind of like a reference frame).
The physicist Francois-Igor Pris points out that if you apply this thinking to quantum mechanics, then the confusion around interpreting it entirely disappears, because the wave function clearly just becomes a way of accounting for the context under which an observer is observing a system, and that value definiteness is just a context variant property, i.e. two people occupying two different contexts will not always describe the system as having the same definite values, but may describe some as indefinite which the other person describes as definite.
“Observation” is just an interaction, and by interacting with a system you are by definition changing your context, and thus you have to change your accounting for your context (i.e. the wave function) in order to make future predictions. Updating the wave function then just becomes like taring a scale, that is to say, it is like re-centering or “zeroing” your coordinate system, and isn’t “collapsing” anything physical. There is no observer-dependence in the sense that observers are somehow fundamental to nature, only that systems depend upon context and so naturally as an observer describing a system you have to take this into account.
Interesting you get downvoted for this when I mocked someone for saying the opposite who claimed that $0.5m was some enormous amount of money we shouldn’t be wasting, and I simply pointed out that we waste literally billions around the world on endless wars killing random people for now reason, so it is silly to come after small bean quantum computing if budgeting is your actual concern. People seemed to really hate me for saying that, or maybe it was because they just actually like wasting moneys on bombs to drop on children and so they want to cut everything but that.