[HN Gopher] Simulation of a 2B-atom cell that metabolizes and gr...
___________________________________________________________________
 
Simulation of a 2B-atom cell that metabolizes and grows like a
living cell
 
Author : jonbaer
Score  : 239 points
Date   : 2022-01-28 12:39 UTC (10 hours ago)
 
web link (blogs.nvidia.com)
w3m dump (blogs.nvidia.com)
 
| Jeff_Brown wrote:
| I wish they detailed how independent the simulated processes
| really are, and what sorts of dynamics are lost by not
| considering the atomic level.
 
| capableweb wrote:
| It seems that the researchers did use NVIDIA GPUs to perform the
| work, but it's not clear what sets the GPUs apart from others and
| why this research wouldn't be possible without NVIDIA's GPUs, as
| the article title and body implies.
 
  | CoastalCoder wrote:
  | Can someone comment on the legality of a 3rd party providing an
  | unauthorized implemention of the CUDA API?
  | 
  | I would think that Oracle's loss of a similar lawsuit with Java
  | would be related.
 
    | my123 wrote:
    | > Can someone comment on the legality of a 3rd party
    | providing an unauthorized implemention of the CUDA API?
    | 
    | NVIDIA said that they would be fine with it in the past, and
    | ROCm HIP is just a (bad) CUDA API clone.
 
  | jabbany wrote:
  | Interesting question. Press article aside, GPGPU applications
  | like scientific compute, ML etc. have all mostly gravitated to
  | Nvidia / CUDA.
  | 
  | Not working in this space, I'm curious why this is the case. Is
  | there something inherently better about CUDA? Or is it that
  | Nvidia's performance is somehow better for these tasks? Or
  | maybe something else?
 
    | mattkrause wrote:
    | The products are good, but NVidia also cleverly bootstrapped
    | a whole ecosystem around them.
    | 
    | One of the other posts mentions 2014 as a turning point. At
    | that time, GPGPU stuff was entering the (scientific)
    | mainstream and NVidia was all over academia, convincing
    | people to try it out in their research. They handed out demo
    | accounts on boxes with beefy GPUs, and ran an extremely
    | generous hardware grant proposal. There was tons of (free)
    | training available: online CUDA MOOCs and in-person training
    | sessions. The first-party tools were pretty decent. As a
    | result, people built a lot of stuff using CUDA. Others,
    | wanting to use those programs, basically had to buy NVidia.
    | Lather, rinse, repeat.
    | 
    | This is in stark contrast to the other "accelerator" vendors.
    | Around the same time, I looked into using Intel Xenon Phi and
    | they were way less aggressive: "here are some benchmarks,
    | here's how it works, your institution has one or two
    | _somewhere_ if you want to contact them and try it out." As
    | for OpenCL...crickets. I don't even remember any AMD events,
    | and the very broad standard made it hard to figure out what
    | would work /work well and you might end up needing to port it
    | too!
 
      | jjoonathan wrote:
      | AMD's GPU OpenCL wasn't just not marketed, it was also a
      | bad product, even for relatively tame scientific purposes,
      | even when AMD made loud, repeated statements to the
      | contrary. Hopefully now that AMD has money they can do
      | better.
      | 
      | I'm sure that NVidia's ecosystem building played a role (I
      | remember events in 2009 and books before that), perhaps
      | even a big role, but it wasn't the only factor. I paid a
      | steep price in 2014 and 2016 for incorrectly assuming that
      | it was.
 
    | carlmr wrote:
    | From my cursory knowledge of the topic, there are competitors
    | like ROCm, but CUDA was the first that had a useable solution
    | here. Also last time I checked ROCm doesn't have broad
    | support on consumer cards, which makes it harder for people
    | to try it out at home.
    | 
    | But it seems ROCm is getting better and it has tensorflow and
    | pytorch support, so there's reasons to be hopeful to see some
    | competition here.
 
    | naavis wrote:
    | NVIDIA provides tools that just mostly do not exist for other
    | GPUs, making it easier to build on CUDA instead of something
    | else.
 
      | hiptobecubic wrote:
      | Absolutely this. When cuda was first making headway it was
      | the only thing even remotely close to a "developer
      | environment" and made things significantly easier than any
      | of the alternatives.
      | 
      | It might be different now, but at that time, many of the
      | users were not computer scientists, they were scientists
      | with computers. Having an easier to use programming model
      | and decent debugging tools means publishing more results,
      | more quickly.
 
    | jjoonathan wrote:
    | Back in 2014 or so I made the unfortunate mistake of buying
    | AMD cards with the thought that I'd just use OpenCL. I knew
    | that some codes wouldn't run, but I had catalogued the ones I
    | really cared about and thought I was up for the challenge. I
    | was so, so wrong.
    | 
    | First of all, software that advertised OpenCL or AMD
    | compatibility often worked poorly or not at all in that mode.
    | Adobe creative suite just rendered solid black outputs
    | whenever acceleration was enabled and forums revealed that it
    | had been that way for years and nobody cared to fix it.
    | Blender supported OpenCL for a while, but it was slower than
    | CPU rendering and for a sticky reason (nvidia did the work to
    | support big kernels with heavy branching and AMD didn't).
    | Ironically, OpenCL mode had decent performance but only if
    | you used it on an nvidia card.
    | 
    | The situation was even worse in scientific codes, where
    | "OpenCL version" typically meant "a half-finished blob of
    | code that was abandoned before ever becoming functional, let
    | alone anywhere near feature-parity."
    | 
    | I quickly learned why this was the case: the OpenCL tooling
    | and drivers weren't just a little behind their CUDA
    | counterparts in terms of features, they were almost unusably
    | bad. For instance, the OpenCL drivers didn't do memory (or
    | other context object?) cleanup, so if your program was less
    | than perfect you would be headed for a hard crash every few
    | runs. Debugging never worked -- hard crashes all around.
    | Basic examples didn't compile, documentation was scattered,
    | and at the end of the day, it was also leagues behind CUDA in
    | terms of features.
    | 
    | After months of putting up with this, I finally bit the
    | bullet, sold my AMD card, bought an NVidia card, ate the
    | spread, the shipping, the eBay fees, and the green tax
    | itself. It hurt, but it meant I was able to start shipping
    | code.
    | 
    | I'm a stubborn bastard so I didn't learn my lesson and
    | repeated this process two years later on the next generation
    | of cards. The second time, the lesson stuck.
 
  | Tenoke wrote:
  | Most likely because the software they use uses CUDA.
 
    | capableweb wrote:
    | I know that CUDA is faster than OpenCL for many tasks, but is
    | there something that is not possible to achieve in OpenCL but
    | possible in CUDA?
 
      | belval wrote:
      | > is there something that is not possible to achieve in
      | OpenCL but possible in CUDA
      | 
      | Developing fast... OpenCL is much harder to learn than
      | CUDA. Take someone who did some programming classes,
      | explain how CUDA works and they'll probably get somewhere.
      | Do the same thing with OpenCL and they'll probably quit.
 
      | TomVDB wrote:
      | I'm subscribed to some CUDA email list with weekly updates.
      | 
      | One thing that strikes me is how it evolves with new
      | features. Not just higher level libraries, but also more
      | fundamental, low level stuff, such as virtual memory,
      | standard memory models, c++ libraries, new compilers,
      | communication with other GPUs, launching dependent kernels,
      | etc.
      | 
      | At their core, OpenCL and CUDA both enable running parallel
      | computing algorithms on a GPU, but CUDA strikes me as much
      | more advanced in terms of peripheral features.
      | 
      | Every few years, I think about writing a CUDA program (it
      | never actually happens), and investigate how to do things,
      | and it's interesting how the old ways of doing things has
      | been superseded by better ways.
      | 
      | None of this should be surprising. As I understand it,
      | OpenCL has been put on life support by the industry in
      | general for years now.
 
        | 4gotunameagain wrote:
        | If you ever need to reap the benefits of CUDA & GPU
        | computations without getting into the details, check out
        | JAX by our corporate overlords(tm)
        | (https://github.com/google/jax), it has a NumPy like
        | syntax and super fast to get started
 
        | p1esk wrote:
        | Why would you suggest JAX? CuPy seems like an obvious
        | choice here (simpler and a lot more mature). Jax is only
        | needed if you want automatic differentiation.
 
      | Cthulhu_ wrote:
      | Possibly, but that's not really the point, the article is
      | part marketing push from nvidia for their HPC department.
 
        | capableweb wrote:
        | > but that's not really the point
        | 
        | That's what I thought as well, so the title on the
        | website ("NVIDIA GPUs Enable Simulation of a Living
        | Cell") is not really truthful then.
 
      | Symmetry wrote:
      | My understanding is that CUDA has a lot of optimized
      | libraries for common tasks, think BLAS, that don't
      | currently exist in OpenCL/Vulkan Compute.
 
  | 01100011 wrote:
  | Nvidia is pushing vertical integration hard. There are all
  | sorts of libraries from Nvidia which build on top of CUDA, from
  | simple cuBLAS to smart cities, autonomous driving, robotics and
  | 5G.
  | 
  | They also provide acceleration of open source libraries like
  | GROMACS, used for molecular dynamics simulation.
 
  | erwincoumans wrote:
  | The fine grain parallelism of this simulation suits the GPU
  | well. It would be possible on multicore CPUs, but possibly
  | slower.
 
  | tmearnest wrote:
  | There are two main reasons to take advantage of the Gpu in
  | lattice microbes. It can simulate the stochastic chemical
  | reaction and diffusion dynamics in parallel: one thread per
  | voxel. For instance, an E. coli sized cell would have ~40000
  | voxels. It's not quite embarrassing parallel, but close.
  | Second, the simulation is totally memory bound so we can take
  | advantage of fast gpu memory. The decision to use CUDA over
  | OpenCL was made in like 2009 or so. Things have changed a lot
  | since then. I don't think anyone has the time or interest to
  | port it over, unfortunately.
 
  | Gehoti wrote:
  | I'm much more aware of slot of things research is doing with
  | Nvidia.
  | 
  | Due to cuda, tools, SDKs etc Nvidia is providing.
  | 
  | I'm not aware of anything similar at any other GPU company
 
| neom wrote:
| I don't see in the article why this is useful or what is it used
| for?
 
  | joejoesvk wrote:
  | i don't even know where to start. you could simulate one cell,
  | then two cells then 4..suddenly you could have an
  | organisms...hell you could see organism that could have lived
  | on earth.
  | 
  | maybe it'll one day help with cancer research.
 
  | JabavuAdams wrote:
  | It's generally not possible to see where everything is in an
  | actual cell, in realtime, due to the sizes of the components.
  | So most of molecular biology relies on very clever lab
  | techniques to indirectly infer what cells are making and doing.
  | 
  | Cells are like little cities in terms of the complexity of
  | their biochemistry. We want to ask questions like "How does
  | this cell respond to this chemical/drug/change in environment."
  | 
  | Imagine trying to understand in detail the gun crime epidemic
  | in a city, if you can only see objects larger than 100 m on a
  | side. You wouldn't see people, cars, or many buildings.
  | 
  | We want to be able to understand, explain, predict, and control
  | cellular process, but so far we have to be quite indirect.
  | Understanding these things at a mechanistic level, in realtime
  | would revolutionize our ability to understand, repair, and
  | build biological systems.
 
  | JabavuAdams wrote:
  | For instance, the cartoon version of DNA that is presented to
  | even lower-year biology undergraduates is of linearized
  | strands. But of course, it's really all spooled and tangled and
  | crunched up into the nucleus of cells. Note that the cell they
  | simulated is of a prokaryote (no nucleus, much simpler cellular
  | processes than e.g. a mammal cell). About 1-2% of our genes
  | make proteins, although the proportion is much larger in
  | single-celled organisms (less redundancy in the genome, no
  | splicing, etc.) So when you hear that e.g. genes turn on or
  | off, this is not a switch. It's literally some sections of DNA
  | being unwound, and large complexes of mutually interacting
  | molecules probabilistically glomming on and off the DNA. The
  | actual 3d layout of this DNA "ramen" matters to e.g. bring
  | promoter regions of genes close to the actual genes they
  | control.
  | 
  | So basically, we have a schematic-level understanding of
  | cellular processes, but to see the actual 3D interactions in
  | realtime would be extremely illuminating.
 
    | JabavuAdams wrote:
    | I should say that this work is not simulating things at this
    | detail. Instead, it's more like a biophysical model of a
    | bunch of chemical reactions with rate information. It
    | probably boils down to a big system of coupled differential
    | equations, at different timescales. So, it's a statistical
    | level of detail, but still very informative.
 
  | madhato wrote:
  | Its simple really. First you simulate a single cell, then a
  | sperm and an egg cell. Then you simulate a virtual a world of
  | virtual captive humans to do our work for us without payment.
 
| barrenko wrote:
| Devs, the prequel.
 
| amelius wrote:
| A two-billion atom cell ... isn't that a bit small for a cell?
 
  | wcoenen wrote:
  | Yes. The cell was first created in the real world as part of
  | research about the minimal set of genes required for life.[1]
  | It is known as "JCVI-syn3.0" or "Mycoplasma laboratorium".'[2]
  | 
  | Still amazing that it can now be fully simulated "in silico".
  | 
  | [1] https://www.science.org/doi/10.1126/science.aad6253
  | 
  | [2] https://en.wikipedia.org/wiki/Mycoplasma_laboratorium
 
    | Traubenfuchs wrote:
    | It says
    | 
    | "In this new organism, the number of genes can only be pared
    | down to 473, 149 of which have functions that are completely
    | unknown."
    | 
    | But if we now can simulate this cell completely, shouldn't it
    | be easy to figure out what those genes are doing? Just start
    | the simulation with them knocked out.
 
      | WJW wrote:
      | Presumably if the number of genes cannot be pared down
      | below 473, it dies very quickly if one of the 149 genes is
      | knocked out. But "it doesn't work without it" is not a very
      | satisfactory answer to "what does it do".
 
        | amelius wrote:
        | Yes, this is similar to opening a radio and saying "I
        | don't know what this transistor does; let's take it out
        | and see what the radio does".
 
        | dekhn wrote:
        | See also "Can a biologist fix a radio"
        | https://www.cell.com/cancer-
        | cell/pdf/S1535-6108(02)00133-2.p... "Doug & Bill"(http://
        | www2.biology.ualberta.ca/locke.hp/dougandbill.htm) "Could
        | a neuroscientist understand a microprocessor"? https://jo
        | urnals.plos.org/ploscompbiol/article?id=10.1371/jo...
        | 
        | The funny thing is if you read the history of Feynman and
        | others, most of them grew up opening up radios and
        | learning how they worked by removing things. fixing them.
        | It's a very common theme (sort of falls off post-
        | transistor tho). I opened up radios as a kid, tried to
        | figure out what parts did what, and eventually gave up.
 
        | breck wrote:
        | That is a great read. Thanks :)
 
      | Jeff_Brown wrote:
      | Before attempting to crack the copy-protection on a game,
      | one might think something similar.
 
      | quickthrower2 wrote:
      | Valgrind that cell!
 
  | vital101 wrote:
  | The article mentions that they use minimal cells. "Minimal
  | cells are simpler than naturally occurring ones, making them
  | easier to recreate digitally."
 
| Jeff_Brown wrote:
| How much emergent behavior arises from the model? The only
| passage I see describing any of it is this one:
| 
| > The model showed that the cell dedicated most of its energy to
| transporting molecules across the cell membrane, which fits its
| profile as a parasitic cell.
| 
| Whether it mimics the behavior of real cells seems like the right
| test. We'll never be able to get it to parallel the outcome of a
| real system, thanks to chaos theory. But if it does lots of
| things that real cells do -- eating, immune system battles,
| reproduction -- we should be pretty happy.
 
| vasili111 wrote:
| Is this simulation on the atomic level with full all interatomic
| physics processes simulation or there were made some
| simplifications?
| 
| All interatomic interactions are simulated separately for each
| atom or they made statistical estimations and used some
| assumptions? Those two are absolutely two different types of
| simulation.
 
  | alpineidyll3 wrote:
  | Absolutely definitely not. It's not even possible to simulate a
  | single protein-molecule interaction to an accuracy such that
  | reaction rates are reproduced at room temperature. Small
  | effects such as the quantum nature of H-motion prevent this
  | from happening with present computational resources.
  | 
  | This research is something like a pixar movie, or one of those
  | blender demos with a lot of balls :P
 
  | perihelions wrote:
  | > _" full all interatomic physics"_
  | 
  | It's certainly not that -- that's a hideously difficult
  | algorithm with exponential complexity.
  | 
  | https://en.wikipedia.org/wiki/Full_configuration_interaction
 
    | sseagull wrote:
    | It's worse than exponential, it's factorial :)
 
    | unemphysbro wrote:
    | verlet list is the standard algo used to reduce the
    | complexity in the number of interatomic calculations
    | 
    | https://en.wikipedia.org/wiki/Verlet_list
 
      | dekhn wrote:
      | that's old tech, these days it's usually some sort of PPPM
      | (particle-particle particle-mesh) which parallelizes
      | better.
      | 
      | But that's for classical simulations. Full configuration
      | interaction is effecftively computing the schrodinger
      | equation at unlimited precision, in principle if you could
      | scale it up you could compute any molecular property
      | desired, assuming QM is an accurate model for reality.
 
        | unemphysbro wrote:
        | p3m, well pme, is exactly what we used for our
        | calculations ;)
        | 
        | i never did any qm work beyond basic parameterization
        | 
        | i'm guessing you are/were also computational physics guy
        | :)
 
        | dekhn wrote:
        | I was a computational biologist for many years, which
        | included a bunch of biophysics. I did extensive work with
        | PME about 20 years ago, on supercomputers. It's a pretty
        | neat technique
        | (https://en.wikipedia.org/wiki/Ewald_summation), once you
        | wrap your head around it!
 
        | unemphysbro wrote:
        | yup, we used PME for non-bonded calculations in our
        | simulations and to calculate things like electric
        | potentials. I finished a biophysics phd back in 2020 and
        | focused mainly on fluid flow.
        | 
        | Pretty cool, what're you up to now?
 
  | gfd wrote:
  | In theory, I don't think there's such a thing as simulation
  | without simplifications. The world seems to be continuous but
  | our computers are discrete. There's a small set of things we
  | know how to solve exactly with math but in general we have no
  | ways to deal infinity. Any given variable you're calculating
  | will be truncated at 32 or 64 bits when in reality they have an
  | infinite number of digits, changing at continuous timesteps,
  | interacting with every other atom in the universe.
  | 
  | In practice, none of this matters though and we can still get
  | very useful results at the resolution we care about.
 
    | merely-unlikely wrote:
    | There's this concept that causation moves at the speed of
    | light. When I first heard that, it sounded very much like a
    | fixed refresh rate to me. Or maybe the "real world" is just
    | another simulation
 
      | Filligree wrote:
      | It does if you put it _that_ way, but another way of
      | putting is that spacetime is hyperbolic (...well,
      | lorentzian), and all (lightspeed) interactions are zero-
      | ranged in 4D.
      | 
      | As in, photons that leave the surface of the sun _always_
      | strike those specific points in space-time which are at a
      | zero spacetime interval from said surface. If you take the
      | described geometry seriously, then  "spacetime interval" is
      | just the square of the physical distance between the
      | events.
      | 
      | (And any FTL path has a negative spacetime interval. If
      | that's still the square of the distance, then I think we
      | can confidently state that FTL is imaginary.)
 
    | qboltz wrote:
    | All simulations have to make the Born-Oppenheimer
    | approximation, nuclei have to be treated as frozen, otherwise
    | electrons don't have a reference point.
    | 
    | There will never be true knowledge of both a particle's
    | location and momentum a la uncertainty principle, and will
    | always have to be estimated.
 
      | chermi wrote:
      | What? This is simply untrue.
 
      | drdeca wrote:
      | But, for a system of two quantum particles which interact
      | according to a central potential, you can express this
      | using two quantum non-interacting particles one of which
      | corresponds to the center of mass of the two, and the other
      | of which corresponds to the relative position, I think?
      | 
      | And, like, there is still uncertainty about the position of
      | the "center of mass" pretend particle, as well as for the
      | position of the "displacement" pretend particle.
      | 
      | (the operators describing these pretend particles can be
      | constructed in terms of the operators describing the actual
      | particles, and visa versa.)
      | 
      | I don't know for sure if this works for many electrons
      | around a nucleus, but I think it is rather likely that it
      | should work as well.
      | 
      | Main thing that seems unclear to me is what the mass of the
      | pretend particles would be in the many electrons case. Oh,
      | also, presumably the different pretend particles would be
      | interacting in this case (though probably just the ones
      | that don't correspond to the center of mass interacting
      | with each-other, not interacting with the one that does
      | represent the center of mass?)
      | 
      | So, I'm not convinced of the "nuclei have to be treated as
      | frozen, otherwise electrons don't have a reference point"
      | claim.
 
      | aeternum wrote:
      | With a quantum computer could one theoretically input the
      | super position of possible locations and momenta and run
      | the simulation based on that?
 
      | phkahler wrote:
      | A simulation can have both.
 
        | Tagbert wrote:
        | Then is it an accurate simulation without the
        | uncertainty?
 
    | dekhn wrote:
    | I doubt it makes sense to assume the unverise is continuous
    | (I'm glad you said "seems"). In particular, space could be
    | spatially quantized (say, around the planck length) or any
    | number of other details.
    | 
    | People have done simulations with quad precision (very slow)
    | but very few terms in molecular dynamics would benefit from
    | that. In fact, most variables in MD can be single precision,
    | exceptt for certain terms like the virial.
 
      | whatshisface wrote:
      | All of our current theories are set in continuous
      | spacetime. At the present, there's no reason to assume
      | anything else.
 
        | dekhn wrote:
        | the issue is that there are no theories based on
        | experimental evidence at very small scales. I agree that
        | in most situations, it would be silly to violate this
        | assumption, unless you were working on advanced physics
        | experiments.
 
        | jacquesm wrote:
        | True, but we do not actually know this for sure. There is
        | a (small) possibility that we are simply looking at this
        | at a scale where all we see is macro effects. It would
        | require the quanta to be much smaller than the Planck
        | distance though.
 
        | webmaven wrote:
        | _> There is a (small) possibility that we are simply
        | looking at this at a scale where all we see is macro
        | effects. It would require the quanta to be much smaller
        | than the Planck distance though._
        | 
        | How much smaller?
 
        | jacquesm wrote:
        | Many orders of magnitude. How many? I do not know, I
        | don't think anybody does.
        | 
        | But photons resulting from the same event but with
        | different energies arrive at detectors an appreciable
        | distance away to all intents and purposes simultaneously,
        | something that would not happen if spacetime were
        | discrete at a level close to the Planck length. So it
        | would have to be quite a big difference for an effect
        | _not_ to show up as a difference in time-of-flight.
 
        | feoren wrote:
        | I wouldn't say that "all our current theories" are set in
        | continuous spacetime. For example, Quantum chromodynamics
        | is set in SU(3), an 8-dimensional group of rotation-like
        | matrices. Electric charge is discrete, spin is discrete,
        | electron orbitals are discrete. In fact position and
        | momentum would seem to be the outlier if they were not
        | also discrete. I hardly call that "no reason".
 
        | whatshisface wrote:
        | SU(3) is a continuous group.
 
        | freemint wrote:
        | Yeah but it is very much not in space time.
 
        | whatshisface wrote:
        | But it is. SU(3) is the group for swapping colors around.
        | It still has spacetime.
 
        | freemint wrote:
        | You can Cartesian product it with space time, yes. But
        | that is possible for any system.
 
        | drdeca wrote:
        | It is based on SU(3), but, does it really make sense to
        | say that it isn't still set in spacetime? Like, quarks
        | still have position operators, yes?
 
      | mensetmanusman wrote:
      | It's definitely fun to think about.
      | 
      | If the universe is discrete, how does one voxel communicate
      | to the neighboring voxel what to update without passage
      | through 'stuff in between' that doesn't exist? Heh
      | 
      | It seems physics is going the opposite way with infinite
      | universes and multiple dimensions to smooth out this
      | information transfer problem and make the discrete go away.
 
        | javajosh wrote:
        | As someone with keen interest in physics (and a bit of
        | training) I find speculation about "discrete space"
        | disquieting. It's the level of abstraction where
        | intuition about space breaks down, and you have to be
        | very careful. Remember that coordinate systems are short-
        | hand for measurement. It's one thing to admit fundamental
        | limits on measurement resolution, and quite another to
        | say that space itself is quantized! Mostly I get around
        | this by not thinking about it; most of these theories are
        | only testable in atrocious and unattainable conditions,
        | doing things like performing delicate QED experiments at
        | the edge of a black hole.
        | 
        | I don't think your "voxel" intuition can be right because
        | it's a small jump from that to (re)introducing an
        | absolute reference frame.
 
        | joshmarlow wrote:
        | > how does one voxel communicate to the neighboring voxel
        | what to update without passage through 'stuff in between'
        | that doesn't exist? Heh
        | 
        | That kind of reminds me of the 'aether' that was once
        | hypothesized as a medium of transmission for light and
        | radio waves [0].
        | 
        | Also, voxel's communicating sounds an awful lot like a
        | higher-dimensioned cellular automata.
        | 
        | [0] - https://en.wikipedia.org/wiki/Aether_theories
 
        | Yajirobe wrote:
        | Stephen Wolfram was right all along
 
  | Frost1x wrote:
  | It doesn't appear to be ab initio simulated (e.g. QED up) if
  | that's what you're asking. They appear to swoop in at higher
  | scales (molecular level) and simulate molecular interactions
  | across "hundreds of molecular species" and "thousands of
  | reactions."
  | 
  | Apparently the interface between molecules uses the Chemical
  | Master Equations (CME) and Reaction-Diffusion Master Equations
  | (RDME) both of which I'm unfamiliar with:
  | http://faculty.scs.illinois.edu/schulten/lm/download/lm23/Us...
 
    | vasili111 wrote:
    | For anyone who is wondering what QED is: Quantum
    | electrodynamics (QED)
    | https://en.wikipedia.org/wiki/Quantum_electrodynamics
 
      | kingcharles wrote:
      | Ah, should have realized when Quad Erat Demonstradum made
      | no sense...!
 
    | dahart wrote:
    | Yes, this appears to be the underlying simulation software.
    | Here's a home page link to the project as well:
    | http://faculty.scs.illinois.edu/schulten/Software2.0.html
    | 
    | "Lattice Microbes is a software package for efficiently
    | sampling trajectories from the chemical and reaction-
    | diffusion master equations (CME/RDME) on high performance
    | computing (HPC) infrastructure using both exact and
    | approximate methods."
 
  | [deleted]
 
  | marcosdumay wrote:
  | The paper (well, the abstract) calls it "fully dynamical
  | kinetic model".
  | 
  | Or, in other words, it doesn't solve the Schrodinger equation
  | at all, but uses well known solutions for parts of the
  | molecules, and focuses on simulating how the molecules interact
  | with one another using mostly classical dynamics.
 
    | blix wrote:
    | I do classical molecular dynamics simulations for a living,
    | and I feel the model using in this paper is pretty
    | dramatically different than what would typically be described
    | as classical dynamics. 2B atoms would be absolutely insane
    | for any sort of simulation that resolves forces between atoms
    | of even groups of atoms, especially in organic systems.
    | 
    | As far as I can tell from their model, molecules don't
    | interact with each other ~at all~ through classical dynamics.
    | Rather, they define concentrations of various molecules on a
    | voxel grid, assign diffusion coffecients for molecules and
    | define reaction rates between each pair of molecules. Within
    | each voxel, concentrations are assumed constant and evolve
    | through a stochastic Monte-Carlo type simulation. Diffusion
    | is solved as a system of ODEs.
    | 
    | This is a cool large scale simulation using this method, but
    | this is a far cry from an actual atomic-level simulation of a
    | cell, even using the crude approximations of classical
    | molecular dynamics. IMO it is kind of disingenuous for them
    | to say 2B atoms simulation when atoms don't really exist in
    | their model, but it's a press release so it should be
    | expected.
 
  | CorrectHorseBat wrote:
  | Of course not, we can't even simulate how one protein folds.
 
    | Cthulhu_ wrote:
    | What does https://foldingathome.org/ do then? That's been
    | going on for nearly two decades.
 
      | echelon wrote:
      | Simulating very expensive to compute protein dynamics.
      | These aren't guaranteed solutions, but it's still useful
      | information.
 
        | vasili111 wrote:
        | So, even one protein cannot be simulated as in real
        | world?
 
        | qboltz wrote:
        | Not if you're going off of ab initio theory such as
        | Hartee Fock, MP2, CC, etc. We're talking amounts of
        | matrix multiplication that wouldn't be enough to finish
        | calculating this decade, even if you had parallel access
        | to all top 500 supercomputers, you get bigger than a
        | single protein, it's beyond universal time scales with
        | current implementations.
 
        | dekhn wrote:
        | Every time some computer scientist interviews me and
        | shows off their O(n) knowledge (it's always an o(n)
        | solution to a naive o(n**2) problem!) I mention that in
        | the Real World, engineers routinely do O(n**7)
        | calculations (n==number of basis functions) on tiny
        | systems (up to about 50 atoms, maybe 100 now?) and if
        | they'd like to help it would be nice to have better,
        | faster approximations that are n**2 or better.
        | Unfrotunately, the process of going from computer
        | scientist to expert in QM is entirely nontrivial so most
        | of them do ads ML instead
 
        | dekhn wrote:
        | A custom supercomputer dedicated to simulating folding
        | proteins (two-state folders with nontrivial secondary and
        | tertiary structure) from unfolded to correctly folded
        | state using only classical force fields _probably_ could
        | work, and DE Shaw has invested a lot of money in that
        | idea: https://en.wikipedia.org/wiki/Anton_(computer)
        | 
        | but, as I pointed out elsewhere, this would not be
        | particularly helpful as it would use an enormous amount
        | of resources to compute something we could probably
        | approximate with a well-trained ML model.
        | 
        | It also wouldn't address questions like biochemistry,
        | enzymatic reactions, and probably wouldn't be able to
        | probe the energetics of interactions accurately enough to
        | do drug discovery.
 
        | beecafe wrote:
        | One single iron atom's electrons - 26 of them - contain
        | more degrees of freedom than atoms in the solar system.
 
        | JabavuAdams wrote:
        | Even one atom of a heavier element cannot be simulated in
        | the real, depending on what level of detail you want.
        | Multi-atom simulations usually treat them as little non-
        | quantum balls moving around in a force-field that may
        | have been approximated from quantum mechanics.
 
      | orangepurple wrote:
      | Full list of achievements
      | https://foldingathome.org/category/fah-achievements/?lng=en
      | 
      | This is the only real update of the year:
      | https://foldingathome.org/2022/01/03/2021-in-review-and-
      | happ...
      | 
      | SARS-CoV-2 has intricate mechanisms for initiating
      | infection, immune evasion/suppression and replication that
      | depend on the structure and dynamics of its constituent
      | proteins. Many protein structures have been solved, but far
      | less is known about their relevant conformational changes.
      | To address this challenge, over a million citizen
      | scientists banded together through the Folding@home
      | distributed computing project to create the first exascale
      | computer and simulate 0.1 seconds of the viral proteome.
      | Our adaptive sampling simulations predict dramatic opening
      | of the apo spike complex, far beyond that seen
      | experimentally, explaining and predicting the existence of
      | 'cryptic' epitopes. Different spike variants modulate the
      | probabilities of open versus closed structures, balancing
      | receptor binding and immune evasion. We also discover
      | dramatic conformational changes across the proteome, which
      | reveal over 50 'cryptic' pockets that expand targeting
      | options for the design of antivirals. All data and models
      | are freely available online, providing a quantitative
      | structural atlas.
 
    | dekhn wrote:
    | Small proteins (one to two alpha helices) can now be
    | routinely folded (that is, starting form a fully unfolded
    | state, to getting stick in the minimum around the final
    | structure) using ab initio simulations that last several
    | multiples of the folding time.
    | 
    | Larger proteins (a few alpha helices and beta sheets), the
    | folding process can be studied if you start with structures
    | near the native state.
    | 
    | None of this means to say that we can routinely take any
    | protein and fold it from unfolded state using simulations and
    | expect any sort of accuracy for the final structure.
 
      | qboltz wrote:
      | When you say ab initio calculations, could you cite the
      | level of theory? I think there could be some ambiguity
      | given differences in scope.
 
        | dekhn wrote:
        | When I say ab initio I mean "classical newtonian force
        | field with approximate classical terms derived from QM",
        | AKA something like https://ambermd.org/AmberModels.php
        | 
        | Other people use ab initio very differently (for example,
        | since you said "level of theory" I think you mean basis
        | set). I don't think something like QM levels of theory
        | provide a great deal of value on top of classical (and at
        | a significant computational cost), but I do like 6-31g*
        | as a simple set.
        | 
        | Other people use ab initio very differently. For example,
        | CASP, the protein structure prediction, uses ab initio
        | very loosely to me: "some level of classicial force
        | field, not using any explicit constraints derived from
        | homology or fragment similarity" which typically involves
        | a really simplified or parameterized function (ROSETTA).
        | 
        | Personally I don't think atomistic simulations of cells
        | really provide a lot of extra value for the detail. I
        | would isntead treat cell objects as centroids with mass
        | and "agent properties" ("sticks to this other type of
        | protein for ~1 microsecond"). A single ribosome is a
        | single entity, even if in reality it's made up of 100
        | proteins and RNAs, and the cell membrane is modelled as a
        | stretchy sheet enclosing an incompressible liquid.
 
        | blix wrote:
        | I would not describe AMBER, or anything using a newtonian
        | force field, as ab initio.
        | 
        | In inorganic materials ab initio means you actually solve
        | Schrodinger's equation (though obviously with aggressive
        | simplifications e.g. Hartree-Fock).
 
        | qboltz wrote:
        | Level of theory as it relates to an-initio QM
        | calculations usually indicates Hartee Fock, MP2 and so
        | on, then the basis set gets specified after.
        | 
        | I also agree that QM doesn't provide much for the cost at
        | this scale, I just wish the term ab initio would be left
        | to QM folks, as everything else is largely just the
        | parameterization you mentioned.
 
        | dekhn wrote:
        | The systemn I work with, AMBER, explains how individual
        | classical terms are derived: https://ambermd.org/tutorial
        | s/advanced/tutorial1/section1.ht... which appears te be
        | MP2/6-31g* (sorry, I never delved deeply into the QM
        | parts). Once those terms are derived, along with various
        | approximated charges (classical fields usually just treat
        | any charge as point-centered on the nucleus, which isn't
        | great for stuff like polarizable bonds), everything is
        | purely classical springs and dihedrals and interatomic
        | potentials based on distance.
        | 
        | I am more than happy to use "ab initio" purely for QM,
        | but unfortunately the term is used widely in protein
        | folding and structure prediction. I've talked
        | exdtensively with David Baker and John Moulton to get
        | them to stop, but they won't.
 
        | [deleted]
 
| intrasight wrote:
| Someone is simulating all of my cells - and yours too ;)
 
  | CapsAdmin wrote:
  | If this is true, there must be some species at some level of
  | simulation who's not being simulated.
  | 
  | I'm not sure if you're being real or not, but if you are, do
  | you think the species running who made our simulation are also
  | being simulated?
 
    | webmaven wrote:
    | _> If this is true, there must be some species at some level
    | of simulation who 's not being simulated._
    | 
    | You can't fool me, it's turtles all the way down!
    | 
    | With that out of the way, I'll observe there is no reason
    | that such a base layer of reality need bear any particular
    | resemblance to ours except in the tautological sense that it
    | would need to be Turing complete in order to be capable of
    | hosting a simulation.
 
      | CapsAdmin wrote:
      | I agree that it would probably not resemble our universe. I
      | would think it has to be a universe that's capable of
      | simulating our universe without consuming all of the host
      | universe's resources as it would need at least some sort of
      | species that would want to simulate our universe. At least
      | initially.
      | 
      | I'm not sure what you (and other people) really mean when
      | you say our universe is simulated.
      | 
      | - Do you mean that the entire universe is simulated down to
      | the planck level? - Do you think there's some sort of
      | optimization going on? - Do you think it's done by a
      | species that evolved to become curious to see what would
      | happen if you simulate the universe (like us)?
      | 
      | I can say that our universe is simulated too, but I have no
      | idea if this simulation was made by someone or if it "just
      | is".
      | 
      | But if you believe the universe is a simulation in some
      | host universe, then it must be possible to have a universe
      | that "just is" / or is Turing complete as you put it.
 
        | webmaven wrote:
        | I mean that such a universe could be so different from
        | ours that the idea of 'species' may not even be sensible.
        | 
        |  _> Do you mean that the entire universe is simulated
        | down to the planck level?_
        | 
        | Unspecified. Perhaps gross approximations are used unless
        | an attempt is made to observe (internally or externally)
        | more detail.
 
        | CapsAdmin wrote:
        | > I mean that such a universe could be so different from
        | ours that the idea of 'species' may not even be sensible.
        | 
        | Alright. I've heard people say they think our universe is
        | being simulated because that's what we would do. For
        | those who think that, the host universe is at least
        | somewhat similar to us.
        | 
        | > Unspecified. Perhaps gross approximations are used
        | unless an attempt is made to observe (internally or
        | externally) more detail.
        | 
        | But if gross approximations are true, that reveals
        | information about the host doesn't it? If they resort to
        | approximations because they don't have enough resources,
        | that tells us they must really want to do this for some
        | reason. Did they want to create our simulation for fun?
        | Out of desperation? Are we made for research purposes?
        | All those questions point to something human-like in my
        | opinion, and thus "species".
 
    | svachalek wrote:
    | Could be an Ouroboros, the entirety of existence being
    | created from nothing in an enormous circular dependency. It
    | sounds farcical but when you think about why the universe
    | exists in the first place, it seems as good a reason as any.
 
      | CapsAdmin wrote:
      | > when you think about why the universe exists in the first
      | place, it seems as good a reason as any.
      | 
      | I think this sums up how I think. If any reason is as good
      | as any then it's equally likely that our universe is not
      | simulated and not an Ouroboros.
      | 
      | It can be a lot of fun to speculate and think about though.
 
| VikingCoder wrote:
| How long did it take to simulate 20 minutes?
| 
| Looks like one NVIDIA Titan V took 10 hours to do it, and one
| NVIDIA Tesla Volta V100 GPU took 8 hours to do it?
| 
| Am I reading that right?
| 
| So the NVIDIA Tesla Volta V100 is 24 times slower than real life?
| Pretty cool.
 
  | Koshkin wrote:
  | Generally speaking, this depends on the size (in terms of the
  | number of constituents) of the piece of "real life" you are
  | simulating.
 
| kingcharles wrote:
| The question is (again) how soon now until I can boot the "ROM"
| file of my DNA in an emulator?
 
  | jacquesm wrote:
  | This won't happen. Computationally inconceivable with all that
  | we know at the moment.
 
| sydthrowaway wrote:
| This seems like the company that will dominate the 2020s. The
| time is ripe to join NVIDIA
 
| rsfern wrote:
| This is really cool, but I don't think it's an atomistic
| simulation so I'm not sure where the title is coming from.
| 
| It seems to be some kind of a (truly impressive) kinetic model
| 
| The paper in Cell is open access
| https://doi.org/10.1016/j.cell.2021.12.025
 
| fefe23 wrote:
| Nvidia GPUs enable nothing, because you can't buy any at
| reasonable prices.
 
  | hiptobecubic wrote:
  | "That place is so crowded that no one goes there anymore."
 
    | HPsquared wrote:
    | Leela: Did you drive much in the 20th century, Fry?
    | 
    | Fry: Nobody in New York drove, there was too much traffic.
 
  | jabbany wrote:
  | This argument makes no sense. Consumer GPU pricing (which I'm
  | assuming is what you're referring to) has very little to do
  | with the pro market (industry, research etc.)
  | 
  | The researchers are using things like the DGX or RTX A-series.
  | These, while quite expensive, are not that unreasonable when it
  | comes to pricing.
 
    | pepemon wrote:
    | An individual could afford computing power for such research
    | activities (not exactly like this one, but e.g. for personal
    | ML experiments) in 2018-2019 for an adequate price. You were
    | able to buy 2 new RTX2080s for the today price of a used
    | single unit. If you want to tinker and need GPU power today,
    | your best option is to rent special datacenter-approved(tm)
    | GPUs for the really expensive $/h. And you don't own anything
    | afterwards (except if you bought GPU before the end of 2020).
    | Does this make no sense? Is this how technological progress
    | should work?
 
      | freemint wrote:
      | If you don't care if some rando who's machine you rented
      | does see what you are doing vast.ai can be a good resource
      | for GPU compute too.
 
      | jabbany wrote:
      | 2080s? With only 8GB of VRAM that's not even ECC backed?
      | 
      | Even for ML model training back then, 8GB was on the small
      | side (a lot of the research repos even had special
      | parameter sets to allow running on consumer level VRAM
      | GPUs). Also, for something like long running bio
      | simulations, you'd probably want to be sure that your
      | memory bits aren't being flipped by other sources -- the
      | extra upfront cost is well worth preventing potentially
      | wrong research results...
      | 
      | Nvidia consumer products have been a better value
      | proposition in the past for sure. But they've always done
      | market segmentation. It's not merely a matter of
      | "datacenter-approved(tm) GPU" (though they do also do
      | driver-based segmentation).
 
| lolive wrote:
| Apart from this article, do we have any visual representation [a
| CGI, may be] of the full activity inside a cell?
 
  | unemphysbro wrote:
  | vmd is a standard biological system simulation rendering
  | software
 
  | 01100011 wrote:
  | I always appreciated the work of David Goodsell at UCSD:
  | https://ccsb.scripps.edu/goodsell/
  | 
  | He paints cell internals.
  | 
  | I also like the Biovisions videos from Harvard:
  | 
  | https://www.youtube.com/watch?v=VdmbpAo9JR4
 
  | vasili111 wrote:
  | We can't visualize what we do not know. Full activity inside
  | cell is not known and we are pretty far from knowing that.
 
  | lolive wrote:
  | I have always been amazed by these 2D representations:
  | https://www.digizyme.com/cst_landscapes.html
 
    | Traubenfuchs wrote:
    | All of this incomprehensible complexity just so our genes can
    | compete against other genes in their mindless drive for
    | survival. It's kind of sad.
 
      | [deleted]
 
      | jjoonathan wrote:
      | Things really got out of hand after that first self-
      | replicating gizmo, didn't they?
 
      | lolive wrote:
      | The boilerplate to make a double click with your mouse do
      | something relevant is also completely mind blowing.
      | #complexity
 
    | dekhn wrote:
    | one note- as lovely as those are, they don't make the point
    | that everything in the cell (all the proteins, etc) is
    | constantly grinding against each other (there's almost no
    | room for water).
 
    | maze-le wrote:
    | That's fascinating, thanks for sharing!
 
| airstrike wrote:
| Now we just need to scale this by a mere 37,200,000,000,000x and
| we'll have simulated the entire human body!
 
  | chroem- wrote:
  | Moore's law suggests it will be possible in 90 years if the
  | historical trend holds true.
 
    | T-A wrote:
    | https://www.cnet.com/tech/computing/moores-law-is-dead-
    | nvidi...
 
      | elil17 wrote:
      | Price and energy use can still go down even if transistor
      | density stays the same
 
      | chroem- wrote:
      | Luckily, there are other means of performing computation
      | than just silicon transistors.
 
| [deleted]
 
  | ajuc wrote:
  | I think it has to reproduce to qualify.
 
| agentultra wrote:
| Permutation City, here we come.
| 
| I wonder if Greg Egan had the foresight to predict this for the
| story or if he invented that part for narrative purposes.
 
  | chinathrow wrote:
  | When I was like 12 or so, I had a thought that if we can
  | calculate everything, we could be living in a full blown
  | simulation.
  | 
  | To be honest, like 30y later, I still go back to that nagging
  | thought _a lot_.
 
    | afshin wrote:
    | This idea has been formalized: https://www.simulation-
    | argument.com/
 
      | sva_ wrote:
      | This idea has also existed for at least 200 years
      | 
      | https://en.m.wikipedia.org/wiki/Laplace%27s_demon
 
        | lelandfe wrote:
        | https://en.wikipedia.org/wiki/Evil_demon
        | 
        | Going further back to the 1600's, Descartes' idea of an
        | evil demon deceiving one's mind with a perfect, fake
        | reality made me think often of simulations in my
        | undergrad philosophy classes
 
        | dekhn wrote:
        | I read that as a teenager, thought it sounded nice, went
        | to grad school and did molecular dynamics simulations
        | (like folding at home) for a decade, then went to google
        | and built the world's largest simulation system
        | (basically, the largest group of nodes running folding at
        | home). Eventually we shut the system down because it was
        | an inefficient way to predict protein structure and
        | sample folding processes (although I got 3-4 excellent
        | papers from it).
        | 
        | The idea is great, it was a wonderful narrative to run my
        | life for a while, but eventually, the more I learned, the
        | more impractical using full atomistic simulations seem
        | for solving any problem. It seems more likely we can
        | train far more efficient networks that encapsulate all
        | the salient rules of folding in a much smaller space, and
        | use far less CPU time to produce useful results.
 
        | sva_ wrote:
        | Yeah, I think the idea of Laplace's Demon is mostly just
        | useful to make a philosophical argument about whether or
        | not the universe is deterministic, and it's implication
        | on free will.
 
        | dekhn wrote:
        | I dunno, I wonder what Laplace would have made of the
        | argument over the meaning of wavefunction collapse. It
        | took me a very long time to come to terms with the idea
        | of a non-deterministic universe.
 
        | mensetmanusman wrote:
        | It's interesting that many things are deterministic to
        | human-relevant time/length scales. If the small stuff is
        | non-deterministic, it's interesting that large ensembles
        | of them are quite deterministic.
        | 
        | It's maddening :)
 
        | sva_ wrote:
        | That's peculiar. Most people probably struggle more with
        | the idea of a deterministic universe, as it'd leave no
        | room for free will, which would make everything kind of
        | meaningless.
        | 
        | I'm also more in the camp of "quantum effects making the
        | universe non-determinstic." It's a nicer way to live.
 
        | dekhn wrote:
        | I've evolved over the years from "determinism implies no
        | free will" to roughly being a compatibilist
        | (https://en.wikipedia.org/wiki/Compatibilism, see also
        | Daniel Dennett). I don't particularly spend much time
        | thinking that (for example) a nondeterministic universe
        | is required for free will. I do think from an objective
        | sense the universe is "meaningless", but that as humans
        | with agency we can make our own meaning.
        | 
        | However, most importantly, we simply have no experimental
        | data around any of this for me to decide. Instead I enjoy
        | my subjective life with apparent free will, regardless of
        | how the machinery of the actual implementation works.
 
    | Taylor_OD wrote:
    | It's a bit naive... But the best argument for me that we are
    | living in a simulation is that we went from Pong to pretty
    | good VR (good enough that if you have a beer or two before
    | using you can forget its VR for some period of time) in 50
    | years. In another 50 years it seems fair to assume that we
    | will be able to create VR that fully immersive and impossible
    | to distinguish from real life.
    | 
    | Even with no other arguments about the benefits of WHY one
    | would want to live in a fully simulated world... It seems
    | probable to me that we are based on the idea that it could be
    | possible.
 
      | iamstupidsimple wrote:
      | > In another 50 years it seems fair to assume that we will
      | be able to create VR that fully immersive and impossible to
      | distinguish from real life.
      | 
      | Technology growth is always non-linear. it's also fair to
      | assume we could stagnate for 50 years also.
 
    | kingofclams wrote:
    | https://qntm.org/responsibility
 
    | joseluis wrote:
    | we don't even need to be able to calculate everything, we
    | just need to fool you! The Truman's show meets the Matrix.
 
    | amself wrote:
    | I went through the same phase at 12. I am nearing 18 now, and
    | I am very thankful for nondeterminism.
 
    | tsol wrote:
    | Looking at it from that view, we're just as likely to be a
    | simulation as we are to have been created by God. I mean I'm
    | a theist, but I don't see many huge differences except the
    | cultural aspect where the theism/atheism debate is something
    | most people have an emotional connection to.
 
      | KarlKemp wrote:
      | A God, not being out for her own amusement, will likely
      | create only one universe.
      | 
      | A player with a simulator will create dozens.
 
        | tsol wrote:
        | >A God, not being out for her own amusement, will likely
        | create only one universe.
        | 
        | Why would that be? I see no reason why God might not
        | create parallel universes
 
        | coolspot wrote:
        | Electrical bill and GPU shortages in God's reality could
        | be a reason.
 
    | reasonabl_human wrote:
    | If you want to solve that nagging thought, pick up Griffith's
    | intro to quantum mechanics textbook. Goes through the
    | philosophical implications of qm alongside learning the
    | physics. The world as we know it is non-deterministic thanks
    | to wave functions and their random collapsing!
 
    | benlivengood wrote:
    | The thought that sticks in my mind is mathematical realism;
    | if we can prove the mathematical existence of the outcome of
    | a simulation (nothing harder than inductively showing that
    | the state of a simulation is well-defined at state S for the
    | first and all successive S) then what's the difference
    | between things in the simulation actually existing v.s.
    | possibly existing? All of the relationships that matter
    | between states of the simulation are already proven to exist
    | if we looked at (calculated) them, so what necessary property
    | can we imagine our Universe having that the possible
    | simulation does not?
 
      | visarga wrote:
      | > so what necessary property can we imagine our Universe
      | having that the possible simulation does not?
      | 
      | It lacks the magical spark, the qualia, the spirit, the
      | transcendent. Or what people like to imagine makes our own
      | reality special. Our own reality cannot be understood
      | because it's such a hard problem, and it "feels like
      | something" (maybe like a bat?), while a simulation is just
      | math on CPUs. Consciousness is a hard problem because it
      | transcends physical sciences, it's so great that it can
      | exist even outside the realm of verification. /s
      | 
      | Hope you forgive the rant, it's just amazing how much
      | philosophy can come from the desire to fly above the
      | mechanics of life. But what they missed is that the reality
      | of what happens inside of us is even more amazing than
      | their imaginary hard problem and special subjective
      | experience. The should look at the whole system, the whole
      | game, not just the neural correlates. What does it take to
      | exist like us?
 
      | VikingCoder wrote:
      | A simulated hurricane doesn't kill anyone.
      | 
      | But it may be possible that there's no such thing as
      | "simulating" intelligence. If you do certain calculations,
      | that is "intelligent." Same for consciousness, etc.
 
        | wrinkl3 wrote:
        | A simulated hurricane would kill simulated people.
 
        | tsol wrote:
        | Think of simulated children! Oh the simulated pain..
 
        | disease wrote:
        | "We live inside a dream."
 
  | agentultra wrote:
  | Specifically, I'm referring to _Autoverse_ , the artificial
  | simulation of a single bacterium down to the atomic level.
  | 
  | It was such a fascinating idea that I found myself more than
  | once trying to mimic the atomic part at a much smaller scale
  | over the years.
 
  | webmaven wrote:
  | _> Permutation City, here we come._
  | 
  | You might enjoy the show Devs:
  | 
  | https://en.m.wikipedia.org/wiki/Devs
  | 
  | Fair non-spoilery warning, there is quite a bit of creepy
  | existential angst.
 
    | stronglikedan wrote:
    | > there is quite a bit of creepy existential angst
    | 
    | seems to be a trend across all genres nowadays
 
      | webmaven wrote:
      | Sure. Though most of it isn't _literally_ existential.
 
___________________________________________________________________
(page generated 2022-01-28 23:00 UTC)