[HN Gopher] NeRF: An eventual successor for deepfakes?
___________________________________________________________________
 
NeRF: An eventual successor for deepfakes?
 
Author : Hard_Space
Score  : 142 points
Date   : 2022-06-27 13:47 UTC (9 hours ago)
 
web link (metaphysic.ai)
w3m dump (metaphysic.ai)
 
| 1970-01-01 wrote:
| Neural Radiance Fields (NeRF)
 
| Workaccount2 wrote:
| I know there are a lot of groups working on how to prevent AI
| disruptions to society, or how to mitigate their impact, but are
| there any groups working on how to adapt society to a full blown
| unchained AI world?
| 
| Like throw out all the safeguards (which seems inevitable) and
| how does society best operate in a world where no media can be
| trusted as authentic? Or where "authentic" is cut from the same
| cloth as "fake"? Is anyone working on this?
 
  | Agamus wrote:
  | One thing we should be doing is supporting critical thinking at
  | the high school and university level. Unfortunately, it seems
  | we have been dedicated to the opposite of this for about 50
  | years, at least in the US.
 
    | Der_Einzige wrote:
    | Critical thinking is the most overrated skill ever.
    | 
    | "Critical theorists" are the people who fetishize "critical
    | thinking" and all it got them was to embrace cultural
    | Marxism.
    | 
    | Constructive thinking is far better than learning how to shit
    | on people, as the skill of critique teaches us...
 
      | gknoy wrote:
      | At the risk of falling for a joke, I'm not sure "critical
      | thinking" means what you think it means. It just means
      | thinking objectively about things before making judgments,
      | it has nothing to do with criticizing people. The things
      | one criticizes are our own beliefs and reasons for
      | believing them.
      | 
      | What do I believe? Why do I believe that? Why do I feel
      | that evidence supports that belief, but not this one? For
      | example, I can explain in a fair bit of detail why I
      | believe that the Apollo landing was not faked. I wouldn't
      | normally bother to explain those reasons, but all of them
      | are based on beliefs and evidence that I've read about, and
      | most of those beliefs are subject to reversal should
      | counter-evidence surface.
 
      | whimsicalism wrote:
      | This is reasoning by word chaining.
 
      | Agamus wrote:
      | I think of critical thinking as the art of being critical
      | toward oneself when one is thinking.
      | 
      | In other words, when I read something and hear myself
      | think, "oh yeah, that sounds right", there is another part
      | of my mind that thinks, "maybe not".
      | 
      | Critical thinking is precisely what could have spared us
      | from all of that 'cultural marxism' you mentioned, or at
      | least, to do it in a way that is... constructive.
 
  | strohwueste wrote:
  | What about nft-based camera recording?
 
    | irjustin wrote:
    | This has been discussed many times and it doesn't work.
    | 
    | Simple answer is I can just record a deep fake and get it
    | cryptographically signed.
 
      | andruby wrote:
      | If you trust a person (or source) and they have a private
      | key that they can properly secure, they could always sign
      | their material with that key. That would prove that the
      | source provided that material.
      | 
      | A blockchain could be a way to store and publish that
      | signature & hash.
      | 
      | It can't say "this is real", it can only say "that
      | signature belongs to source X".
 
        | blamestross wrote:
        | > A blockchain could be a way to store and publish that
        | signature & hash.
        | 
        | yes but it would be a bad one. We have multiple key
        | distribution mechanism that are better for the use case.
 
        | intrasight wrote:
        | I disagree. The key alone is not sufficient nor secure.
        | We will need crowdsourced validity data as well. We need
        | a zero-trust model - and I too believe that blockchains
        | will play a role.
 
        | notfed wrote:
        | If a video is already cryptographcally signed, then you
        | can safely distribute the signature on an untrusted
        | channel.
        | 
        | Adding Blockchain into the mix is superfluous, and
        | destroys scalability.
 
        | intrasight wrote:
        | We don't watch signature keys - we watch videos.
        | 
        | The TV will have to match every short segment - perhaps 5
        | seconds of video - against a blockchain which scores the
        | validity of that segment - and of course looking back to
        | it's original source. Signing the whole video is
        | necessary but not sufficient.
        | 
        | But yes, this is going to be resource-intensive.
 
        | blamestross wrote:
        | > against a blockchain which scores the validity of that
        | segment
        | 
        | Why not just allow a cert with that information to be
        | delivered alongside the video? Where would the "score"
        | come from?
 
        | blamestross wrote:
        | Zero trust models don't exist and the laws of physics
        | (probably) don't provide for them. (Materialism is a real
        | problem in physics nowadays)
 
  | endtime wrote:
  | I think most of those people believe that humans will no longer
  | exist in a "full blown unchained AI world".
 
  | cameronh90 wrote:
  | I suspect we'll need to return to the idea of getting our news
  | from trusted sources, rather than being able to rely on videos
  | on social media being trustworthy.
  | 
  | Technically, we could try and build a trusted computing-like
  | system to ensure trust from the sensor all the way to a signed
  | video file output, but keeping that perfectly secure is likely
  | to be virtually impossible except in narrow situations, such as
  | CCTV installations. I believe Apple may be attempting to do
  | things like this with how Face ID is implemented on iPhone, but
  | I suspect we'll always find ways to trick any such device.
 
    | wongarsu wrote:
    | 80% of the problem could be solved with a reliable signature
    | scheme that allows some remixing of video content. So if CNN
    | publishes a video, signed with their key so it's verifiably
    | CNN, we need the ability to take a 20 second bit of it and
    | still have a valid key attached that verifies that the source
    | is CNN (without trusting the editor). Then you can share
    | clips, remix it, etc, and have integration in social media
    | that attests the source.
 
      | _tom_ wrote:
      | Once you remix it, it's no longer reliable. So, you don't
      | want it signed if it's modified.
 
      | intrasight wrote:
      | My plan to solve this "20 second bit of it" is that it's
      | done at the analog hole. Whatever is painting those pixels,
      | a smart TV for instance, will be coordinating with cloud
      | services to fingerprint at a relatively high temporal
      | resolution - maybe 5 seconds. The video itself is the
      | signature. But we will need either trusted analog hole
      | vendors or some trusted non-profit organization - or likely
      | both. I think that "viewing" will be delayed by perhaps 30
      | seconds to allow for that signature analysis. These smart
      | TVs will overlay a scorecard for all displayed content, and
      | owners will be able to set device scorecard thresholds such
      | that low-scoring content will be fuzzed out.
 
        | dTal wrote:
        | I sincerely hope this dystopian vision of the future is
        | satire, but it's already a worrying sign of the times
        | that I'm not sure.
 
| wazoox wrote:
| There are more dangereous AI than deepfakes. Blackrock 10
| trillions investments are driven by an AI, Aladdin. It was also
| sold to some other investors, and controls about 21$ trillion
| globally. It has basically the power to drive markets worldwide.
| It's a systemic problem nobody talks about...
 
  | astrange wrote:
  | Blackrock sells passive index funds. There's no room for an AI
  | to make decisions there, so it probably doesn't do anything
  | interesting.
 
  | smiddereens wrote:
 
  | nix0n wrote:
  | That's not an AI problem, it's a concentration of wealth
  | problem.
  | 
  | Giving that power to a person or group of people would be
  | almost as bad.
 
| kobalsky wrote:
| why do websites feel the need to hijack the browser's scrolling
| logic?
| 
| this is very annoying to browse on chrome, but it works well on
| firefox.
 
  | sigspec wrote:
  | Agree. It's a jolt to scrolling expectations
 
  | arky527 wrote:
  | Very much agreed. It just results in a terrible UX when 98% of
  | other websites have a standard scrolling mechanism
 
| [deleted]
 
| jdthedisciple wrote:
| The bad news: This can and obviously will be abused - be it by
| the secret services or hackers.
| 
| The good news, I suppose: As fast and scarily as the tech to fake
| things is evolving, so is presumably the tech that detects fakes.
 
  | _tom_ wrote:
  | Technology to detect fakes necessarily lags a bit from
  | technology to create fakes.
  | 
  | In general, you need to have examples of a type of fake to
  | detect it.
 
| RubyRidgeRandy wrote:
| Something I've wondered lately is what will life be like in a
| post-truth society? we already see examples of this now where a
| large number of people get their news from fake memes on
| facebook. There are huge swathes of people who live in their own
| make-believe world, like those believing wholeheartedly that the
| 2020 election was stolen.
| 
| What will life be like when you can't trust any video or
| interview you see because it could be completely fake? How long
| before someone uses this technology to frame someone for a crime?
| Could the FBI create a deepfake of a cartel leader meeting with
| them and leak it so they think he's a snitch?
| 
| I don't think we'll have the ability to handle this kind of tech
| responsibly.
 
  | hutzlibu wrote:
  | "I don't think we'll have the ability to handle this kind of
  | tech responsibly."
  | 
  | I do not think so either, but so far we survived 75+years with
  | nukes around.
  | 
  | But you can argue, it was mainly by chance. Technological
  | progress _is_ awesome, but our societies cannot keep up yet.
  | They will have to do heavy transition anyway, or perish. Or
  | rather, we are in the process of transition. 20 years ago most
  | people did not really know, what the internet is, now most are
  | always online. Data mining, personalised algorithms for ad
  | exposing, ..
  | 
  | So deepfakes are a concern, but not my biggest. Rather the
  | contrary, when people see how easy it is to fake things, they
  | might start developing a healthy sceptism to illuminating
  | youtube videos.
 
  | munificent wrote:
  | I think we'll solve it the same way we solved similar
  | transitions when text and image faking became easy: provenance.
  | 
  | For many years now, most have understood that you can't take
  | text and images as truth because they can easily be simulated
  | or modified. In other words, the media itself is not self-
  | verifying. Instead, we rely on knowing where a piece of media
  | came from, and we associate truth value with those
  | institutions. (Of course, people disagree on which institutions
  | to trust, but that's a separate issue.)
 
    | kbenson wrote:
    | In other words, the same way we dealt with information before
    | photographs and videos were invented. The answer to how to we
    | deal with the fact that images and videos can't be trusted is
    | to look at what we did before we relied on them. If we're
    | smart about it we'll try to pick out the good things that
    | worked and try to build in safeguards (as much as possible)
    | against the things that didn't, but I won't hold my breath.
    | We're already heading back towards some of the more
    | problematic behavior, such as popularity or celebrity
    | equating to trust.
 
  | pfisherman wrote:
  | I think that people will adapt. Humans are very clever and have
  | been evolutionarily successful because of the ability to adapt
  | to a wide range of environments.
  | 
  | Think about how devastatingly effective print, radio, and
  | television propaganda at the time each medium was widely
  | adopted compared to how effective they are now. They still
  | work, but for the most part people have caught on to the game
  | and adjusted their cognitive filters.
  | 
  | My guess is that we will see a bifurcation of society into
  | those who are able to successfully weed out bullshit from those
  | who can't. The people who are able to process information and
  | build better internal models of the world will be more
  | successful, and eventually people will start imitating what
  | they do.
  | 
  | Edit: I do think that these tools coupled with widespread
  | surveillance and persuasion tech (aka ad networks), have set up
  | conditions where the few can rule over the many min a way that
  | was not possible before. I do think some of the decentralized /
  | autonomous organization tech - scaling bottom up decision
  | making to make it more transparent and efficient - is a
  | possible counter. Imo, this struggle between technological
  | mediated centralization and top down enforcement and control vs
  | decentralization and bottom up consensus will be the defining
  | ideological struggle of our time.
 
    | azinman2 wrote:
    | I think you're underestimating the power of cognitive biases
    | and violence from those who only believe the information that
    | want to hear.
 
      | pfisherman wrote:
      | I think there are quite a few recent historical examples of
      | this - WW2, US invasion of Iraq, Russian Invasion of
      | Ukraine, etc.
      | 
      | However there is a price to pay for operating on beliefs
      | that do not align with reality. It's why almost all
      | organizations engage in some form of intelligence
      | gathering. Those who are at an information disadvantage get
      | weeded out.
      | 
      | Philip K Dick has a great quote "Reality is that which,
      | when you stop believing in it, doesn't go away."
 
  | adhesive_wombat wrote:
  | A skeptic (i.e. someone who cares to verify) not being able to
  | trust media because it might be fake is only a minor problem as
  | long as you have at least one trusted channel.
  | 
  | The president, say, can just release the statement on that
  | channel and it can be verified there (including
  | cryptographically, say by signing the file or even using
  | HTTPS).
  | 
  | If you lose that channel, then you're pretty much screwed
  | because you'll never know which one is the real president. But
  | there are physical access controls on some channels, say the
  | Emergency Alert System, which can be used to bootstrap a trust
  | chain.
  | 
  | What will be much more possible is that someone who will not
  | check the veracity of the message will take it at face value
  | without bothering to validate it. This is your news-via-
  | Facebook crowd.
  | 
  | At that point, it's less a technical issue than simply people
  | don't want to know the truth. No amount of fact-checking and
  | secure tamper-proofing of information chains of custody will
  | help that.
 
    | toss1 wrote:
    | Agree, and it's even worse than that
    | 
    | An incredibly small minority of people even understand your
    | phrase with any actual fidelity and depth of meaning:
    | 
    | >>it can be verified there (including cryptographically, say
    | by signing the file or even using HTTPS)
    | 
    | Even fewer of that microscopic minority have and understand
    | how to use the tools required to verify the video
    | cryptographically, AND even fewer know how to fully validate
    | that the tools themselves are valie (e.g., not compromised by
    | a bogus cert).
    | 
    | Worse yet, even in the good case where everyone is properly
    | skeptical, and 90+% of us figure out that no source is
    | trustworthy, the criminals have won.
    | 
    | The goal of disinformation is _not_ to get people to believe
    | your lie (although the few useful idiots who do may be a nice
    | bonus).
    | 
    | The goal of disinformation is to get people to give up an
    | even seeking the truth - just give up and say "we can't know
    | who's right or what's real" -- that is the opening that
    | authoritarians need to take over governments and end
    | democracies.
    | 
    | So yes, this is next-level-screwed material.
 
      | adhesive_wombat wrote:
      | > AND even fewer know how to fully validate that the tools
      | themselves are valie (e.g., not compromised by a bogus
      | cert).
      | 
      | Kind of, but once you have _a_ single verifiable channel
      | back to the source (in this case, some statement by the
      | president) it 's now possible for anyone to construct a web
      | of trust that leads back to that source. For example,
      | multiple trustworthy independent outlets reporting on the
      | same statement in the same way, providing some way to
      | locate the original source. This is why new articles that
      | do not link to (on the web) or otherwise unambiguously
      | identify a source wind me up. "Scientists say" is a common
      | one. It's so hard to find the original source from such
      | things.
      | 
      | This falls over in two ways: sources become non-independent
      | and/or non-trustworthy as an ensemble. Then you can't use
      | them as an information proxy. This is what is often claimed
      | about "the mainstream media" _and_ the  "non-mainstream
      | media" by the adherents if the other. All the fact checks
      | in the world are worthless if they are immediately written
      | of y those they are aimed at as lies-from-the-system.
      | 
      | The second way is that people simply do not care. It was
      | said, it sounds plausible, and they want to believe it.
      | 
      | So I would say that actually the risks here are social, not
      | technological. Granted, perhaps a deepfake-2'd video might
      | convince more people than a Photoshopped photo. The core
      | issue isn't the quality of the fake, it's that a
      | significant number of people simply wouldn't care if it
      | _were_ fake.
      | 
      | Doesn't mean we're not screwed, just not specifically and
      | proximally because of falsification technology, that's
      | accelerant but not the fuel.
 
        | toss1 wrote:
        | >>that's accelerant but not the fuel.
        | 
        | Yes, indeed! Which is why I'm having so much trouble with
        | ppl proposing technological solutions - technically it
        | might solve the problem in some situations, but the
        | bigger problem is indeed some combination of general
        | confusion, highly adversarial information environment
        | laden with disinformation, and people's all-too-frequent
        | love of confirmation bias and willingly believing BS and
        | overlooking warning signs.
        | 
        | I hope we can sort it...
 
  | efrbwrh wrote:
 
  | gadders wrote:
  | >> There are huge swathes of people who live in their own make-
  | believe world, like those believing wholeheartedly that the
  | 2020 election was stolen
  | 
  | There are also those that wholeheartedly believe Trump colluded
  | with Russia to win the 2016 election, or that the Steele
  | dossier was factual.
 
    | dTal wrote:
    | These are not equivalent. Russia _did_ interfere, there
    | _were_ links between Trump and Russia, therefore there is
    | circumstantial evidence that collusion occurred, sufficient
    | to trigger a widely publicized investigation. The allegations
    | of election fraud in 2020 however are 100% alternate universe
    | yarns spun for political gain with no basis in fact
    | whatsoever.
 
  | decafmomma wrote:
  | Here's the thing though: in theory, we should already be
  | skeptical of video and audio evidence on its own.
  | 
  | Most of our institutions, in theory, do not focus on single
  | mediums for assessing veracity of truth. The strength of claims
  | and our ability to split the difference between noise and truth
  | comes down to corroboration. How many other sources strength
  | and work consistently with a claim? That's, in theory, how law
  | enforcement, intelligence, and reporting should work.
  | 
  | In practice, there are massive gaps here and people's attention
  | -> decision is lower than ever.
  | 
  | I don't think it's impossible for us to handle deep fakes, but
  | I sense the same fear you have. I think ultimately it is more
  | about our attention spans, and the "urgency" we feel to act
  | quickly, that will be more of our down fall than the ability to
  | produce fakes more easily.
  | 
  | You don't in fact need a convincing fake to create a powerful
  | conspiracy theory. Honestly you only need an emotional
  | provocation, maybe even some green text on an anonymous web
  | form.
 
  | berdon wrote:
  | Reminds me of Stephenson's "Fall; Or Dodge in Hell" where all
  | digital media is signed by their anonymous author and public
  | keys become synonymous with identities. An entire industry of
  | media curation existed in the book to handle bucketing media as
  | spam, "fake", "true", interesting, etc.
 
  | narrator wrote:
  | Surprise Plot Twist: Maybe we're already living in a post-truth
  | society and you are still sure you know what the truth is. How
  | would you even know that what you were ferociously defending as
  | the truth wasn't a lie? What makes you think you're not smart
  | enough to not fall for lies?
  | 
  | Largely, I think most people's means of finding the truth is
  | just to take a vote of the information sources they find
  | credible and go with whatever they say. I was talking with some
  | friends about the California propositions a while back. Some of
  | them were not clear cut which way we should vote on them.
  | Instead of discussing the actual issue, people just wanted to
  | know what various authority figures thought. These were not
  | dumb people I was talking to, and I used to remember an era in
  | the 90s maybe where you could actually have a reasoned debate
  | and come to the truth that way. It seems that's obsolete these
  | days since nobody seems to agree on the basic facts about
  | anything.
 
    | cwkoss wrote:
    | Disinformation is very common in traditional news media. This
    | technology just democratizes this tool and allows anyone to
    | engage in it.
    | 
    | There will probably be a net increase in disinformation, but
    | citizens will likely also get better at being skeptical of
    | currently unquestioned modes of disinformation.
 
      | corrral wrote:
      | > There will probably be a net increase in disinformation,
      | but citizens will likely also get better at being skeptical
      | of currently unquestioned modes of disinformation.
      | 
      | Russia seems to be farther along this path than we are and
      | every account I've read of their experience of disinfo
      | isn't that they got better at seeking the truth, but
      | instead just assume everything's a lie & nothing's
      | trustworthy, and disappear into apathy.
 
  | wildmanx wrote:
  | > I don't think we'll have the ability to handle this kind of
  | tech responsibly.
  | 
  | Makes you also think whether anybody from the Hacker News crowd
  | working on any contributing tech is acting ethically
  | responsibly. For myself, I have answered this question with
  | "no", which rules out many jobs for me, but at least my kids
  | won't eventually look me in the eye and ask "how could you?"
  | 
  | Sure it's cool tech. But so was what eventually brought us
  | nuclear war heads.
 
  | brightball wrote:
  | Honestly, when people have gone so far as to redefine common
  | words it makes it really difficult to have conversations with
  | people.
  | 
  | 1. Hate going from one of the most visceral and obsessive
  | emotions that exists to being tossed around at everything
  | 
  | 2. The advent of your truth instead of the truth
  | 
  | 3. Constant injections of "x" into every existing word
  | apparently?
  | 
  | "Womxn in Mathematics at Berkeley" - https://wim.berkeley.edu/
  | 
  | This is all before we get people to understand that having a
  | discussion where some of their points might not be a strong as
  | they think they are...somehow means you're attacking.
  | 
  | The world that we have created for ourselves over the last 20
  | years is weird.
 
  | jkaptur wrote:
  | Wasn't this bridge crossed when Photoshop became popular?
 
    | bberrry wrote:
    | It takes some level of skill to produce a convincing
    | Photoshopped image.
 
      | BoorishBears wrote:
      | Does that matter when they stakes are as high as these
      | arguments always claim?
      | 
      | If we're doomsaying about a "post-truth society", we're
      | talking about high-stakes society-scaled skullduggery.
      | 
      | If you're aiming for that level of disruption, easy
      | deepfakes vs hard video/photo editing is not an issue,
      | getting people to trust your made up chain of custody is.
      | 
      | -
      | 
      | This is like when people worry about general AI becoming
      | self-aware and enslaving mankind... the "boring version" of
      | the danger is already happening: ML models being trained on
      | biased data are getting embedded in products that are core
      | to our society (policing, credit ratings, etc.), and that's
      | really dangerous.
      | 
      | Likewise, people worry about being able to easily make fake
      | news, when the real danger is people not being equipped to
      | evaluate the trustworthiness of a source... and that's
      | already happening.
      | 
      | You don't even need a deepfake, you tweet that so and so
      | said X, write a fake article saying they said X, amplify it
      | all with some bots, and suddenly millions of people believe
      | you.
 
  | kache_ wrote:
  | The one unwavering thing about technology is that it doesn't
  | stop advancing, and we can't use it responsibly.
  | 
  | The good news is that we've been going through rapid, rapid
  | tech advancements the past 50 years and we're still here.
 
    | mckirk wrote:
    | The thing I don't like about these 'well people have been
    | complaining about this forever' arguments is that, it's
    | entirely possible to have a) people pointing at an issue for
    | a long time and b) still have that issue get progressively,
    | objectively worse over time.
    | 
    | There's that example of people pointing out smartphones might
    | be bad for children, then someone counters with 'well thirty
    | years ago people complained about children reading too much
    | instead of playing outside', with the implication being:
    | adults of all ages will find some fault with newer
    | generations, and not to worry so much.
    | 
    | But just because it is true that adults will probably always
    | worry about 'new, evil things' corrupting the youth, this
    | does not mean that the 'new, evil things' aren't getting
    | _objectively more dangerous_ over time. Today adults would be
    | happy if children still had the attention span and motivation
    | necessary to read a book. They'd be happy if they themselves
    | still had it, actually.
    | 
    | Graphing the progress of a sinking ship and pointing out that
    | the downwards gradient has been stable for a while now and we
    | should therefore be okay is generally not a useful
    | extrapolation, I would say.
 
      | cupofpython wrote:
      | >Graphing the progress of a sinking ship and pointing out
      | that the downwards gradient has been stable for a while now
      | and we should therefore be okay is generally not a useful
      | extrapolation,
      | 
      | I like this analogy. I've had similar thoughts for a while
      | too. Granted I also saw some research that society has been
      | objectively _getting better_ in a lot of areas people
      | _think_ is getting worse (like violence, specifically
      | police abuse) compared to the past. theoretically this is
      | because we have a lot more information now than before, so
      | smaller occurrences are generating a larger impression.
      | 
      | that said, I still very much agree with your point and that
      | it is very applicable to _specific_ individualized issues.
      | Saying that people have been concerned for a while and
      | nothing bad has happened yet is accurate for the situation
      | where nothing bad will happen, AND the situation that it
      | was bad then and is worse now, AND the situation where we
      | are approaching a tipping point  / threshold where the bad
      | will start.
 
    | andruby wrote:
    | > The one unwavering thing about technology is that it
    | doesn't stop advancing, and we can't use it responsibly.
    | 
    | While I think that is true in general, I am optimistic that
    | we've seen at least one technology where we were able to
    | constraint ourselves from self-destruction: nuclear weapons.
    | 
    | Of course, nuclear weapons tech is not in reach of
    | individuals or corporations, which means there are only a
    | handful of players in this game-theory setting.
 
    | sitkack wrote:
    | > rapid tech advancements the past 50 years and we're still
    | here
    | 
    | This is a tautology. At some point the music stops and you
    | aren't here to make the argument that we are still here.
 
      | cupofpython wrote:
      | not entirely tautological. the probability that something
      | bad happens tomorrow if we do X today for the first time is
      | very different than the probability that something bad
      | happens tomorrow if we do X today GIVEN weve been doing X
      | every day for 50 years.
      | 
      | It is still insufficient to say nothing bad will happen, of
      | course
 
        | sitkack wrote:
        | Thats not the argument. The one you are making is the
        | same one people make when they conflate weather and
        | climate.
 
        | cupofpython wrote:
        | Conditional probability applies to many things
 
  | beisner wrote:
  | So there are information theoretic ways to certify that media
  | is genuine, if you assume trust at least somewhere in the
  | process. Basically just cryptographic signing.
  | 
  | For instance, a camera sensor could be designed such that every
  | image that is captured on the sensor gets signed by the sensor
  | at the hardware level, with a certificate that is embedded by
  | the manufacturer. Then any video released could be verified
  | against a certificate provided by the manufacturer. Of course,
  | you have to trust the manufacturer, but that's an easier pill
  | to swallow (and better supported by our legal framework) than
  | having to try and authenticate each video you watch
  | independently.
  | 
  | There are issues that can arise (what if I put a screen in
  | front of a real camera??, what if the CIA compromises the
  | supply chain???), but at the end of the day it makes attacks
  | much more challenging than just running some deepfake software.
  | So there are things that can be done, we're not destined for a
  | post truth world where we can't trust any media we see.
 
    | notfed wrote:
    | > a camera sensor could be designed such that every image
    | that is captured on the sensor gets signed by the sensor at
    | the hardware level
    | 
    | A hardware-based private key like this will inevitably be
    | leaked.
 
      | beisner wrote:
      | Each sensor could have a unique cert
 
    | [deleted]
 
    | pahn wrote:
    | I'll probably get downvoted into oblivion for mentioning
    | blockchain tech, but this might at least help, maybe not in
    | its current form but... I did not follow that project, but
    | there do exist some concepts in this direction, e.g. this:
    | https://techcrunch.com/2021/01/12/numbers-protocols-
    | blockcha...
 
      | beisner wrote:
      | The idea of having a public record that attests to _when_
      | an event happened is interesting, although not sure it has
      | to be blockchain for it to be useful.
 
    | bogwog wrote:
    | That's helpful for the legal system, but it's not going to
    | help for attacks designed to cause mass panic/unrest/revolts.
    | If another US president wants to attempt a coup, it'll be
    | much more successful if they're competent and determined
    | enough to produce deepfakes that support their narrative.
    | 
    | The only way to prevent stuff like that is to educate the
    | public and teach people how important it is to be skeptical
    | of anything they see on the internet. Even then, human
    | emotions are a hell of a drug so idk how much it'd help.
 
      | notahacker wrote:
      | US Presidents have had the ability to make false claims
      | based on video of something completely different, create
      | material using actors and/or compromised communications,
      | stage events or use testimony that information has been
      | obtained via secret channels from appointees heading up
      | agencies whose job it is to obtain information via secret
      | channels for a long time now.
      | 
      | If anything, recent events suggests the opposite: deepfakes
      | can't be _that_ much of a game changer when an election
      | candidate doesn 't even have to _try_ to manufacture
      | evidence to get half the people who voted for him to
      | believe his most outlandish claims.
 
  | [deleted]
 
  | Zababa wrote:
  | > What will life be like when you can't trust any video or
  | interview you see because it could be completely fake?
  | 
  | I don't understand your point. This has been the case for a
  | while. People were editing photos to remove people in the time
  | of Stalin. And even before that, you can lie, write false
  | records, destroy them.
 
  | tomgraham wrote:
  | The good news is that public awareness of potentially
  | manipulated media is on the rise. Coupled with good laws, good
  | detection tech - public awareness and media literacy is
  | important. At Metaphysic.ai, we created the @DeepTomCruise
  | account on Tiktok to raise awareness.
  | 
  | We also created www.Everyany.one to help regular people claim
  | their hyperreal ID and protect their biometric face and voice
  | data. We think that the metaverse of the future will feature
  | the hyperreal likenesses of regular people - so we all have to
  | work hard today to empower people to be in control of their
  | identity.
 
    | SantalBlush wrote:
    | Creating yet another product to monetize is not a solution,
    | it's just more of the problem. It incentivizes a perpetual
    | arms race between fabrication and verification at the cost of
    | everyday users. No thanks.
 
      | tomgraham wrote:
      | It is free. Protecting individuals' right is more important
      | that making money!
 
        | random-human wrote:
        | Free but collecting and storing peoples biometric data on
        | your servers (per the FAQ). How do I know it's not a
        | clearview ai clone but with easier data gathering? What
        | is that saying about what the real product is if
        | something is free?
 
  | fxtentacle wrote:
  | I believe we'll go back to trusting local experts that you can
  | meet in person to confirm that they are not a bot.
  | 
  | Because anything online will be known to be untrustworthy. Most
  | blogs, chat groups and social media posts will be spam bots.
  | And it'll be impossible for the average person to tell the
  | difference between chatting with a bot and chatting with a
  | human. But humans crave social connections and intimate
  | physical contact. So people will get used to the fact that
  | whoever you meet online is likely fake and so they'll start
  | meeting people in the real world again.
  | 
  | I also predict that some advanced AIs will be classified as
  | drugs, because people get so hooked on them that it destroys
  | their life. We've already banned abusive loot box gambling
  | mechanics in some EU countries, and I think abusive AI systems
  | are next. We'll probably also age-limit generative AI models
  | like DALL-E, due to their ability to generate naughty and/or
  | disturbing images.
  | 
  | But overall, I believe we will just starting to treat
  | everything online as fake, except in the rare case that you
  | message a person which you have previously met in real life (to
  | confirm their human-ness).
 
    | newswasboring wrote:
    | Your second paragraph is very intriguing. I never really
    | thought about this. I wonder if people will actually be able
    | to restrict usage though. Its software, and historically it
    | has been hard to restrict it. Of course cloud based systems
    | have two advantages, software is hidden behind the API and
    | they have really powerful systems. But the former requires a
    | single lapse in security to leak and latter just requires
    | time till consumer hardware can catch up. If I use past data
    | to predict future (which might be a bad idea in this case),
    | it might be almost impossible to restrict AI software.
 
      | formerkrogemp wrote:
      | I've heard this for years, but software will eventually
      | face its own regulation and barriers to entry much as
      | healthcare and accounting have theirs.
 
    | userabchn wrote:
    | I suspect that many chat groups (such as Facebook groups),
    | even small niche ones, already have GPT-3-like bots posting
    | messages that seem to fit into the group but that are trained
    | to provide opinions on certain topics that align with the
    | message that the organisation/country controlling them wishes
    | to push, or to nudge conversations in that direction.
 
      | fxtentacle wrote:
      | Aww, that reminds me of the good old IRC days where
      | everyone would start their visit with !l to get a XDCC bot
      | listing.
 
    | fartcannon wrote:
    | I want to agree with you, deeply, but the number of people
    | who fall for simple PR/advertising in today's world suggests
    | otherwise.
    | 
    | I think we'd have a chance if they taught PR tricks in
    | schools starting at a young age. Or at minimum, if websites
    | that aggregate news would identify sources that financially
    | benefit from you believing what they're saying.
 
      | corrral wrote:
      | I've long thought that high school should require _at
      | least_ one course that I like to call  "defense against the
      | dark arts" (kids still dig Harry Potter, right? Hahaha).
      | 
      | The curriculum would mostly be reasoning, how to spot
      | people lying with graphs and statistics, some rhetoric, and
      | extensive coverage of Cialdini's _Influence_. The entire
      | focus would be studying, and then learning to spot and
      | resist, tricks, liars, and scam artists.
 
    | jayd16 wrote:
    | When you say "everything online" do you mean every untrusted
    | source? Surely the genie is out of the bottle on
    | communication over the web. That local source will have a
    | website. Because of that I feel like we'll always just have
    | to be vigilant, just like we always should have been. After
    | all, local scams still exist. Real humans are behind the
    | bots.
 
      | fxtentacle wrote:
      | > Real humans are behind the bots.
      | 
      | Yes, but those humans are usually anonymous and on the
      | other side of the planet which makes them feel safe. And
      | that allows them to be evil without repercussions.
      | 
      | Back in the days, I went to LAN parties. If someone spotted
      | a cheater, they would gang up with their friends and
      | literally throw the offender out of the building. That was
      | a pretty reliable deterrent. But now with all games being
      | played online, cheating is rampant.
      | 
      | Similarly, imagine if those Indian call centers that scam
      | old ladies out of their life savings were located just a
      | quick drive away from their victims' families. I'm pretty
      | sure they would have enough painful family visits such that
      | nobody would want to work there.
      | 
      | Accordingly, I'm pretty sure the local expert would have
      | strong incentives to behave better than an anonymous online
      | expert would.
 
        | jayd16 wrote:
        | To argue that scams didn't exist or weren't a problem
        | before the internet is pretty indefensible, no matter the
        | anecdotes.
 
        | fxtentacle wrote:
        | I was merely trying to argue that scams within a local
        | community would be less severe than scams between
        | strangers, because they are easier to punish and/or
        | deter.
 
    | wongarsu wrote:
    | I'm not sure the experts have to be local. I can't be sure
    | that a random twitter account isn't a bot, but I can be
    | pretty sure that tweets from @nasa are reasonably
    | trustworthy. People will form webs-of-trust: they trust one
    | source, the people viewed as trustworthy by them, etc. Anyone
    | outside of that will be untrustworthy.
    | 
    | That's not too dissimilar from what we do today, after all
    | people have always been able to lie. The problem is just that
    | if you start trusting one wrong person this quickly sucks you
    | into a world of misinformation.
    | 
    | I find your point about regulating AI interesting. We already
    | see some of this, with good recommendation systems being
    | harmful to vulnerable people (and to a lesser degree most of
    | us). This will probably explode once we get chatbots that can
    | provide a strong personal connection, replacing real human
    | relationships for people.
 
      | freetinker wrote:
      | HN is a good example (or precursor) of webs-of-trust. Nice
      | phrase.
 
  | mrshadowgoose wrote:
  | The concerns in your second paragraph can be mostly mititgated
  | using a combination of trusted timestamping, PKI,
  | cryptographically chained logs and trusted hardware. Recordings
  | from normal hardware will increasingly approach complete
  | untrustworthiness as time goes on.
  | 
  | The concerns raised in the first paragraph however... the next
  | few decades are going to be a wild ride. Hopefully humanity
  | eventually reaches an AI-supported utopic state where people
  | can wrap themselves in their own realities, without it
  | meaningfuly affecting anyone else. Perception of reality is
  | already highly subjective, most of the fundamental issues are
  | due to resource scarcity/inequality. Most other issues
  | evaporate once that's solved.
 
    | FastMonkey wrote:
    | I think you can technically mitigate some concerns for the
    | people who understand that, but practically it's going to be
    | a very different story. People will believe who/what they
    | believe, and an expert opinion on trustworthiness is unlikely
    | to change that.
    | 
    | I think being in the real world and meeting real people is
    | the only way to create a real, functional society. Allowing
    | people to drift away into their own AI supported worlds would
    | eventually make cooperation very difficult. I think it would
    | just accelerate the tendency we've seen with social media,
    | creating ever more extreme positions and ideologies.
 
  | thedorkknight wrote:
  | I don't think it'll be way too much different than it has been
  | for most of human history. We really only had a brief blip of
  | having video, which was generally trustable, but keep in mind
  | that before that for thousands of years it was just as hard to
  | know truth.
  | 
  | Someone told you stuff about the outside world, and you either
  | had the skepticism to take it with a grain of salt, or you
  | didn't.
 
  | gernb wrote:
  | I was listening to "This American Life" and they had a segment
  | on someone who setup an site to give you a random number to
  | call in Russia where you were supposed to give them info about
  | what's happening in the Ukraine. It was someone shocking to
  | hear their side of the story, that Russia is a hero for helping
  | oppresed Russians in Ukraine.
  | 
  | But then I stepped back and wondered, I'm assuming that the
  | story I've been told is also 100% correct. What proof do I have
  | it is? I get my news from sources who've been wrong before or
  | who have a record of reporting only the official line. My gut
  | still tells me the story I'm being told in the west is correct,
  | but still, the bigger picture is how do I know who to trust?
  | 
  | I see this all over the news. I think/assume the news I get
  | about Ukraine in the west is correct but then I see so much
  | spinning on every other topic that it's hard to know how much
  | spinning is going on here too.
 
    | RobertRoberts wrote:
    | I was asked "What are we going to do about Ukraine!?" And I
    | said, "It's a civil war that's been going on for almost 10
    | years, what is different now?" and their response was "what?
    | I'd never heard that." And I added, "In 2014 there was an
    | overthrow of an elected president there and it started the
    | war." Blank stares.
    | 
    | I have a friend who traveled to Europe regularly for tech
    | training, including Ukraine, and he was surprised about how
    | little people know what is going because people's news
    | sources are so limited. (mostly by choice I assume)
    | 
    | No special tech needed to manipulate people, just lack of
    | multiple information sources?
 
      | mcphage wrote:
      | > what is different now?
      | 
      | Um. Is this a serious question?
 
        | RobertRoberts wrote:
        | Yes, I didn't know there was an invasion when asked about
        | Ukraine, but I knew about the past history. (some at
        | least)
 
      | HyperSane wrote:
      | The president that lost power in 2014, Viktor Yanukovych,
      | was a Russian puppet who refused to sign the European
      | Union-Ukraine Association Agreement in favor of closer ties
      | to Russa. The Ukrainian parliament voted to remove him from
      | office by 328 to 0. He then fled to Russia.
 
        | synu wrote:
        | It's hard to fathom believing there's nothing new or
        | relevant happening with the 2022 invasion, or why if
        | there was a lead-in to the conflict that would be on its
        | own a reason to conclude that there's nothing to be done
        | now.
 
        | RobertRoberts wrote:
        | See this is the problem. While I follow plenty of
        | international news, I didn't know this.
        | 
        | There is often times just too much to know to fully
        | understand a situation. So how can anyone form a valid
        | opinion?
        | 
        | As a follow up, was the election of Victor Yanukovych
        | lawful? If not, then why not point out he was a puppet
        | from a manipulated election? That would be worth a coupe,
        | but not because you disagree with his politics, that's
        | just insanity. Look what Trump believed and supported, we
        | didn't start civil war because Trump wouldn't sign a
        | treaty and he was accused of being a Russian puppet too.
        | There is just more to this story than you are letting on.
 
  | idontwantthis wrote:
  | This doesn't bother me that much because evidence isn't
  | required to convince millions of people that a lie is true. We
  | already know this. Why make fake evidence that could be
  | debunked when you can just have no evidence instead?
 
    | [deleted]
 
    | fleetwoodsnack wrote:
    | Different instruments can be used to capture different
    | segments of the population. You're right there are gullible
    | people who are more likely to believe things with limited or
    | no evidence. But it isn't necessarily about the most
    | impressionable people, nor is it about installing sincerely
    | held beliefs.
    | 
    | Instead, what may be a cause for concern is simply the
    | installation of doubt in an otherwise reasonable person
    | because of the perceived validity of contrived evidence. Not
    | so much that it becomes a sincerely held belief, but but just
    | enough that it paralyzes action and encourages indifference
    | due to ambiguity.
 
      | astrange wrote:
      | This discussion isn't useful because you're assuming people
      | actually care if something is true before they "believe"
      | it, which they don't, so they don't need evidence.
      | "Believing" doesn't even mean people actually hold beliefs.
      | It means they're willing to agree with something in public,
      | and that's just tribal affiliation.
 
        | [deleted]
 
        | fleetwoodsnack wrote:
        | >you're assuming people actually care if something is
        | true before they "believe" it, which they don't
        | 
        | This seems like an assumption too. I know there are
        | instances like you've described but they're not absolute
        | nor universal and I accounted for that in my original
        | comment.
 
      | idontwantthis wrote:
      | Think about how few people believe in Bigfoot when video,
      | photographs, footprints, eye witness testimony all exist.
      | 
      | Think about how many people believe in Jesus without any of
      | that physical evidence.
      | 
      | If anything, the physical evidence turns most people off.
      | And I'd argue that most Bigfooters don't even believe in
      | the physical evidence, but use it as a tool to hopelessly
      | attempt to convince other people to believe in what they
      | already believe is true.
 
      | mgkimsal wrote:
      | > You're right there are gullible people who are more
      | likely to believe things with limited or no evidence
      | 
      | Often the lack of evidence _is the proof_ of whatever is
      | being peddled.  "No evidence for $foo? OF COURSE NOT!
      | Because 'they' scrubbed it so you wouldn't be any wiser!
      | But _I_ have the  'truth' here... just sign up for my
      | newsletter..."
 
  | mc32 wrote:
  | I think a bigger question is whether reputable sources --those
  | people trust for whatever reason, would use this technology to
  | prop up ideas and or to create narratives.
  | 
  | I don't think it's far-fetched. We've already seen where videos
  | are misattributed[1] to stoke fear or to promote narratives
  | --by widely trusted news sources.
  | 
  | [1] This was foreshadowed with "Wag the Dog" but happens often
  | enough in the media today that I don't think use of "deepfake"
  | technology is beyond the pale for any of them.
 
    | ketralnis wrote:
    | It almost doesn't matter now that people have fractured on
    | which sources they consider reputable. Trump called a CNN
    | reporter "fake news" and presumably his followers think of
    | them the same way I think of Fox. I absolutely think that Fox
    | would use technology to lie, and I'm sure Fox fans think that
    | "the liberal media" would. So people are going to think that
    | reporting is fake whether or not it is
 
      | mc32 wrote:
      | Wasn't 'The Ghost of Kiev' almost entirely fake but the
      | news carried it as real?
 
        | ketralnis wrote:
        | I don't know. How would you "prove" it? Google it, and
        | look for what other people that agree with you think?
 
        | mc32 wrote:
        | Well...
        | https://www.washingtonpost.com/world/2022/05/01/ghost-of-
        | kyi... admitted by Ukrainians themselves...
 
        | ketralnis wrote:
        | According to an article you found online. That's exactly
        | my point, if we can't trust news sources then we can't
        | really know anything. Because of my aforementioned
        | distrust of Fox News, if it were written by them I'd
        | dismiss that article out of hand placing no truth value
        | on it either way. I'd expect somebody that distrusts CNN
        | to do the same if it were written by them.
        | 
        | "It's confirmed!", "They admitted it!", and other
        | unprovable internet turns of phrase in comments sections
        | are really just "I believe it because somebody I trust
        | said so" and that only has weight as long as trust has
        | weight.
 
        | mc32 wrote:
        | If the accused admit to something it's more believable
        | than the alternative (that they were forced into false
        | admission).
        | 
        | So in this case if the Ukrainian government admit to
        | making things up then I would think it's believable that
        | they made something up for the sake of propaganda. We can
        | also check more independent sources --read Japanese news,
        | or Indian news sources, etc.
 
        | ketralnis wrote:
        | We don't know that the accused admitted to anything. We
        | know that the Washington Post says that they did. The
        | world becomes very solipsistic when you lose trust in
        | reporting.
 
  | GuB-42 wrote:
  | There has never been a truth society.
  | 
  | This tech will certainly be used to frame someone for a crime,
  | like I am sure Photoshop was used in such a way, and thousands
  | of other techniques. And modern technology offers counters. It
  | is an arms race but because of the sheer amount of data that is
  | collected, I think that truth is more accessible than ever. The
  | more data you have, the harder it is to fake and keep
  | consistent.
 
    | fxtentacle wrote:
    | "Cheerleader's mom created deepfake videos to allegedly
    | harass her daughter's rivals"
    | 
    | "The girl and her coaches were sent photos that appeared to
    | depict her naked, drinking and smoking a vape pen, according
    | to police"
    | 
    | https://abcnews.go.com/US/cheerleaders-mom-created-
    | deepfake-...
 
    | jl6 wrote:
    | I don't know, it seems like the existence of widespread, easy
    | photo/video/audio faking technology could be a really strong
    | argument for dismissing any purported photo/video/audio
    | evidence.
    | 
    | Wouldn't it be funny if deepfakes destroyed the blackmail
    | industry?
 
  | kleer001 wrote:
  | Thankfully things that a real are very cheap to follow up on.
  | Questioning some security footage? No worries there's 100+
  | hours of it to cross check with the footage and three other
  | cameras too.
  | 
  | IMHO, consilience and parsimony will save us.
 
  | VanillaCafe wrote:
  | The real problem isn't the veracity of the information, but the
  | consensus protocol we use to agree on what's true. Before the
  | internet, we were more likely to debate with their neighbors to
  | come to an understanding. Now, with the large bubbles we can
  | find ourselves in, afforded by the internet social media, we
  | can find a community to agree on anything, true or not. It's
  | that lack of challenge that allows false information to
  | flourish and is the real problem we need to solve.
 
    | whimsicalism wrote:
    | I would be curious if false information is actually more
    | common now. It seems like people regularly believed all sorts
    | of false things not too long ago.
 
  | redox99 wrote:
  | People with some degree of knowledge already know that any
  | photo could be photoshopped. People that don't care will
  | blindly trust a picture of someone with a quote or caption
  | saying whatever, as long as it fits their narrative.
  | 
  | This has been the case for photos for almost 2 decades. The
  | fact that you can now do it with video or audio doesn't change
  | that much IMO.
 
    | hutrdvnj wrote:
    | I think it does, because while you obviously couldn't trust
    | images since two decades or so, you could resort to video
    | which wasn't easy to believably deep fake until recently. But
    | if everything online could be a deep fake, how can you find
    | out the truth?
 
      | whimsicalism wrote:
      | Videos can be faked too, it is just cheaper now.
 
        | makapuf wrote:
        | It's called special fx, is more than a century old
        | (people are now aware it's fake but remember the word is
        | that the train coming in la ciotat movie made people run
        | out of the movie theatre).
 
    | micromacrofoot wrote:
    | Photos have been altered for much longer than 2 decades.
    | Think of airbrushing models in magazines (used to be literal
    | airbrushes painting over photos). This has had a serious
    | impact on our perception of beauty and reality.
 
  | MarcoZavala wrote:
 
  | WanderPanda wrote:
  | I think society will adapt within 1 generation. The tech is
  | already there (signing messages with asymmetric encryption)
 
    | toss1 wrote:
    | And how many use or will use the tech, and how many of those
    | will use it competently, and how many of those are competent
    | to validate and know that their checking technology has not
    | been compromised (e.g., hacking or distributing bad
    | authenticity checkers and/or certs like hacking or
    | distributing bad crypto-wallets)?
 
      | bee_rider wrote:
      | End-to-end encryption was a giant pain in the butt that
      | required dinking around with PGP or whatever, but now it is
      | a pretty mainstream feature for chat apps (once they
      | figured out how to monetize despite it). Tech takes a while
      | to trickle down to mainstream applications, but it'll get
      | there if the problem becomes well known enough.
 
        | toss1 wrote:
        | I agree that e2e encryption is becoming more widespread
        | and "user friendly".
        | 
        | However, the friendliness seems inversely proportional
        | with the ability of the users to detect that their tool
        | is failing/corrupted/hacked/etc. So while we might have
        | more widespread tools, we also have a more widespread
        | ability to give a _false_ sense of security.
 
  | nathias wrote:
  | you could never trust it, now you'll know you can't trust it
 
| EGreg wrote:
| Is this what Matterport 2 app uses?
| 
| When you take many photos of a scene or indoor place and it's
| stitched together?
| 
| Can this be used for metaverses?
| 
| ALSO why not synchronize videos of the same event and make an
| animated 3d movie from 2d ones !!! Great as a "disney ride
| metaverse"
| 
| Who is doing that in our space?
 
| getcrunk wrote:
| I just started playing cyberpunk 2077. Spoilers:
| 
| The idea of the "black wall" to keep out bad ai comes to mind.
| Not arguing for it but just acknowledging that maybe one day will
| all have to live in walled gardens to stay safe from rouge ai's
| or rather rouge actors using powerful ai's
 
  | henriquecm8 wrote:
  | One thing I've always wondered about that, they explained where
  | those are running, but how they still have autonomy? Are they
  | producing they own power? How about when they need to replace
  | hardware?
  | 
  | I am not saying it's impossible, but I would like to see that
  | part being explored, even if it's in the tie-ins comics.
 
| echelon wrote:
| What's the difference in NeRF from a classical photogrammetry
| pointcloud workflow? It seems like the representation and outputs
| are identical.
| 
| Why would you prefer NeRF to photogrammetry? Or vice versa?
 
  | randyrand wrote:
  | Nerfs can represent reflection, speculars, refractions.
  | 
  | They also are proving to be faster, more accurate, etc
  | 
  | The input data is the same. Nerfs have the chance of requiring
  | less input data.
 
  | flor1s wrote:
  | Neural Radiance Fields are a technique from the neural
  | rendering research field, while photogrammetry is a research
  | field on its own. However these are just turf wars and in
  | practice there is a lot of overlap between both fields.
  | 
  | For example, most NeRF implementations recommend the use of
  | COLMAP (traditionally a photogrammetry tool) to obtain camera
  | positions/rotations that are used alongside their images. So
  | this multi-view stereo step is shared between both NeRF (except
  | a few research works that also optimize for camera
  | positions/rotations through a neural network) and
  | photogrammetry.
  | 
  | After the multi-view stereo step in NeRF you train a neural
  | renderer, while in photogrammetry you would run a multi-view
  | geometry step/package that uses more traditional optimization
  | algorithms.
  | 
  | The expected output of both techniques is slightly different.
  | NeRF produces renderings and can optionally export a mesh
  | (using the marching cubes algorithm). Photogrammetry produces
  | meshes and in the process might render the scene for editting
  | purposes.
 
| natly wrote:
| I was initially annoyed by this title but now I'm gonna switch my
| perspective to being happy that ideas like this are floating
| around since it acts as a really cheap signal to tell if someone
| knows what they're talking about or not when it comes to ML.
 
| nathias wrote:
| can't wait until deepfakes completely revolutionizes people's
| relation to information
 
___________________________________________________________________
(page generated 2022-06-27 23:01 UTC)