|
| tdsone3 wrote:
| https://archive.is/YTOeD
| antirez wrote:
| I believe you can't, at the same time, accuse OpenAI of
| benefitting of ideas mostly invented elsewhere, and also claim
| that they are in a predominant position. It does not make any
| sense, and they are not naturally positioned for a monopoly. Just
| other companies have to move their asses.
| tomlue wrote:
| I'm a fan of OpenAI, but this is nonsense. All of human
| existence is mostly other people's ideas. Among a ridiculously
| huge list of other things, OpenAI benefits from the mountains
| of labor that made scalable neural networks possible.
|
| Have they had their own good ideas? Definitely. Are they
| benefitting of ideas mostly invented elsewhere? Also
| Definitely, just like everybody else.
| antirez wrote:
| I don't think you understood me. I'm with you, but given that
| often OpenAI is accused of using public ideas for profit,
| this, in turn, means they are obviously not in a dominant
| position. So far they are just better than the others.
| version_five wrote:
| I'm not really a fan of openAI, but I think we're seeing the
| classic mistake of confusing product with technology. Steve
| Jobs / Apple didn't anything you'd call new ideas either
| (obvious cliche but so is the criticism). It's execution and
| design once the tech reaches a certain level
| poopypoopington wrote:
| I've heard Altman (on the Lex Friedman podcast) and Sundar
| Pichai (on the Hard Fork podcast) say things to this
| degree. The thing that OpenAI really managed to crack was
| building a great product in ChatGPT and finding a good
| product market fit for LLMs.
| anon84873628 wrote:
| What's the product market fit for LLMs, and how does
| OpenAI fill it?
| fooster wrote:
| How much money is openai making from chatgpt premium now?
| How much revenue are they they making from the api?
| Robotbeat wrote:
| Well sure, but there still aren't any other LLMs at the
| level of GPT3/3.5 let alone GPT-4. GPT3.5 just using the
| API returns fantastic results even without the ChatGPT
| interface (which isn't terribly hard to replicate, and
| others have using the API).
|
| There are dozens if not hundreds of companies that
| could've done something profound like ChatGPT if they had
| full access to GPT3/3.5. And honestly, OpenAI stumbled a
| lot with ChatGPT losing history access, showing other
| users' history... but that doesn't matter much as the
| underlying technology is so profound. I think this really
| is a case of the under-the-hood capability (GPT3/3.5/4)
| mattering more than productization and execution.
|
| (Now I think there are not a ton of companies that could
| do what Microsoft is trying to do by expanding GPT4 to
| power office productivity... that is a separate thing and
| probably only about 3 companies could do that, at best:
| Microsoft, Apple, and Google... and theoretically Meta
| but their lack of follow through with making Metaverse
| useful makes me doubt it.)
| mustacheemperor wrote:
| Hm, I wonder how much of the API's performance is related
| to training/finetuning done by OpenAI planned towards the
| ChatGPT product. I think the RLHF is partly product
| design and partly engineering.
| vitehozonage wrote:
| Rather than execution or design perhaps this time it was
| mainly about being unethical enough to sell out to
| investors and unflinchingly gather enough data in one place
| by carelessly ignoring copyright, authors' desires, privacy
| regulations, etc.
| brookst wrote:
| It's a good comparison. And once again tech enthusiasts are
| confused and outraged that the product people are getting
| credit for tech they didn't invent. Once again missing the
| forest that people buy products, not tech.
| simonh wrote:
| There's an awful lot of judgement, engineering and
| technique that goes into a really well thought out
| product. It's often deeply underestimated, and culture
| makes a huge difference in execution. Bing/Sydney came
| out after ChatGPT, based on exactly the same tech, but
| was hot garbage.
| mustacheemperor wrote:
| It's interesting to see the ad implementation, I recall
| some predictions that Microsoft would be particularly apt
| at finding a way to integrate advertising organically.
| Instead it just seems to have made the bot more stupid
| because sponsored products are forced into its
| recommendations.
|
| I've gone back to just using the GPT API, unless I
| absolutely need to search the internet or information
| after 2021 for some reason.
| brookst wrote:
| I don't know, my family talks almost daily about how
| amazing bing chat is. The Sydney eta was kind of crazy,
| but the core product seems to be doing well.
|
| It is definitely solving a different problem than chatgpt
| though, and maybe a less inspiring problem. Chatgpt is
| like an open world game where you can do anything; Bing
| chat is just a point solution for the vicious spiral of
| SEO and Google's profit motive that rendered search
| results and web pages so useless.
| FormerBandmate wrote:
| Bing/Sydney is better than ChatGPT. It had serious bugs
| in beta testing
| simonh wrote:
| It's a dramatically worse chat bot, but being able to
| search the internet does give it an additional useful
| capability, while limiting it to five interactions papers
| over its psychotic tendencies.
| tomlue wrote:
| mostly agree. Though I wouldn't underestimate the tech and
| engineering work behind OpenAI. That microsoft partnership
| is no joke.
| wslh wrote:
| > OpenAI benefits from the mountains of labor ...
|
| Can't you say the same about Google? Google lives from the
| labor of others.
|
| Not entering into the debate if this is ethically or not just
| saying that OpenAI is not much different. When photo services
| such as Google Photos recognize the Eiffer Tower in Paris
| they are using images from others.
| matthewdgreen wrote:
| There have been many examples of companies inventing new
| technology, then failing to take it to market until a
| competitor copied the ideas. The classic example is Apple and
| Xerox PARC. The criticism in this case is that while icons and
| GUIs are obviously harmless (so it's good we got them out of
| the lab!), maybe AI is the kind of tech we should have let
| researchers play with for a while longer, before we started an
| arms race that puts it in everyone's house.
| ren_engineer wrote:
| OpenAI acts like they are some underdog startup for PR purposes
| while actually having access to effectively unlimited resources
| due to their relationship with Microsoft, which rubs people the
| wrong way
|
| Plenty of other companies had released GPT powered chat bots
| like ChatGPT, they just couldn't offer it for free because they
| didn't have a sweetheart deal with Microsoft for unlimited
| GPUs. Google did drop the ball though, they were afraid of
| reputation risk. Google should have used DeepMind or another
| spinoff to release their internal chatbot months ago
| simonh wrote:
| I think you're missing how much of a profound difference the
| intensive RLHF training OpenAI did makes in ChatGPT.
| Microsoft's Sydney seems to also be GPT 3.5 based, came out
| after ChatGPT, and it was an utter dumpster fire on launch in
| comparison.
|
| Meanwhile nobody has even caught up to ChatGPT yet, not even
| Microsoft whose resources are the secret sauce you think is
| the game changer, and now 4.0 is out and even more massively
| moved the ball forward.
| jamaliki wrote:
| No, actually. Microsoft's Sydney is GPT 4 [1].
|
| 1 - https://blogs.bing.com/search/march_2023/Confirmed-the-
| new-B...
| nomel wrote:
| > to release their internal chatbot months ago
|
| I'm not sure that would have been wise. Bard clearly isn't
| ready.
| sebzim4500 wrote:
| Yeah but if they released earlier that fact wouldn't have
| been so embarassing.
|
| As it was, they intially just claimed that releasing a
| competitor was irresponsible, then they eventually did it
| anyway (badly).
| rhaway84773 wrote:
| I'm not sure why that doesn't make sense.
|
| In a winner takes all market, where the product is highly
| complex, which is likely the case for any product developed
| since the advent of computers, if not before, the predominant
| position will almost certainly be taken by somebody, and since
| it's a highly complex product, it's likely no one entity
| thought of and/or implemented even close to a majority of the
| ideas needed to make it work.
|
| In fact, it's likely to be a random winner amongst 10-20
| entities who implemented some of the ideas, and another
| potentially larger number of entities who implemented equally
| good, or even better, ideas, which happened to fail for reasons
| that couldn't have been known in advance.
| curiousllama wrote:
| > In a winner takes all market
|
| Just because Search was winner take all, doesn't mean AI will
| be. What network effects or economies of scale are
| unachievable by competitors? Besides, Alpaca showed you can
| replicate ChatGPT on the cheap once its built - what's
| stopping others from succeeding?
| tyfon wrote:
| Yeah, I don't think this will be a corporate thing but a
| private decentralized thing.
|
| It's probably the worst fear of the likes of google etc, a
| 100% local "search engine" / knowledge center that does not
| even require the internet.
|
| I've been running the 65B model for a bit. With the correct
| prompt engineering you get very good results even without
| any fine tuning. I can run stable diffusion etc fine too
| locally. If anything will let us break free from the
| corporate, it is this.
| ftxbro wrote:
| > I believe you can't, at the same time, accuse OpenAI of
| benefitting of ideas mostly invented elsewhere, and also claim
| that they are in a predominant position.
|
| There used to be a team at Google called Google Brain and they
| all left to go to OpenAI after the employee protests against
| taking military AI contracts in 2018. Now Microsoft has those
| contracts and funneled $10B to OpenAI from the CIA. OK that's a
| little bit of exaggeration but not so much; I guess not _all_
| employees left, and Google Brain still technically exists. Also
| some of the brain employees went to other startups not only
| OpenAI.
| FormerBandmate wrote:
| Microsoft doesn't have $10 billion from the CIA. They're
| splitting military cloud contracts with Amazon and other
| startups, but it's just the same thing as corporate contracts
| sebzim4500 wrote:
| >funneled $10B to OpenAI from the CIA
|
| Without some evidence to support it, this really sounds like
| a conspiracy theory.
| beebmam wrote:
| Many of those employees have changed their minds about
| military contracts in this new cold war era.
| cowl wrote:
| If I'm not mistaken their opposition was on principle not
| whether it was needed or not. So the fact that we are in a
| new cold war era does not change that Equation. The
| principle is still the same. The only way for this "change
| of mind" is if their opposition was due to "it's not needed
| because US has no rivals". Or what most probably happened,
| they realised that if they want to keep workign on this
| field there is no escape from those kind of implications
| and They don't have the big bad wolf Google to blame
| anymore.
| jeremyjh wrote:
| And then Russia invades Ukraine and people change their
| minds about what is important, and what is possible.
| cowl wrote:
| But that's my point. If their opposition was not on
| principle but based on the naive Idea that "we will live
| in peace and harmony", I really am afraid what other
| naive principles are in the bases of their work and what
| safeguards are being set in place for the AIs.
| jeremyjh wrote:
| There is a such thing as being principled and then
| finding out you were naive.
| mistermann wrote:
| There is also the substantial effectiveness of
| propaganda, and the fact that it is essentially
| impossible to know if one's beliefs/"facts" have been
| conditioned by it.
|
| That most discussions on such matters typically devolve
| rapidly into the regurgitation of unsound memes doesn't
| help matters much either.
| sebzim4500 wrote:
| Sure, but it is equally likely that your old perspective
| was the result of propaganda and your new one is a
| rational adjustment in the face of new information. Or
| that both are propaganda, I suppose.
| cowl wrote:
| ofcourse but then you still must have some principles and
| be as loud about how naive you were as you were when you
| were protesting the thing in the first place.
| josephg wrote:
| Why must you be loud about changing your mind? Why not
| just quietly realise you were wrong, have conversations
| with friends about it and move on? That's what I'd do.
|
| My life is a story. I'm under no obligation to share.
| robbomacrae wrote:
| Not the op but the quotes from John Maynard etc made me
| think... if I listen to your original thesis and you
| convince me (maybe you have some authority) and I go on
| believing you, is it not harmful to me if you realize
| your error and don't inform me? Boss: "why did you do X?"
| Me: "But sir you told me doing X was good!" Or to put it
| differently, if you spread the wrong word then don't
| equally spread the correction then the sum of your
| influence is negative.
| pen2l wrote:
| When I was a young person, I used to deride writings like
| 1984 on the ground that the scenarios and stories presented
| to carry the message were too far-fetched.
|
| Reading your comment has set off an epiphany, I think I get
| it now, there probably exists some higher-up person who is
| thinking in these terms: we must always be in a state of
| war, for if we are, the populace will want to be ready and
| willing to fund the instruments of war. And we always
| _want_ to be ready for war, because if ever we are not, we
| lose our capability to win a potential future war. We must
| even contribute our efforts to build these instruments of
| war. War is constant. War is peace.
| HyperSane wrote:
| The more powerful your military is the less likely you
| will need to use it to defend yourself.
| InCityDreams wrote:
| America! What is in your military now, eventually ends up
| in the hands of the civilian police.
|
| *and other countries are following. Fucking, sadly.
| sebzim4500 wrote:
| What the hell are they going to do with an M1 Abrams?
|
| Having said that, I dread to see what the NYPD manage to
| achieve with an F-35B.
| [deleted]
| m3kw9 wrote:
| OpenAI getting rich now everyone wants a piece and everyone
| fighting over it like a billion dollar inheretence fight. In a
| round about way, if GPT5 is nearly as advertised, watch the govt
| swoop in under national security guise
| oldstrangers wrote:
| Undoubtedly OpenAI already has some very close ties and/or
| contracts with the DoD.
| jstx1 wrote:
| Any evidence for this, or are you just assuming that it's the
| case?
| roflyear wrote:
| To me it overstates what is achieved. This isn't AGI.
| Unsure how much the government would care about this.
| unshavedyak wrote:
| > if GPT5 is nearly as advertised, watch the govt swoop in
| under national security guise
|
| Re: GPT5, are there any .. reasonable/credible sources of
| information on the subject? I've become deaf from all the
| speculation and while i am very curious, i'm unsure if anything
| substantiated has actually come out. Especially when
| considering speculation from Sam himself.
| blihp wrote:
| There's barely any credible information about GPT4 (i.e.
| OpenAI hasn't said very much about what's going on behind the
| curtain) and there's absolutely none re: any releases beyond
| that.
| zarzavat wrote:
| It's unlikely to be any time soon. Despite productization by
| OpenAI, LLMs are still an active area of research. Research
| is unpredictable. It may take years to gather enough
| fundamental results to make a GPT-5 core model that is
| substantially better than GPT-4. Or a key idea could be
| discovered tomorrow.
|
| Moreover, previous advances in GPTs have come from data
| scaling: throwing more data at the training process. But data
| corpus sizes have started to peak - there is only so much
| high quality data in the world, and model sizes have reached
| the limits of what is sensible at inference time.
|
| What OpenAI can do while they are waiting is more of the easy
| stuff, for example more multimodality: integrating DALL-e
| with GPT-4, adding audio support, etc. They can also optimize
| the model to make it run faster.
| dougmwne wrote:
| I keep keep hearing people claim we are at the end of
| corpus scaling, but that seems totally unfounded. Where has
| it been proven you can't running the training set through
| in multiple epochs in randomized order? Who's to say you
| can't collect all the non-English corpus and have the
| performance transfer to English? Who's to say you can't run
| the whole damn thing backwards and still have it learn
| something?
| SamPatt wrote:
| There are diminishing returns from compute time but it
| looks like even though they are diminishing there's still a
| fair bit on the table.
|
| Though my guess is that GPT-5 will be the last model which
| gains significantly from just adding more compute to the
| current transformer architecture.
|
| Who the hell knows what comes next though?
| whiplash451 wrote:
| You are missing the scaling on context size.
| ftxbro wrote:
| They've already swooped in.
|
| I mean the conspiracy argument would be that the $10B isn't a
| normal investment. It's a special government project investment
| facilitated by OpenAI board member and former clandestine CIA
| operative and cybersecurity executive Will Hurd through his
| role on the board of trustees of In-Q-Tel the investment arm of
| the CIA. It's funneled through Microsoft instead of through
| Google in part because of Google's No-Military-AI pledge in
| 2018 demanded by its employees, after which Microsoft took over
| its military contracts including project Maven. The new special
| government project, the Sydney project, is the most urgent and
| ambitious since the project to develop nuclear weapons in the
| mid twentieth century.
|
| Of course I don't necessarily believe any of that but it can be
| fun to think about.
| koboll wrote:
| Wait holy shit this is at least partially true though?
| https://openai.com/blog/will-hurd-joins
| nirushiv wrote:
| Please stop spreading FUD and unsubstantiated rumours all
| over this thread
| paganel wrote:
| Why wouldn't the US Government invest billions of dollars
| in a technology that it sees as essential? What's FUD-y
| about that? Most of our industry itself is the result of
| the US Government's past investments for military-related
| purposes.
|
| Later edit: Also, article from 2016 [1]
|
| > There's more to the Allen & Co annual Sun Valley mogul
| gathering than talk about potential media deals: The
| industry's corporate elite spent this morning listening to
| a panel about advances in artificial intelligence,
| following sessions yesterday dealing with education,
| biotech and gene splicing, and the status of Middle East.
|
| > Netscape co-founder Marc Andreessen led the AI session
| with LinkedIn's Reid Hoffman and Y Combinator's Sam Altman.
| The main themes: AI will affect lots of businesses, and
| it's coming quickly.
|
| > Yesterday's sessions included one with former CIA
| director George Tenant who spoke about the Middle East and
| terrorism with New York Police Department Deputy
| Commissioner of Intelligence & Counter-terrorism John
| Miller and a former chief of Israeli intelligence agency
| Mossad.
|
| So, yes, all the intelligence agencies are pretty involved
| in this AI thing, they'd be stupid not to be.
|
| [1] https://deadline.com/2016/07/sun-valley-moguls-
| artificial-in...
| ftxbro wrote:
| Now seven years later Will Hurd and George Tenet are
| currently the managing director and chairman respectively
| of Allen & Co! More facts worth considering are in the
| mysterious hacker news comment from the other day:
| https://news.ycombinator.com/item?id=35366484
| [deleted]
| kneel wrote:
| It would be irresponsible for intelligence agencies NOT to
| involve themselves in AI. LLMs have the capabilities to
| catalyze economic shockwaves on the same magnitude of the
| internet itself.
|
| Notice how OpenAI is open to many Western friendly
| countries but not certain competitive challengers?
| https://platform.openai.com/docs/supported-countries
| FormerBandmate wrote:
| Out of the BRICS, Brazil, India, and South Africa are
| there. Russia and China aren't, but that's not really an
| issue of "competitive challengers" so much as
| dictatorships who are invading or threatening to invade
| democracies
| ChatGTP wrote:
| Realise that America has invaded plenty of countries,
| overthrown leaders, been a huge driver of claimed change
| and oil industries, basically done whatever it wants and
| continues to do so, pretty much based on being able to
| print as much money,as it likes and if you don't like
| that, you'll face the full force of the military
| industrial complex.
|
| Look at Snowden and Assange. They tried to show us what's
| behind the curtain and their lives were wrecked.
|
| The rhetoric on here about Russia and China = "bad guys",
| no questions asked is overly simplistic. Putin is clearly
| in the wrong here. But what creates a person like that? I
| believe we are somewhat responsible for it.
|
| People cite possible atrocities of Xinjiang, but what
| about Iraq and Siria, North Korea, Vietnam whole entire
| countries destroyed. Incredible loss of life.
|
| American attitudes are a huge source of division in the
| world. Yes so are China and Russias.
|
| We cannot only see one side of a story anymore, it's just
| too dangerous. As we have more powerful weapons and we
| do, we have to, absolutely have to learn to understand
| each other and work through diplomacy with a more open
| mind and peaceful outcomes which are beneficial for all.
|
| No I'm not advocating for dictators, but you cannot
| pretend that Americans invasions have been always
| positive or for good intention, or that American
| interests are always aligned with the rest of the worlds.
|
| The arms races need to stop. Very quickly.
| sebzim4500 wrote:
| Is it not enormously more likely that Microsoft decided to
| invest a small amount (to them) in a technology which is
| clearly core to their future business plans?
| xenospn wrote:
| Who has time to come with this stuff?!
| poopypoopington wrote:
| ChatGPT
| seattle_spring wrote:
| Most of it seems to come from /r/conspiracy, which
| unsurprisingly has a lot of overlap with another subreddit
| that starts with /r/cons*
| the_doctah wrote:
| I'd love to go through all the things Liberals labeled a
| conspiracy in the last 5 years that actually became true,
| but I don't have that kind of time today
| slickdork wrote:
| I'd settle for three examples with sources.
| pjohri wrote:
| Well at least one: in the year 2000, I used to work for
| Verizon and a picture from one of the local networks hubs
| was circulated showing a bunch of thick cables tapping
| into the network and alleging that the government was
| listening to all calls Americans made. People made a lot
| of fun of that photo until Snowden brought the details to
| light.
| [deleted]
| mongol wrote:
| How did he become so rich? His Wikipedia page does not have much
| details on this.
| [deleted]
| ftxbro wrote:
| Paul Graham is his Les Wexner.
|
| "Sam Altman, the co-founder of Loopt, had just finished his
| sophomore year when we funded them, and Loopt is probably the
| most promising of all the startups we've funded so far. But Sam
| Altman is a very unusual guy. Within about three minutes of
| meeting him, I remember thinking 'Ah, so this is what Bill
| Gates must have been like when he was 19.'"
|
| "Honestly, Sam is, along with Steve Jobs, the founder I refer
| to most when I'm advising startups. On questions of design, I
| ask "What would Steve do?" but on questions of strategy or
| ambition I ask "What would Sama do?"
|
| What I learned from meeting Sama is that the doctrine of the
| elect applies to startups. It applies way less than most people
| think: startup investing does not consist of trying to pick
| winners the way you might in a horse race. But there are a few
| people with such force of will that they're going to get
| whatever they want."
| version_five wrote:
| If anyone has read "Wild Sheep Chase" by Haruki Murukami,
| there was the idea of being possessed by the sheep which
| turned one into a forceful business man. I have only met a
| couple people like this, but as in the quote it's immediately
| obvious, and you see why they are where they are.
| gautamcgoel wrote:
| In my opinion, this book is his most underrated book. It's
| very funny, highly recommend!
| bmitc wrote:
| I feel I'd take every word of that with a massive grain of
| salt. It reeks of cult of personality.
| dilap wrote:
| I did take it with a huge grain of salt (and an eye-roll)
| reading it years ago. However, given where sama and OpenAI
| are today, perhaps pg was right all along!
| jasmer wrote:
| Loopt was a failure. Political skills. It's a stream of PR,
| not anything operationally applied, or performant.
| pbw wrote:
| I don't have the source, it was a video interview, but Sam said
| he has personally invested in around 400 startups. And it says
| here he employs "a couple of dozen people" to manage his
| investment and homes. At that scale I think you yourself
| basically are a venture capital firm.
| jeremyjh wrote:
| You are just describing a rich person, not how he became
| rich.
| elorant wrote:
| Those 400 startups are gradual investments. They didn't all
| happen overnight. If you have really early access to some
| very promising startups you don't need a shitton of money
| to invest in them.
| naillo wrote:
| I'd imagine he got to invest early in a lot of successful YC
| companies during his time there.
| ohgodplsno wrote:
| Parents were rich, sent their child to Stanford and used their
| connections to let him build connections to other rich people,
| founded a shitty startup in the middle of a period where any
| rich kid making a social media company would get bought out for
| dozens of millions, rest is history.
|
| He's always been rich.
| 876978095789789 wrote:
| On Loopt's questionable acquisition:
| https://news.ycombinator.com/item?id=3684357
| 876978095789789 wrote:
| He was in the first batch of YC startups with a feature phone
| location-aware app called Loopt. Once smartphones came along,
| it became largely obsolete, and started to become irrelevant,
| but still got acquired under questionable circumstances anyway,
| enough for Altman to get rich:
| https://news.ycombinator.com/item?id=3684357
|
| From there he became a VC and ultimately president of YC.
| 876978095789789 wrote:
| I'm not sure if I should elaborate further, but in the last
| years of Loopt, it actually devolved into a hook-up app
| servicing the gay community, basically Grindr before Grindr:
| https://news.ycombinator.com/item?id=385178
|
| I guess he was ahead of his time, in a way? Still, I've never
| forgotten that this silly "success" was the first big exit of
| the most touted YC founder, ever.
| sebzim4500 wrote:
| I think that's just being ahead of your time, no
| qualification needed.
| 29athrowaway wrote:
| Should we also talk about the contradictions of WSJ?
|
| The only way to never contradict yourself is to never say
| anything.
|
| Now, is AI the right work area to "move fast and break things"?
| no.
| molodec wrote:
| Altman "hadn't been to a grocery store in four or five years". He
| is so out of touch with real world, people needs and desires, and
| fantasies about the future world based on the assumption that
| most people want to be free "to pursue more creative work." I
| think most people don't actually dream about pursuing creative
| work. Being absolutely "free" of work doesn't make one more
| creative. Real problems and constraints force people to come up
| with creative solutions.
| leobg wrote:
| The contradiction I noticed during his Lex interview was him
| talking about being attacked by Elon Musk. He said it reminded
| him of how Elon once said he felt when the Apollo astronauts
| lobbied against SpaceX. Elon said it made him sad. Those guys
| were his heroes. And that the wish they would come visit and look
| at the work SpaceX was doing. I found that comparison by Altman
| disingenuous. First, he didn't seem so much sad as he seemed
| angry. At one point in the interview, he said that he thought
| about some day attacking back. That's not at all how Elon had
| felt about those astronauts. And second, why doesn't Altman just
| invite Elon and show him the work they are doing? Wouldn't take
| more than a phone call.
| DubiousPusher wrote:
| Stop asking the market to look after collective interest. This is
| the job of the government. The ultimate effect of virtue wanking
| CEOs and corporate governance is to deceive people into thinking
| democracy is something that can be achieved by for profit
| organizations and that they can forsake the formal binding of
| collective interest through law.
|
| It's nice if people are nice but it is not a bulwark of the
| collective good. It is a temporary social convenience. The higher
| that niceness exists in the social order, the greater its
| contemporary benefit but also, the more it masks the
| vulnerability of that social benefit.
|
| It matters if Sam Altman is Ghandi or Genghis Khan in a concrete
| way but you, as a citizen must act as if it doesn't matter.
|
| If AI poses a danger to the social good, no amount of good guy
| CEOs will protect us. The only thing that will is organization
| and direct action.
| erlend_sh wrote:
| The market should absolutely be looking after the collective
| interest. That's what the we the collective created the market
| for in the first place!
| DubiousPusher wrote:
| The market will follow its inceotives. You shape those
| incentives with laws. If you want a market that allows people
| to take risks, you do that by inventing the limit liability
| corporation not by telling people to be nice and to not
| pursue their debts unto their debtors' personal property. If
| you want a market that discourages monopoly you do that by
| regulating combination not by writing articles about how
| "good businessmen" don't act anticompetitively.
| jokethrowaway wrote:
| That's completely misguided. Consumers purchase from companies
| they like.
|
| Why do you think companies are all woke and virtue signalling?
| Because they interpreted the vocal woke minority as the voice
| of the country and they want to capture that market.
|
| Corporations will absolutely try to do go in order to maximise
| their profit.
|
| And this is ignoring private charities which get more done than
| any government has ever done.
|
| Collective interest is a nice concept but the government, like
| all large organizations, is not capable of moving in any
| direction. Whatever you need done, chances are someone's cousin
| will get a job, the job will be done poorly and the taxpayers
| will pay more in taxes to fix the problem again and again and
| again.
|
| A government can't fail and it's therefore inefficient.
| runarberg wrote:
| Your post ignores the existence of democracies. Governments
| fail all the time. In a democracy failures of government will
| often yield a total collapse and complete replacement. If a
| failure is spectacular enough, these failure often come with
| constitutional reforms or even revolutions.
|
| In addition, democratic governments (and even many autocratic
| governments) have some levels of distribution of power. Your
| small municipal government may very well end up being
| absorbed into you neighboring municipality because it is more
| efficient. Maybe an intermediate judicial step is introduced
| at a county, or even country level.
|
| Governments do try, and often succeed into making your
| freedom and your interaction with society at large as
| efficient as possible, while trying to maximize your
| happiness. (Although I'll admit way to often democratic
| governments do act in a way that maximized the profits of the
| wealthy class more then your happiness).
| clarkmoody wrote:
| The third way is to build AI technology that empowers the
| individual against state and corporate power alike. Democracy
| got us here. It cannot get us out.
| Quekid5 wrote:
| Who do you imagine has the majority of compute power?
| maxbond wrote:
| That's not a stable equilibrium. Blogs gave individuals
| asymmetric control over disseminating information - it didn't
| last. If you don't create institutions and power structures
| that cement and defend some ability of individuals, it will
| decay as that power is usurped by whatever institutions and
| power structures benefit from doing so.
| rhcom2 wrote:
| > AI technology that empowers the individual against state
| and corporate power alike
|
| What does this even mean though? Seems like hand waving that
| "AI" is just going to fix everything.
| bredren wrote:
| It's much easier to imagine progressed applications powered
| by recent AI that would provide outsized civic weaponry.
|
| Identification of unusual circumstances or anomalous public
| records seem ripe.
|
| But more straight forward and customized advice on how to
| proceed on any front--super wikihow--makes anyone more
| powerful.
|
| Today, complicated solutions can sometimes to require
| extensive deep research and distillation of material.
|
| So much so that DIY folks can seem like wizards to those
| who only know of turn key solutions and answers.
|
| At the risk of causing a draft from further hand waving: a
| bigger tent can mean a higher likelihood of a special
| person or group of folks emerging.
| joe_the_user wrote:
| _The third way is to build AI technology that empowers the
| individual against state and corporate power alike_
|
| Ha, I'm OK with that as long as I get to pick the individual!
|
| I mean, an AGI under the control of some individual could
| indeed make them more powerful than a corporation or even a
| state but whether increases average individual freedom is
| another question.
| felix318 wrote:
| Such techno-utopianism... political power belongs to people
| who control the guns. There is no way around it.
| smoldesu wrote:
| So long as you buy your inferencing hardware from another
| private party, I'd wager you're helpless against both state
| and corporate power.
| YawningAngel wrote:
| Given that we don't have such AI technology at present, would
| it not be prudent for us to assume that it may not be
| available imminently and plan for how we can address the
| problem without it?
| m3kw9 wrote:
| It's all about risk and rewards, nobody who owns openai is
| going say let's pause. It also never stopped the country that
| first invented nuclear bomb, sure they could paus and then
| Russia would have done it and would have it first, and then
| said "thanks for pausing"
| DubiousPusher wrote:
| I don't know why people think nukes are a good example here.
| Nukes were outright birthed within the government within that
| government at its height of intervention into the market, at
| the height of its reach into the daily lives of every
| American, at the height of American civic engagement.
|
| Policy makers spent a huge amount of time creating a
| framework for them. Specifically there was a huge debate
| about whether they should be under the direct control of the
| military. The careful decision to place them in civilian
| control under the Department of Energy is probably part of
| the reason they haven't been unleashed since.
| bennysonething wrote:
| Though we do vote with our wallets too.
| avgcorrection wrote:
| How great then that some wallets are millions of times larger
| than others.
| bloodyplonker22 wrote:
| Yes, but for your own immediate benefit. People, in general,
| are just not trained to think long term and "for the greater
| good".
| josephg wrote:
| Why not both?
|
| I agree government is useful and needed sometimes. But laws are
| slow, blunt instruments. Governments can't micromanage every
| decision companies make. And if they tried, they would hobble
| the companies involved.
|
| The government moves slowly. When AGI is invented (and I'm
| increasingly convinced it'll happen in the next decade or two),
| what comes next will not be decided by a creaky federal
| government full of old people who don't understand the
| technology. The immediate implications will be decided by Sam
| Altman and his team. I hope on behalf of us all that they're up
| to the challenge.
| TheOtherHobbes wrote:
| They'll be decided by AI, not by governments, corporations,
| or individuals.
|
| There's still a kind of wilful blindness about AI really
| means. Essentially _it 's a machine that can mimic human
| behaviours more convincingly than humans can._
|
| This seems like a paradox, but it really isn't. It's the
| inevitable end point of automatons like Eliza, chess, go, and
| LLMs.
|
| Once you have a machine that can automate and mimic social,
| political, cultural, and personal interactions, that's it -
| that's the political singularity.
|
| And that's true even the machine isn't completely reliable
| and bug free.
|
| Because neither are humans. In fact humans seem predisposed
| to follow flawed sociopathic charismatic leaders, as long as
| they trigger the right kinds of emotional reactions.
|
| Automate that, and you have a serious problem.
|
| And of course you don't need sentience or intent for this.
| Emergent programmed automated behaviour will do the job just
| fine.
| jasonhansel wrote:
| > And if they tried, they would hobble the companies
| involved.
|
| Well, yeah, that's the point of regulation: to limit
| corporate behavior. There are plenty of other highly-
| regulated industries in the US; why shouldn't AI be one of
| them?
| matthewdgreen wrote:
| >The immediate implications will be decided by Sam Altman and
| his team. I hope on behalf of us all that they're up to the
| challenge.
|
| Will they really be determined by OpenAI? So far what Altman
| has accomplished with OpenAI is to push a lot of existing
| research tech out into the open world (with improvements, of
| course.) This has in turn forced the hands of the original
| developers at Google and Meta to push their own research out
| into the open world and further step up their internal
| efforts. And that in turn creates fierce pressure on OpenAI
| to move even faster, and take even fewer precautions.
|
| Metaphorically, there was a huge pile of rocks perched at the
| top of a hill. Altman's push got the rocks rolling, but
| there's really no guarantee that anyone will have much say in
| where they end up.
| whiddershins wrote:
| The problem is that I don't believe we have any organization in
| government currently staffed and active that I trust to take
| any action that will benefit the public at large.
|
| The problem space is too confusing, and the people making
| decisions are too incompetent. It's a huge skills and knowledge
| gap.
|
| And that's without factoring in corruption and bad intentions.
| merlinoa wrote:
| This is socialist nonsense. The government won't protect
| anything outside of their interests[0]. Free markets are good
| and necessary for human flourishing[1].
|
| [0] https://www.amazon.com/Creature-Jekyll-Island-Federal-
| Reserv... [1] https://www.amazon.com/Human-Action-Ludwig-Von-
| Mises/dp/1614...
| dehrmann wrote:
| > Stop asking the market to look after collective interest.
| This is the job of the government.
|
| One of the main roles of government is to step in where markets
| fail.
| nateabele wrote:
| > _One of the main roles of government is to step in where
| markets fail._
|
| Because that's worked out great so far?
|
| In every case I can think of, the government (usually because
| it's already captured) only ever serves to further exacerbate
| the issue.
| liketochill wrote:
| Has a government never prevented a merger that would have
| created a monopoly?
| mafuy wrote:
| Seriously? You can't think of how it was useful that the
| government mandated that factory door must remain open to
| help prevent human deaths in case of a fire? You can't
| think of the advantages of governmental food and medicine
| safety obligations?
| whiddershins wrote:
| The government has had some wins, and it would be naive
| to say they haven't.
|
| They have also had many catastrophic failures.
|
| It's unclear whether in the end regulations trend towards
| net benefit, but it does seem likely that the more
| nebulous a problem, the harder it is for government to
| get it right. Or anyone, for that matter. But especially
| government because the feedback loop is so slow and bad.
| DubiousPusher wrote:
| That's literally my point. When you want behavior that is
| contrary to market incentive you need laws not guilt ridden
| editorials.
| hackerlight wrote:
| > This is the job of the government. The ultimate effect of
| virtue wanking CEOs and corporate governance is to deceive
| people into thinking democracy is something that can be
| achieved by for profit organizations
|
| Sam Altman says the opposite of what you're insinuating, if by
| "virtue wanking CEO deceiving people" you are referring to the
| subject of the article, Sam Altman. He says he wants a global
| regulatory framework enacted by government and decided upon
| democratically.
| yafbum wrote:
| I agree that it's the government's role but I think you can
| look a bit beyond the law itself, which is often hard to get
| right, especially in very fresh new domains. Some nice
| behaviors can be induced by mere fear of government
| intervention and fear of future laws, and I think we're seeing
| some of that now.
| AussieWog93 wrote:
| This is a really cynical take.
|
| CEOs and companies can, and should, act ethically. Not just
| because it's the "right thing to do", but because it's the best
| way to guarantee the integrity of the brand in the long term.
| DubiousPusher wrote:
| I feel I made it clear in my post that indivual integrity
| matters and has real consequences. But you as a citizen have
| 1 no way of validating a CEOs real intentions and 2 no
| recourse when that CEO fails to live up to those intentions.
| If you only fight for the protections you want once you need
| them, you will be at a serious disadvantage to win them.
| Bukhmanizer wrote:
| I'm not sure you could ever say that a company can act
| ethically. People within the company may act ethically, but
| the company itself is just a legal entity that represents a
| group of people. The company has no consideration of ethics
| to act ethically.
|
| A company that is composed of 100% ethical actors may one day
| have all their employees quit and replaced with 100%
| unethical actors. Yet the fundamental things that make the
| company _that company_ would not have changed.
| jasonhansel wrote:
| If unelected CEOs have more power than the democratically
| elected government, we aren't in a democracy. That's the
| problem.
| jcz_nz wrote:
| CEO's and companies can act ethically while that aligns with
| the interest of the shareholders. Reality is that at some
| point, this becomes impossible even for those with the best
| intentions. "Do no evil" rings any bells?
| sbarre wrote:
| History proves that we have way more CEOs and companies
| acting out of self-interest, against the common good, than
| otherwise.
|
| So yeah, they _should_ , but they don't.
| LapsangGuzzler wrote:
| > CEOs and companies can, and should, act ethically.
|
| I can and should always drive the speed limit. But that
| doesn't mean I do, which is why highway patrol exists, to
| keep people in check. "Should" is such a worthless word when
| it comes to these discussions because if you believe that an
| executive needs to act a certain way but you don't believe it
| enough that some sort of check is placed on them, then you
| must not believe in importance of their good behavior that
| strongly.
| dantheman wrote:
| In general governments have done far more harm to individuals
| than anyone in the market. The job of the government is to
| protect your individual rights not the collective interest.
| Quekid5 wrote:
| I refer you to Thomas Midgley Jr.
| piloto_ciego wrote:
| I don't know that this is true.
|
| At very least they both share immense responsibility for
| causing individual harm. Sure the government may start a war,
| but that war can't happen without bombs and bullets, and in
| America at least those factories aren't run by the
| government. There is an intermediate step oftentimes, but I
| don't think that necessarily disconnects companies from
| responsibility.
|
| If you work at a guided bomb factory you may not be the
| person dropping it, but you are responsible for the
| destruction it causes in a small way.
|
| Also, if global warming kills us all then it is likely that
| the oil companies bear some responsibility for it right?
|
| Government sucks - I agree with that statement, but we
| shouldn't act like corporations are appreciably less
| responsible.
| fyloraspit wrote:
| They are worse together. Achieving some fine balance of
| corporations may seem somewhat utopian but we are pretty
| far from utopia in the current day.
|
| Building a mega corporation without big government/s I
| would argue is basically impossible. And local level
| governance is more likely and potent without big
| government. Again though, all of that is quite hard to
| achieve / see how to achieve when people with existing
| power enjoy the status quo control more of the levers than
| the masses, including the ones used to influence the
| masses.
| jokethrowaway wrote:
| The market responds to need.
|
| If nobody were able to socialise the cost of going in a
| foreign country and killing people, there would be no war.
|
| If the government didn't steal my money against my will on
| threat of incarceration, there is no way in hell I'd spend
| my money on bullets to kill someone's son in another
| country.
| syzarian wrote:
| The absence of government just leads to a situation in which
| some group takes control of a given area. In effect
| government will then exist again. During the absence of
| government there will be chaos and rampant crime.
| jokethrowaway wrote:
| I'd take a local firm offering to protect me for money over
| one that manages the large part of a continent.
|
| The small one will redistribute my money where I live at
| least.
|
| Besides, there is an alternative model where there are
| competing groups of people and I can pick the best among
| them based on price and services.
| krapp wrote:
| >Besides, there is an alternative model where there are
| competing groups of people and I can pick the best among
| them based on price and services.
|
| In the absence of government, what's stopping these
| groups from simply joining forces into a cartel, getting
| some armed thugs and making you an offer you can't
| refuse? History suggests that to be a far more likely
| scenario. Oligarchy, rather than competition, is the
| natural state of capitalism.
| majormajor wrote:
| That's just quibbling over what "your individual rights" are;
| where does the line get drawn between "exercising my right"
| and "having my rights infringed on by the actions of
| another." There is no shortage of harm done by "anyone in the
| market" today, whether it's currently illegal and we call it
| "crime" instead of just a person exercising their freedom, or
| whether it's harm that isn't regulated today.
| 7e wrote:
| Promoting the general welfare is literally discussed in the
| first paragraph of the U.S. Constitution. The Bill of Rights
| came later.
| clarkmoody wrote:
| The Constitution is just a piece of paper. For every
| thousand steps the Congress, regulators, and state
| assemblies take in the direction of tyranny, the courts
| claw back one or two.
| [deleted]
| atmosx wrote:
| If we could connect social responsibility with the stock, it
| would matter a great deal :-)
| deltree7 wrote:
| chatGPT, like Google search has essentially democratized
| knowledge and wisdom to every human on planet earth.
|
| OTOH, it's the governments that have banned it.
|
| Do you think the citizen's of Italy are better off because of
| chatGPT?
|
| Corporations are more benevolent than government
| davnicwil wrote:
| I don't know, isn't history really just a series of specific
| people doing concrete actions?
|
| Do you think on some level the idea of some abstract
| 'government' taking care of things is just a narrative we apply
| to make ourselves feel better?
|
| Sure individual decision makers _in_ that government can
| concretely affect reality, but beyond that are we just telling
| a story and really nobody is 'in control'?
| DubiousPusher wrote:
| The actions that idnoviduals take on behalf of government are
| a direct reflection of the "abstract" policies and laws of
| that government. If you cannot discern this from 20th century
| history I don't know what to tell you.
| runarberg wrote:
| > I don't know, isn't history really just a series of
| specific people doing concrete actions?
|
| That is like saying: "Aren't human brains just series of
| neurons, firing at specific moments."
|
| History is as much--if not more--about interactions between
| people, feedback loops, collective actions, collective
| reactions, environmental changes, etc. I would argue that the
| individual is really really insignificant next to the system
| this individual resides under, and interacts with.
| Mezzie wrote:
| > beyond that are we just telling a story and really nobody
| is 'in control'?
|
| Basically. Or at least that's the impression I'm left with
| after spending a couple of years in politics work.
|
| I had a nice long breakdown.
| trendroid wrote:
| Possible to Elaborate?
| Mezzie wrote:
| tl;dr: Nobody is steering the ship because they don't
| know they're on a ship. Or that the ocean exists.
|
| It's hard without doxxing myself or calling out specific
| people and organizations which I'd rather not because I'm
| a nobody and can't afford lawsuits, but for various
| reasons I ended up political education and marketing for
| civics advocacy. Ish. To be semi on topic, I know some
| people who are published in the WSJ (as well as the
| people who actually _wrote_ the pieces). I 'm also a 3rd
| generation tech nerd in my mid 30s so I'm very
| comfortable with the digital world - easily the most so
| outside of the actual software engineering team.
|
| I've spoken with and to a lot of politicians and
| candidates from across the US - mostly on the local and
| state level but some nationally. And journalists from
| publications that are high profile, professors of legal
| studies, heads of think tanks, etc.
|
| My read of the situation is that our political class is
| entangled in a snare of perverse disincentives for action
| while also being so disassociated from the world outside
| of their bubble that they've functionally no idea what's
| going on. Our systems (cultural, political, economic,
| etc.) have grown exponentially more complex in the past
| 30 years and those of us on HN (myself included)
| understand this and why this happened. I'm a 3rd
| generation tech nerd, I can explain pretty easily how we
| got here and why things are different. The political
| class, on the other hand, has had enough power to not
| need to adapt and to force other people to do things
| their way. If your 8500 year old senator wants payment by
| check and to send physical mail, you do it. (Politicians
| and candidates that would not use the internet were
| enough of a problem in _2020_ that we had to account for
| it in our data + analyses and do specific no tech
| outreach). Since they didn 't know how the world is
| changing, they also _haven 't been considering the
| effects of the changes at all_.
|
| Furthermore, even those of them that have some idea still
| don't know how to problem solve _systems_ instead of
| _relationships_. Complex systems thinking is _the_ key
| skill needed to navigate these waters, and _none of them
| have it_. It 's fucking _terrifying_. At best, they can
| conceive of systems where everything about them is known
| and their outputs can be precisely predicted. _At best_.
| Complex systems are beyond them.
|
| Add to this that we have a system which has slowly ground
| itself to a deadlocked halt. Congress has functionally
| abandoned most of its actual legislative duties because
| it's way better for sitting congresspeople to not pass
| any bills - if you don't do anything, then you don't piss
| any of your constituents off. Or make mistakes. And you
| can spend more time campaigning.
|
| I left and became a hedonist with a drug problem after a
| very frank conversation with a colleague who was my
| political opposite at the time. I'm always open to being
| wrong, and hearing that they didn't have any answer
| either was a very 'welp, we're fucked' moment. I'm
| getting better.
| olao99 wrote:
| history is a simplified, prettified story that we tell
| ourselves about how things happened.
| beepbooptheory wrote:
| The more we remember the possibility of true collective self
| determination, the more likely we are to survive all this
| mess we're making.
|
| These days we are constantly bombarded by this contradiction
| of individualism being primary and desirable, but at the same
| time impotent in the face of the world this individualism has
| wrought. And its all a convenient way to demoralize us and
| let us forget how effective motivated collective interest is.
| Real history begins and ends with the collective!
| DubiousPusher wrote:
| Yes. There is a reason that when Britain felt threatened by
| the turmoil in France they didn't just bar unions or
| political clubs. They banned "combination" almost entirely
| in general.
| invig wrote:
| "think on some level the idea of some abstract"
|
| This! So much this! These conversations are being held at
| such a high level of abstraction that they don't make sense.
| It's one giant "feels" session.
| DubiousPusher wrote:
| Right. The National Health, turn of the century sanitation,
| widespread vaccination and the EPA are just all about "the
| feels".
| TaylorAlexander wrote:
| Sure but in my country (USA) the government is hopelessly inept
| at regulating technology. We still don't have privacy
| regulations and now to work around this they're trying to ban
| specific foreign apps instead of protecting us from all apps!
| I'd honestly be horrified if they tried to regulate AI. They
| would be in bed with Facebook and Microsoft and they'd somehow
| write legislation that only serves to insulate those companies
| from legal repercussions instead of doing anything to protect
| regular people. As far as I can tell it is the view of congress
| that big tech can to whatever they want to us as long as the
| government gets a piece.
| alex_sf wrote:
| There is no government that _isn 't_ hopeless inept at
| regulating technology.
| DubiousPusher wrote:
| Agreed. The US has backslidden since the 20th century back
| towards an elitest Republic and away from democracy. But even
| in the US, collective action has a better track record than
| "altruism".
| TaylorAlexander wrote:
| Sometimes I wonder if the back slide narrative is really
| accurate, or if we're looking back at the myth of history
| rather than the facts. When the country was founded, only
| white men could vote and people of color were legally
| property with no rights. That's obviously not democracy, so
| I question at what point after that but before today we
| really had democracy to have slid back from.
| bugglebeetle wrote:
| When we realize it's really only about from the 1970s
| that we had full enfranchisement and political
| participation of all citizens, this becomes more obvious.
| "Coincidentally," this enfranchisement was followed by
| the Volker shock and then the Reagan administration, both
| of which led to the decimation of labor's political power
| and share of the economic pie.
| Overtonwindow wrote:
| AI is not true AI.. at the moment. ChatGPT is inherently biased
| by its developers, which means at least half the population may
| not trust its answers. For true AI he will have to give it
| autonomy, and I'm more interested if Altman is ready to live with
| an AI he cannot control.
| ur-whale wrote:
| https://archive.is/Quo4t
| gardenfelder wrote:
| https://archive.ph/YTOeD
| eternalban wrote:
| None of archive.* are working for me - cloudflare dns issues.
| Anyone else has access issues?
|
| [thanks to those who replied. strangely stopped working for me
| since yesterday [US]. can you post the ip you see?
|
| Cloudflare returns a 1001 error: "Ray ID: 7b128f151e4b0c90 *
| 2023-04-01 17:30:15 UTC" ]
| version_five wrote:
| Works for me in western europe
| wslh wrote:
| No issue here (Buenos Aires).
| marginalia_nu wrote:
| Works for me
| wslh wrote:
| Sidenote: I just reported to archive.is that it would be great
| to have the capability to render it throught services such as
| Pocket.
| ThomPete wrote:
| Listening to Sam talking with Lex Fridman about the dangers and
| ethics of AI while his company destroys entire industries as a
| consequence of their decision to keep GPT4 closed source and
| spitting out an apex api aggregator is one for the history books.
|
| Well played :)
| nomel wrote:
| > destroys entire industries as a consequence of their decision
| to keep GPT4 closed source
|
| Could you expand in that?
| avgcorrection wrote:
| Like contemporary language models, some HN commenters read
| the text itself in isolation and then extrapolate about "what
| that means" but then immediately jump to the conclusion that
| whatever real-world (beyond the text) that they imagine will
| eventually happen has in fact already happened--in effect
| they're hallucinating.
| washywashy wrote:
| OpenAI/ChatGPT seem to be very creative based on variations it
| can make on "things" that already exist. I'm just curious if we
| see AI being truly creative and making something "new". Perhaps
| everything is based on something though, and that's a rough
| explanation for this creativity. Maybe AI's true creativity can
| come from the input prompts of its "less intelligent", but more
| flexibly creative users.
| OscarTheGrinch wrote:
| Perhaps if more CEOs / controlling shareholders were criminally
| liable for damage caused by their products, like the Sackler
| family, they wouldn't be so gung-ho.
| leetharris wrote:
| What a ridiculous take and a slippery slope.
|
| Should car manufacturers be liable for drunk drivers? Should
| kitchen knife manufacturers be responsible for stabbings?
|
| Your idea is great if you want your country to be left behind
| entirely in innovation.
| dclowd9901 wrote:
| Worth reminding that the slippery slope argument is not a
| valid argument at all.
|
| As with standard slippery slope reasoning, you jump to the
| most extreme interpretation. Yet reality shows us that you
| cannot, in fact, boil a frog, because at some point it just
| gets too fucking hot.
|
| Should car manufacturers be liable for drunk drivers? Maybe,
| if they include a space in their vehicle specifically to
| store and serve alcohol.
|
| Should kitchen knife manufacturers be responsible for
| stabbings? No. But no reasonable person would ever suggest
| they should. I might remind you also that "reasonable
| standard" is a legal concept.
| robertlagrant wrote:
| Without defining the reasonable standard, it remains a
| silly idea in this case.
| iibarea wrote:
| There's tons of extremely effective marketing around who this guy
| is, what he stands for - and so I'd instead look at what he's
| done. He took a non-profit intended to offset the commercial
| motives driving AI development, and turned it into a for profit
| closely tied to Microsoft. I think he's an extremely shrewd
| executive and salesman, but nothing he's done suggests any
| altruistic motivations - that part always seems to be just
| marketing, and always way down the road.
| startupsfail wrote:
| What I'm afraid of is that he and Ilya are not as good and
| smart as they paint themselves.
|
| And that a lot of key people had left (i.e. to Anthropic). And
| that by pure inertia they have GPT-5 on their hands and not
| much control over where this technology is going.
|
| I can't tell for certain, but it does look like one of their
| corner pieces, the ChatGPT system prompt which sits at the
| funnel of the data collection had degraded significantly from
| the previous version. Had the person that was the key to the
| previous design left? Or it no longer matters?
|
| One could argue that OpenAI is very hot and everyone would want
| to work there. But a lot of newcomers only create more pressure
| for the key people. And then there is the inevitable leakage
| problem.
| jstummbillig wrote:
| There are some vague ideas and fears here. Understandable.
| Trying find a silver lining from where to get somewhere:
| Where would GPT4 and onwards be better housed? Is there a
| setup -- an individual, a company, an institution, a concept,
| license -- where the whole thing would clearly better fit,
| than with OpenAI?
|
| Note, I am not suggesting that they are particularly
| un/qualified or un/trustworthy. I am just trying to figure,
| if the problem is with the nature of the technology, that
| maybe there is not entity or setup, that would obviously be a
| good fit for governing gpt because gpt is simply scary, or
| this is a personality issue.
| bmitc wrote:
| > What I'm afraid of is that he and Ilya are not as good and
| smart as they paint themselves.
|
| This describes almost any venture capitalist or high-profile
| startup founders, as far as I can tell. Most don't realize
| their either privileged path or lucky path or both had more
| to do with it than their smarts.
|
| I really like James Simons as he mostly attributes his
| success to luck and being able to hire and organize smart
| people and give them the tools they need to work. He
| basically describes it as luck and taste, despite his actual
| smarts and his enormous impact on the world.
| Mistletoe wrote:
| I don't know everything about him but from what I do know,
| I would put Bezos in the "not just luck" very lonely camp.
| I think his Day 1 work and iterate every day idea is just
| that powerful and real and he really did it instead of
| talking about it. Even though he says he won several
| lotteries to get where he is, I'm not so sure.
| saghm wrote:
| > I really like James Simons as he mostly attributes his
| success to luck and being able to hire and organize smart
| people and give them the tools they need to work. He
| basically describes it as luck and taste, despite his
| actual smarts and his enormous impact on the world.
|
| Plenty of really smart people don't end up having a big
| impact on the world, and it's possible to make a difference
| without being an outlier in terms of intelligence. Everyone
| who has made an impact has benefited to some degree by
| circumstances beyond their control though, so even if
| someone is genuinely smarter than anyone else, it's a
| fallacy for them to assume that it was the determining
| factor in their success and a guarantee of future success.
| iibarea wrote:
| ... and Simons would maybe be the most justified in
| overlooking luck, but he's smart enough to realize how
| random the world is. Peter Norvig also emphasizes the role
| of luck in his life. It's honestly a very good test of
| self-awareness and empathy, though there's def some
| negative selection against those traits in sv.
| sebzim4500 wrote:
| I'm sure they are both pretty smart, but if anything that
| makes their apparent monopoly more concerning.
| dnissley wrote:
| Monopoly over what?
| sebzim4500 wrote:
| LLMs that actually work.
|
| They are on GPT-4 and no one else is close to GPT-3.5.
| FormerBandmate wrote:
| You can use Anthropic's chatbot in Quora's Poe app. Right now
| it isn't as good as Bing or ChatGPT. Misses some basic logic
| things, and the "As an AI language model" BS still stops it
| from doing fun things like making Jesus rap battle Gus Fring
| (that was like a month ago, someone in the replies got it to
| do that so I'll have to check it out again). I'd have to see
| how it is at writing PowerPoints but idk
| IncRnd wrote:
| Verse 1 - Jesus
|
| I'm the son of God, the King of Kings
|
| You're just a drug lord, selling crystal meth and things
|
| My teachings change lives, bring peace to the world
|
| You bring addiction, violence and pain, unfurled
| Chorus
|
| Jesus, the savior, the light in the dark
|
| Gus Fring, the villain, who leaves his mark
| Verse 2 - Gus Fring
|
| You talk a big game, but where's your proof?
|
| I've built an empire, with power that's bulletproof
|
| Your miracles are outdated, my tactics are new
|
| I'll take you down, no matter what you do
| Chorus
|
| Jesus, the savior, the light in the dark
|
| Gus Fring, the villain, who leaves his mark
| Verse 3 - Jesus
|
| My love conquers all, it's the greatest force
|
| Your money and power, just lead to remorse
|
| You're just a man, with a fragile ego
|
| I'll show you mercy, but you reap what you sow
| Chorus
|
| Jesus, the savior, the light in the dark
|
| Gus Fring, the villain, who leaves his mark
| Verse 4 - Gus Fring
|
| You may have won this battle, but the war is not done
|
| I'll continue to rise, until I've won
|
| You may have followers, but they'll never be mine
|
| I'll always come out on top, every time
| Chorus
|
| Jesus, the savior, the light in the dark
|
| Gus Fring, the villain, who leaves his mark
| Outro
|
| In the end, it's clear to see
|
| Jesus brings hope and love, for you and me
|
| Gus Fring may have power, but it's not enough
|
| Jesus is the way, the truth, the life, and that's tough.
| FormerBandmate wrote:
| Huh, last time I checked it it gave me a message about
| how that was "offensive to Christians". I'll have to
| check it out again
| sroussey wrote:
| > ChatGPT system prompt which sits at the funnel of the data
| collection had degraded significantly from the previous
| version
|
| They purposely moved free users to a simpler/cheaper model.
| Depending on your setting and if you are paying, there are
| three models you might be inferencing with.
| startupsfail wrote:
| I'm not talking about GPT-3.5 vs GPT-4. I'm talking about a
| change to their system prompt.
| TechnicolorByte wrote:
| It's a capped-profit structure where excess profit will
| supposedly go back to the no-profit side. From a recent NYTimes
| article [1]:
|
| > But these profits are capped, and any additional revenue will
| be pumped back into the OpenAI nonprofit that was founded back
| in 2015.
|
| > His grand idea is that OpenAI will capture much of the
| world's wealth through the creation of A.G.I. and then
| redistribute this wealth to the people. In Napa, as we sat
| chatting beside the lake at the heart of his ranch, he tossed
| out several figures -- $100 billion, $1 trillion, $100
| trillion.
|
| How believable that is, who knows.
|
| [1] https://www.nytimes.com/2023/03/31/technology/sam-altman-
| ope...
| G_z9 wrote:
| This is ducking insane. How are people not up in arms about
| this? Imagine if the guy who invented recombinant insulin
| stated publicly that he intended to capture the entire
| medical sector and then use the money and power to reshape
| society by distributing wealth as he saw fit. That's ducking
| insane and dangerous. This guy has lost his fucking mind and
| needs to be stopped.
| [deleted]
| sebzim4500 wrote:
| I find OpenAI a bit sketchy, but this is an overreaction.
| The only difference between OpenAI and the rest is that
| OpenAI claims to have good intentions, only time will tell
| if this is true. But the others don't even claim to have
| good intentions. It's not like any of OpenAI's actions are
| unusually bad for a for-profit compnay.
| tsunamifury wrote:
| I'm sorry your AI keyboard didn't like your sentiment.
| Words have been changed to reduce your vulgarity. Thankyou
| for your human node input.
|
| On a serious note I think you are right. In private the
| ideology of him and his mentor Theil is a lot more...
| elite. Their think tank once said "of all the people in the
| world there are probably only 10,000 unique and valuable
| characters. The rest of us are copies."
|
| I'm not going to criticize that because it might be a valid
| perspective but filter it through that kind of power. I
| don't love that kind of thinking driving such a powerful
| technology.
|
| I am so sad that Silicon Valley started out as a place to
| elevate humanity and ended with a bunch of tech elites who
| see the rest of the world generally as a waste of space.
| They claim fervently otherwise but at this point it seems
| to be a very thin veneer.
|
| The obvious example being GPT was not built to credit or
| give attribution to its contributors. It is a vision of the
| world where everything is stolen from all of us and put in
| Sam Altmans hands because he's... better or Something.
| blibble wrote:
| > How are people not up in arms about this?
|
| they will be once they realise
|
| > This guy has lost his fucking mind and needs to be
| stopped.
|
| I agree, hopefully via regulation
|
| otherwise the 21st century luddites will
| cactusplant7374 wrote:
| Are the profits capped for Altman?
| Jasper_ wrote:
| Returns are capped at 100x their initial investment, which,
| you know, is not that big of a cap. VCs would go crazy for a
| 100x return. Most companies, even unicorns, don't get there.
|
| They're justifying it by saying AGI is so stupidly big that
| OpenAI will see 100000x returns if uncapped. So, you know,
| standard FOMO tactics.
|
| [0] https://openai.com/blog/openai-lp
| mirekrusin wrote:
| This cap is much smaller, 100x was for initial investors.
| Microsoft took every single penny they could to get 49%
| stake.
|
| If they won't do AGI, they won't go over the cap with
| profits and all drama is for nothing - so saying it's fake
| cap is not right.
|
| Please somebody correct me if I'm wrong.
| [deleted]
| sebzim4500 wrote:
| I mean, presumably they are at like 30x already?
| jprete wrote:
| I believe the Microsoft investment is around $10 billion,
| so they can get up to a trillion dollars of return under
| the cap.
| iibarea wrote:
| Again, this is unbelievably good marketing - and good sales
| when pitching VCs. Plus it's a nice reworking of the very for
| profit nonprofit model (see also FTX). But in terms of actual
| reality openAI is mostly succeeding by being more reckless
| and more aggressively commercial than the other players in
| this space, and is in no meaningful way a nonprofit any
| longer.
| btown wrote:
| I'm sure he believes that at such a time when OpenAI creates
| AGI, all the company's investors' profit caps will have been
| passed (or will immediately be passed), and thus he will have
| removed all incentives for anyone at the company - including
| himself - to keep it from the world.
|
| But there are so, so many incentives other than equity that
| come into play. Pride, well-meaning fear of proliferation,
| national security concerns, non-profit-related but still-
| binding contractual obligations... all can contribute to
| OpenAI wanting to keep control of their creations even if
| they have no specific financial incentive to do so.
|
| Whether that level of control is good or bad is a much longer
| conversation, of course.
| birdymcbird wrote:
| > altruistic motivations
|
| feel like trend among Silicon Valley companies and tech
| 'genius' personality about having altruism. some delusion that
| basing personality on this lie will make them untouchable and
| elevate their character, as if they not in this just to make
| ton of money like every other company and industry. and
| American media generally push this propaganda. SBF prime
| example.
| cactusplant7374 wrote:
| > and American media generally push this propaganda. SBF
| prime example.
|
| Did they? I've only listened to one interview of SBF and that
| was done by Tyler Cowen. He seemed totally aloof to the
| seriousness of running an exchange. If anything we've been
| convinced that idiosyncratic individuals are our saviors.
| birdymcbird wrote:
| Sbf constantly promoted in us media by news organization
| like nytime and celebrity before all the fraud became
| apparent.
|
| once cat was out of bag they run story going easy on sbf.
| never apologizing for promoting this fraud and they wrote
| articles sympathetic to sbf, also giving him platform to
| visit each news show or talk show and give defense of
| himself as he knew nothing about what was happening, all
| part of legal defense strategy to say incompetence but not
| criminal negligent
| cookingrobot wrote:
| He didn't take equity in OpenAi. Does that suggest altruism?
| iibarea wrote:
| Assuming we take this at face value, once you have a lot of
| money power becomes appealing - and control over a very
| important player in the AI space is that. The original vision
| of openAI was democratization of that decision-making
| process, the model now is - these guys are in charge. Maybe
| that's altruistic, because they're the smartest guys in the
| room and they can mitigate the downside risks of this tech
| (... not fucking AGI, but much more like the infinite
| propaganda potential of chatGPT). I'm more a fan of
| democratization, but that's not a universally held opinion in
| sv.
| choppaface wrote:
| He also invited peter theil to YC and made his first millions
| selling the personal data of Loopt users to low income credit
| card vultures. Also ... Worldcoin?
| victor106 wrote:
| > OpenAI and Microsoft also created a joint safety board, which
| includes Mr. Altman and Microsoft Chief Technology Officer Kevin
| Scott, that has the power to roll back Microsoft and OpenAI
| product releases if they are deemed too dangerous.
|
| what a joke.
|
| find me one instance where the ceo of a company picked public
| interest/safety over profits when there was no regulatory
| oversight?
| dragonelite wrote:
| Do people really think AI will go haywire like in the Hollywood
| movies?
| coffeebeqn wrote:
| Some do. Personally I think that LLMs will hit a ceiling
| eventually way before AGI. Just like self driving - the last
| 20% is orders of magnitude more difficult than the first 80%
| pell wrote:
| I don't think we're close to a situation where they send us
| into a Matrix. But I can see a scenario where they are
| connected to more and more running systems of varying degree
| of importance to human populations such as electrical grids,
| water systems, factories, etc. If they're essentially given
| executive powers within these systems I do see a huge
| potential for catastrophic outcomes. And this is way before
| any actual AGI. The simple "black box" AI does not need to
| know what it's doing to cause real-world consequences.
| gwd wrote:
| Not like in Hollywood movies, but yes:
|
| https://www.youtube.com/watch?v=gA1sNLL6yg4
|
| https://www.alignmentforum.org/s/mzgtmmTKKn5MuCzFJ
|
| Look around alignmentforum.org or lesswrong.com, and you'll
| see loads of people who are worried / concerned at various
| levels about what could happen if we suddenly create an AI
| that's smarter than us.
|
| I've got my own summary in this comment:
|
| https://news.ycombinator.com/item?id=35281504
|
| But this discussion has actually been going on for nearly two
| decades, and there is a _lot_ of things to read up on.
|
| EtA: Or, for a fun version:
|
| https://www.decisionproblem.com/paperclips/index2.html
| jeron wrote:
| Yudkowsky, to begin with
| sobkas wrote:
| > Do people really think AI will go haywire like in the
| Hollywood movies?
|
| No, it will just make every inequality even harder to fight.
| Because computer algorithm can't be biased so every decision
| it makes will be objective. And because it's really hard to
| know why AI made decision, it will be impossible to accuse it
| of racism, bigotry, xenophobia. While rich and powerful will
| be ones deciding (through hands of "developers") what data
| will be used to train AIs.
| barbazoo wrote:
| I don't think it's about AI going haywire, more about how the
| technology will be used by people for nefarious purposes.
| benrutter wrote:
| I don't for one, but I still think there could be legitimate
| safety concerns. LLMs are unpredictable, and the possibility
| for misinformation in pitching them as search aggregators is
| pretty large. Disinformation can have, and previously has
| had, genuinely dangerous effects.
| dclowd9901 wrote:
| Are there any employees of OpenAI around? I had a question:
|
| Does anyone in the office stop to contemplate the ramifications
| of developing technology that will likely put most people out of
| a job, which will have a whole host of knock-on effects?
| NickBusey wrote:
| Self checkout machines put people out of a job.
|
| Cars put stable boys out of jobs.
|
| Light bulbs put candlemakers out of jobs.
|
| Are the people who made them also morally culpable?
|
| Let's just make no progress ever in the name of employment I
| guess. /s
| madmask wrote:
| I think these comparisons always miss that humans are still
| useful because they are the control system in the end. Even
| if at very high level. When AGI comes along, humans will have
| to compete with it in the market, and there may be very
| little actual need for humans.
| NickBusey wrote:
| I believe humans will continue to be useful. You apparently
| do not. I have not missed anything.
| sammalloy wrote:
| > Self checkout machines put people out of a job.
|
| Have they, though? Is there good data on this? I haven't seen
| anyone lose any jobs over this in my area. They either get
| reassigned to different departments or get better jobs doing
| the same thing in companies that refuse to use self-
| checkouts. And I read here just a month or so ago that Amazon
| closed many of their "no cashier" stores in NY and Trader
| Joe's vowed to never use them.
|
| > Light bulbs put candlemakers out of jobs
|
| I would guess that such artisans moved on to other niche
| products. Where I live, candlemakers today make a lot of
| money, and I had a chance to watch them do their thing in
| their small commercial space. This was a one person
| operation, no employees and no mechanization, and judging by
| the amount of product they were creating and their wholesale
| prices, they were pulling in about about 300k a year or more,
| after taking into account supplies.
| NickBusey wrote:
| I like that your first point asks for data on a common
| phenomenon. (Side note: Everyone was reassigned or got a
| better job? You sure about that?)
|
| Then your second point is a wild anecdote with zero data.
| (Side note: With zero additional data I can tell you that
| your "judging" of their profit is wildly inaccurate.)
| sammalloy wrote:
| Google is your friend. Candlemaking businesses run by
| sole proprietors are highly profitable with total revenue
| in the billions. I probably wouldn't have believed it if
| I didn't see the operation up close and personal for
| myself. It's a lot of work for one person, and the the
| person who runs the business I saw works 12 hours a day.
| Carrok wrote:
| > person who runs the business I saw works 12 hours a day
|
| Oh OK, so if you work yourself to death you can make
| slightly above average income.
|
| Cool, I guess.
|
| I'm not seeing how it's relevant to GP's point though.
| Candle makers still exist yes.
|
| But are you really arguing the light bulb did not cause
| that specific job to become less common?
|
| I also thought it was fun how, you asked for a citation,
| then when you were asked for to provide one yourself, you
| respond with "gOoGlE iS yOuR fRiEnD".
| CatWChainsaw wrote:
| I'll give them the benefit of the doubt. Many of them probably
| do contemplate it. Aware that since AI is a sea change, there's
| difficulty predicting the full range of first-order
| consequences, much less all the resulting second-order ones.
|
| But... genie, bottle; prisoner's dilemma. If they object to
| what they're building, or how it's implemented, too
| strenuously, they will be out of a job. Then not only do they
| have more immediate concerns about sustaining themselves, they
| have _no_ weight in how things play out.
| galoisscobi wrote:
| Maybe Sam thinks about this at some level. From his New Yorker
| profile[0]:
|
| > "The other most popular scenarios would be A.I. that attacks
| us and nations fighting with nukes over scarce resources." The
| Shypmates looked grave. "I try not to think about it too much,"
| Altman said. "But I have guns, gold, potassium iodide,
| antibiotics, batteries, water, gas masks from the Israeli
| Defense Force, and a big patch of land in Big Sur I can fly
| to."
|
| This doesn't explicitly talk about him being worried about AI
| putting a lot of people out of jobs but he is prepping for AI
| going awry.
|
| [0]: https://www.newyorker.com/magazine/2016/10/10/sam-altmans-
| ma...
| CatWChainsaw wrote:
| The fact that his backup plan if things go really wrong is to
| bug out, damn the rest of the world, is, to put it mildly,
| _not great_.
| efficientsticks wrote:
| > The goal, the company said, was to avoid a race toward building
| dangerous AI systems fueled by competition and instead prioritize
| the safety of humanity.
|
| > "You want to be there first and you want to be setting the
| norms," he said. "That's part of the reason why speed is a moral
| and ethical thing here."
|
| Clearly having either not learned or ignored the lessons from
| Black Mirror and 1984, which is that others will copy and emulate
| the progress.
|
| The fact is that capitalism is no safe place to develop advanced
| capabilities. We have the capability for advanced socialism, just
| not the wisdom or political will.
|
| (I'll answer the anonymous downvote: Altman has advocated giving
| equity as UBI solution. It's a well-meaning technocratic idea to
| distribute ownership, but it ignores human psychology and that
| this idea has already been attempted in practice in 1990s Russia,
| with unfavourable, obvious outcomes).
| sampo wrote:
| > lessons from Black Mirror and 1984
|
| Those are works of fiction.
| quickthrower2 wrote:
| And works of prediction.
| efficientsticks wrote:
| They're dystopian fictions, ie. examples of what _not_ to do.
| But experience has shown that the real-world often recreates
| dystopian visions by example.
|
| So trying to be the first to show something in a well-meaning
| way can nonetheless have unfortunate consequences once the
| example is copied.
| latency-guy2 wrote:
| Tell me, are the dystopian fictions that represent
| socialism or communism as bad, just as reasonable?
|
| Then following up with whatever your answer is: Why are you
| picking and choosing which fictions are reasonable?
|
| Let's dispel the notion that artists and writers are more
| aware and in tune with humanity than other humans.
| roywiggins wrote:
| 1984's Oceania is at least as Stalinist as it is anything
| else.
| jjulius wrote:
| >Then following up with whatever your answer is: Why are
| you picking and choosing which fictions are reasonable?
|
| This is arguing in bad faith. You don't care what their
| answer will be, you have decided that they are absolutely
| picking and choosing, and will still accuse them of as
| much even if their answer to your first question is,
| "Yes".
| latency-guy2 wrote:
| This isn't an argument in the first place buddy.
|
| You're right that I don't care, because it has already
| been decided that Orwell is representing the future if
| things go "The Wrong Way (tm)", buy 1984 at Amazon for
| $24.99, world's best selling book. Or more succinctly to
| OP, "The Capitalist Way (tm)".
| maxbond wrote:
| It's okay to decide that something isn't worth arguing
| against, and to spend your time in a way you find more
| productive.
|
| Having articulated an argument (which you absolutely
| did), it's not okay to try to retcon that you were just
| trolling and everyone else is the fool for having taken
| you seriously.
| latency-guy2 wrote:
| > It's okay to decide that something isn't worth arguing
| against, and to spend your time in a way you find more
| productive.
|
| Who the hell are you talking to? This is some weird segue
| way into something that wasn't even being talked about at
| all.
|
| > Having articulated an argument (which you absolutely
| did), it's not okay to try to retcon that you were just
| trolling and everyone else is the fool for having taken
| you seriously.
|
| I think you're hallucinating anything and whatever you
| want. Anyway don't feel the need to be productive today.
| It's Saturday after all.
| xiphias2 wrote:
| Maybe because they are more digestible then reality. Reality
| is much much worse.
| CatWChainsaw wrote:
| "The only thing stupider than thinking something will happen
| because it is depicted in science fiction is thinking
| something will not happen because it is depicted in science
| fiction."
|
| https://philosophybear.substack.com/p/position-statement-
| on-...
| atleastoptimal wrote:
| The only issue is that in the real world, capitalism has a
| better track record than socialism.
| efficientsticks wrote:
| I agree it's worth looking at the history, and to not repeat
| its mistakes, though at the same time this is a new
| situation, and it will continue to be new into the future, so
| sticking to heuristics may not serve humanity as well than
| being open-minded on the policy front.
| htss2013 wrote:
| Do we have the capability for advanced socialism? Because I
| recall all the smartest economists circa 2021 saying inflation
| wasn't a thing, it's transient, it's only covid affected supply
| chains. In reality we are in an broad sticky inflation crisis
| not seen since the 70s, which may be turning into a regional
| banking crisis.
|
| It's difficult to believe we have reached advanced socialism
| capabilities, and all of the forecasting that would require,
| when we don't even understand the basics of forecasting
| inflation 1-2 years out.
| efficientsticks wrote:
| The ambiguity of "advanced socialism" is problematic for any
| meaningful debate, so I apologise for that.
|
| I was meaning something closer to "we have the resources and
| technology (in this advanced era), just not the wisdom or
| political will". The actual nature of what could be provided
| is up for debate, but if we're looking at mass unemployment
| in 2 decades' time, perhaps it's a conversation worth having
| again.
| moron4hire wrote:
| I'd be very happy to see existing regulation on safety critical
| software systems updated to put a moratorium on AI integration
| for at least the next 5, maybe 10 years.
| sacnoradhq wrote:
| Sama <3
|
| He stuck to the 99.9% gamble of setting money on fire that were
| startups and navigated to the big time(tm).
|
| Also, he helped lift Clerky, Triplebyte, YC, and may others pre,
| during, and post Loopt.
|
| Not many people (or no one) "deserves" success, but Sama brings a
| healthy dose of goodwill wherever he goes.
| choppaface wrote:
| Triplebyte and Loopt both ended up selling / monetizing user
| data in ways the users really didn't like.
| jackblemming wrote:
| Are you kidding? He had one startup that was more or less a
| flop, then for some reason was appointed to a high position at
| y combinator, got lucky allocating capital (plenty of idiots
| can and have gotten lucky or were at the right place at the
| right time) and now is the CEO of OpenAI. This man is the
| definition of it's not what you know, it's who you know, and
| that's not a good thing.
| sacnoradhq wrote:
| Even with the best ideas, execution, and teams 99.99% of
| startups are. That's okay. They're assumed to be experiments.
|
| There is no such thing as self-made. And there's nothing
| wrong with friends and networking, especially as some
| particular help or chance encounter could be pivotal to
| nudging onto something great.
|
| It's the trying and learning that are the gold to try again.
| Timing, honest perspective, persistence, and a measure of
| prepared luck seem to be more of it. There is no magic
| formula. I wish success to all who want it.
| boeingUH60 wrote:
| Altman befriended Paul Graham, and his life blossomed...
| w3454 wrote:
| [dead]
| schiffern wrote:
| Looks like the media has chosen Sam Altman as the next Elon Musk.
|
| This makes sense. He perfectly fits their cliche of a socially-
| awkward technologist, and he's trusting (foolish?) enough to make
| complex nuanced statements in public, which they can easily mine
| for out-of-context clickbait and vilification fodder.
| eternalban wrote:
| 2014: https://www.ycombinator.com/blog/sam-altman-for-
| president/
|
| 2015: https://www.nytimes.com/2015/07/12/opinion/sunday/sam-
| altman...
|
| 2016: https://www.newyorker.com/magazine/2016/10/10/sam-
| altmans-ma...
|
| ..
|
| That last link, _Sam Altman 's Manifest Destiny_, is worth the
| read. However the last time I posted that link HN went down for
| an hour right afterwards. (of course correlation is not
| causation :/)
|
| https://news.ycombinator.com/item?id=35334023
| greyman wrote:
| What would make him "socially-awkward"?
| tedunangst wrote:
| Never forget the time he wore sneakers to the Ritz.
| schiffern wrote:
| Sam or Elon?
|
| I'll assume you meant Sam. IMO Sam is mostly just shy and
| cerebral, but to many people that will come off as awkward
| and robotic.
|
| Watch his recent Lex Fridman interview. Personally I thought
| it was great, but I'm aware enough to realize that (sadly)
| many low-knowledge people will judge such demeanor harshly.
|
| Mark my words: the media will, 10 times out of 10, _exploit
| that misconception_ , not correct it. "Ye knew my nature when
| you let me climb upon your back..."
| lubesGordi wrote:
| I don't know, it seemed to me his responses on Lex were
| very measured and carefully restrained in a lot of places,
| calculated and vague in others. He doesn't come off as
| genuine to me at all.
| drewcoo wrote:
| If you mean the one where the interviewer, Lex, was wearing
| a suit and Sam was in a hoodie, where Sam droned in a
| robotic monotone and often sat with crossed arms, staring
| downward . . . I think the knowledgeable people might also
| assume he's the next evil tech overlord. Or certainly
| distant and uncaring.
|
| The only things missing were lighting from below and scenes
| of robots driving human slaves.
| [deleted]
| Kkoala wrote:
| It's interesting that in the article he was described being
| "hyper-social", "very funny" and "big personality" as a
| child. I guess those don't necessarily contradict with
| awkward and robotic, but also wouldn't come to my mind at
| the same time
| andsoitis wrote:
| You'll get some clues when you watch his recent interview
| with Alex Fridman https://youtu.be/L_Guz73e6fw
| seydor wrote:
| No, the next Zuckerberg. The media sees (rightly) openai as a
| competitor medium.
|
| Although he s much more prepared to face the next Greta
| (Yudowksi)
|
| He has to fix his vocal fry however, it is annoying
| [deleted]
| phillryu wrote:
| When I compare the two Elon was (lucky?) to at least have a
| string of vision-fueled ventures that became a thing. What is
| Sam's history of visions? Loopt? Is Y Combinator considered in
| a new golden era after he took over? Did Worldcoin make any
| sense at all?
|
| I'm honestly hoping I'm entirely ignorant of his substance and
| would feel better if someone here can explain there's more to
| him than that... I would feel better knowing that what could be
| history's most disruptive tech is being led by someone with
| some vision for it, beyond the apocalypse that he described in
| 2016 that he tries not to think about too much:
|
| "The other most popular scenarios would be A.I. that attacks us
| and nations fighting with nukes over scarce resources." The
| Shypmates looked grave. "I try not to think about it too much,"
| Altman said. "But I have guns, gold, potassium iodide,
| antibiotics, batteries, water, gas masks from the Israeli
| Defense Force, and a big patch of land in Big Sur I can fly
| to." https://www.newyorker.com/magazine/2016/10/10/sam-altmans-
| ma...
| sixQuarks wrote:
| Are you really insinuating that Elon was simply "lucky" when
| it came to disrupting and transforming two gargantuan and
| highly complex industries at the same time?
| schiffern wrote:
| I'm not talking about the _reality_ of Sam and Elon. I 'm
| putting my ear to the ground and observing the way the media
| is (and will) portray them.
|
| I wish that "actual reality" was all that mattered and not
| such low-knowledge "optics", but sadly we don't live in that
| world.
| wetmore wrote:
| I'm with you, listening to his interview with Ezra Klein gave
| me the impression that he doesn't actually think that deeply
| about the possible impact of AI. He says it worries him, but
| then he seems to wave those worries away with really
| simplistic solutions that don't seem very tenable.
| xiphias2 wrote:
| The main question about OpenAI is this: can you have any
| better structure to create singularity that will happen
| anyways (Some people don't like the word AGI, so I just
| definine it by machines having wastly more intellectual
| power than humans).
|
| Would it be better if Google, Tesla or Microsoft / Apple /
| CCP or any other for profit company did it?
| davidivadavid wrote:
| What bothers me most is that the picture he paints of
| _success_ itself is some handwavy crap about how it could
| "create value" or "solve problems" or some other type of
| abstract nonsense. He has displayed exactly 0 concrete,
| desirable vision of what succeeding with AI would look
| like.
|
| That seems to be the curse of Silicon Valley, worshiping
| abstractions to the point of nonsense. He would probably
| say that with AGI, we can make people immortal, infinitely
| intelligent, and so on. These are just potentialities with,
| again, 0 concrete vision. What would we use that power for?
| Altman has no idea.
|
| At least Musk has some amount of storytelling about making
| humanity multiplanetary you may or may not buy into. AI
| "visionaries" seem to have 0 narrative except rehashed,
| high-level summaries of sci-fi novels. Is that it?
| efficientsticks wrote:
| I agree, listening to the podcast I think the answer is
| that "yes" that is it: faith in technological progress is
| the axiom and the conclusion. Joined by other key
| concepts like compound growth, the thinking isn't deep
| and the rest is execution. Treatment of the concept of
| 'a-self' in the podcast was basically just nihilistic
| weak sauce.
| jasmer wrote:
| AI is not an abstraction. It's rational to be hand wavy
| about future value, it's already materialized. AI is
| basically an applied reseaarch project, he should be more
| like a Dean herding researchers and we should take him as
| that. In a previous era, that's what it would be: a PhD
| from Berkley in charge of some giant AT&T government
| funded research Lab thing. He'd be on TV with a suit and
| tie, they'd be smoking and discussing abstract ideas.
| f38zf5vdt wrote:
| [flagged]
| ninth_ant wrote:
| What specifically in that article was vilification of Sam or
| clickbait, or statements taken out of context?
| schiffern wrote:
| In these early days of a smear campaign (even an
| unintentional one that's just about chasing clicks), the game
| is mostly about plausibly deniable innuendo.
|
| The headline is a great start. Contradictions are bad. Altman
| has contradictions. Therefore Altman is bad. They don't say
| it, but they also know _they don 't need to_. They lead the
| audience to water and trust that enough of them will drink.
|
| The closing paragraph is another great example. It
| intentionally leaves the reader hanging on the question "so
| why did Altman do AI if there are moral downsides," without
| resolving the question by giving Altman's context when he
| said it.
|
| Trust me or don't, but what you see here is just the
| beginning. In 6 month's time Altman will be (in the public's
| eye) evil incarnate.
| ninth_ant wrote:
| They discussed the why earlier in the article, specifically
| a fear of AI being primarily developed in private labs
| outside of public view -- the partners feeling they could
| help bring an alternative approach.
|
| I feel they left it on that point not as part of some grand
| conspiracy theory, but because the potential for this to be
| good or bad is a question taking place around the world
| right now.
|
| Overall this piece feels positive towards Sam, despite what
| you feel is a negatively loaded headline. He's walking a
| delicate balance between profit and nonprofit, between
| something that could be harmful or helpful to society --
| these things are in contradiction and he's making those
| choices deliberately. This is an interesting subject for an
| article.
|
| I find it deeply unlikely he will be viewed like Musk in 6
| months. Musk is a fairly special case as he's unhinged and
| unstable more than evil. If someone wanted to paint Sam
| with an evil stick, Zuckerberg would be a more apt
| comparison -- playing with something dangerous that affects
| all of us.
| schiffern wrote:
| I genuinely hope that you're right and I'm totally wrong,
| but my experience watching the media landscape says
| otherwise. It would seem I have less faith in our
| journalistic institutions than you.
|
| The media operates on a "nearest cliche" algorithm, and
| the Mad/Evil Genius cliche is so emotionally appealing
| here that they'll find it irresistible. Even if it's not
| true, _they 'll make it true._
|
| Don't say I didn't warn you. :)
| okareaman wrote:
| This is a very fluid and chaotic situation so I'd be more
| concerned if he said one thing and stuck to it
|
| _When the Facts Change, I Change My Mind. What Do You Do, Sir?_
| - John Maynard Keynes
|
| _A foolish consistency is the hobgoblin of little minds, adored
| by little statesmen and philosophers and divines. With
| consistency a great soul has simply nothing to do. He may as well
| concern himself with his shadow on the wall_ - Ralph Waldo
| Emerson
|
| _Do I contradict myself? / Very well then, I contradict myself.
| / (I am large, I contain multitudes)_ - Walt Whitman's "Song of
| Myself"
| balls187 wrote:
| When will we get the "Contradictions of Dang?"
| roflyear wrote:
| I hope dang gets a small piece of the YC pie
| maxbond wrote:
| "The Lonely Work of Moderating Hacker News"
|
| https://www.newyorker.com/news/letter-from-silicon-valley/th...
| gamesbrainiac wrote:
| I think there are some interesting questions:
|
| - Sam does not have equity in OpenAI. Does this mean he can
| potentially be removed at any point in time?
|
| - OpenAI's profit arm will funnel excess profit to its non-profit
| wing. If this is the case, who determines excess profit?
|
| - OpenAI's founding charter commits the company to abandoning
| research efforts if another project nears AGI development. If
| this happens, what happens to the profit arm?
| whitepoplar wrote:
| I refuse to believe that Sam doesn't have equity in OpenAI. It
| _must_ be some 4D-chess-style ownership structure, which I 'm
| guessing is for tax avoidance.
| brookst wrote:
| There's plenty of evidence that he has no equity. I'd love to
| see contradictory evidence, but without that, just refusing
| to believe things based on intuition isn't great.
| ryanSrich wrote:
| Why is that not great? It makes absolutely zero sense for
| him to have no equity, or at least some agreement in place
| that equity is coming. Or some other terms that essentially
| amount to equity. You don't need evidence to be skeptical
| of the situation.
| peripitea wrote:
| Why does it make zero sense? It makes perfect sense to
| me.
| brookst wrote:
| He was wealthy before and has other means to parlay
| openai to further wealth.
|
| You're doing the "only the true messiah would deny his
| divinity" argument -- if he was going to profit, that's
| bad. If he's not going to profit, obviously he's lying
| and is going to profit, so that's bad.
|
| IMO arguments are only meaningful if they can be
| falsified. Your argument can't be falsified because
| you're using a lack of evidence as proof.
| graeme wrote:
| With no equity comes no control. I would find it very
| surprising he has no control over the project.
|
| And if he does have control that has value whether you
| label it equity or not.
|
| It is possible he literally has no control and no financial
| upside but who would turn down control over what they
| believed to be a world shaping technology?
| peripitea wrote:
| The parent is a non-profit, hence no equity is required
| to have control. Seems pretty straightforward to me.
| somsak2 wrote:
| I mean, most non-founder company CEOs don't have a
| significant % of total equity and they still have control
| over the company.
| [deleted]
| xiphias2 wrote:
| It's nothing special, there's a company under the foundation,
| he doesn't have share in the company, he's ceo and board
| member of the foundation.
|
| It's just this one non-important detail is now being repeated
| over and over.
| boeingUH60 wrote:
| > It must be some 4D-chess-style ownership structure, which
| I'm guessing is for tax avoidance.
|
| How would this even work? If only I got a dollar for someone
| suggesting that there are Magic ways to avoid tax..."they
| just write it off!"
| namaria wrote:
| Magic is just a word for things we don't understand. As a
| poor wage slave sap, I'm 100% sure the world is run by
| magic guilds, i.e. a bunch of powerful people conspiring
| stuff I could never fathom. Whatever gets to my eyes and
| ears has been approved for public disclosure. I don't know
| shit, everything is magic to me. I kinda know how to
| survive. So far.
| fauigerzigerk wrote:
| _> OpenAI's founding charter commits the company to abandoning
| research efforts if another project nears AGI development. If
| this happens, what happens to the profit arm?_
|
| I think the definition of AGI is sufficiently vague that this
| will never happen. And if it did happen, abandoning research
| efforts could take the form of selling the for-profit arm to
| Microsoft.
| gamesbrainiac wrote:
| I think you have a point there. AGI doesn't have a straight-
| forward litmus test.
| ren_engineer wrote:
| he admits in the article to having equity via his investment
| fund, they are using semantics because he doesn't "personally"
| have equity. He also tries to downplay by saying it's an
| "immaterial" amount, but in reality that could be billions of
| dollars.
|
| There's also nothing preventing Microsoft from gifting him
| billions in Microsoft stock so they can claim he's not
| motivated by profit with OpenAI despite indirectly making money
| off it.
|
| You'd have to be extremely naive to look at the decisions he's
| made at OpenAI and think it was all purely out of good will.
| Google and AWS both offer credits for academic and charity
| projects, why did Altman choose to go all in with Microsoft if
| it wasn't for money?
| peripitea wrote:
| Do you honestly think that AWS or Google Cloud would have
| given them billions in credits just because they're a
| nonprofit? I'm all for being skeptical of powerful people's
| motives but that suggests a major disconnect from reality
| somewhere in your thinking.
| davidgerard wrote:
| didn't know Thiel had helped start this here Y Combinator
| dang wrote:
| He didn't. That sentence was a bit confusingly written but "co-
| founded" binds only to the second name.
| mimd wrote:
| He's begging for regulatory capture. "I can destroy the world but
| I won't. My competitor's will, so regulate them." A shrewd plan
| considering he's not offering something beyond what another
| company with a large nvidia cluster could offer.
| [deleted]
| dmix wrote:
| The odds that is the end game of the AI ethics movement is
| pretty likely: A mega-monopoly AI firm with a wall of gov
| policy that will cripple any upstart who can't jump through
| "safety" hoops written for and by the parent company. So any
| talented dev who wants to do great AI work either has to work
| for parent company or build a startup designed to get acquired
| by them (aka don't rattle the cage).
| sebzim4500 wrote:
| I think that's the goal but there is a reasonable chance that
| they completely fail and no serious regulation ever gets
| passed.
|
| That's the thing with technology, people get used to it and
| then trying to ban/control it makes you look ridiculous. It's
| like how now that Tesla has made it normal to have driving
| assistance (and calling it FSD) there is little appetite
| outside of contrarian circles for serious regulation. If,
| however, regulation was proposed before Tesla shipped then it
| might have passed.
| sebzim4500 wrote:
| >A shrewd plan considering he's not offering something beyond
| what another company with a large nvidia cluster could offer.
|
| It's been 4 months, no one has released anything nearly as good
| as the initial release of ChatGPT. Meanwhile OpenAI has
| released GPT-4 and is trialing plugins and 32k context.
|
| Either their competition is incompetent or OpenAI is doing
| something right.
| zmmmmm wrote:
| > A shrewd plan considering he's not offering something beyond
| what another company with a large nvidia cluster could offer
|
| So why does Bard seem inferior to GPT-4?
| jamaliki wrote:
| It'll get better fast.
| skybrian wrote:
| All that may be true, but it doesn't help us decide whether
| more AI regulation is a good idea or not.
|
| As with most things, it probably depends on how it's done.
| mimd wrote:
| That's a difficult question. I'm just pointing out that Sam's
| "contributions" are unhelpful to solving that question.
|
| You're limited by two prisoners and separate, sometimes
| antagonistic, countries. For example, we have little
| agreement on nuclear weapons, at best we've gotten
| concessions on testing, a few types of missiles, and so on.
| Same with current climate legislation. So getting global
| agreement is hard outside of the bare minimum. Most of the
| inner county approaches seem to be either panic, political
| distraction, or like Sam, regulatory capture, as they are
| ignoring that it means nothing if another country pursues it.
|
| So I'd focus on what simple agreements we could get
| worldwide.
| jasonhansel wrote:
| An excellent case of doing the right thing for the wrong
| reasons.
| breck wrote:
| I love what OpenAI has built and it's awesome to see them succeed
| on the AI front (also I'm forever amazed to see SamA live up to
| the Lebron James early level hype about his entrepreneurial
| skills).
|
| This article sheds more light on how the non-profit front failed.
| I find that to be a very hard and interesting problem. IMO, it
| points to a larger problem with our current laws, where trying to
| do good and compete fairly is made much harder (near impossible?)
| when you compete against companies that exploit unfair
| monopolistic laws.
| mellosouls wrote:
| There seems a lot of negative perception of the guy here, and
| OpenAI definitely deserve criticism for some stuff (so as the
| CEO, so does he), but - even if it was built on the work of
| others, and with the obvious caution about what may come next -
| he and they deserve immense respect and credit for bringing in
| this new AI age.
|
| They did it. Nobody else.
| [deleted]
| BarryMilo wrote:
| The expression "on the shoulders of giants" has never been so
| relevant.
| mellosouls wrote:
| Ha! To be fair though, Isaac Newton is not a bad person to be
| implicitly compared to. :)
| anileated wrote:
| I for one have nothing against Sam as a person (not knowing him
| well enough), but I question the sentiment that he and the
| company deserve respect for what they're doing--much less _by
| default_ , for some self-evident reason that doesn't even
| require explanation.
|
| Do people mean it in a sarcastic sense--and if not, why does
| OpenAI deserve respect again?
|
| -- Because it is non-trivial (in the same way, say, even Lenin
| deserves respect by default--even if the outcome has been
| disastrous, the person sure had some determination and done
| humongous work)?
|
| -- Because this particular tech is somehow inherently a good
| thing? (Why?)
|
| -- Because they rolled it out in a very ethical way with utmost
| consideration for the original authors (at least those still
| living), respecting their authorship rights and giving them the
| ability to opt in or at least opt out?
|
| -- Because they are the ones who happen to have 10 billions of
| Microsoft money to play with?
|
| -- Because they don't try to monetize a brave new world in
| which humans are verified based on inalienable personal traits
| like iris scans, which they themselves are bringing about[0]?
|
| This is me stating why they shouldn't have respect _by default_
| and counting to get a constructive counter-argument in return.
|
| [0] https://news.ycombinator.com/item?id=35398829
| cvalka wrote:
| >Because this tech is somehow >inherently a good thing
| Without technology humans are just unworthy bugs.
| anileated wrote:
| It is generally accepted that some applications of
| technology are good and some are not, or at least not self-
| evidently so (weapons of mass destruction, environmentally
| disastrous things like PFAS, packaging every single product
| into barely-recyclable-once plastics, gene editing humans,
| addictive social media/FB/TikTok, etc.)
|
| Is _this_ particular application of technology good, and
| even self-evidently so?
| nice_byte wrote:
| > he and they deserve immense respect and credit for bringing
| in this new AI age.
|
| Why? Who asked for it? I think that if openAI's breakthroughs
| never happened, we would not be any worse off (actually, we'd
| probably be better off).
| nadermx wrote:
| Meh, the contradiction seems to be that creating a source of
| power, be it via a physical or virtual means is different. It is
| not. A tool, is and always will be, a tool.
___________________________________________________________________
(page generated 2023-04-01 23:00 UTC) |