
Hacking: The Art of Exploitation, 2nd Edition
Jon Erickson
4.7 on Amazon
19 HN comments

Bitcoin: Hard Money You Can't F*ck With: Why Bitcoin Will Be the Next Global Reserve Currency
Jason A. Williams and Jessica Walker
4.8 on Amazon
19 HN comments

Grokking Algorithms: An Illustrated Guide for Programmers and Other Curious People
Aditya Bhargava
4.6 on Amazon
18 HN comments

The Effective Engineer: How to Leverage Your Efforts In Software Engineering to Make a Disproportionate and Meaningful Impact
Edmond Lau and Bret Taylor
4.5 on Amazon
18 HN comments

About Face: The Essentials of Interaction Design
Alan Cooper , Robert Reimann , et al.
4.5 on Amazon
18 HN comments

The Web Application Hacker's Handbook: Finding and Exploiting Security Flaws
Dafydd Stuttard and Marcus Pinto
4.6 on Amazon
17 HN comments

The Art of Game Design: A Book of Lenses, Third Edition
Jesse Schell
4.7 on Amazon
17 HN comments

Think Bayes: Bayesian Statistics in Python
Allen B. Downey
? on Amazon
15 HN comments

Mastering Bitcoin: Programming the Open Blockchain
Andreas M. Antonopoulos
4.7 on Amazon
15 HN comments

Working in Public: The Making and Maintenance of Open Source Software
Nadia Eghbal
4.6 on Amazon
15 HN comments

Rocket Surgery Made Easy: The Do-It-Yourself Guide to Finding and Fixing Usability Problems
Steve Krug
4.5 on Amazon
14 HN comments

Software Engineering
Ian Sommerville
4.3 on Amazon
14 HN comments

The Making of Prince of Persia: Journals 1985-1993--Illustrated Edition
Jordan Mechner
4.8 on Amazon
13 HN comments

Python Machine Learning: Machine Learning and Deep Learning with Python, scikit-learn, and TensorFlow 2, 3rd Edition
Sebastian Raschka and Vahid Mirjalili
4.5 on Amazon
12 HN comments

Life 3.0: Being Human in the Age of Artificial Intelligence
Max Tegmark, Rob Shapiro, et al.
4.5 on Amazon
12 HN comments
agi_prometheusonApr 22, 2021
If you have a science background then read
1) Good to Great
2) Life 3.0
If you don't have a science background
1) Future of Capitalism.
2) Good to Great
"Good to great" is a fantastic book for every entrepreneur out there.
Go! read them right now.
_sy_onDec 7, 2017
arkanoonSep 22, 2019
__sy__onMay 12, 2020
dtujmeronOct 7, 2018
The "true" AI ethical question is related to ensuring that the team that develops the AI is aware of AI alignment efforts and has a "security mindset" (meaning: don't just try stuff and repair the damage if something happens - ensure in advance, with mathematical proof, that a damaging thing won't happen). This is important because in a catastrophic superintelligent AI scenario, the damage is irreparable (e.g. all humanity dies in 12 hours).
For a good intro to these topics, Life 3.0 by Max Tegmark is a good resource. Superintelligence by Nick Bostrom as well.
For a shorter read, see this blog post: https://waitbutwhy.com/2015/01/artificial-intelligence-revol...
For general information about AI ethical principles, see FHI's website, they have publications there that you could also read: https://www.fhi.ox.ac.uk/governance-ai-program/
gfodoronMay 10, 2018
My point was that a suboptimal solution is one where there was no careful, methodical thought to the design of AI systems. If AI designers and researchers blindly react to public outcry after 24 hours from a demo (which, in general, is something I would expect Google to do) the kind of thinking you mention above is just as unlikely to happen as if they trudged on forward without any consideration to these things at all. In both cases, this is a suboptimal outcome for society.
In one, we get a fairly random, undesigned world of AI systems that don't serve anyone well and generally are underutilized because nobody is willing to push the boundaries. In the other, we get dystopian AI hell. It's important that researchers be given the space to think you describe. The right way to allow that is to foster an informed and open-minded public about the emergence of AI and the decisions we need to make about it as a society, and not have prominent voices writing about knee-jerk reactions making authoritative demands to a specific public demo after just a day. (See Max Tegmark's book Life 3.0 for the kinds of stuff we need more of being put into the world imho.)
klenwellonMay 31, 2018
https://www.newyorker.com/magazine/2018/05/14/how-frightened...
China's social credit system is glossed in the article.
Doesn't seem like there are a lot of good outcomes where AI is involved. A passage near the end of the article:
In the meantime, we need a Plan B. Bostrom’s [author of book Superintelligence] starts with an effort to slow the race to create an A.G.I. [Artificial General Intelligence] in order to allow more time for precautionary trouble-shooting. Astoundingly, however, he advises that, once the A.G.I. arrives, we give it the utmost possible deference. Not only should we listen to the machine; we should ask it to figure out what we want. The misalignment-of-goals problem would seem to make that extremely risky, but Bostrom believes that trying to negotiate the terms of our surrender is better than the alternative, which is relying on ourselves, “foolish, ignorant, and narrow-minded that we are.” Tegmark [author of book Life 3.0: Being Human in the Age of Artificial Intelligence] also concludes that we should inch toward an A.G.I. It’s the only way to extend meaning in the universe that gave life to us: “Without technology, our human extinction is imminent in the cosmic context of tens of billions of years, rendering the entire drama of life in our Universe merely a brief and transient flash of beauty.” We are the analog prelude to the digital main event.
Takes the idea of moving fast and breaking things to the next level.
spaceknarfonNov 12, 2018
matthalvorsononMar 16, 2021
milansmonJuly 2, 2019
> When they launched, Prometheus was slightly worse than them at programming AI systems, but made up for this by being vastly faster, spending the equivalent of thousands of person-years chugging away at the problem while they chugged a Red Bull. By 10 a.m., it had completed the first redesign of itself, v2.0, which was slightly better but still subhuman. By the time Prometheus 5.0 launched at 2 p.m., however, the Omegas were awestruck: it had blown their performance benchmarks out of the water, and the rate of progress seemed to be accelerating. By nightfall, they decided to deploy Prometheus 10.0 to start phase 2 of their plan: making money.
marrowgarionDec 12, 2018
Leonardo da Vinci by Walter Issacson - fascinating look into the real life of Leonardo, demystifying the genius
Excession by Iain M Banks - a bit of a let down
Bluets by Maggie Nelson - lyrical and philosophical and explicit ruminations on the color blue
How to Change Your Mind by Michael Pollan - a lot of already known and rehashed info on psychedelics
Lost City of the Incas by Hirham Bingham - Yale professor who discovered Machu Pichu. Good history of the Incas and region
Farenheit 451 by Ray Bradbuy - Classic!
2041 by Kim Stanley Robinson - NYC underwater in the future. A bit of a let down compared to his Mars series
Shiver by Junji Ito - short stories from the king of Japanese horror manga
Lenin: The man, the dictator, the master of terror by Victor Sebestyen - great bio on Vladimir Lenin. Knew very little about him before reading this. Fantastic!
Deep Learning by Ian Goodfellow, Yoshua Bengio, and Aaron Courville - definitive text book on Deep Learning
The Curse of Bigness by Tim Wu - interesting read into the history of Antitrust and the Sherman Act and how they relate to modern tech giants like Amazon, Google, Facebook
Connecting the Dots by John Chambers - a bit dry. Lessons Chambers learned while CEO of Cisco
amasadonOct 6, 2017
When he got involved in the AI Risk community I thought it might be good thing that an actual scientist is involved, maybe to ground the community's heavy speculation in scientific thinking. However, what happened was exactly the opposite -- Max turned into a fiction author! (ergo this piece). Now, of course there is a role for fiction in expanding our understanding of the future but the AI Risk community is already heavily fictionalized. The singularity, intelligence explosion, mind uploads, simulations, etc are nothing but idle prophecies.
Karl Popper, the famous philosopher of science, made a distinction between scientific predictions which usually takes the form "If X then Y will happen" and scientific prophecies which usually takes the form "Y will happen" which is exactly what Max and the rest of the AI Risk community is involved in.
Now back to Max's San Francisco talk, I actually asked him this question: "Who is doing the hard scientific work around AI Risk?" and after a long pause he said (abridged): "I don't think there is hard scientific work to be done but that doesn't mean that we shouldn't think about it. We're trying to predict the future and if you told me that my house will burn down then of course I'll go look into it".
This doesn't inspire much confidence in the AI Risk community, where scientists need to leave their tools at the door to enter The Fantastic World of AI Risk and where fact and fiction interweave liberally -- or as Douglas Hofstadter put it when describing the singularitarians: "a lot of very good food and some dog excrements".