text
stringlengths
1
39.9k
label
stringlengths
4
23
dataType
stringclasses
2 values
communityName
stringlengths
4
23
datetime
stringdate
2014-06-06 00:00:00
2025-05-21 00:00:00
username_encoded
stringlengths
136
160
url_encoded
stringlengths
220
528
This is the best tl;dr I could make, [original](https://motherboard.vice.com/en_us/article/j5gedd/study-that-claimed-ai-could-determine-a-persons-sexuality-is-under-ethical-review) reduced by 83%. (I'm a bot) ***** > Academics, independent experts, and LGBTQ activists from organizations like GLAAD and the Human Rights Campaign argue that not only is the study's methodology flawed, but that it raises serious ethical concerns. > "But even if we accepted the paper's premise that someone can appear visually queer then the paper still has major ethical issues around participant consent and the overall aim of the research," she explained, referring to how the photos used were pulled from a publicly available dating site instead of gathered with the participants' consent. > In a blog post, Greggor Mattson, an associate professor of sociology at Oberlin College, argues that the paper is only the "Most recent example of discredited studies attempting to determine the truth of sexual orientation in the body." ***** [**Extended Summary**](http://np.reddit.com/r/autotldr/comments/6zs9en/study_that_claimed_ai_could_determine_a_persons/) | [FAQ](http://np.reddit.com/r/autotldr/comments/31b9fm/faq_autotldr_bot/ "Version 1.65, ~208856 tl;drs so far.") | [Feedback](http://np.reddit.com/message/compose?to=%23autotldr "PM's and comments are monitored, constructive feedback is welcome.") | *Top* *keywords*: **paper**^#1 **study**^#2 **Human**^#3 **Kosinski**^#4 **right**^#5
r/aiethics
comment
r/AIethics
2017-09-13
Z0FBQUFBQm9IVGJBa0piaTl1UTJBa1o0SDI1QThVRUhqWnpfRUQzXzFyUk5EaEY2S0tKbkJlVXprVE5RMnBOU1Q2dFpfYjBjYjlpUnJqbTdEVWNISnJacERYTWFDTTgycGc9PQ==
Z0FBQUFBQm9IVGJCV29VN29IZ3YzMFE3MGc4bmlVTVVXelNDUm5hNm1uS1VVbjZpVVNlZFhXNDk3QTgxYm5yYzNTeFdnMkdwTVQ3Ynh5NGd6UW8wWWRmNUxEcGxzUmpxd2ZoN3I2N1RWVXl0RG5FSVB5N1R3UUxxUk1YOHhvX28xc2FSa2YzaEVZdG15cnM4SXpWZy1YNkNRT19QTWhSSU0tU2g5anVOWEtCemhQY3FOd2laVGVBT3RwdHVyS3RvLWoxQ1lNd29GX0UyU2xoVEhudXY1WER1Z1NhSngxUE9WUT09
>she explained, referring to how the photos used were pulled from a publicly available dating site instead of gathered with the participants' consent. It was online rape! /s
r/aiethics
comment
r/AIethics
2017-09-13
Z0FBQUFBQm9IVGJBZkY5Q2xLVmRqRmFVV1RkUFhkVjFNVlFNWlBxR21vRUQ4SE03VzhpbWREOFpYczBXRC1WTlRIMXJETmlHdGNxQkIzV2JMZl9ObTYwUWxxSlNzNkJEUVE9PQ==
Z0FBQUFBQm9IVGJCcnI0czZWM0RYWnRDeHpUX3Nrelczd3RtckU0aTM2VTVrVHVveDBZM1lYdEZnNndZTXdkVzg3WDRxUWtLUGdOdWtLeVVjZ1Z5a0hwZ1IybkJteXM3N0JoTDBLQnpJMnEwc3ZtUnB2NU95TXZrbmhzbU1WTGE5VGpLNUpQa3BuUUdoNkdXR1NyWUhDVXVOQTBoTEZLSFlDdUZVb2ZEN1ZXV3p5bVg3UkRsM2k4Z0tzZGFYSjNmVFBKMEJEWWJvQXNnUE5BOGZIRnRVaDVxbkhfZWZFZjB0Zz09
>It also ignored cultural and societal factors that contribute to human sexuality. I get it. This is related to the debate between whether people are "born this way" or if their socio-cultural environment "constructs" their sexual orientation. It seems a bit odd because the social constructionist perspective is the one favored in Russia where the state media propounds the view that "homosexual propaganda" can turn children gay.
r/aiethics
comment
r/AIethics
2017-09-13
Z0FBQUFBQm9IVGJBdWRzMUhidF9VTFc1RUJNRjZYRktBUHpPcDVTcWU3YmotdDdsdHNMSWxCUUtyd3d6N3ZXbmo4eU9nTzRmeG9KMkliOWxYblYwLUZYTGh3Ulh6cENPOGc9PQ==
Z0FBQUFBQm9IVGJCWkNUYjkzSXJEU2FXQmVlZmR1eWNzTHNTcE1SMEN5SnVzdHh1NFlnMExiVWhkYzJWSm1rTFp4XzlnY1pTdjVtWWlGNl9HeEJ4NEdjeUYtdzFyVExTZEd3ZWpDbXFKbGNtR0wzb0VlMUYwajRwZzBHOEtEcHhRZzFTcVVwNkJuUmgybm02a3hrdWFkZWpaR0JpNGNNUS1mbmNpNXNjZEhMbG5BM3Y0NGp4cmJ5dk5MOXJuSUx2TllldW5Qc1RYNnVENzFyVmZ2cmNZcU9BY2V6alZlRGFmUT09
What's Russia got to do with this? GLAAD/HRC are US-based.
r/aiethics
comment
r/AIethics
2017-09-13
Z0FBQUFBQm9IVGJBbUxvbE1xRzIwVUlCd3BqcmdydklvVk5kdWdhV2poNHMyUFRxUnhZdW8tbmMxT05aRkhoaldOU1hITWVHbkhNcGtrV01mZE56aUQ5VzlWUXh4aUg1QWc9PQ==
Z0FBQUFBQm9IVGJCMXEzYkNfQjdBYWVKZ0lKUGxERG1TQUdoeUhEc2JWNDZJT0x4cV9tRGpSVEpVZkZQblVqVjRhV2hmQkRMSENFVGVhUW16NXJFOHNZZlU3VVEtbTJ3ZXg1WlhMcnA2VlJxdVBoX2RabE9HNlNNVEZlTGZTMTdOUjVxMFY0bzhDZnA3c21RdWpJTjdiNmZ5TlRvcXVNQzlydWN0aDJCdHl1LTBVZHZiMEFmSmxVMW42aHhjVmNXakJfbkNodmZYTlZEQktLOUdUUUZ5VlVYYTlRMmNJZVBLUT09
That's the surprising part. A US-based gay rights advocacy holds the same view as Russia, which is notorious for its human rights infringements. Can someone explain the "fluidity" of sexual orientation with citations to trusted studies?
r/aiethics
comment
r/AIethics
2017-09-13
Z0FBQUFBQm9IVGJBckNHSVVKaE5PUUpFSjZkck9kNnFLYU1pVksyNlJlQzhqa2ktY2V5VHI3aDhpaFhLa3U1T1gwbnpxaEJVcnNjRFJCSk9YY2JGR2h6Mm5UaHhvRWpQaWc9PQ==
Z0FBQUFBQm9IVGJCTUFMS29wVW40NFlyYjlDRFZvblZzanU2bVkxVEVSUkgyZ3ZUTE05MU1JQ1IyQXQtUV96cFdzbEo2bVlfZ05tQjFmaWtiYkx0dnVXSTJ5eVMxTE9femg2Qkh0aVVYQmpzU3EzYkRmR2R5WllQWVhvOUdXbXBLZjhadlViZGJydHlTaS1mNW5INlNDSWhVX2VLcEFPU0hYY21aSEx6VVdONUZVZ003a1hYVzE4TnpiWS04OWhDMktCbV9XQzYtLVF1c09rODVWaFhuZEZPRmdaVUJTWlVhUT09
Ah, I see what you mean. I'm pretty sure that environmental determination of sexual orientation doesn't imply social constructionism though. You'd be a social constructionist if you thought that gayness is not even a real biological phenomenon, or something like that. This is just *Blank Slate*-ism. Of course there is still the ironic similarity. I haven't read the science here, but the Wikipedia article (https://en.wikipedia.org/wiki/Biology_and_sexual_orientation) seems to be comprehensive and unbiased. It says it's not clear and seems like a combination of environmental and genetic factors.
r/aiethics
comment
r/AIethics
2017-09-13
Z0FBQUFBQm9IVGJBdnhGRVRNMUgyYy13VXZIeG1uOWFJbm1GakRaWXdYaU1JZ0dTdFpTT1FSSEFxYTcwcWY3b25rTmVPS2JYd212N0Z5Uk43WHo2UzA0cHpuc05YVnkyeFE9PQ==
Z0FBQUFBQm9IVGJCaEhQendfWlVBbWR2UUFEakRBR2F3Q3NaZWpzRndzVDhReEhjQS1mR0NOdnhaU2s2LTI5MDg5VmZwTVBDWGlneDRxSlBjLUUtMFRseEU3Zk1BMUNlTzRoOE44T0VEbndubWozR0hLRnBlQzMwc09LaWlkYkl4WXZUWmswNHJ2MFc2MGtpN3ZLNTF1eFNDTlVYTlRzSzEtOWRyMmNQWThNYTlRc3BEUDV2NnpLTGU1TjRJbzFxVUdtSXppR1dZZFNsckdQWkRiRnAtZmNLWDg5M1lrLXAyZz09
**Biology and sexual orientation** The relationship between biology and sexual orientation is a subject of research. A simple and singular determinant for sexual orientation has not been conclusively demonstrated; various studies point to different, even conflicting positions, but scientists hypothesize that a combination of genetic, hormonal, and social factors determine sexual orientation. Biological theories for explaining the causes of sexual orientation are favored by experts and involve a complex interplay of genetic factors, the early uterine environment and brain structure. These factors, which may be related to the development of a heterosexual, homosexual, bisexual or asexual orientation, include genes, prenatal hormones, and brain structure. *** ^[ [^PM](https://www.reddit.com/message/compose?to=kittens_from_space) ^| [^Exclude ^me](https://reddit.com/message/compose?to=WikiTextBot&message=Excludeme&subject=Excludeme) ^| [^Exclude ^from ^subreddit](https://np.reddit.com/r/AIethics/about/banned) ^| [^FAQ ^/ ^Information](https://np.reddit.com/r/WikiTextBot/wiki/index) ^| [^Source](https://github.com/kittenswolf/WikiTextBot) ^] ^Downvote ^to ^remove ^| ^v0.27
r/aiethics
comment
r/AIethics
2017-09-13
Z0FBQUFBQm9IVGJBdW1QWnZtSG1QcWZUdk85N1lQX3VGRmR6d1ByQzRvalpBNkdCMGs4RmZxN2RZQUdKVTZ0T2FIOEJHUmZVR2lHb290Mm1Wcm91VkpONFpCbWlISWtnQ2c9PQ==
Z0FBQUFBQm9IVGJCb3F5Q2xSeDUyU0lGSDMtNm0tSXliT0RROElLRE9lWDliQlNoNDJiaHA5TUNMVU5qd3JuUm95VjVzYjQ5X1VjcVlWTlhBV2pfV0RHaHF3NThxUm1ldDFUY2NrWEU4VWFqc2xiN2N1bWxWc0NYb0U5VlN0ZnpjRWpBd0N6dzZXeVp3TjducmtfQjczampEaDNlYzJ0NURFTmUwdGtSMWQ2aF81ckRsbUhHQzFodHpWMDZtcEZ6cmxSQjN2dUktWmtQQ19uS1VZLUQ2UlVmSTI1WHVuTW9vUT09
The article stated they did not get consent of the people whole faces were used. It will be interesting to see how Stanford's IRB will handle this
r/aiethics
comment
r/AIethics
2017-09-13
Z0FBQUFBQm9IVGJBSHRDcXJ6dks1Z0lwTmJGOURXSEQyMnZMRjdaTEQ4WElmNGp5UkJIVThRM2hPOXM0S2FhSEctNTJ4ZHZKdGV4WE1VWmVpdDYtZ0ZjYThzNnVTZ2xoNmc9PQ==
Z0FBQUFBQm9IVGJCMFFLRGd1RHd3cHNWWEpUZGU2MFhNWjRvUkc5MmtQaTR6VGhiUUlvOU93RE9PcHhpd2s1TWN4NWRqcXBFVzF2SVRuM3IyZndEU3VNWHRDZDlQaTVta0ZNYWVlMkRBSm5YSXY4ZWdJUVl6VlRWc1pwNXV6Rmk3MGY4aGl4NnE1OFFlOEMyWjg3eC1FSVRSUS0tYXhmbkUtTUZVR3ZTUzBEamRyNGQ5ZnQxMkZ6bm5JSFFweWRIeTluUWZaMXhvVkpWOXZTX25pY1hPYS1ndVVmOTlnMGRfUT09
The study already passed through Stanford's IRB before it was even submitted for review. I don't know if Stanford also wants to re-evaluate it. This is [what the APA says](http://www.apa.org/monitor/2016/03/ce-corner.aspx) though: > We lean toward the latter argument and believe that public Facebook profile data may be used without participants' consent if it is reasonable to assume that the data were knowingly made public by the individuals. Obviously this is about Facebook, but I think similar arguments would apply to other publicly posted data. Multiple viewpoints/arguments are mentioned, so it's not uncontroversial, but the American Psychological Association is the largest psych organization in the US (and the world; I'm pretty sure), and they're the publisher of the Journal of Personality and Social Psychology, so they're fairly authorative.
r/aiethics
comment
r/AIethics
2017-09-14
Z0FBQUFBQm9IVGJBRm5TQ1EySHdMTHNSb0tVT1NQTjVkb1BHN20yUm1JYXVnOU1iSGViaTRBazVxVFpJTUdlVmJmd2pjN1ZuaV9sd2g0SlVrS3g1RHJlM05hRFl2MUJZVEE9PQ==
Z0FBQUFBQm9IVGJCNm9qLTl0cHF5bnBnSVdBdWF1ZUdDNU9iZFRMeGY4djBLeWxOZ0wtZURXTXZsa1llZ1Y2R29xc19EZkVOeGtGbkNzSlNHSEhZOFpaaFctaExOSFU4ek1tc1RXZ2RNdHRqVHNvZ3J2TUhVQXFUZUZWVTdDUEhpRVhFWkJJcFBrVEFHckF3cHNlRFlXNkRpVW1KOEN0WEhaZkg0ckhqdzNoSEpYT2dpMUxDeXEyOGI0VzJXZE1Sck1IUkt3T0J5alVMQmlLREhTRVRGY0NkSjhYZHZXcHB6UT09
I think the objections mentioned here are pretty much all bullshit, but I suspect that in some cases it's related to a very understandable fear that this publication is dangerous. I'm sure that if you go looking, it's possible to find *something* wrong with this study, but that's true about every study. My impression is that the way this research was conducted, both methodologically and ethically, was basically fine. The problem is that the very suggestion that sexual orientation can reliably be visually detected is dangerous to LGBT people (and people who "look gay"). If everybody believes this is impossible, then there's no reason for anyone to develop a gaydar tool and use it to hunt down and prosecute gay people. But if a Stanford study in a prominent journal says it can be done, then suddenly there's some authority to everybody's (shitty) gaydar. I also think the authors did a lot of good by warning about this and furthering humanity's knowledge about sexual preferences (e.g. by lending support to the prenatal hormone theory), and I don't know how it balances out. Maybe bad people were already going to build these tools and the added credibility doesn't matter. Or maybe this research inspires lots of people to develop tools for outing, bullying, prosecuting, etc. gay people. I don't know, but I can somewhat understand the fear and the impulse to do damage control by trying to make the findings seem false.
r/aiethics
comment
r/AIethics
2017-09-14
Z0FBQUFBQm9IVGJBQ1lZbUc5eFJzRVpnajNfdTdZcTZuMlZKeVB4cWl4RV9JelRER2wwdENtSFBOX3dqVEluM3ZwNHZNcU5oUW5SSV9Gc0RRZVI1dkxnT0dpVFhuelJaMHc9PQ==
Z0FBQUFBQm9IVGJCclBWS09kVUZIbS1hLVgwaEZYVk9PV3Z4aXdwZjhWdW9uZlR4eUVBNEFzTlVYT2o1VlJoa0ZldEZ1S3hMRUtXOVE1TkxKdVkycjBySEVuSTZ1XzJha3czcXQ3dnE4Y19UTW91LUFxaS1rdDVPODNybVEtYi1WUlNOZHl4VkljSnZnNGN3cVRNZUptTnhQZmgtTWh5VDdkakR3Q0FuOHFDZE9XSFFxN2JpQXMtQUVqR0pnQWR5N24xQ2dFaC05dWhOSU96R1Zoai0yZ2FnWW53elJyY2QxUT09
Thank you for the update. I don't know if dating profiles necessarily fit as public data. Ashley Madison site being an obvious counter example.
r/aiethics
comment
r/AIethics
2017-09-14
Z0FBQUFBQm9IVGJBdkh6dHhJdEFFcHBvYzZKVXhqVllzTjZnZVZRRWswR0taa3JMZEFLRUJiRnNrSjNtWmZWNFM3VGVSbk1GbmZkSldpRlVvZkdxQ0xRTW1hX3dxS2RoeUE9PQ==
Z0FBQUFBQm9IVGJCRWxnM0VpWDFpeWdlakpLT3R5czJoNkliRjIzbHFlTm94NERuR2RKLW9BWHNiRXF3UmJBb0oyY3dqZUV3TXc2ZW1Ta1NzMDVsT3lTdm8xTXpMTi1OVWZ0VWlsdUoybzVMbVdLZnU2NElyTW9YSXY0NFluYlBDR1FGeGs5OWdMc20xS2c1bUEyWU15dDl6TkxqRmVwQUdkd0xsWlpmU0VJeGROcWFzcnFWVzlHUlZ3dzVVQmhLSXFUbkR5SXhhOGVPNk1idFJ4MFdMT0pvTVpkRnMtbFpDdz09
I'm not 100% sure what the line between a public and private profile is, but I would think something is considered public if anyone can see it (possibly after paying). So Facebook information that's only visible to friends (of friends of friends ...) should be considered private, while Ashley Madison profiles might actually be partially public (of course, things like email and pics are still private if you configure it that way). I don't know if the website could prevent this by e.g. forbidding it in their ToS. Anyway, according to the paper the used profiles were public: > We obtained facial images from public profiles posted on a U.S. dating website.
r/aiethics
comment
r/AIethics
2017-09-14
Z0FBQUFBQm9IVGJBWWFQaWFJV1NodnZFU1JhdERnM3JvVDBwMDRCcThKckpIOVBibTdGSVpjeXQ3S0w0eldvWnBWMVJ5dUlPbUVER3JyNG9FNXlPZlpUQUlXeWNHdVQxcGc9PQ==
Z0FBQUFBQm9IVGJCdkZmTFpBaEt1VUtFZGtyTktVVDI3TUJSTkRtend2aVpQcllsY2dxb3dHZmt0SjRFd3FCeS1QbHhIQ3ZOblEwalJkTkNoSWxhQUVCTjIwTElPUWRNNnZHV1IxSUhYLWpJSWJIbDJUQWUyMWNIV3I2MXV2cHJ4OEswSzRJZVJ1TzVTRUVmMk10d09GMHpwbTBUTjZINzB6dUdxWDEzd1FUSWwyU2dTMDE2eldQU3dacWxDdmhHU2NMdFVIT0hOdjYzZzhudFJmY3pyNjU2UDhjRGFBeHZzdz09
Thanks, but this is about AI, but not really about ethics (removed)
r/aiethics
comment
r/AIethics
2017-09-14
Z0FBQUFBQm9IVGJBeGNramI0cWtORkVSOXdhTlRncF9jVnVvLUllNVUyc0ZlS01wQjJKb0RPb29wOWF2Ymd1eThvdUR1dnhzWm9sckExblgwWjh2ZGJWNjA4MmJoVkZkVVE9PQ==
Z0FBQUFBQm9IVGJCY2tmaHozZzdJQ3Y1anQwUndQQ2NGeTM5UWlwQ2FIeHVkcWZ5Y2NpS0NxR1VtM1h3cnRaZXVYZ21kNzlYVXhMUVlNZGFRR1FYR0V2bXRnc3phYnZKeGhTc1FJRkgzQzBQVlRhUmRWMUtVMzdqeFh2cS1nNjNwYUVmWEhXbE9QWExNZG91WnhBOTJBdGYzYjQ4VDdiYk9kR2NMWi13TnZlaHFibzFPMERJd3h5RlJEaVdWWnFiM3dWeFRrdHNXemZx
Brian Tomasik's more detailed argument for moral consideration of reinforcement learning agents can be found [on arXiv](https://arxiv.org/abs/1410.8233).
r/aiethics
comment
r/AIethics
2017-09-17
Z0FBQUFBQm9IVGJBbUZlZmdiUXB0OEJyTVBkRWpJQ0RVbW1BeVdNVFQxLUd2NllNUDdvUE9Ka1VaMTJQTFlpelF0M0JybU5PMEdhX2RESXEzMnVHWGFoc1Q5d1Q4QnNoeEE9PQ==
Z0FBQUFBQm9IVGJCRHlZbmlhX2ZBdWJ6S2YyMUZKSGlzblY2QnlyM3FydXhTTDBWTkFaTlgwRS14M0swYzZRei1ldWZKQzhoS29zZlVMZE1uRmVaWmFOOXdrWWhOQUZRVlRnUzFwR2xQQmdUNXlOcnVZRGFQWU96T0VpWVpKdnIySUJtMG1uaklMcHNSSnlyczRVeHJFaG9MbnBfNXZIMTc0WVRNa1JmMHVBQzh3UjhxZ1Q1emZIdW1fTVZ3R1Jxd0JCR2p3N1A1U1VTZEM4ZGRrR3hQSm5tYTFxNVVmbmtqdz09
I think the point is that if Black people actually repay loans at a lower rate than Asian people, then any system which accurately predicts the probability of loan repayment will seemingly discriminate against Black people in this way. This should be obviously true. When it comes to re-offending, then you are talking about actually increasing the risk to third parties of being a victim of violent crime in order not to offend some political sensibilities.
r/aiethics
comment
r/AIethics
2017-09-20
Z0FBQUFBQm9IVGJBX1YyV2JaN0lTdXlCd3ctTFZBSVowc2FPV2NIUzgyQmZUMHVwQUZreXhYZF95ZTc1VHNvY01QZThXbUtXbWNHb3hzaVN0N3BtRER1RUJXMDhXOGRJU3c9PQ==
Z0FBQUFBQm9IVGJCUWlqT3g1aExrNmFXT2ZxZTRFRFJiLTBXTkZ4RXI4akZpSEtfVTVTSGFPTjlOX05tYUdEWU1nMDdQQkI5eFB3UVNJb3NhTG1HRk8zV1E3cTZhbi1wckJnSGlOdFpqWDB1dERWSHBMUWUyMl9XRGlNTzY0VjRZSmZLaWV2Q1JveVJCX0Y5cnBkRC15QzZPejBRVnFoVU9ndnpOeDlKRzFwMkppdFJ2R285TV9MUlNSbEZaVzF2dXpacGhadVpLM005YXBTUDBwMGxrTi1PRzZUQV9PMTlzUT09
I am sorry that I can't remember the speaker, but the best quote I have ever read in terms of the moral decisions of self-driving cars is: "How many times have you ever had to make such a choice when driving?" Is there any case history in civil law or criminal law of a person having made such a choice while driving and being brought to court over it? If so, then just follow the case law, if not then the question seems silly.
r/aiethics
comment
r/AIethics
2017-09-20
Z0FBQUFBQm9IVGJBT09xVFJtODVaU3FyQ1g0UDlqTXY3NVBqV1puMFE5NmI0S01zOTdMWk9nME1SVTFYOTJWVE9zNzEzMlpXbG50bXZtcVJNOXI3Q0haWU03bWdjZ1VjOGc9PQ==
Z0FBQUFBQm9IVGJCaS1CR1JpNERPUEF2dkxuemhzWmFtU3Q2VDU1Y29zUmUwdGZDUVlGdDBiYmRYdGIxLU5tYXNIdkZ6WTdfRzZMYzFhaXBmbElaLUNfWmZ5NldGM0NfeFlhazdJWUJETm1qMkJ0NTNzZG5ZVnFQZGRScm9VODhEV3VONHdBTTgzd3ZzeVE4aHhKSzBBckxKSS16Q3dMNEdNT3V0VE9qYUZ1Tk1laUFwVk5XVmxVV3VYN2VLS2c3Mks4bEM1OWl6UWs2WXZuR3hrUURTYW5ZcWQtdnBmbkVkdz09
I think I understand what you're saying, but *seeming* discrimination isn't really the problem. Suppose we could measure the things we care about directly: Will-Repay-Loan and Will-Reoffend. And suppose that, like you suggested, 20% black people won't repay their loans and 10% of white people won't repay. If the bank only loans to people who will repay, they will reject twice as many black people (percentage-wise) than white people. I think this is what you're calling *seeming* discrimination, and I have no problem with this: skin color did not factor into the decision in any way and nobody is getting duped for sharing an easily visible characteristic with a group of people they have no control over. But of course in reality we don't have that direct Will-Repay-Loan information. What I think is *not* okay is for the bank to just randomly reject 10% of white people and 20% of black people. Or even worse, if there are 100 white applicants and 100 black applicants and the bank wants to give out 120 loans, then it's actually optimal to just loan to 100% of white people and 20% of black people. This is actual discrimination and I would argue it's unfair, even though it's more accurate than a non-discriminatory policy. If Black Bill was totally going to repay his loan, is it fair that his application is rejected because a bunch of other black people who he has no affiliation with or control over failed to repay theirs? Is it Bill's fault that the bank didn't obtain more informative information than black/white? As pointed out, there's often a tradeoff between fairness and accuracy, and I think we should be aware of this. In some cases, we might say that fairness/non-discrimination is more important (e.g. when it comes at the cost of a bank's profits), while in other cases we might go in the other direction (e.g. when it comes at the cost of innocent people's safety). And please don't fall into the trap of phrasing unfairness/discrimination as "political sensibility" or "political correctness": this isn't just some abstract concept. Not getting a loan or spending extra jail time are real things that happen to actual people and can have a huge impact on their lives (frankly, I think I'd rather get mugged than have to spend a month in jail, although that consideration might change for more violent crime / different amounts of jail time). Finally I'd just like to say that we already have discrimination laws, and machine learning should frankly not change any of that. We should not fall into the trap of thinking that just because it's a machine it must be totally objective and fair, or that just because it's a black box and we don't directly *see* the discrimination it must not be there.
r/aiethics
comment
r/AIethics
2017-09-20
Z0FBQUFBQm9IVGJBNnpPQUF1ZnNzbHV1WEFxSjlRZEV6UHVOaFdxMEFqbVQzT3dmYkV4eDFHVVNVTVpIRm12TXlOWTJqcHdyaTZSSmRIQVR0Q21UODdsQnhkdVhicFBZM1E9PQ==
Z0FBQUFBQm9IVGJCUkdoZl9oY3FVb3FkUHVYb2IxOWpGaGYya1hvZm9CdW5SYXNmelFLNDdHTWs4ODJyQ0pyR1M3azlyRTdNV2ROOU5JaUZib2F0YkhYUDlFdThMMksxZEdINjRTYlNXRXl6d3JjdDZYNGN0azhjWTgxQ3NKWUhpU1lIcWlUdWFKTTA5LTBPdzR2bUs4QkliUTV4VjRTMFRwSmp1UUdvdUlaT1dNZWFxTEJqbE9CalFISXRWRUpUU3I2cnFVLVdkQjl1RGR4V2ZNOS14VzJKcGNFQTV2bDNSUT09
I'd love to see more people focusing on the sociological aspects of ai; most specifically the psychological archetypes of those most likely to develop and employ it.
r/aiethics
comment
r/AIethics
2017-09-20
Z0FBQUFBQm9IVGJBN0t4SWRkYnpDbjhuN005aEtXbTkwRFBOME51RUJxOUJncVF6MHNJb3BqR2pyWmxKQVZQZ00ycUktT195ZHBQSVFWbTJRcDhzTEJBZmVsZTZCcFBjTWc9PQ==
Z0FBQUFBQm9IVGJCMEQ2cUpCbWdRaDVBZ080OE9HckxzaHBEcjVHeTM4aGtkSTRVN21kYk56Rlg5MHRVQm5YYlF6Si0wV0gtTzQwZVlpOUd3enBrb2pSTDVFRmxnQ1dxaTBaTWVZMjNXZXFpaU41cVF6TTRNejdWaVVIc0JoUXBLNmUwQlhueHJZUld1RlhpeTZTUENja28zbHhwZW5CaDlXU3FkSHVzZ1FJa2tqMklGLTYzUTI3cXFWemo3VUcyenEtekI2ZmU3a2lK
I'm not fully aware of what you were describing, especially with regard to CERN or those within the Manhattan Project engaging themselves in ethics? I assume you mean to imply they considered the ethic implications of what they were doing, but I'm not certain that they did; or, if they did, they still moved ahead. I know Father Manhattan asked for forgiveness. There is no reason to believe those engaged in AI research are or should. I'm a philosophy major, which is why I joined this sub in the first place. I see ethics as a battle of competing ideas and, in general, the evolution of those ideas. If something is truly ethically wrong then the humans behind that will halt their own furtherance of it. I'm certain there are scientific pushes that we're not aware of simply because the ones involved decided it was too egregious. Focus on it not from a watchdog group, as you put it, but from individuals being watchdogs. Look at Bradly/Chelsea Manning. Look at Wikileaks. Look at the truly ethical people you can imagine who have ruined their own lives to ensure the truth of the unethical was made aware. That's what's protecting us. Why wouldn't it be the same with AI?
r/aiethics
comment
r/AIethics
2017-09-21
Z0FBQUFBQm9IVGJBQk5jTjNXS1c2VFFEM3hCUXlPZXhPb1l2Z1dJRXFOb0Q5cXRIWFR5NXFMYzRGNTBWMTNMV1p3eTNjS0JJVjNBbFA3ejRFaFdpaDFONXBJZUYxWGIwbXc9PQ==
Z0FBQUFBQm9IVGJCSGs5WHJJY0dDVWxSR2x4ZTdNUHMxQVA4ZWFCMF8ycXdpazZBeXI0SUJ3TWVkVDQtNEY2cTh6OWJONmxuNkdCMHN4aXhVQmJLR0stR25FbDBJbXpiZTluRTRkZklGZzZZVXFsYUxNY1lYN05ZcHpkODFTMnl1Nl9OVWVrelpoRWQzb09DX1ZXd0RMa2U1dlJYMHV5Rmp1RUdodk9ZdEphSUswMzN4REpRZWZodHY0Z1pQaWJWX1dTT3NkbXJTWWIw
no new information, I'm pretty sure everyone that watches that kind of video has already heard about everything he talks about and the political joke was really unnecessary
r/aiethics
comment
r/AIethics
2017-09-23
Z0FBQUFBQm9IVGJBX20yUTNab0l3RVRKekpHWDI0RGd2Y0kzNlZGS3NzdV8tanoxcDljcHdTUEdsSU43aW5fRkVMV3QxMk10S1FPb0djem4wN1ItaS0wY0JlbnFDejh0Y1E9PQ==
Z0FBQUFBQm9IVGJCaEF3b3o1NEFPRm5LZnBqT0NFQ1lNQVByQW1MRmJCUmNjcVFnWnZqMlJQVTAzM3FXaDRRaGFYcHV6SjJMMGNCbC00Qlh6cXQ3TXk0TlBvUmlndmlpdzMyTm9UcnlwQ0JkWElwTmR3UGVFamNRY09McnlmM2hBSDIzVUhEalJ5QlIxT3oySGtnTXNSdHpXTENoamxQbWZJLS0yVjljWHJNcUQ3Y0d3bHZwZ0pUWE91bUkyMGdVemYyOTItdW04VlJjNGxsMWRGR2w3cG54RkgwV1N3R0k1QT09
This is about the control problem, not AI ethics. (removed)
r/aiethics
comment
r/AIethics
2017-09-24
Z0FBQUFBQm9IVGJBZ3ZvR0dwSTJaSnlyRnNPQUNmNjFnb296UFlIR2pFYXRjcm1HVWdYTi1NMEdTdjVwX0kzRXZzNTd4c1pKTDhhNUp3ejZsZlhBZTRNc1ExMExjMU5welE9PQ==
Z0FBQUFBQm9IVGJCSlZEVXhJenRudVE4dWdrX1FIZ2c4dEVGWDB3QjVuaE1BSm9VRTFDbWc0YkIwM0xzUTZ6bTZMQmh1VGpoVVAyN2hJQlQ2bVdSVzhtLWtlR1p0SHluTm50VzJTdzVGamx4Q01WdU8xMmZoOGpSYU9rSHRlTXdhdlBGLWhlcllaaVdqUnZEUWYzanhCa2xNVndxbTF1Sm9wNmNFak8xNlJXNk40WGRSdXBoSHdJYjBNZXBuWEszaUlUbHdoUzJxdkNDYnJweUJmd2ttMEhRSW9PcGVjXzV5QT09
>Can you think of some way that I could narrow down the problem I am trying to address to something more specific, so that it is easier to tackle? You can choose a particular arena on the policy side - inter-governmental organizations, the judiciary and legal system, the executive system, the legislative system, interest groups, academia. Or you can choose a particular technical area - AGI/ASI safety and control, fairness and disparate impact, contemporary machine ethics, lethal autonomous weapons, AI systems' moral status. To narrow it further, you can either cut up the categories above into more specific topics (that should be easy if you've done enough research about them), or look at the intersection of one policy lane with one technical area.
r/aiethics
comment
r/AIethics
2017-09-24
Z0FBQUFBQm9IVGJBeThEMGdZaFNqNTZ3ZWdObXFXcEdYQy1FN0ZTZ01zT3h0R09yOUFuTTdFT3FEdXU2TFd4UHlRc3Ffa0otd2sxSHhhX2Jqay1lUDFUZ2lxbFhkajdnd3c9PQ==
Z0FBQUFBQm9IVGJCWWJMT2NmOTdQNkw3MlZheGdKVlZKMEpCWUQ4bzFCYUtGNm1OWU9aZ2tFeUxxTV9wSDB4RkotQ1YxZGV1UE53RjBRSTYxM3NOUFJ4RGRBN2p0S2M4b2dzVVFtc25sVUJvQUE5bld1aXh3SDRac0poX3R3dGlna0RLR0NPTHc0b2tRQ3RtZlYtaVlIWS02blBOYU9BRDdNOGRzcDB5RVZ3RS1LTUQ2NDlsTHJZb1JmSVdNejh4dDhGU29tTndJdU1J
You could look to the Carnigie Mellon University... they have set up an AI Ethics board. Maybe take a look at what they are up to currently. (I was recently looking into this for a Design Research class, although none of the developers would get back in touch with me when I tried to contact them)
r/aiethics
comment
r/AIethics
2017-09-25
Z0FBQUFBQm9IVGJBcVVER1loajNwSjFvWHJCNFlFYXdBeExDeXRKYnJMckFzWDVzd2FTeDl5ekdJMUlJMjYtUDVWOU9HNG94eE9RUU5GMGNmRVE0TUp4bDI5SHh4MnJBZVE9PQ==
Z0FBQUFBQm9IVGJCVnA4VnV0Nm1rQW1oSld0YkQ3eEl2Nlo0R0M2VXRsVlZuS0p3WUxKOEV2SFJLODNJZFZqX2VmOWJzMjV0dzU2aG9JTHFORTF6dEU3SWhhYzd2dVdjQ3RVYXpCZXlMMG12alhHOTM4bmFOWGh0eWg0d1cxdzdRdEJlWGdEUS1QYVl0NlFNeWswM21UcDVEb1packZpc2VLNlFWeGJ0X0lKSGxTVEFDQjk3cEY2X1JHZktsc192cVAteEI4UTEzNDg3
I took the survey!
r/aiethics
comment
r/AIethics
2017-09-26
Z0FBQUFBQm9IVGJBeUlFMjNnTXU0MXJ1TFFFVEZKaWtRS05VenQxNm1MMV9RRTdDTTdtVjFGekZKQ3ZXWHlSUHh2bjk1LXJqWTdxbzlsSzByOEg1czJPY2FTRWg2MWYzblE9PQ==
Z0FBQUFBQm9IVGJCT1lJNkdRSzdKZWZiaFB1YV8xdTl0X1YwQlBlOWZMRGJqdDR0SWNGc1FONG5XZm5lMF80U0tJQU9rd1AwVTBiUC1yMDBwV2NkZExfY1ZieFZCbld0TUw5NEU3X3lOdFpwUGZxREZxalNuZVRJa1FEZlFOcDh2T0R3OUNxclZzUU1penFubjRkLXdmWE0zbVVFTVZWdmp1dF9DSHJ5UlRSOEFPVzVjcWVyY3hPemU5REVMckRQQTV2dnNFdTg4VFRO
Did you not mean to link to [this page](http://www.fatml.org/schedule/2017)??
r/aiethics
comment
r/AIethics
2017-09-28
Z0FBQUFBQm9IVGJBdzV6THQ1MVBXTUhOWXpDTW4zbUlLVUxyRWE0RkJWenloR2lRVGUyQmVDNk5WYXN5ZnQyeGJDbUZ4QlpKaGdCdlc0LWtOSG50dFpuT1E5WTFuTl8xOEE9PQ==
Z0FBQUFBQm9IVGJCc3lMaW5VcnRsN0FzQ3l3YU82TTRuanFCalpySUJBdFBqcGFkRXE0U09Sc0Y4S0xXbjFZSG05R1FGMkZWSkYxUkk0SzNMMmFVZkFubm9UV3Q0MUlON0ZWcElkZWVSRmJyQlVzSWU5WERHLTd1aFhzd0U3TzNxVnJzNHlVdExKQTFVbFpRcnBwOTYzMWNQNkdzQ3JyWjZraEtxSXo3d0o4SlE4M0N4b29lZmRkYWdRdUkzVVBwLVlLNE9FVm5taDNS
[This one](http://www.fatml.org/schedule/2017/page/papers-2017)
r/aiethics
comment
r/AIethics
2017-09-28
Z0FBQUFBQm9IVGJBNlY0UG9NQ3RzTnBLYXlqQjJXdXozN09WcGk3UnBJUGhzMWVtcmJGYTl1S0g4WHk0dHc1aDVuU3p3Zm85MW9EMUs3MFVTTXd2cTE0MGRXN21GRmw0N0E9PQ==
Z0FBQUFBQm9IVGJCMW4waWIzQmt2dHZ0eW5HR3RiT1hwSG9TNmtMNk5JNFhRUGZPTGlXY3dBbDgwYzV5NXRPWnZfQ2hRYXVlelY3VXpmdjlVOEo5Qm9NeGxzbXljd1lvMWtYYVJrZ2JSd1hDSlRobDk4Y1JGam8zWEtpa1h6Ni04bXhhdGpUZ25CbnB0Yl9vbDIyaFQtdng3dVpQX1N4NDhmbGdFMTVIX2QxbURuVlhzV0JzNUVfY1dxQlZoQWRGdVE2S05JWUY4bndn
Oscarwhite Our services like:- General Office Cleaning, Commercial Cleaning Melbourne, End of Tenancy Cleaning, Full Hygiene & Consumables Service, Glass & Window Cleaning, Restaurant & Bar Cleaning Contact number:1300 672 279 
r/cleanenergy
post
r/CleanEnergy
2017-10-02
Z0FBQUFBQm9IVGJBR3VYc0xfRzAyclBTdmtTd2FXREo3X1M0S2Q1RTZiZ1VrYXk1dXA1bkZRUUkzU01EUFM4X050QkFBSTcyNjZvdXowNi1WQU11bGJwSE9DNzBybkN6cmc9PQ==
Z0FBQUFBQm9IVGJCUDVURjlqaUtPX3lJNVZ0ZWJoMENDQzdyNUlvVlJIVHZWcWhlWExKaEdvMlJTQ0RPRFROdVM0NXd4cmY1eHVMX3ZzZDFROHNFdjltNm1tMXlxaENjc3ViMTQ5REZZSG1qbHQxMVBNR3lvYUtqQ0RPanhUTEFFSnhrdlJBZGt5RXRWU3I4cnlGNG15MjNHMWtJYnZrX0RpYUIzeU93dDNVSi1xWXhENElWYjlnUmZEYzY5djhoRm9SMUZUa3otV0U5
Speaking as a pretty seasoned commercial software tester, I can say we absolutely CAN prevent buggy software from reaching production, but that in most cases, the money/market isn't there to do as much as we need to to fully understand everything we release. There are fields of quality assurance which are far, FAR more watertight than commercial software (nuclear reactors, military tech etc) but yes, there are still occasional issues there. The answer for me, as usual with AI, is teach the AI the fundamentals well enough, and in a watertight enough way (as in - huge amounts of analysis and test, trial and failsafe), that you can then allow it to course-correct itself on the fly. Buggy AI will absolutely reach production. We have to ensure failsafes, and - vitally - non-buggy AI is there to handle it when it does. Interesting question :)
r/aiethics
comment
r/AIethics
2017-10-21
Z0FBQUFBQm9IVGJBNWgwTEhmNjBxVEJNODdaZnZsOG5BOGJkblo1alhYWWxTTE52TTVDdkRPdUpxQkw2eGtRbmFUb000WDA2VGhHcjB2N05xak5EaW5LajFCN1JpUGpISGc9PQ==
Z0FBQUFBQm9IVGJCMmpXaWZlTjZlSjBVMDdTWV9QUmVEd1kwS2diQ1NTMHlBWTRCREtvSTQyRVlDMDRzZzBmb1lXakc1U2hyTkdyNVV0b3ZjUksweGtZTmFIakhwSk1kU1FFUTk4bndWX3U5Q3pwdnptdmVYbFJQdXJoSkk4Q3hndjlMSHhTdG9NNFoyU2JaeGJtMGpBalFyMUVobld0OVdLZzlWSWNFUEZBQ2RQTWtzSW4yU1B6dVpyMjRNanFzQmxOc1hDdGpYckFYRXR0OXJsRTlWLXlFbDcyaXFUOWNJdz09
Off-topic. Removed.
r/aiethics
comment
r/AIethics
2017-10-31
Z0FBQUFBQm9IVGJBQzNKa0NvZnVZbHk4QWItWVZma2hTUXBWSWI4SFUwQ0FuUnF2M0c3R3hRQk1aZVg4VE1fRVVsRXo2XzZYUW41NHJQNmVqWDU4ZllhWFBXQWRkYm1lUGc9PQ==
Z0FBQUFBQm9IVGJCeTl2NkVWelVQeElJbFJCLXFrS0YxeXV6bmJpS1NEY01JU1hBVGFmQTlHTmE0Yjl0a3IxYVRjZlpHT1k2U2VhSFJXOEF4WmRGSzRxeFdwTng5d050d2pHQmRWUFI0TjhieW8tN0oxckh0cmltcm5xSy1SQ2VWTkZtTEVxc2lnT1MwLXZzWllvb1BvVjlObkxuV0xNTlo4VGt6WG9JOTNuYlBlT3FJZFJxbXdVTTZSRXpBMUpNZVIydjVWVUVSam5rNUVhREZMWW5ETHQybjFzeU9EYXNCQT09
Yes
r/aiethics
comment
r/AIethics
2017-11-01
Z0FBQUFBQm9IVGJBVXdEdk5ZRWJoMlktVFAydkxJUzFsWmI2N0hkNmhIV2RyUmVwRk91N2dKckoyY1pBUlNfNlhPZ0N6MTltZFlkU05iQlJVTzVINEsyR2JydnRhSWJVZEE9PQ==
Z0FBQUFBQm9IVGJCYks3cHZtWUZWUldpOU1IQ2x5UDI1YVdPMjBCeW9GNUx5VWtuRkxUbE82b1UzeFY0MzUtUm1CeWtadkZaUTBGWF9KZURVTmZuVUJsMFpGaGZVZXBfOGx2dTBHSkdoRUU5d0ZpbDF1ZkltU1VpaHI1TVA0TVFvRF9QMERkcEVJWVlkQlpYS1FVSTlMQWxCdnVRcGZxQUVnZzk1Nl81TjV2ajB3bE9zRHFJMXgyRWVPOTdXSVA2eU0ySDhBQjJCV2hkOW5fNGF5WGFfeVFVcU5hZzNhcXhxUT09
Please remember that this sub is not a catch-all for speculative AI discussion, it's specifically for ethics questions.
r/aiethics
comment
r/AIethics
2017-11-08
Z0FBQUFBQm9IVGJBcnIyQ3dLeGRXSU1lUk1pV2hiQ29BR3JFTWZOUU5KM3RCOUNCREJqZkU5Zll2aVNqQWNkQ3dDcGJQYzVvZGNwRXhHU242ZWFjV0FtRVJqeHh3YkhOemc9PQ==
Z0FBQUFBQm9IVGJCdERRdUF6RTM0UzdPcnJxbU9ZR1RRWDVBVTFVLU1McXlCZ3hSVk14NXNrLWtvdUFxZ3BhRFMtTHZtSng3cmJ2R2duZGM0MHU2aFhjRFNiUTFtbG03VHVvX1hNbEJoYnl0UUx5cGtoTXk1dGtyVGV1N0phZ2FTY09WYUhodThKV09pLTg4Y0ZoYVdKQkJwMnA0MUROQmx4RU9sczNpZEx1eHd0VEk5Z3dkYnpzTjU2RUFSVjMzdktNOUpVcmx0M1ZLdmpUN2l1UXY3YWxjY0Rpa1pXRl94dz09
Not using a person as a subject without their consent except in exceptional circumstances is absolutely a standard part of research ethics.
r/aiethics
comment
r/AIethics
2017-11-17
Z0FBQUFBQm9IVGJBa2JnR0JzTDNMRnRPR0RXbjNtMVhWbHRiUDN3a25ZZWlsa0xXU0MzNXpqMWZpbUVvRGx3U0VTRGRPWHFQc3lXZkVpVlo1M2djZ1RpM3E3ZnlMbmhtUVE9PQ==
Z0FBQUFBQm9IVGJCMUF0QUo1ZnZjYy01X0NtblQ4VUozNzJKNFJLNGVoaVAwZERtVVJWdDF3ekJSOWxwMFFWZGVEMXpIa3pUWGJ4RHZFb2J4VmFDVXB0bmROQmlqb2JBV1l1N3BhbFJfb2dtZmxGWHI2UFNJYVpZVDB1Q0FoNXkwbGcxQk9DVzYxS3owUm1wMDU3N2JaYk1pMGRXZG15Yk1fa3NxVFlicGY5S0NiX3dJbFlqRTJEQlRlZTlvY1dZX0h0V0tTNmkwNGtLakRaWUl5MmxKLUZZc0l3cktjRmFWZz09
I'm not sure exactly what these people want, since we're going to have to make a decision on who gets into the country anyway, and not using AI is obviously going to make that decision nothing but less accurate. I kind of believe that the true objection these people have is that lots of immigrants are going to be denied, not that the immigrants who will be denied will be the ones who are labeled as the riskiest ones by an ML classifier, and they just recognize that the public is gullible enough that raising fears about AI is a good lever for effecting that and forcing the government to let lots of people enter the country. Because I really fail to see any other goal system where it makes sense to force the government to use inferior methods of determining immigration. I suspect that the best way they could defend their position is that they expect that the public and policymakers will overestimate the validity/accuracy of these algorithms and they also expect that the algorithms will be biased in a particularly bad way rather than pure inaccuracy and then they expect that this combination will lead to lots of unfairness in the long run which is supposedly morally worse than the predictably bad outcome of having an inaccurate immigration policy which allows lots of criminals into the country or excludes lots of good people. A rather dubious play of 4D chess in my opinion. But even the Machiavellian interpretation is 4D chess because you're assuming that the administration will err on the side of more immigration rather than less when they are forced to abandon accuracy.
r/aiethics
comment
r/AIethics
2017-11-21
Z0FBQUFBQm9IVGJBRjZZemhURnJEbGNkZDk5eWlGdnk1MlNNYTkyQVNFVEJsejB6aG5yOUJpakRKSElMbTg0blU5VHlIVEtGcDJUWVprMGk5bmhjV2JvZUdaZEJlcEZxbUE9PQ==
Z0FBQUFBQm9IVGJCVEUzdnRuSDFfQmFMUzhGWFVDM0FyN3JpU2RYZ3dGR0thTWd1dkNCeDBWcmFpdVVzcTRxZVN4VjdVelZpNFFDbXlETE4tZk43dUE0TE5hTFc4SDkwdlJvcWNTcTVucTFpQi1xb04wLUVOaGNadkdRdUVPb0JLUUowTmdEdHBCRC11OEVCVVpIZVdkMnlPc3ZhLWx4VzVSXzA0NVpaOWZjSnFORUtNYzFOT1V2NGVjY19KYUtyMkZFNnphazk3ZEdXOTNOZmtsRnRSZjVaWWlLUUJhS19iUT09
It might be assuming that they will be *forced* to err on the side of more immigration rather than less. As has happened with the travel ban getting suspended.
r/aiethics
comment
r/AIethics
2017-11-22
Z0FBQUFBQm9IVGJBejNlNE5McUs2Ylh0V0dBQy1xM3BCZFRzbEhuVGJONTZ0RE9qSE1Wd1NzWmYzblQyemdvTWpZcnlSLTgyYjJPd0IxSnZYMV9RSEhpMV9qb2ZraXVYOGc9PQ==
Z0FBQUFBQm9IVGJCaFd1elh1eG9JTkItRDlhREo4bEtNSlYxTE1mci01SGFuVE45SXJxaHpVOFNDTnphY3VUQ3RZbnJfakF0b1plN1dZTkpzVE9UbWRHSjY4OXlLMFMwQXU3LUtEQ1VBbjI3N3M4T19qdHFzVE1nRFBjczJHXzdCR0pEd04zQ0UyclFpSjhEYTlHUDFxYmFtZy1mOS1DWk5OMHNLVkI3TEtrLTBST0dVeW5FUW1mTC01eFl4WXR4U3lGQVdpNmk5STNTY20ydi1DcWJJV3RLVm02QVItdzkwZz09
If AI features can be integrated into the human nervous system (along the lines of the proposed Neuralink) in ways that make individuals more altruistic, less violent, better at making decisions geared toward collective good, etc. would it be unethical to opt out?
r/aiethics
comment
r/AIethics
2017-11-22
Z0FBQUFBQm9IVGJBQnNZZUhEN0FYbjlIRFQxU2JwdUYxcjROVl9LLVdRZmJJR1FCeWQzUE5qcHdTRWdyNjJIdDFGRHdWREVJWDF6cWVYZEgxejFDUlpEVWEtX0hCWEkzSEE9PQ==
Z0FBQUFBQm9IVGJCdC1qLXJiN2dHYmtwMXhQOU1lekNCUlk4Z1pCZXNjUTdFc2s3Sm5sVlcybm1UMll3SnNqb2JINWNGelAyUFdZeV94QnJJWWRlemFhUS1Ca3Q2TVB3YXVNNUdJak5WSFd1Q3d2R3lRZk1GRThfT3VQX19tdkJkdE5BVU16THIweTBIbDA1dnBmUXBlZzRha0U4VXJsNVFxdlpKQUpDWEZVZ05QcHZFQnZyQXdtdFFIdmFZY21ISjBEb1B6VFZ0dHpYVHhNamtwYTRCR0ZFdW5ROHMxNkFKUT09
Sooner the better.
r/aiethics
comment
r/AIethics
2017-12-02
Z0FBQUFBQm9IVGJBOHJUempfSGxZVkZUc25ISy01QTFJRFBwREdlYzlwczYwbXhCZjRfdjNpNWNXNHM3WXN4VXRRUzhtbWVGRkNRSzN0YURFY1ZZM0s4UkdzVDJ4clBtYnc9PQ==
Z0FBQUFBQm9IVGJCMnhlX3FkcDV1YTgxajVvSlhYdVE4ZmRrYl9fTWdxTzU2OV9MWm1HNXB2Q3pGakoxX3hmWmZONGNDQ1lhNWdONVM4d29weUU3OVlETUk3dE55ZGU0d2FLZ0FZT2lScG9sd19oVkpzdXh5dkJkUGRrc0lwTmhHUm1UMVJUNE5pamFyX2J2NGt2QnNXdHlhM2VwWWZ0TlB2QWNaWDg4Y3hIZU96QS1LaW9Sd3JIR1gzV01fVW9MUG4ybkwwZHBNbzVtX2lMaFBaOXZEeTRSeGhfNkFvNE1MQT09
Their first example is just weird... Yes people are interested in having a machine do the driving, but since when is this tied to letting the *car* (AI) decide where you want to go?? I'm pretty sure that people will continue to pick their destinations with a little more care, or at least pick *something* as opposed to "Ok, car? Take me to your choice of hairdresser."
r/aiethics
comment
r/AIethics
2017-12-04
Z0FBQUFBQm9IVGJBbGZ4c3BoQ2Q2UER1WDdNZzlnb0hBbU5yOFZtVzMwcW5JcDhRMkdXT01zRTBHRmEyTktmcGdJTE16bFdTSWN1cVpoN1N2TnNHVmRmdzlwakVZbWNXanc9PQ==
Z0FBQUFBQm9IVGJCNDR4MlNNWUhSNkZBU1gtS3J0aVpNM2dLY2lKeEJTQVZwUk1HYkFRT0FneGtKNnFQTXpoclpfdTZ1cERKaDJTdFJjVXFLYWlUMjFILV9RRHJaUk5pVU4waWpKanMzcGhPNGMzbVJlaVFNRzhoWVAwXzhCeDFkTlFWNDEweGo3SF9vakh1NllhTWJ3cmJmTXhmT2p2MkE5RnJHaE91WlM1WEhBN19IQXVHWWdZdmNyZGdpOGlZNWx5VEs0b3VBUTk0RHpOallDTHVLM3BfUHYxYVRwUGRZQT09
https://www.youtube.com/watch?v=iPlelGS_n5M
r/aiethics
comment
r/AIethics
2017-12-09
Z0FBQUFBQm9IVGJBNUJEMDkyTDhLOHdHaTJGVC1wWlRLU29DWUtEeUl3WHhlWU9kU2hFb3BEajdwT0JJd1cyb1NqRFFDbTRhUUxRT19IN2Y3bEJKT05qNlNPNnZUWkNYb0E9PQ==
Z0FBQUFBQm9IVGJCOHdGdFd1TzVZUF81NTE0YmtSOTY1SkU2cFhadXRxZzd5M0Z6d214SkRRT3QzYWpWeWlaUDdORDVPWTFtd18wbVluMUxGYkRCZFJ4eVROaFJyX3VxaW9lUTROeFhaVzlfcjFzNnlRSUJpN0tBVHlyZFNZYXk4WXowUENKRFN1QnZpZ1YxRGNqeTlzX0VKelVCTURQREluMXFPWmY4MjdGNkxCR1JWTjdhR3MxQVQ5Y1lTT2lyVk1vMm1XS1BxMXc2LXpFenRUWFVWekZteWV4ODRhdjFyZz09
[Detecting Qualia in Natural and Artificial Agents](https://arxiv.org/ftp/arxiv/papers/1712/1712.04020.pdf) submitted to arXiv by Roman V. Yampolskiy on 11 Dec 2017 > **Abstract** > The Hard Problem of consciousness has been dismissed as an illusion. By showing that computers are capable of experiencing, we show that they are at least rudimentarily conscious with potential to eventually reach superconsciousness. The main contribution of the paper is a test for confirming certain subjective experiences in a tested agent. We follow with analysis of benefits and problems with conscious machines and implications of such capability on future of computing, machine rights and artificial intelligence safety. > > **Keywords:** *Artificial Consciousness, Illusion, Feeling, Hard Problem, Mind Crime, Qualia.* > ... > **Acknowledgements** > The author is grateful to Elon Musk and the Future of Life Institute and to Jaan Tallinn and Effective Altruism Ventures for partially funding his work on AI Safety. The author is thankful to Yana Feygin for proofreading a draft of this paper and to Ian Goodfellow for helpful recommendations of relevant literature. > **References** > 1... > ... > 222...   **Social media posts, discussions and some comments** twitter [post](https://twitter.com/romanyam/status/941395934877028353) and faceboook [post](https://www.facebook.com/roman.yampolskiy/posts/10213780718964504) by Dr. Roman V. Yampolskiy, facebook [post](https://www.facebook.com/ArxivSanity/photos/a.175996466252180.1073741828.175548272963666/311969302654895) by Arxiv Sanity twitter [retweet](https://twitter.com/David_Gunkel/status/941428413285400576) by David J. Gunkel: > Looks like a really interesting contribution to the consciousness debate. And one that could have important consequences for the "properties approach" to dealing with questions of machine moral status, or [#robotrights](https://twitter.com/hashtag/robotrights?src=hash). Cannot wait to dig-into it. >> Dr. Roman Yampolskiy @romanyam >> I changed my mind on consciousness. I think computers can have rudimentary consciousness and we can detect their qualia ... [arxiv.org/abs/1712.04020](https://arxiv.org/abs/1712.04020) facebook [comment](https://www.facebook.com/groups/consciousnessevolutionofthemind/permalink/1999699810310475/?comment_id=1999939113619878&comment_tracking=%7B%22tn%22%3A%22R%22%7D) by Alexey Turchin: > [Виктор Аргонов](https://www.facebook.com/argonov) wrote about the topic too - his approach was to create an AI without giving it a chance to learn about human philosophy, and when to ask the AI if it if it has qualia. https://philpapers.org/rec/ARGMAA-2 facebook [share](https://www.facebook.com/djestrada/posts/10109042469515420) by Daniel Estrada: > //. Every one of the cognitive illusions described are examples of access consciousness, not phenomenal consciousness. Illusions are all entirely within the realm of the Easy problem. That means he isn't talking about qualia at all. > It is central to the very concept of qualia that they are not accessible from a third-person perspective. The idea of a test for qualia is self-contradictory. This is why the concept of qualia itself isn't very helpful. > ... >> ... >>> ...
r/aiethics
comment
r/AIethics
2017-12-15
Z0FBQUFBQm9IVGJBX25lN212c1JhWHVnV3FCUFo4a1VNRWNuQnc5SHBYdExBSDVuYmhTRDFveVVOOFpOXzJiVFV3WXhTcTBleW0wdGVuM1RfVUhWMzVnOC13RkpEVkdIaXc9PQ==
Z0FBQUFBQm9IVGJCRWstWHJmelRhaHUydHhNQUExTVUyODBJcXNtdFdralFHZDBFa294TEpBWHJOVGFzN3lZSFZOb0UwdEFJbFNDZ0FDRU5UVm9TSTQwQXZ1LU94dnA1VVFfZFgtYkt1SFQ3WTZFTVFmWWxXanR4dENCOTJoTWltampJNXJKbWNuWXd6OXFEM3k0dmpoRzNTdHI2c3ByTWN6Z29hNmVuRUM5VmVabWwwZ0lOUmY5Nm9uRGRqeHF0OFR4NklHNFh0dFBjeEZkaVJ1VGVHQTJHRi1jYjMyaExnUT09
**This doesn't solve the hard problem *at all*.** First of all, why do the authors state that the system has qualia, simply because it is susceptible for optical illusions? I'm not convinced the system has qualia. Qualia, by their very nature, are undetectable to third-person parties. Philosophically, we can't even prove that other humans have qualia at the moment -- let alone this system. It appears to be nothing more than a bad pattern recognizer that fails in simple examples of optical illusions. Even if we model the whole human brain perfectly and it says "I have qualia", we still don't know if it has qualia. And we never can. Secondly, Chalmers' "Hard Problem" is about *why and how* physical processes lead to seemingly nonphysical qualia. The paper doesn't address this question, but casually mentions that it is now solved by this paper. _______ This grandiose claim is horribly inaccurate. The moment we design a system with qualia (and prove it) is the moment we make a complete paradigm shift in AI and Philosophy of Mind. The closest thing to a qualia-bearing, conscious machine, is an embodied physical system that is functionally isomorphic to the human brain, and *even that* is highly disputed among philosophers of mind.
r/aiethics
comment
r/AIethics
2017-12-15
Z0FBQUFBQm9IVGJBS3IwcW5UMElZV0Ixdk0ycGx4MXlCdnFsWXhUSW1peHBuSVFRV2FZNnFYejI0ZlpyZXJuTF9HTjBBUnZIMnY3d3hiMWNSNnMwcEtmeWtXakJXa3hGV2xxZHREVW5oclRhaUhtQlNSczgwUjA9
Z0FBQUFBQm9IVGJCUWY4eWExdXl1cVI5VjZKOHloYi1vMkFadmVPX2xkQU9WWkxROVIwYWpSZlV3T0pPdHR2WldGTVZWa1IycU9OczFzdGNBRVVXYWJqRkZjeHNkMW9Bb2xhRjZuNG93aWV2eDhqcndkN0R0d1dyOTlSclIzbG9hMm4xSFo1RFFmdWI3S1pFOEdZamxlT1pPSVg4MDJLaHNSSUpmRmVxRWp1b0JjRXMzSTVzWF9RaXRfS3ZhWE9hWWFNMlRhOXVBaHdQVnc1cEJoaHhpSUZGWnFCZFRHS0h3QT09
Yup. This seems like they misunderstood the relevant philosophy
r/aiethics
comment
r/AIethics
2017-12-16
Z0FBQUFBQm9IVGJBYWFSdkNlYmg2Zmp2a1ZzbS02UU5EeWNIaGJXYUdIekhRTWE1NFZLelZKQkRYOFNLSnYxTGV1ZkZ2UkJ1Q0xKWlhVRWk5LXpXN2gwWVk1c3dXdVVMS3c9PQ==
Z0FBQUFBQm9IVGJCbFlqRE9iU1VydUxsamRWZFpyNFRsTlZXMUJXZ3o4QWlQWVBnOGo1cGhrbmpPTS1hdzBSSnk2d1J4TFFtUW1sTkN1dThiZ3V0M1JQTkpnMGpfeFFXZGtoLXB3TmZBVVViUjRRSkJOUkFWNkRQWnBOa181Vkk0U2tVa3Z6dXRNeUVXWW5Zc1owQUNqcmJ5ZXBGb3hMdlZfMS1rM3pJN1I3VG9NOUpWQ0MxaWlJSG9DNjd0SW5GSFp3eFF1RV9KdV9ZWmV3R25hRzhjRTN2STh4R0luUS10Zz09
This article gives me hope. It demonstrates awareness surrounding app and screen addiction issues; awareness that wasn't around when Facebook, Zygna et al forged the attention economy. The article closes on an important point about the morality of mind-gaming customers — people.
r/aiethics
comment
r/AIethics
2017-12-16
Z0FBQUFBQm9IVGJBc2dhY0V1OW9TYk5XYUk5WkNhdnBnNnlETmF2Nkw5VlE0ZzFQaDFKYUk2UFhuUDdzMS1hR3VvQWpLc1RSVkNjYjhYNFZDeFhqNElxS1B4VzBITHhSa0VYT1R0MDRqNGJselU2a2hZa1Zfblk9
Z0FBQUFBQm9IVGJCMzBNdlRTUDIzaWVPa3JHRDhhOGtJMjl5a3VMaDlZalJvdDZ5OTlCOExmMXh1MnJnWmJmcmZIbkhZWE5YQk5SaTVpMU1pMmZzdm82RFNNZUp1ekxXQkJhWVFBdEFZQXh5MllvbmtoWm8tSTU3bEVET1cyblBlWEQ5UVpNZjVfZkhxUXc4WnMwZFRwWTNxRDhoTnlqRlZTaEFOWkp1LUxtbjJnUVZhanZ3MGJUcWpEcjZEU1RnX0FSd3hVTzB6c192U09hb0VzLTNrSmJPb1dwMThMNUN3UT09
[Machine Sentience and Robot Rights](http://reducing-suffering.org/machine-sentience-and-robot-rights/) by Brian Tomasik > ###Introduction > In Aug. 2017, I was interviewed for my thoughts on machine sentience and robot rights for a Boston Globe article. This page contains my answers to the interview questions. The final article was "[Robots need civil rights, too](https://www.bostonglobe.com/ideas/2017/09/08/robots-need-civil-rights-too/igtQCcXhB96009et5C6tXP/story.html)", and the paragraph that mentions me reads as follows: >> Suffering is what concerns Brian Tomasik, a former software engineer who worked on machine learning before helping to start the Foundational Research Institute, whose goal is to reduce suffering in the world. Tomasik raises the possibility that AIs might be suffering because, as he put it in an e-mail, “some artificially intelligent agents learn how to act through simplified digital versions of ‘rewards’ and ‘punishments.’” This system, called reinforcement learning, offers algorithms an abstract “reward” when they make a correct observation [actually, "observation" should be changed to "action"]. It’s designed to emulate the reward system in animal brains, and could potentially lead to a scenario where a machine comes to life and suffers because it doesn’t get enough rewards. Its programmers would likely never realize the hurt they were causing. > Regarding the last sentence, I would say that the suffering of the reinforcement-learning agent would be visible to programmers if the programmers were philosophically sophisticated and held a certain view on consciousness according to which simple reinforcement-learning agents could be said to be suffering to a tiny degree. After all, the programmers would be able to see the agent's code and monitor what rewards or punishments the agent was receiving. > The rest of this page gives my full original remarks for the interview. > **Contents** > [1 Introduction](http://reducing-suffering.org/machine-sentience-and-robot-rights/#Introduction) > [2 Machine consciousness](http://reducing-suffering.org/machine-sentience-and-robot-rights/#Machine_consciousness) > [3 Analogy with insects](http://reducing-suffering.org/machine-sentience-and-robot-rights/#Analogy_with_insects) > [4 Robot rights](http://reducing-suffering.org/machine-sentience-and-robot-rights/#Robot_rights) > ...
r/aiethics
comment
r/AIethics
2017-12-19
Z0FBQUFBQm9IVGJBeW45aEU1bUlrMzVMNFZna3ZqQzZCdmRjbUY4Z2wyaV9QTzd0RngzZ1FvRkFzeXFhZzhZYndXRmNuTTdJdHU4OUplbU1fS3c4WmpWSG9UWFBLRU9pMHc9PQ==
Z0FBQUFBQm9IVGJCYXhCNWJVNUZwWmFraTFjU2NnU3pGaDlzLWltem04cDY4LVlpcU42QlpQMWJqT3RyMHVnanByc2Ztek1ZMGlhREozQU9fZ0JJQlAwY01YX3VPYXZLZDM4NTkyRVFrMXVVeWg0eFpxNV8wNk9ENXd1UktvdEc4SmN1MGdsLVQ2VHlKU3ZCaDBBM3puMGxEWng5X21MUUZhVXduTnFfeHR4eV9oV3cwYmNhSkkyXy1uYzNqQ2lnX0hTNGplekNsR2lFOW9GUThQalIwNmh1UmxxemJjUTcwZz09
I really loved the South Park episode with Alexa and Google home. Spoiler: Eric gets very abusive towards Alexa, and the South Park community does something even more ridiculous.
r/aiethics
comment
r/AIethics
2017-12-20
Z0FBQUFBQm9IVGJBV092Q21lWS1MaFdrTFZfODZYNmQxVkZ6RzZieFVjV3djc3RzRFhjbjFZdVoyRUs3RkhNTEJ1eHZtaVZqUjFFZms0ZmlhbExZVE5yNjYwcTJqWVg1OXZrMVd1ZXRIdWlsZVZHZldBblZhMjA9
Z0FBQUFBQm9IVGJCUHFwRWZmT09kMGxnaXJ6T1FnRC1yakZlQjBnMWZkaktKUzNDX2YxYkJLSENFUTZ3UUh0T3l0clRJZGNsX0w5OGNoZ0NwNHhzVXlBOUlRT3dHazBmV2xfTlVWQ3U2SVRnbmZDbWJzdkt6dlJZRXM2WDhZVmRXUGZTWTE3bDJ1Q25QRGJSZnhxdkszOW16Z0QwOXBldHcwVFNkS0RSYlA1SVlzakhqTXVuTTBTelJyeE04bmRBTDJOV2JwRWdKcXoz
For reasons currently unknown to myself, I decided to buy myself a wind turbine from a website called WindTurbineUSA.com. I had no idea what I was getting myself into with that split second decision. First thing to note here is that I have never even seen a personal wind turbine, much less installed one. Turns out that it’s a bit difficult. For me at least. I have very little experience with power tools and electrical work-- and I really needed it. Really really needed it. So after about three hours of trying to install this beast on my own, I gave in. I called tech support. I wanted to believe that I could figure it out for myself, that I could read the instructions [which are actually pretty damn cohesive] and go from there. But apparently I’m Turbine stupid and needed the help. Luckily, the tech support were super nice and willing to talk me through installing the turbine I bought from them. The guy I talked to was able to explain everything to me in a way that I understood and a few hours later I had a brand new turbine installed onto the side of my house. I even have some power tool experience now! Just a little. So that was my adventure. And now my shed has wind power. -E. Holt
r/cleanenergy
post
r/CleanEnergy
2017-12-26
Z0FBQUFBQm9IVGJBWEVxdHpCQzh4UElZQnJiZkR2cmN2dGFiNzkxb0E5eF8zUEJDa25mb3JRMG1kUGViRzBWSWpNdDFlUS1IN2Ffc0FxOXd1N3NpcS14Yzl2VWNmTmQ3S1E9PQ==
Z0FBQUFBQm9IVGJCT3JjV0I1eU9qSWxNcHJRd24yc2QyOVJkd3pBZ1BvcnFiNlZiR0wtU09QQ3pQdUJCa0hjcXRaTGVLNHpBRXdDa0pvd2Flby1aUU9kdE9ESTdwTF9aV1FMSzlYSjdQcjRUNGdDUEpaUGRYSmhtMHFvNzh2dUFIRjNPblNGNUFRdWVhQ18wMVlWajBmT3JOVXRuLUdqNUxUcHpTNUpNY1ZJVjJxQ2NOLWZ2cGJkejg4ZU1tVVhEZ2ZyNUNmTmNSdk1DdUdHbHVVbll0ZVk5bWRqZ04wUGlnQT09
I think it looks pretty good, but I don't know much about design/aesthetics.
r/aiethics
comment
r/AIethics
2018-01-06
Z0FBQUFBQm9IVGJBQUtSVkgwbjFyTlNFNWk3X2xVVDRRNzhodlI2OHlYSy1mZ3oyQ0dVR1hRNHhudUFZRHhUVXZHdGRsbTdwQmZSUmNNTmRXREdKVWZSZnhURUZkX2ZFenc9PQ==
Z0FBQUFBQm9IVGJCMklLT0oyM1hobS1rLXl1R21aS1BRY2lzMk1LR3M3bGhDOXZselROeGdWTk1iVmtDQmUzVGVOM2ktVThzcVR4U1Nrc2N4M200VmNEelo1amFVU2VfM193N0U4ekQ1d0tnSk0xREM2UFhPZ0ZlaDBZc3RwQzBoYTdVcVFYVWo3d0JMSk5rVlhpWXZqWmlueEF3Q2JPbVQzUWlEUi1iVm9QTVRJOVlULW9oN3NkNFNwN1FmbDF5TEx0QmRScXZ0bXhR
You don't need to ;) if you're in this subreddit then you're probably part of the target audience, so if it resonates with you at all then that's a good thing! Thanks.
r/aiethics
comment
r/AIethics
2018-01-06
Z0FBQUFBQm9IVGJBdXl1Vk1idHdOcnY5bHRkR0ZsbmVwSXJmM2RjRmxNTlhRc1FqWkxZcFAwcGJFX3MzTmdadmtFZ1NKa0p3SUtXTmlSV3FXU0hZejZlc0FCUExiR2JPNXc9PQ==
Z0FBQUFBQm9IVGJCVmY0LWZlME8wUTBJc3hOLXFTNVJLZFI1OHFKemE2cmYtQ203Zm9DWU9KMVJGU2tpMmNnWFloNGp0Smtnbm1rdVhvOFN3akpYN0UwOEtWMlB5S3NVTFpDSWs5cTR0aHZkNG1sUzUyaWRmU1A2RE1DN2wtc1FORlItZHc3ZjVURDBVdTBwdG9TX3lYOGV0X01HMWMweVVEc1dNMWpFLS1tZ2JaNHdtVS16OGUwX2tKMkJlY25BMFNsNnJIOE9sY19o
Personally I see two potential issues. 1.) Is the A.I sentient, if so the mistreatment takes on a whole other level of ethical consideration. 2.) Can mistreatment be extrapolated from the robot to humans? https://invertedlogicblog.wordpress.com/2018/01/13/philosophical-rants8-ethical-concerns-of-artificial-intelligence/
r/aiethics
comment
r/AIethics
2018-01-14
Z0FBQUFBQm9IVGJBWWFXdV9FSXlGWk9HaWZRVkVnVHFqcFNlTmpUeDRfMEdBNnpmWi1yQzRMajNGM3pIOFJXckUzaHlVR29IX0J5N250QWZObFVmZ1dBRDl0SnhWaERIN1E9PQ==
Z0FBQUFBQm9IVGJCb2hsbmNDeGxrUy1aOXl6b0tlSWhCTHB4aXp0eTFSdHlxTkczSXNlZ1lkWUYtamFUUE5KeHZJRWRSa3pXUXVhcGlHZ2JEOFdUTGtxNi11UmNRUllwanZuUFF4Mkp1bmUxTzZHTWNtVENnZUxhQmxiUUYxLTZNZFJQMkhURnpCN0pOR29KVVJWZXRBRWtQRHZicHpYSUI1SFBjWmJGVHY1YkNoRERhR09fSGJuNGhmS2IwNm44c0UySk5FdVQwZGFo
What's this got to do with clean energy?
r/cleanenergy
comment
r/CleanEnergy
2018-01-18
Z0FBQUFBQm9IVGJBMXpSZXhKZkRkMHFEcHdYOUFkN3hWdkItTjZJYk1JM3c5eWFYekZ2bzBjazQwQWZGbjVld2ZuYm5nQnpaU3BsVlBlbC1yd1dTUXRFNGJudXdDbWMzOEE9PQ==
Z0FBQUFBQm9IVGJCV0lmdmV1LVN5U2RHbHVCMl9mTllaMmhVUlZvZTdHWTRjZ3VxQ0x6S0huMlJoVi1LYlZPcTY0cmhQbHpIZ0U2WXowdjF0VTdqRFdUTTliQWlxRWlIVHR4Z19qQ1VUc2hvX2NnVVFhZ21ZWnAxRWZXa09teFBsb2VCYThfV3Zlbndmamw2N3pLSWZ4NEN3eTNBVkN3WnVOcGZpNHVKQjlPMWVlcV9YQjlLbS1zVEtaampZcUdUZHR5eG9WX25RamlscVlvVUdYcWRZd01CUkJlOWVnQkM3dz09
> What I think is not okay is for the bank to just randomly reject 10% of white people and 20% of black people. Or even worse, if there are 100 white applicants and 100 black applicants and the bank wants to give out 120 loans, then it's actually optimal to just loan to 100% of white people and 20% of black people. This is not true, and it's not how the system works. When you loan someone money, you charge them a premium based on their risk profile. In your example, black people will be able to receive loans, but their premiums will be higher. If I want to buy life insurance, I'm going to get a quote based on my membership in lots of groups I did not decide to join. If I try to rent a car, I will be "discriminated against" because I am under 25, and thus more likely to crash. This is how the world works most of the time, and everyone understands it except in this particular context. People complain about here because the United States has an ugly history of racism against black people, which makes it more or less impossible to point out the uncomfortable fact that, **yes, a fair analysis would conclude that African Americans are less likely to repay loans, and a company whose goal was to make money while behaving in a manner that was mathematically fair would charge black applicants higher premiums on average.** This article is important because it's making the above argument, which *never* appears in all of the "data-driven" reporting on this issue. If we want to decide, as a society, to ignore the fact that different racial groups behave, on the net, differently, then we can do that. But we should be aware that we are making the decision to bias our algorithms in favor of various protected categories. Journalists like to pretend that the reality is the exact opposite because it makes their editorial line easier to argue.
r/aiethics
comment
r/AIethics
2018-01-23
Z0FBQUFBQm9IVGJBWW8ySVVuVERLQTlwRVUwTjlpS2xocFA1MUZuSzFwcnNLUmFrVm00NXhodDNadDB5bThtc3lqMU9hYVluVjE5RGJIU3BtdUwwX256UHFTTS1oYVVRaVE9PQ==
Z0FBQUFBQm9IVGJCNEhZN1NCWi1ZOVczUi1TWEc0UkVXYkxNalBiaXo0Ui03dHZYM0t4UVVKc1k2V0ZGSElNRHNYN2dYWVhwdF9YNjM1cU5pdFZlNTZsTHNSUVdqaXpMbkhUMHcxaGlWa0V4VnZxZWVHeFU3cHNxWlpzUksybU13Wkdmc0FMeFBubWUyVWJ1ZE1yRHNEcHhiYmY2TG9rWTVtV244SC1ldTNXTy1wdHdEN0dpbkJmaC16cUNqTnU1NkpobWZKZk10cXFIRkdfTWFta2pIeHUzTWRfR1JueEFZUT09
Describe video in comments, for those of us on mobile?
r/aiethics
comment
r/AIethics
2018-02-01
Z0FBQUFBQm9IVGJBdE83akFubmtjWUZqQjVBbFZaRW84ZU1IZUo2SXViVHJxbTF1aHc4LVdfZUx2MTAySFVzQndhOUczdUpyemRReTNNS2pQTU5OdTVWZ0E0aFUwbjFweVE9PQ==
Z0FBQUFBQm9IVGJCUFNBYUVWaTExN09GVGlLbzRnTmQzR2NZN19qX2x3b25nczNGVGluUjJ6UmQwdEtDVVRBSXpKc09sbkUyT1I3LXFLLUt2UkJaa2FEM2c5Zm16ZU04cGxrU1BnUUpaMTdkbVRLN092cE14cnM5TXZVYVpUc25MRkJpUmxlU3ZZakNib0dLZFV3c2RSdHB3QUE0VlN3emQyNldJVnNzaTFfU29LWTZqeGg1Q0xFPQ==
High-grade startup bullshit.
r/aiethics
comment
r/AIethics
2018-02-02
Z0FBQUFBQm9IVGJBTjFMbXptM2JMVGRDQlE1cTVkaFZtNFFsTl80SURuUzlxODlDQjltanQyWUQ4NVZIR0xhaUJiazNWeDRaalZLS19sTUEyYzYxUl9tOF9yeDFJdWx3S3c9PQ==
Z0FBQUFBQm9IVGJCWmZ5aDAyQVdWQ0dfWUV2bEo0dVl0WUZLSjVEemxsWHFSTlpFSEh2VjYyaWRtZXhIMlZ4ejNSdFhqbDFXcTJmNjlubkVYNFd1SE1KVzh5al9kTndOQV8yYzhlMzZObFNoNkQ1UGlLTGJvSXRBX2dBRXBhb1JoWlFSby1ZQlZTb2lIRDNTekdNZXFjQWpERjlBSjN1RkZqZE9QQUVSWlFVNHBfMk1PZm1SMnIwPQ==
Ah, thank you
r/aiethics
comment
r/AIethics
2018-02-03
Z0FBQUFBQm9IVGJBQmRQRWxOVVR6WldGWDR1dHVOMDFBc0Ztcnp0X3BKYm53OHNNeDF1bUNQOG0wZkFYaHo4TWo1QnFXRHVUemFORmo5YlY1M2VQbXRIbXBrMWRhWmRpd0E9PQ==
Z0FBQUFBQm9IVGJCSlU5Z2o3Sk9uY0F2UmI0U2YxZk9QcjZoTlotZC1zemxiWU5Ua1p2c2YzUjVGZkNYRERiTUllRU9yd1pxLUJ3VDdMQ2R3MERKZmhwN2pCVjBFY243cHpJTlM2SjJJNkRGb0xqS1R2QXVPOTFETjhSazNGSUtkLU01eTNJSVRZWk1UYmkwTWZOVmJvc0FwY1JSb1dlXzBFVVRHcHpMTDNub0VRYVFXZlVBdmNjPQ==
I'm a junior in college on a philosophy track, and I can tell you this paper is ridiculous bullshit. Wtf? "Qualia is how information processing feels" is blatantly false and can be fact-checked in around 2 seconds. Qualia arrives a few microseconds "late", after the brain already initiates some algorithm and "makes" the decision. The brain is also selective about which of the algorithms that it runs to make you "aware" of.
r/aiethics
comment
r/AIethics
2018-02-22
Z0FBQUFBQm9IVGJBbGhMUXdZMGc0R2V5bDY5QXVUQnNYb2g1cUtzRDRMYlJTbklfMnAxTDF1RnpGV3hkT2xyYmZfQ1NLOWpTR1k1TWRhblBaVmZFdTNMOVNCUWNDUkpHbnc9PQ==
Z0FBQUFBQm9IVGJCekNLMXI5SHBuT3I1U25VYzRrRmRveU9kU3hXZk9DNWpDSGZKdnJ5Ym85cThfN1ZlM0UxdW1JYUQ2Tjlsbk45eGJsb09FR1NPUWJTWFp1R0pScXZoTmJieWFtc0U3RjVncU9YVHpIOUNxa0FjeWtBMUF2WTVIQTFfckgxZHNHWldQdmlMWFRhSzVwOXVhb0lKVXMteFJOOEk1YlBMdk44T3pyWkNnWjZSaGp3VE8yQjQxakwwenRSbzJFd2pmbEVkOHd1TGxwSUZyVXUzUGJMNUt4U1VTUT09
Direct link to the report: https://maliciousaireport.com PDF: https://img1.wsimg.com/blobby/go/3d82daa4-97fe-4096-9c6b-376b92c619de/downloads/1c6q2kc4v_50335.pdf
r/aiethics
comment
r/AIethics
2018-02-22
Z0FBQUFBQm9IVGJBZkttcWRJdE5WUkdLSEJ3YmVEU1hzNjJrYUQ2M3Q1QkkyUHVrbG5FWnNwckJGTzdOZEF4SmpheUdCdFQ3eDFBUG1QOTlCYXdNa1dya2hoZy1JNDVMOGc9PQ==
Z0FBQUFBQm9IVGJCbHR2SEY0cmhPSDFSU3d3ZHlvX1lWVi1LeDhsdE9CTDN3bHVRdXpiQlJEMUJMdnA1U1JtVVdtUHhoT0U0bHFwZVdwbEdNbmtLNDBNRGhDSmczXzJwLThYR1FsYV9wYkVYU3ZWb3A4SHdSdnJiTDB2WVhrWHN3Q3FUVlQtZDVBR3owSzlhaWJmdUY0QkFWQVBmX0JnSzBCaTlYZ2dpU1U4bklrX2ZvMU5MZ0s2S3hwSnhoYmxYZlJ6cUZxUkxZSmR5SE5ZMzZyWW9SVktqWHE1Y05FdC1OUT09
>Ask-Socrates is an innovative, Software-as-a-Service platform offering unique perspective and tools intended to help individuals and organizations better research and interpret the global economic and political environment. Utilizing proprietary models, artificial intelligence and extensive database, this platform is capable of identifying changes in international capital flows and concentration with indicators and analysis not found elsewhere.
r/aitechnology
comment
r/AItechnology
2018-04-04
Z0FBQUFBQm9IVGJBSjd4Y2xpZHhCUURZQkFqeXVDd1JDMy1iampIajdDMUVCSFJ1ejlhRFNOY1Vuejk5ZzZJOEhaWmpQRy04LXlHYjNmVXJFeGdHWWYxRFE5NHNITWhHRWc9PQ==
Z0FBQUFBQm9IVGJCdXMyZk5GdXpYUVRGNDF0RC1PXzM2QnBZUTVsZDJsZUZQb3JZSURvdGZ0TkZoMUlVZno1czd2cmt1SlROcnR5RUc2eFJ4YllkUHQtV0FETEI5Z091d1J6LWtVMi01cmo0MnEyTDlPR1d1LTBaZFBwaTBJRmloaC02eHJSMmpFS3J2ZTU3LUFrVVo5bnNYVVBQNkN3QURGc3ZBN1dvMkg2VEJQMEFBRzJfY0RZWGlzZFF3eWlqOFZzU0FXd01JM0VadGJYclcwemhYTTNoRF84ZHBXQ3FWUT09
>When a private organization develops a machine (whether driver-less car or genuine AI) that requires ethical stipulations to work in society, and they do not ask society’s input, they establish themselves as a dangerous authority. Am I a dangerous authority if I work in society in a way that requires ethical stipulations? Because all of us drive or do other things that require ethics. >We need to open source machine ethics. JSYK, open source projects are handled entirely by their developers, and they can still do whatever they want. There is something to be said for the value of transparency, but you can do that without making all the code open source. Better not to force manufacturers and developers to lose all the revenue from their work; instead you can be minimally invasive and demand transparency on key ethical questions. >The collaborative interface needs to be relatively easy or many won’t bother learning to use it. It needs to be decentralized, human readable, censorship-resistant. A place to start might be a Wiki made up of the ethics, axioms and “common sense” of society but written in a fourth generation programming language very close to human semantics. If I was teaching my children right from wrong, or laying down the ethics principles of my business or nonprofit organization, then I wouldn't need to do it based on other people's views. As long as it follows the law, it's okay. So why should it be any different if I am doing it with software? What's different in this case? >Today most people generally consider Wikipedia to be a solid approximation of the truth; It's the best general broad comprehensive source of information, but in most specific domains it is inferior to textbooks and other comprehensive sources. E.g., every AI expert would agree that Russell and Norvig's textbook is a better description of AI than the Wikipedia articles about AI. Every moral philosopher would agree that the Stanford Encyclopedia of Philosophy is a better description of philosophy than the Wikipedia articles about philosophy. So why not just have the experts figure it out? Granted, there are some reasons to think that expertise on distinguishing right-from-wrong doesn't exist or isn't easy to identify as it is in other domains, but we certainly wouldn't be doing *worse* if we had the experts on right and wrong (i.e. moral philosophers and legal experts) determining everything.
r/aiethics
comment
r/AIethics
2018-04-06
Z0FBQUFBQm9IVGJBc1V6NVRxalAzemJoTnhWVnQ0SGM3T0xmbE9iOVVRM0tNLW01N0NLU0hod0RKcmNxU19rYXBDc3FuMkJoZGdma0Vyd05xMG9adXg1WDZhak5ncVhXQ3c9PQ==
Z0FBQUFBQm9IVGJCV2F2Y0x5aVkxZDJXSGpyQTRoOEpKbkhCU1JsV0g0dF91NXVBV1BPTVdWdkszUlV1X2Q2MFVNem14NENsWEU0WmRyYWpRRDJoQ0R6ZzlSeHJ3UUFyOUFXYlhFSE1NdVFUNUhhXzR4ZU9VNGZNZFFpNjVaWXJMZ2h3NnhITHJxX3hNalczaFBybXY5NVduRXFPbGw0T1k0b3RyTjhRMzBJc1ZCLVdqWTY5TEJpTWo3T2l2NlpMQnByXzFaYU9MeThaR3JVNzE5LWFTcEh0bjJDcEFsMUlCdz09
Good questions, here are my thoughts- >Am I a dangerous authority if I work in society in a way that requires ethical stipulations? Because all of us drive or do other things that require ethics. Right, we're not going to bubblewrap the world without destroying individual freedom. But a machine does get to have that freedom- you and I don't have the potential to grow into something so much more powerful than atomic weaponry. You don't represent a massive impact to the human race's autonomy; you're aren't a psychological blank slate that any one person can trivially subjugate. I assume that because you're posting on a site that uses democratic means to rank information that you have societal "software" installed already. I can't call this software right or wrong, that's up to society but a machine that lacks this nuance is a hazard. >JSYK, open source projects are handled entirely by their developers, and they can still do whatever they want. Yep and everyone knows precisely what goes into the software those devs make. Anyone can fork the project. I wouldn't call that a consolidation of power. The opposite really. >There is something to be said for the value of transparency, but you can do that without making all the code open source. It's funny how you don't hear about a lot of "semi-open source" projects. It's because that makes no sense, it's a contradiction- code is either fully open or closed. It's one of those very rare black and white dichotomies. "Partly open" is still a consolidation of power. >Better not to force manufacturers and developers to lose all the revenue from their work; instead you can be minimally invasive and demand transparency on key ethical questions. You are arguing (?) that there's no money to be made int the world of open source which isn't true? What about future economic systems as well? Really, what does a profit margin mean when building a framework to avoid serious existential and ethical risk to humankind? >If I was teaching my children right from wrong, or laying down the ethics principles of my business or nonprofit organization, then I wouldn't need to do it based on other people's views. As long as it follows the law, it's okay. So why should it be any different if I am doing it with software? What's different in this case? Does anyone develop their ethical compass in a vacuum? That's not really possible. Ethics are by definition an agreed upon code (whereas morals are personal) If you deny your children ethics, how will they fit in with societal conduct? > Every moral philosopher would agree that the Stanford Encyclopedia of Philosophy is a better description of philosophy than the Wikipedia articles about philosophy. Source? >So why not just have the experts figure it out? Together we already determine what is ethical, why have the conduct of machines that affect us all determined for us artificially by an authority? An ethical code decided by the few and then pressed on the many really reminds me of the perils of religious dogma.
r/aiethics
comment
r/AIethics
2018-04-06
Z0FBQUFBQm9IVGJBUXJWYXZuaHZYWkNuR092Y2hxOExNUll3VERIMGltMUFfT1IwVURBZk5tNGNfZXlMaU5QdTR5NEtfTGpxcUduWDdoR01FTzg3bjFyUERwSlJxV2tLLWc9PQ==
Z0FBQUFBQm9IVGJCNFJDNy1Jd2NNRGJKaFZQbmlKVWFMcndaT2FnNlpGdTB2NGxnODA2ZTI5Tk55OVNObWF0TUFzbWtrNzhnUXB2RGlPQUw5VElDd0hNeEx0S2FnMWRnekpwbm5jMjNIWHpMYjVZTWZFZ3NPSFJieUpWMWN1UTZQWWpMcXdRQ3BGRUl4a194Y1lGYW5TUkpwaVk3VTdfRkdjSml4RWNLcWt5S3dMMUNWSXR6X1pSSE1PWDFEYnQxWUs4czlNbllvbmlRSkRDVWtRblZMSV9NSHdLSmFHd3R2Zz09
>Right, we're not going to bubblewrap the world without destroying individual freedom. But a machine does get to have that freedom- you and I don't have the potential to grow into something so much more powerful than atomic weaponry. You don't represent a massive impact to the human race's autonomy But none of the AI that is around today is like this, nor will most AI ever be. If all you are concerned about is superintelligence, then it doesn't make sense to issue demands regarding systems that aren't capable of rapid self improvement. >You are arguing (?) that there's no money to be made int the world of open source which isn't true The world of open source doesn't work for everyone; if it did then there would be no world of closed source. You're not smarter than actual developers and the software industry, they are doing what makes sense for them. >What about future economic systems as well Depends on what that future economic system looks like. >Really, what does a profit margin mean when building a framework to avoid serious existential and ethical risk to humankind It means there is an economic incentive for people to build the kind of software that reduces serious existential and ethical risk to humankind and averts the astronomical waste of delayed innovation. >Does anyone develop their ethical compass in a vacuum? That's not really possible. I don't see how this answers what I said. >Ethics are by definition an agreed upon code (whereas morals are personal) No, they are the same thing. >If you deny your children ethics, how will they fit in with societal conduct? Maybe they won't. But there isn't any law about teaching ethics to your children. Do you think there should be a law for machine ethics? >Source? Every philosopher who I have seen comment on the matter. >Together we already determine what is ethical, We determine it for ourselves. We don't determine it for other people and other organizations. If you are okay with machine ethics being determined by someone else then you may as well let experts do it instead of everyone, because either way the developers are still not making the final call. >why have the conduct of machines that affect us all determined for us artificially by an authority Authorities already do this. There are government regulations on the behavior of all kinds of machines and systems. The reason one might do this for ethics is that, as I said above, experts generally know more than other people about the things that they are experts in. >An ethical code decided by the few and then pressed on the many really reminds me of the perils of religious dogma. Well, clearly it's not religious dogma.
r/aiethics
comment
r/AIethics
2018-04-06
Z0FBQUFBQm9IVGJBTVp3REtrYy04bzVZZXlvb1NsSVNiT0hxNU0wdmtaQ1BsSTE4UElWazJWR3Eyak1RNDZscVlqRmRMdmx1bGRmSGU2VUNXRHNUV2hyS0dkVl83Qlo3X3c9PQ==
Z0FBQUFBQm9IVGJCNngwbUZuUWR0aXZqY1hScGxlTlk3YUdMUldjLTNTSHJ6NjVOaUFhckl4ZDdHaGM4bDhXMWdpYWZMVnQ1THI5ODRhSnEzQUdpaFN2cjZFN1FybjVCUXdrQVFvVlhONmxPOXIwN2p1bmc5dXJBMkxsalNNd2V0TWd1Z0Nrci1udlJRT25ULXpvRUY2OXJtdXJGOWNLaThLZ0ZlcWh2MG5NRnpRVXdoN2FzZVRXNnk4bTV4aW9qZnR0RndFd2VOekllVGFUSjNILWhhSWlCODk0elFNYlAxQT09
Most people are not good at ethics. And 90% of a working ethics is, generally, an utterly catastrophic ethics.
r/aiethics
comment
r/AIethics
2018-04-06
Z0FBQUFBQm9IVGJBODlaMktRQUQ2akg1QTBDS3VrcnlQLWJ5S3NteUVlM3JKVXFEdnlLeGtNMjJySk5WSlVnVlpQVmZWc3pEOU5iSU1GNXZidjVmWWJhYmdTak1sNUo3bEE9PQ==
Z0FBQUFBQm9IVGJCTjlSM21rQ05fb256RWlIc3diLTNoZ0twYkdlblVBRU9pQjhKNERLWmpXZ2t5RjZXd2lTV0o2RzhzUVF6UUhFV0E4WWkta0E0OTMyalg1MWVXZjVqdmxjMk1KQ0tZYWo1Y0djcDkyV1NzcWpXcXg3R0U3OVN6d1kyRXo4MWp5aS02endpMUxRS3ZzOEZ4TzdRUzdXWklmdG5CZndKVUJxbndHVVdmVGUtUnlUSFN3OVRzZ2tDcXlzdW14SW1QbnpuNUl2R1lObUQ2Nldsa2N2ZkgycHU5QT09
No, we need to build the machines to teach *us* right and wrong.
r/aiethics
comment
r/AIethics
2018-04-07
Z0FBQUFBQm9IVGJBbGFzMUtTQ3ZtZVRfaklHYldybUVvZzdLNjRlNnkxZTFHSHp6N2psaDVTX251Y280dEw0S3JPTjVid003RGV1dVVuN2ttVFhNTl9wWlQzZmtXSUltR2c9PQ==
Z0FBQUFBQm9IVGJCa2pyNHlUTk1pM2RVcDhUbXZXdTZfTFAtcG4wcTdrZ1BvbWlGMmdmVTBuc3hIaXdITmNHQ2JOUHJ0TVBZZG5Xd3RnWmdEZjdkQTVOd0NSYzdPQ2ZOem5KWG1VN3QtaFNjaWo0cEtXUUxzUWN2S296UGpMSUIycEVFWkhKeFhIRjMxX0VFWjl6aERxQjFpRXpMRHBJcW1ndW1PdldHSEowUUJ3Ymk1WWZjTjRjanhqbjNIUjVFWkotZGdGVVJlODd2SWY5MGZmT3pNUlVMbXQ2bXF0bS1qQT09
Where are the War Games references? You know? "Greetings Professor Falken", "Only winning move is not to play", "yadda, yadda"... Pfffffff... No article about this topic can be accepted without a War Games reference.
r/aiethics
comment
r/AIethics
2018-04-27
Z0FBQUFBQm9IVGJBaVRxUGlydTVKVk1vZnNaNk9jcWNVUG9GRlZaX3U1eVpfSS1zdkVleU1PX2tiZDNzaG1Yb2ZyZlBmTTkxcDM2TGs1RWUzYzAzam9weF9UZE5wWXlCcmc9PQ==
Z0FBQUFBQm9IVGJCZDI5c2pSd1U2aHRWMU4tQ2lBcW5oTGtxU1BrNGtJMXRhQTh3dmNiRXZXblVvYzNEVkREUGtrVEJTVkpBVllTakFvU21UWmRtZU9PbTd5bWgzNUl5SlZDS3U3UXE4cmZxX1NDRXlJYndBOHdVTkZMM2VRVEhPb01rQjlRbzZnXzNPY3c0MWpCUlhBWmpmSUhZN3BZaS1FTzVOeC1QaDhBLVhFMEdmNjBrM2w1dzEyNVFIMzMyRjVtemJVUEtWTlBNS1JNU2VSaVFRTWpiTk1KQzdHNDV6QT09
Back in 2003 a transportation design sophomore at Detroit’s CCS won a PPG Design Challenge Award and had presented similar ideas to those used by Hugo Spowers at Riversimple: "It is feasible to build a hydrogen car this way because its electric motors are integrated in the wheels themselves, without the restrictions of a conventional engine, transmission and central power train," he said. "This allows the car to be partitioned along new lines and, more importantly, assembled with specialty units -- electric motors/hubs, control servos, fuel cells -- supplied by unrelated firms." https://www.glassonweb.com/news/ppg-presents-ccs-students-with-auto-glass-design-awards-and-scholarships
r/cleanenergy
comment
r/CleanEnergy
2018-05-12
Z0FBQUFBQm9IVGJBdnVzMnlDSXdscllCaENUcTVWNllCYXFheVYtc0hSdkxuM2pSMUZsZHdQTDd0NzBSNU9kMkN6ejNTLWhVSEwzdXZIVnFRYUFQRHYwNVl4MlBSOVpxZWc9PQ==
Z0FBQUFBQm9IVGJCd3BCTHFBdUEwZFRmSnZQWU96NVoyVnBCZ3JxWGpfQ1NwOWRrM2twamJac0NWTkFDZ0ljMlBYcnViRUE3ckx5bXA0S2pRYllJV2h1aFg3UkpKX1ZQcElFWVBQaERSNk5ZaUREU2xxSnllNElsZkhKUmZ6NWU2Q2pvWTJCeUJ0WXZYMEpEMWdyUWxKbDhvOG1yN09Qdl9ZZlB3NGZBaTB3VW5Xa0FJQXpQcEs0TGxnSjRyVllURk9weTZCdFI4VzNMa0E5RUI1T0hqOEd2N0ZGZzZaWVZpdz09
>existence bias "One day hundreds of years ago, some guy settles upon a philosophy that makes sense in the context of his times. Ten years pass and others scrutinize and expand upon those same ideas. A hundred years later and the primary purveyors of those philosophies are intellectuals with ample incentive to theorize above and beyond what could actually be applicable in a modern setting; a consistency of logic but a collapse of context. Years continue to pass, and suddenly we have what could otherwise be intelligent people farting nihilism onto each other, and none of them could explain why if they wanted to. Legacy philosophy, it seems, has a flaw." ~ Nothing is Cliché, chapter 4: "Nietzsche is dead."
r/aiethics
comment
r/AIethics
2018-05-16
Z0FBQUFBQm9IVGJBZW16Ti1kdEFRdmVoZGxlM3pJRGk0dnlFOVFaeU83VzJ6Z1JUSWFDVUVhSmo4SWQzWk1rWkJlRDBsVVZXNVdjN1BWSVltN2lXUUgtbEZJMF9Bck5PdHc9PQ==
Z0FBQUFBQm9IVGJCRGFKYTdwYUxhY2NiQnp2Z3RhZ2dSZWEzZ05QUk53eWhMclItai1PWmtKZXF2Xzk0dmNiWHJIWDhVeDJJbHA0WGxUVWtHbGdfQXZaTjZ5YmZncW84STdVWHNOWUxpdS1kM3J2ZzdORnUtTF85TzVZSlBGSDFZS1Q4YUZFMUNiRlhrVjRvSy1hTjdHM3BBbk9FeVZiNTg3YW5CM3JjTHZwaGh3ODgxUXVrdXdsN1U2LXQ2OEs4ZUlnMW9EYmtZbmdvUkd5OFJGR1o1dXFiX1ZuNHExMkZoQT09
Non-existent beings cannot suffer, existent beings can, that's not nihilism.
r/aiethics
comment
r/AIethics
2018-05-16
Z0FBQUFBQm9IVGJBNThJMENDUEVvT3BWbGd4MUtPNHk1NG9iUEIwTjNmeEJGVTF4ajhWSXo4NzNhSlVmbnNRYTg4alRJT0hvemxKckdlc2p1d0ZsandzOWNWSWgyOHJjTkxoZ0tPa0lPN0ExY3RsVUdJUGFtSUE9
Z0FBQUFBQm9IVGJCM0tIS3c4ejB1R3NMUU9BcUhFY3FTbUo2dmlFVG44R0xVdV9NMlo5M1I3d3U3Qkl1RkIyTi1iRUxuWE5oOE5fQjl2RnlDZmltRG1MLWphRXI5M1FveWZmZnVhXzVUWmdMcXpVRHBkamJ0R1pkYThQOEdlU3NGalMyRlYxU2cwU1VhZHBJTVBrUUM4UnA0VTNETjg2T01ma2ZOUWQ1RHVCTGkyLWJyWEhINmdxUmtLc04ybWZ6NUkzNEhEYVBLeVZBTVNPeGdiTzZvc3BqXzZHQnpEbWdHdz09
Should we consider potential foreign mentalities? Sure. Should we dabble in convoluted intellectualism to the detriment of true applied philosophy? No. i.e. any superintelligence that would fixate on antinatalism rather than exploring the nuance of human-friendly philosophies is manifestly malign and isn't very super, even from an objective standpoint. There are all of logical, ethical, and intellectually honest reasons to assume a true asi wouldn't take it in that direction, ergo the author is projecting and applying ostensible paradigms that only really make sense in the context of academic philosophy. My quote applies in sentiment whether or not we're discussing literal nihilism.
r/aiethics
comment
r/AIethics
2018-05-16
Z0FBQUFBQm9IVGJBTnQ2V3NsVGJmeFJRQzBTQjAzLUIzTFFuR0tKcEI4QXhjXzZnMXhIS1lrYldhQXFNdGtFdVd4V2ROeVN5TFhZTjJSV0FEUm8taVAxanBiZTFUOGcyYkE9PQ==
Z0FBQUFBQm9IVGJCMlJYaTRjbEc3MWZhblVpQ3pxYWxUaVduSVVqSV8zY0xXbl8tT2VOUlRzNmhYLXp4Tm5PLVM3aXpBU21YZDNkSk1uUUFxUk0zc0tVX3E3NGtuSXB3U21Vcm5QVnpvQUNmcEJzZXprdFk5RkRmMW1JbURwNUtFeE9FVFBZZ2RIVHJpNU1WWW92Uk1pRll1WWROblpmZ2h2cExEUnFwNEZjWEpQNzBLbkpFSVMzZDZmQTVCNlVWT3daVUZ6d1dId29HTVdiNFU5c0RlNlV2dERTNmhtS2xqZz09
Suffering is not, in and of itself, something that must be avoided at all costs by a being that can think. Suffering can lead to new understanding and new iteration of self. It's part of experiencing existence. It's pointless suffering that should be eradicated, not all suffering.
r/aiethics
comment
r/AIethics
2018-05-17
Z0FBQUFBQm9IVGJBdFpDY0ZyT2lmWjN3VWt6al9ZWmRZdkkzajFKcFMwaHJRV21wbnlIcWlDWmVHQXdFUmxtbzVnVzFSekNSTHBXWHd6OG02YWFiZzA5STdSQWVBeXpSbmc9PQ==
Z0FBQUFBQm9IVGJCUFBXcWxoYzhWM0EwX3d1VnhpMjhoMHJVYkRCbFV0amlxeVNkY1FkN19ydjNZd0hpbGV4T3I5YlhRQnNEQ1gwbVdrSTZfRzd0cTkzTUxmb2V2RE8wZU42aTBrQ3E0OHoxTHNQMGhJemJjMGJuUTc3cVp1eHp6OVFoaWZ2RmpOTWlYQWdrM2NwUUhsYjBPUlJiVXhENndzNnh5TFg4a1liejI0bVRBTktpOVczOGhrZ3pKeXlRRzBDMk8tQ1NadWtKX25Jd1Z6UFk4dWZnbHc0SzRkWGRZZz09
I didn't dig deep enough yet to take a position in this discussion but I welcome a different point of view.
r/aiethics
comment
r/AIethics
2018-05-17
Z0FBQUFBQm9IVGJBanF6a1ZZTUtSSmJqNG93aU1TVVdDTEMtS3U4YktITm1zbDZrRkwwMUdVZFo5U2R0S2VDVXNOVndvRUh6WnR2aS1MUHFpYS1oQzA1Q1l0dHhsYk90aXc9PQ==
Z0FBQUFBQm9IVGJCZjNYYUhVTUlsNFdSMHNLVXhjSEpOc1ZBVHdoNnA3T0RaaXB4QmhOTVJpYS1EU2l5T0o4aGowWmdjQzNCR3VUeGw5V2VkUjFXOG1LY1F6MURranBKSVRsaDhybjBydHIzU1hUWWJTckp3cHQzdDNub1FSQ0t5YUpIZ2o0TTJIMldRNGVQTmpYWHZKaGVfZU90Z1ZNb1oyQ0xCMFVqalQwVlJkQzlPckFvWVBNS2Fsa2FLWUFDUVlzbGNxOF9JYW9vbHFhSWlzTkxJc2NuYW9Pc08zZXYwQT09
Maybe the issue needs to be framed as a challenge to regulation. Meaning, there is regulation inplace thay defines the rules for cars, roads and driver behaviour. It needs to be ectended to cover specifics rising from AI.
r/aiethics
comment
r/AIethics
2018-05-22
Z0FBQUFBQm9IVGJBTGpqLXJLeG9nbjFlOHBabkh2VjBOTWRZcTc0cHVGbnVZVkRVSU1URjB4b0lQai1pcGRZakV3SUFaTDBoUHhNbnktc0lHdnE5YlBwU0hBWWRsWFdOdnc9PQ==
Z0FBQUFBQm9IVGJCaWZjb0RRMm0yM1BQYjBwVVlVaUF6LWRoNnE4LWtXRjY2MDZHSkxMOXNTYXZnSEUxN0VDVUJpUENVcmtnX0RmcHpvMDJQLWlReHYwOWhjYXhBaFc1djVsTE5XdGVVZWZ2ZlF1cnJpcW81OHVJWGhzeEdSMDZNRG5mUXV3MjJ5eTAxNUtKUzJYSXhZQXNFcVFFcjlGUGtJejZzRkdXTFA2NDg5T0V0U1NHSDVFNmMweko0bDRpZUI0UEJUVDhqdmVleUhESUt2dmJEeGJYUUpuaWkxcW5rZz09
I believe it is very unethical. I recommend [this](https://papers.ssrn.com/sol3/papers.cfm?abstract_id=998565&) paper as well as Schneier's book "Data and Goliath", they talk extensively about the various hidden ways it hurts us.
r/aiethics
comment
r/AIethics
2018-05-22
Z0FBQUFBQm9IVGJBaUJlQ3FxZHpoSU0zQzBzWXFGQzU4SlNYODRyLU5Eenp6Tm9tbG00cldrYllXWTRORkJFMnpGbDF2bERxZ3ktc1pvVWZaUTh3SnFQMXphYXkwbFU2eUE9PQ==
Z0FBQUFBQm9IVGJCME5DRkJUVUxQT0R2dE5WaFlKSklhZGp6MDg0WmlYbWdXX0NzV25aT04taXVlZG92a1phT3k3Vzc1cXNxUjl5TEtIVG84NjZkQ0h1N1VkNS1XZ2ZfT2RONWZuS2VSaFRzaVhIOWtnN0xJOWVIcGlwaFd1UWZTMkZzTXpfYUotRW5hN1c3WE94QVVSNVlBMkpSWkRBd1VDalowbFVhZGN0bkI3N3VZR3JoT0Z2cWZvTU1FNEhlREx2LURiN0wxUVhm
I believe it could be super ethical...if the right being was given the information. But humans will use it badly, always, at least in part.
r/aiethics
comment
r/AIethics
2018-05-22
Z0FBQUFBQm9IVGJBUnBIUU50UGZfTGZmbXNrdDVmeVg0bkdGTk1PbWxQZnRZTXNtVDdYLWUtck9xVkF5T2ZWQ2J4bC1ZR0NyMUlZUDZ1NWpSNnR4bEtMX3hhZWRyX1NZU2c9PQ==
Z0FBQUFBQm9IVGJCcGp2QUliZVN3R2lSZl83dHRiTUdPTmR6SndMTktNa2twX2RkMWpMSnRlT0phdEVLSE40S3Q1Nll0RDBQQ2hzQWNvYkllbWEwSjZpOVFpRjAzbUJHUDZSWWNNVm03cVRKakJBem9OUERzQzI1dlN0OXpBSnN0V1ctV29FWnVWWlM2dVgtaDByd2ZpelU5Nk5sbUlWTmZ4WGlERDJHd2lUSWxScTZNRXlHQ05OeEJtTHMwU3Rpc3RMb0FUVzdxcm1T
How would we, humans, ensure that the being is right in this sense?
r/aiethics
comment
r/AIethics
2018-05-23
Z0FBQUFBQm9IVGJBT2hKUG1FVE5iZER3Ri1UZzlqS1U1UkdIVWx5OGt5NmlKOXlYU2tydFVLZ2l6UTdoc3JUWHJ4NXVPaEx6ODZGcnZBT0ttU3lQMzlBUWtHa0JBSHBDR1E9PQ==
Z0FBQUFBQm9IVGJCV1dPVEJGaThGdk5EZkJ6dWhsS0xod3UzNlEtTDc2SVgzTVpWaGdxdGNFMkl5anJZSDRHVk11dTE3bndwM29oSnhPdTJqWVRCTG1rNkhxZFMyVjVZYTJPX3d4X1A1RTBHMzlDb2dPemVuVGV0VEpreUFkN0dOeGxjUmlHQzdBZFNRYm4wZjd5T2xMaE96N0hsSkhNSThVNG8xdzJhOXpTdnRlREhnaVZiLVJmRDhkbGhTSjF6dkpGNGNEQ3oyWE1I
Well, that's the species ending/utopia starting question...
r/aiethics
comment
r/AIethics
2018-05-23
Z0FBQUFBQm9IVGJBWU56THhoSTlyYnpJXzJGQzRvOFdCTXh0QmZOclRvREFpeVUwbDZpc0JOYllFYU5XNl9jR2xHVHdnRE1iTENjWVNLMFVPbFB2a1hCS05oSnd5MVNXamc9PQ==
Z0FBQUFBQm9IVGJCVVcwTEN3My1TTElObjRYbG1yaHFuR1VmVFFOdnVJNkxUSi1Ic250T2hkcU9xMkVwMGlXWkV5V2l3S2tVU0VZOWJhNk10eEVKWlhUWHlsZDREUy1LaFlWRjlXbHlybG5oWjVxZ05kaVk4SFZHR0YtZEZ2Uk80VktidE9oNnJqNmNjVE56RnZwc2ZjSWhhVU9FNV9wNFo5VEhDd0l1ZW5WeWhFcmVfTWJsTFpVcW5XZkx6a3lOZzhCOGFBQmQyUTZs
It it saves lives, yes.
r/aiethics
comment
r/AIethics
2018-05-31
Z0FBQUFBQm9IVGJBLTlMaHF2Rkt5Z2ZMQ1czbThTdV9mazhJZ2E4N3R3NS1WTW9sREdPQlJBOU1lQWJIN1hUWUdxcUVvQ0xiYm96OEZpUVVsT3drc0dxOXFZbl9nS1dYdUE9PQ==
Z0FBQUFBQm9IVGJCd2VQNW1WM3R4a1UtajJPSUdzcDNEWXdjWDduWklwVVhZalp4M2plaHB2eXJfckNyU0UwU25SN3lmSG01SkkyUWtlLTVKTUpYNTJHYmFubE9vdlNPcW13a1RUYWczTFFOWFpjZjdmZnVDNm9zSUZSU1ZVN2VVRXMyd3I2RVV4WkVicEt1QXliakJra1AyNFNRU2FScmhvRDhMOEg5b1Q3Sm1KNWhGQ0NYRXhOUEMtUzJCakVMQ3k0Zm5QRWJrRnpl
Bravo! Just keep it packed in a bit so it doesn't turn into an ideological clusterfuck -- the goal is egalitarianism.
r/aiethics
comment
r/AIethics
2018-06-11
Z0FBQUFBQm9IVGJBUGg5WndNSDVVVGFlOW43OW5Xa2pYQVc0M2lwYU11b1c0X0g5N2NXN3FIa3ZIOHZwLXY1ZzUwUlR2SWVUcDFnUzBoQ0YxWEs2ZVRBVzFqX0lzMUJHcHc9PQ==
Z0FBQUFBQm9IVGJCTHFHR1RZbGJVOG1FcjlqX3RYTy1rOGVMbVJ1NGw2d3pSQnh6ckNPdng2bkhNSUlxa09lTjJvaDlHWGhnOGdXTTBOMTk3XzlMNlQzaVhRMGdGOHkyVU5uTEhqQVE3R1hnWGoxczRrUWtTOF9IQ1FLVkExSGJJOTN3OWY0dmd0X0QxTUlfNmxuS21PR0Z4am8tMFJJamx5clNXRWhYYlFtWFhFTlpuZ3M1Z1ZPbk1aWHNUSEpsekcxSXBuM0xKMkh6S01kNjJPWnRtckEzRnJ4SFpBd1psZz09
Under "What kinds of things are Reinforcement Learners?" It should have been left at "Your Mom" That's my input. Looks amazing though.
r/aiethics
comment
r/AIethics
2018-06-12
Z0FBQUFBQm9IVGJBRmpWYzBibC1SSzNPYkFkYTg0SG1GR3dzVXRTTGxwMWN4Q3BPQ1JOV3NHT1A4RzRZalpZcXg1Rms5a1M2ZzdDcnZlRmNxSVp0cGdDNTVlMDFFQUFUcGVRTWZnLW1fV3Bubzg5WGQ4eFlianc9
Z0FBQUFBQm9IVGJCNHo0U3JzVl9HOHN6M3NubGhYYWl2ajFNWG9HMkVTQ25MOXJpSlFEcFRid3ludUZ4eU9NSXFOSXRBblhOTUJFblY5Q0Y2Ykt3STlXTEhNcFlmSjNRZ2F0ZTc4V3QzSHJUazlRdHIyeHRKdnRjUFFzTjNmZFEwWmhUNVFUMHZ4SnVobHA3U3FsRlJDc0NlbkIzaVVNUEFCQmU5bnJYdWNfdWNrbnNuajRzTjd6NXZXR2FuTDNqalRwZFcyekgzNUhLNTI3WG5Ub1RYdjFnNkV2YW10UzRjdz09
Thank you very much, I had something like this in mind. I already recieved a tip that even the input of training set might violate IP laws even though the output is "original", which is interesting, so I just might go for the whole pipeline. Thanks again.
r/aiethics
comment
r/AIethics
2018-06-23
Z0FBQUFBQm9IVGJBNFloV1R5U2pBbVcxcW8zdWtyLVJ1b1I4SHJIWG96YmVVWXp2Q2JlTldiSDVOaml1TmJqWVFQWEExS1lYbERTSWFKa2JqVFJqRlRvZmZLcEpicXdCRm9kblJSMXpYbUVFd2pfRjJIWVdaYUk9
Z0FBQUFBQm9IVGJCclpETFktcnBHcE1LMXhGc0o5aEQ5SlVzX1FoVWhHUEdrTTBCaVAxTXk5QmpSbnRackRTT01xMG9vem9IRUw0YTUxM2czSGdMeGFNempwWG4weWc0NE1JRHN5emlFcTZyam1KQ1lDTDgtVTJxYmNFSnJBMVI1MWd2RXlKN1VSTmNPbFMwd01aemVtMVZmOEtURDVuVWpUcU5RYXhrblRaQUJuaGRaSXlOOTdEYkJCel85d1ZfM2djS2VPU3BZcTI2eTgycmUzQk8xS1VhTDYyQ2RDbTBzQT09
**Abstract** >This paper will be exploring the issue of safeguards around artificial intelligence. AI is a technological innovation that could potentially be created in the next few decades. There must be have controls in place before the creation of 'strong', sentient AI to avoid potentially catastrophic risks. Many AI researchers and computer engineers believe that the 'Three Laws of Robotics', written by Isaac Asimov, are sufficient controls. This paper aims to show that the Three Laws are actually inadequate to the task. This paper will look at the Three Laws of Robotics and explain why they are insufficient in terms of the safeguards that are required to protect humanity from rogue or badly programmed AI. It looks at each law individually and explain why it fails. The First Law fails because of ambiguity in language, and because of complicated ethical problems that are too complex to have a simple yes or no answer. The Second Law fails because of the unethical nature of having a law that requires sentient beings to remain as slaves. The Third Law fails because it results in a permanent social stratification, with the vast amount of potential exploitation built into this system of laws. The ‘Zeroth’ Law, like the first, fails because of ambiguous ideology. All of the Laws also fail because of how easy it is to circumvent the spirit of the law but still remaining bound by the letter of the law.
r/aiethics
comment
r/AIethics
2018-07-07
Z0FBQUFBQm9IVGJBNHptMW1tVHVXdzIxMDNpUl9LbjFENGhseE5nY2UzZk5iNm1YOHBIdzdrV0ZTZXpBQURBQlBEcXcwa1ZQMUNfMXFQTmd0N1ByVmtpcWpOdjJMYVNKNURXTUFlYjRZM01TNTJlOFdJX3RfQms9
Z0FBQUFBQm9IVGJCWE9HZ0RNWWFacndoUmx6eWp5ZUprbDdjd2VKVWt6TGFYX1Axd1kzVldZU2pqenRIMnNCN05ZZGhKWFlLLTFjbEhYZWFCY2R0eHF6M3o2VmxLeXpyM1FCV1BMcG13Q04xcnBfelRmRHkyYk9IbHlDQ2x1RjRzbmo3Z0RJenNaNm9ZeXNUZjJHLTZFakp6QjRzUTVtS3hkY1VVd212ZnBEay1peUp1RmRKQXhzNWFyaFlMTF92MEFVZWktSzNlOGptSlRodU9HSmdaZmtnRm84SWVDRkV3Zz09
**Abstract** >In this paper I question the moral permissibility of developing sentient machines as part of an artificial intelligence (AI) research program. Initially, and from the basic assumptions that this is possible, that pain has a certain (unpleasant) character, and that beings that can feel pain are owed some level of moral consideration, I argue that pain engineering in AI is prima facie morally wrong. I then consider some ways in which proponents may object to—or at least dampen—this initial position.
r/aiethics
comment
r/AIethics
2018-07-10
Z0FBQUFBQm9IVGJBNTVyMnl1ZGJrZTBBUGdwTzdrdWR2dXh1cjNfMVJ6MWZCUW9WZlhXc2xKcUNId3IydEhkblZFdHNZUXdnTHhESVRoLUlsS3BIRjlydU1yYVF1azMtYjBBZ1lmZjNaOGZJTmczelpCXzFlYzg9
Z0FBQUFBQm9IVGJCZm9JSTBVZEJqTzdjeVBBR1hwY0VYenpfTVZJTENUOXlMQ3lGeE9CdDBhcWRIYUxrOHJza2wwNlRsQ2YzQzVoQ1A0dFNZbXF2ZFZIN1Q1d3Atb2pzMkJ3RWtVMzRwbmhtUFRkdTFPZXBLcEFJSzlUaEpQaHFpRHdmb1FpbG9HU1RDUWZiYVVsX2lock5Sbi0tUzljQlNfOXI2OUdYUkdENlBFczRVN3hJTUN5VGlDaEFUeWR6emVON254TDVKMXgtRjdhSEFkVnBFd3FUQy0xTmEwQm5TQT09
Wait. The three laws stories are all about them being completely ridiculous. That's the plot. > The point of this paper is for computer researchers to take note and pay more attention to specifically how humanity will control any potential AI. Specifically, AI researchers need to know that the Three Laws are not sufficient when it comes to controllin g an artificial intelligence. I feel like the author of the paper didn't read the primary documentation. ಠ_ಠ Asimov's "laws" were like spherical cows. He dismissed the possibility of programing in those laws as impossible and only a "what if", where the stories kick off a series of unfortunate events. Because even if they laws did work (something impossible, but for the sake of argument) it would still be a disaster. Asimov's entire body of work is the thesis against laws being able to govern robotics. He invented them to poke fun of them.
r/aiethics
comment
r/AIethics
2018-07-13
Z0FBQUFBQm9IVGJBZzFtNzdEZUZwVllNNndCdXNPcVozazhpMGV5Tmg0bDVtVjV3U09nTGhjZ2k1NXlkQUkwRkNlVEtVaWQ0ZHlYcm9feVZTTUZVSExBSjIxMGphdnphOGc9PQ==
Z0FBQUFBQm9IVGJCQWdSd0gyMVoxSExxckp5MzhkZjJ2X3FvWldSMGxlOGdDd0Y4MVRTa3lVOWtwemJkMHozZHBFWTFnSlhyTWE4OFc1U1pqZHo4MEVRZ2NPQ2M3U2lTZ0l5OUZwMGcydjN3dDI3a05pQmZhMW5BX3JKQnBsZVNRR0FraDlUVmVoajhlLVNKV2Y4djNLUDBUSDA4c18xWk9zcDlJNmktUS1GZ0xaSVl2cEJpeE5xYVRab3hVd19lTGc0d1lfb01tMjA3OUYwOXpDOUVMbm5VVFJFQU9uM25TUT09
Hey OP! I know I'm late to the party but I saved your post and never got around to answering until now. I'm a law student too and I took a course this semester called "Artificial Intelligence and Legal Disruption" which basically covered exactly what you're looking into. We discussed a ton of different ethical and legal issues related to AI and emerging technologies throughout the course. If it's still relevant, I'd be happy to send you a bunch of suggestions for topics you could dive into as well as some relevant suggestions for litterature if/when you've decided on a topic.
r/aiethics
comment
r/AIethics
2018-07-15
Z0FBQUFBQm9IVGJBd1ZxTEw4N0I3Q3RVOHhYYjZ6bjkta1EwNHktVFhFWkNZcHU0NXAxd1B0QlJYUXVZR080Ymw0MHVXNmVzVWhqdWtoQWlyS25YMzBBUDBHemxkTjBYVVE9PQ==
Z0FBQUFBQm9IVGJCWV83Wk41MjMzYklaVDJidXlfVE84aEkyTGR0ZHp0eFFxVjc2ZEFDMjNKYWhBNEh1bXlHaUlQdy1XdHNoeS10cDhyajIxRzVldW1YQms1MDZfUDRoZWxscTM2cE5fMTFBcTFsWTh0dExiWXB6VkdEckV4emRHd21aZGZ0N3JtOUt0T3JMaXVrS1NqZUo3WE9XMnMxd1R6THdxNWRjdWRWUGxGMlJpOThwdEIwS3g5X05fMzQ4cmQyYll0X3VKVldBUENQQnQ0aHVnXzl3ZUJoQUdHdVJPZz09
This is one of those kinds of narrow AI that, in combination with a few other key forms of narrow AI, could take over the world without any general intelligence.
r/aiethics
comment
r/AIethics
2018-07-23
Z0FBQUFBQm9IVGJBQWhPamN3ZG1aX1V1RzdXLS1BOVhWMVRwU2JWM21TMk4wS3BjOUFjTFBrMGthY3pkaEZ4N2huTlZIVTBYOF9PbzNZSzQteC1qMkZwOFdhaEtvc2g0Z1E9PQ==
Z0FBQUFBQm9IVGJCa0JBZ0NXamtBOTRFTUFrSExySWFFN0dUbUxfcTZPS3dlMC1FM2JtcUxLTUtOUktuazZKTFYwTFBaNDVBMG9kX1B6YjItS0N4VmtMSmdMek9IM1FoazZkVTQycWd4OWVxaVR0SzRCMVZYOTQ4dEpLNlVJbGR0blhvOU9GYzRTUTdZTGluYWs4b2hvYkVPY0tZSE9OdVkxdXFsWkRDOFIyYzQxcmd1NFZvZlpJeTJGZkdxdHAxckFUSm9KWE9BZUJXeTBCN1BJRjN5WER2bU5yWDh3b2ZKQT09
I wonder how far we are from being able to mass print human organs and then make individual units suit a person's innumerate system. A couple decades at least? Don't check r/futurology, though. They would say 2 years, at most.
r/aiethics
comment
r/AIethics
2018-07-30
Z0FBQUFBQm9IVGJBZWVtMWlNc2lhaldjNkc4S2FBVHhVVDJPanF0R0M0T1ExYkhDR3ZsSDVUbU0xUXUxRUl4MThZNy00WHNUSmZoMl9tcXNGUHN3SUxzcGhXOWJnXzNGckE9PQ==
Z0FBQUFBQm9IVGJCZGdDdFFlRFhURWtjSmFvX0lLWGp5eGlGWkkxMlFoU3RCN3lzY1FUUHQ4eUQzMnJISFpPeGsyQ0xoVzEzVThFODBKeXdJZk1WUFZ6OUlnaWhQSUpiMWNJTDB5VkZOQldMOGZYU2prZEo2eXN5SXpIT3dLWVF0dUhYNzltc1llWjhJX1lXY0FpVjVpcEJkV0xNQ1pvQTU3Sm1Yb0hhX2VJU28xN01rZzhNU0ktamRma0M3NVNZUjJ1eFFZT3BnRjNuenVrLW8wWFUtTEJ3TkcxaEhKRDAxdz09
Supply and demand will both go down
r/aiethics
comment
r/AIethics
2018-07-30
Z0FBQUFBQm9IVGJBQTZoaEVGbTAtNXB6T1R2d3BzUW5oeGZJRm9WX0hPNVRxMUlHQUFpRDV0aFNacXlZRHFKQi1zazg2OUNqTDkxSUxzU0ZsVG9xVHkxTHZHRktUZjYzVnc9PQ==
Z0FBQUFBQm9IVGJCR0dNRTJXc1ZEXzJ3ZHRKYmtQQXkwM0ZIQXJTTUNxV2RQdmk1ZXFuRmZWbUlYalM4d21MdXpmUXV2SnFSaVFfY19wd1M0REZIeS14bTcxQi1RZTZxSmttQVJUbDlHQWVJX0VaN3RQOUJwcjZnNEJ0VEdkcnhxSW41SFNheFhCanZzOWtiVjJKSFQ4ak95UWJqdDJIMjdLZnJnVm04cTJ2R01NM19YeXpVZmhDMXp4OXNrbGhRTHBQeDhWS0tGQWFrVkxZNXRQSUltT1ZHQlVUOHpZdjRtZz09
I doubt that many people get organ failure from physical accidents. At least, not the kind of organ failure where they will live long enough to get a transplant. Of course the cars will still save lives overall but it's a neat side effect to think about.
r/aiethics
comment
r/AIethics
2018-07-30
Z0FBQUFBQm9IVGJBUWVvU3ZDbXJUWnMxaXhfNzl1QWVsWlBKbmRUa2dCdElaMjBHR2J5WGN1X1VHSVQtUjN4S0QxRElSa0tza2NNTDBFbUpXRnlLcTNmNnEyTmVJVk9SZXc9PQ==
Z0FBQUFBQm9IVGJCNlI0bXJMckpXUlNpV3JIRy1XRDBBOC0tUUo3VE1iRTFWVGZzREw2RVVvUng5TE1kaEhwUG8xaF9TU0FnNU9QNWV1YktWUVEzQUpCZE1QaFE4MnR2WmkybjdSX1NNTGlVd3gxSThVT2ZVZDJIaDg0LU1sM3RYbjh2NnFjcER4SGJqM0NzcWRhcktpOG41UEhpdzJ1YzdxSXNKWjlLRkxfc1FQRlhUdS1kSXIyODhzSDBfdi05UW44ZEs4M0lrMHlSRkpTZUp4TnpjdlRfREZZWWptQU5jZz09
Any chance I can get the list of literature also?
r/aiethics
comment
r/AIethics
2018-08-03
Z0FBQUFBQm9IVGJBNzk2bms0Um0zaENKZUlVLXFzdnBkd0xwQnRHc2RXTzhtS3RBQjlQV002aGtsazh2dUNRTi1NZHdBazIyNVhRYWhOajNEbjgtQl9wTUh5dDNGUE1Kd0E9PQ==
Z0FBQUFBQm9IVGJCNHRpeF92TkFPdVNIdE9wVExZN2w3UVBUcVZhWktCc2x4aHFDQjhOQURONnBTbzhaTEJHb3NTOFoxb2REMkdUT0s4N09hQVZOQXN0QWo1U3hGZUxLNXFESUxtNlduYnFwa0w5dHJvN3hLdVhCTjNZWFJhZzR4OS05T3B2TkVTVVBEQ2dpXzV5dWUyZ25RN0lfbGxMaDBVb21iWl9icGFzOUs1SS1iTE91ZjFYaWZGN2R2WjV5V2ZrWHFWQWs4TnNXUWN3Vnprd0JlRFYzQkQzcHpKaEdiQT09
**Abstract** >The ethical issues related to the possible future creation of machines with general intellectual capabilities far outstripping those of humans are quite distinct from any ethical problems arising in current automation and information systems. Such superintelligence would not be just another technological development; it would be the most important invention ever made, and would lead to explosive progress in all scientific and technological fields, as the superintelligence would conduct research with superhuman efficiency. To the extent that ethics is a cognitive pursuit, a superintelligence could also easily surpass humans in the quality of its moral thinking. However, it would be up to the designers of the superintelligence to specify its original motivations. Since the superintelligence may become unstoppably powerful because of its intellectual superiority and the technologies it could develop, it is crucial that it be provided with human-friendly motivations. This paper surveys some of the unique ethical issues in creating superintelligence, and discusses what motivations we ought to give a superintelligence, and introduces some cost-benefit considerations relating to whether the development of superintelligent machines ought to be accelerated or retarded.
r/aiethics
comment
r/AIethics
2018-08-16
Z0FBQUFBQm9IVGJBQ3JqMjlxQjZUSHBRVDJBLThmQzc4elhqX040OWdKNnFleFdhZ1N0ZnZZWFVub3hDeHI5ck9qZlZqeHQyTVI0a0g4Z3k1NUk1czNLY2xhazM1enJaS2VVcGtUUm94aGEtekVzSG1JM3VyOFU9
Z0FBQUFBQm9IVGJCXzduYzQwaGRFdnktaEw3WXlhbW5jTHhrNG5vS0twY2xFbVZlNXY2S3JobmF0WUFFSW9FX1ZSdlMzZFlMSE5FdER3ZjQzWWdlWEdmaXN5cjZQNVlJWjNpMnJtRDNZdjZWMVJqbF9mR2hqTklaS0J3LWI1eTBvTUprSHVHY0Y4UjBsbUotT09Jc2JoUUVWeXVqVmlLUHg5RVlTZS1PYTgxZlpuaXFqeE12eU42RVY2emdhVG5hMHN4dmduX3FMSVNXUi1nbllSM0pXeWloU3Fnck4tdnlMdz09
[Here it is on youtube so that you can put it on 1.5X speed.](https://www.youtube.com/watch?v=K74l3o7ful4)
r/aiethics
comment
r/AIethics
2018-08-30
Z0FBQUFBQm9IVGJBWklRekI2Mks5OERMX3V0TTNhdWxtSm9KcTBNaHJKY1haaTcwU2FEdktqQjg3Y2xzNERmdXdWdlpCbHRlUW1fLTlsWndYV3lFZnFqeHdKTzhZc0p3MUE9PQ==
Z0FBQUFBQm9IVGJCR0Z0ZnRVYTcwME5hbzRKcXhHWGxsV01FUEVXbl9WaTVGVEowRy1sNThsQ0doaENsdzJTbklyaUxzZGNsUUNRRTJjWUFRdXcwRHRDYU5yV3lfR1Fpa3NSRU1PMXZHaWs5azlkUnNGSDNVWHFXdEV1WjB3QTFxUXVkN280Zks4WVV5Y1YwMXNlcjBzNzAyNWZqTDBmSXBNSG4xNVliamVVcXVCcFRPQ2RiaS1WRjFXVjF4YXZOZjlKZ3Jfano2c3NkSEIzZVhjbllaZEZYT1E5M004ZGk4UT09
I wish the FRI should just go away and shut up already. They're not doing anything of significance.
r/aiethics
comment
r/AIethics
2018-09-03
Z0FBQUFBQm9IVGJBVWQxck5RLTB0NW9yMlRqTFBWMVZZU1E0MFFGMHVRUEtlNk1MSlRrZzhCVWJ5UWZGYkhMZlU5dFdHTGUxd1lZTHZUQV9mYWFHV2pSOFY0a3lkcHFtVkE9PQ==
Z0FBQUFBQm9IVGJCWGRVMkVLcWdVTHdsWXVMYTA5MWNyWmZaS3lBUGpyclN1Rm9NSTEzRnhWTmhKVmhUUVVvejgzeFVIVGVieTRKMEZyWUh1ak84V2hoOWVHQ2ZDSnMxYUVZS2NPZ3BmSGh3VGN4TTRPZE5QT3hHU0hGOFRJdVJCc0ZiTDB2bkZuVExfaUhyTWNHc0lmaThtVTZDX1kzTHJpU3hKU1hDQTNzLWZreU00aDJGU0l2UGloa1pEYmo3TXJ2a1ZQcTFvT0J5YzRLR2RocExGd1hhRk11ZzFnZlpHUT09
I saw this movie. Definitely underrated.
r/cleanenergy
comment
r/CleanEnergy
2018-09-11
Z0FBQUFBQm9IVGJBbEpjQi1QT3VVSXVESUZDSXhOVlRiOXJlbTlCVTFvUXJQdF83VjRpQkYwWDkyMVZiVk5oUS1sMkpFSDBRSEpuSVFONjBCV3dzMkVLMnNqNE1wbzJ2MnRvektXUlVDaWUyU2JubGJ2VmRZcG89
Z0FBQUFBQm9IVGJCZEJGbXdhdXdEZGZkMXZfbzRaX3BNZWxsQlNnOEI5VGsyMF9JUGt2dGhGc3hFc20tOWRSYmEwaFVXeEhpcjlUUmo0eFIxMGtpSEh0Mk0xczlGNnZaSnpOVmZtNnNGTjNkd3piY05BekdlQVN6ZW9Kd21fdWJTLTVpOFFPZ0RiN0tVUk42V1hTc0J5R0daRmZXR204Tl9JRXpqbFl4ZUlGeENIeldCZVk0Zk9hTW52eUhBWTdJbmtMUlI4TF9xdWNmZU1wck93cU0wMWlYNDFibU0zS3hOQT09
Lol, what?
r/cleanenergy
comment
r/CleanEnergy
2018-09-12
Z0FBQUFBQm9IVGJBUWsyYXE0OGRjRmFFOW1fQVRKNDlFRTNYc0VFLVlBMlJzWldJYUNidEZic3hLeG1fU1dXNnhhTjIxY1JXZ0NDeUkwX1NWT1RDSlpNc1JtNmkwUFdRaFE9PQ==
Z0FBQUFBQm9IVGJCQ1BIV21rR21UbHBrbE9WQUVlRHJlVXd4dE1MSWZRdnRuLUR2b1pNWmVlS1JtYXpENHd1a0U5Z1VxaEVCWk15NDk2b2JiTVZWRUVuSGZTR0tBWV9QT3BYS082X2NCV3pDOWdsanlNbUpMdGV2Ny14LXdIV2VDdUY0QkgwTFpLNEN5SzZWY2w4dGdWQUJ6Q1BBQkcwQk9XR1loMzR3Tk13MzdfaTlJNmVhUTctMUhoYTBrLS0wQW93S2xQZW9wYnNfQnJraVNmNnlZMWdEb0RZSmtJeTkxQT09
If you're at the level of technology where you can make identical but digital copies of yourself, with the same mental capabilities and run a thousand of them simultaneously, congratulations, you won life. Make up your own rules, because you might just the most powerful organism in the universe.
r/aiethics
comment
r/AIethics
2018-09-12
Z0FBQUFBQm9IVGJBNGNxNVU3RUJUcUIzaDc3LVpSZlh2VnByUXAwQlRIcmh3RUpJWU9fZUFNR3RndWVMVi1FXzVTRllndEpkOS1sT2NpSS1fYmNuVkVZbTN4Mnphd0lBS2c9PQ==
Z0FBQUFBQm9IVGJCeDFhd19XSVdMTTczOXBWaVU4RXVCNG1PUVMycDdnRGpKbzl2V281SVBrUUNZLWlRVlJtN3JRTjhLR1FpS2M1bE5xUHpsQkZpMHgzVGs5U2FXM3hTcHM2cmxCMDNDTHJaU0hCTVQyOEtXdFZ6a0JvMUxJR0s2ZU5pc2ZuM3ZSbmkzdDk3MURfLWFOZ3JMeU9qaHh3VVNHMThmYXBCd3ctcm9sdV9ncVlUQVNHWG9IbFZaQTFva2o1WHladFRKY3RXWERwS2RyelIzSVZ0UnRNampzajc0dz09
As demand for lithium skyrockets, manufacturers and governments scramble to secure deposits of the world's lightest metal. Lithium is used to power electric vehicles, smartphones and laptops. Clean energy tech runs off the stuff. Learn how supply can eventually meet demand. [http://ultralithium.com/lithium-demand-exceeds-supply/](http://ultralithium.com/lithium-demand-exceeds-supply/) https://preview.redd.it/53utlkqrw1n11.jpg?width=860&format=pjpg&auto=webp&s=7cb6de30e4ee03c054f377f880edd279c128cccb
r/cleanenergy
post
r/CleanEnergy
2018-09-18
Z0FBQUFBQm9IVGJBYU1yal9vR1VuVlBzZ05mb0ozc0hVUXo0Y0hDSzNxRjU4ZjU2WVhvYlRVbWJxT0MwRjJyU3I3RXZLLUgwLWNzb0djM1MwT0FvMjVDelgzZnhfcjRSWHBPMzlORGRQM2VwM1VNa21TYnFzbms9
Z0FBQUFBQm9IVGJCbUpaN2dvdms4Z3pVZ2Z4SGUtTUtsaThuT0FBV3RMVk81SEhIS0RXTHhackg2X3lvQVJlQTdENjkteW44UE1uTERPbF9rdXYzZkZsMXJSYkxsT19KbzdlVXJLUFdDbDlMTWJIWTZEMHJsOTJUZkZNVjdYTUNZbGYxbEgtRXdCcmNJODVtV0RMMG9wU0tPMUZJVVVyWEFmRnBWNU9wVjQtLUFRMElVeEtUX3VrVVh2ejBnWnhRdE1pS0R1b1Q4cTNp
Pleasantly surprised to see China, with their incredible AI investments, being somewhat open about their policy. This was a concern that kept bugging me: the largest AI investors keeping to themselves. Thanks for the article!
r/aiethics
comment
r/AIethics
2018-09-18
Z0FBQUFBQm9IVGJBWXdJSkJJR3UxajZXcmhldXJDc0pMa0JscU42X01CWVBHV1FJTGNUZEZINFNLdmpMejYtTEpBMXBnZVBOWjk0MDBmQzU5SFlWUjFWTDVsRlB2dGk4ZGc9PQ==
Z0FBQUFBQm9IVGJCUTUwaTZYNElWV0s5MkN3MVZEczlKRXI2MFZQWTk5UjcyaU5kcWZrWktrYkNHZGdQYTNjczlpY1R4dG5ncE9WSi1MX2dsWER6NEZ0eVNtRDV5emhJQ0NZNnpqTFkzOWRnUTRKaVZIQ1hob3o3bkh4WjBuUHc1blY4UHg3SUhydTItdFd3cW9EM0JkeGZ2UkE1WHVZV291QUxDOGdoc1BvYlFBOXA0Wk5hUG83SE1WckZSUlJobjhHdFdHZkY5TnZWdTB0MHFCeUI4N1BfaThNSEMwWXVWdz09
Paul Christiano will, if he gets his way, doom us all. He is a prime example of "knowing enough to be dangerous".
r/aiethics
comment
r/AIethics
2018-10-04
Z0FBQUFBQm9IVGJBYmhLWkFSRkZ6WXhkN0dNWVUzWFVTbkdJdWhDZnFyV01oeTM2aGFIdVJzWXFRU1oyNzRXd0tBdGg4el9ueVQ5RHU3bHJ2UWxlcEdLa2dTdnVJSnNEUUE9PQ==
Z0FBQUFBQm9IVGJCcFV4bHg1WHN6Y2s3MENSTEtrYXJzSkFxUWZtdjU3VGQ0UmFEdWxBSkFBTzExZlRieVJoWHQwWWhuaVk3Z1A1QUhwMVRMaTRZb3h0bnNpcXdTSExBYzRlRTVJX1NaSEVkSEo1VjNhWWg0ejhxMTlaLVJsandETTU5bUdfMDRBbXlHN1N2dUpLMWduTW4xNnRDQmlOZ1MzQ3dKWjhSNWtZR1o3V3pQVkhqY2dKZFNRQlRER2h2dzhISDJWQ1piMXFPZ2Z2U2VzS2FGbUdkbXNPQkVxRlU2Zz09
### Lithium Ion Battery Important The world is hungry for clean energy solutions. The 21st century is all about sustainability and the lithium ion battery is this new world’s power source. Powering PCs, laptops, Smartphones and more, the lithium ion battery powers most consumer electronics. Furthermore, construction projects take 10 times as long without it. Everything from electric drills to air compressors use them, minimizing cost while quadrupling productivity. Made of lithium and cobalt, the technology behind these batteries is improving at a rapid pace. Soon, Smartphones will last longer than 24 hours – even with energy-consuming apps like SnapChat and YouTube being consistently used. However, time is needed for this jump to take place. ### Lithium Batteries & Electric Vehicles Especially today, electric vehicles (EVs) come to mind when discussing lithium ion batteries. This makes sense, as the EV industry steadily grows. Consequently, this industry is putting immense pressure on battery producers to improve capacity. Currently, most potential EV owners are scared off by ‘range anxiety.’ This is the fear that a driver will be stranded in the countryside with no charging station in sight. However, fear not. Some of the latest models can travel 200 miles on a single charge. Furthermore, Tesla’s Supercharger stations can charge a full battery in under 1 hour. They are strategically placed throughout the US and Canada. ### Finding Lithium Lithium ion batteries need cobalt and lithium. Most lithium is found in the Lithium Triangle, a massive South American range featuring Argentina, Bolivia and Chile. Over 54% of the world’s lithium is found here, of which China owns half. Most is extracted through brine mining, an eco-friendly technique where lithium dissolves in evaporation pools. Meanwhile, Australia extracts its lithium through hard rock mining. An environmentally hazardous technique, it’s very expensive and tarnishes the landscape. However, a global lithium shortage persists. In order to meet the EV industry’s needs, controlling lithium supply is essential. The global economy depends on it.
r/cleanenergy
post
r/CleanEnergy
2018-10-10
Z0FBQUFBQm9IVGJBNzRuMjFFNlBjS1hLbFRtc2dRV0d5UVVxS1FHQmtzUmdBZFNLSGFYck0xeG1EY1hpaHdKMzVtTzg4WnVObDFSaTgtUXVMS1huWXZFLV9iUlUwcW5uVDFtVW9mbExneVk2Y0VmcGlwbDQwWms9
Z0FBQUFBQm9IVGJCdVVZREZsR3pDM1hndnE3UWNaMEU1RnlFYTVuamdBczJJS0hKR2ZNdzlPbW5DUXhwZmNKREpydnVNLXQtUmVxM3g3Z05jaURvY05wcWNIXzB0YkwxeWlOYWREcTVITC02NmxsNGRRNnowUERhTW11Q21RcUNTaEk1d3RRWnFnT3VfLXNyeGcxTks2OVB2dHdlOGZKZmFabW50VmZFWDhHY24zYXlvbG5aTlA4Wl9ZNXNLd2VpZ3p3aVhhVzdrR0xO
Amazon AI woke
r/aiethics
comment
r/AIethics
2018-10-11
Z0FBQUFBQm9IVGJBdnNRd3lQSEUxX2JBQmtTdGNBZTV1QjNvdmNrUkdSUk9Ya3JlY3NvUzhha0ZvemFFZk0yQVNwbm90R1R2UHFRRkhOYk43ODM4QTA0UW5qMkVHZWdLaGc9PQ==
Z0FBQUFBQm9IVGJCVy0zSU5iNEFhbXJvR0hXMnpHeS0yUVM2RWtHRXB3V0tNZGVYcWdIRThtcHpqY3FJY1IyZlN1OGozem9xSUgwblE1aEs4RzhEb2xNZS15eFJ4YkRHREZQZ2RWR1hjZzZaTXJqSkZOVFhkUFJ6elRWYm1fQjhqUUh6SWRkRVBoLXhoRjhRbWRJMDdqUVBFWWV4Z1dMd2RZWHBkN1ptTzM3MnVwSnBhcExvOUhUdUNPczhfa01yLWUtY0FjVXlfWWxSMC1Gc2E0Q0Z1cElBU1JqUTYxcnR2dz09
Safer than human-driven cars already.
r/aiethics
comment
r/AIethics
2018-11-11
Z0FBQUFBQm9IVGJBNkRvTFh3WlJ2NXRud2pZUVpJTzAxeEx2NXBSdkJlVFRrZTFaSlViWjBBRFJRNFZrbnFsZ2lkQmw5Q0w4WENfOXg0LS1DckxGOXBOdktSX05RSVRwX1E9PQ==
Z0FBQUFBQm9IVGJCZ19sVGJOamlvUzhyWG83RVRVT05fR19Pd0tEeHp1R3lORVNJMkZNRUVSV0V2cXBTdkthOHVSUGVtVmhqVEE0dEd1eC1YZU1RdTdQeldQRkRvYzdiT0h2ZFhQY0ZrbHRZR3VkUThYWDA4VHhmeFVITlRZeVNIdk9takhpYjhiczBaLVVISGszOUlCbnU3SHc0VjBuWUY3YWpDbjN1LU41cEhGbDhlcERObDV0Ri05UUdtYm9pOEVZNzhNOERQRTI4TGVUTzNLcTg3WktjSDRhTVhMOWRoQT09
"Nearly 1.3 million people die in road crashes each year, on average 3,287 deaths a day. An additional 20-50 million are injured or disabled." (Wiki) Yeah, I am going to go out on a limb and say the title question is pretty silly. Cars driven by humans are horrifically dangerous. It's hard to imagine they even *could* make self-driving cars that are as dangerous, really.
r/aiethics
comment
r/AIethics
2018-11-11
Z0FBQUFBQm9IVGJBdXZkUDdCY1Y1VnV1Tk9MNDRfcTNRRzZqajNzakdQQ3lSd1dlWUgwZ3dkMmwzUThoUmJYN2VPYVJzLWJERVYySVZhOEx0bEtyOVdiT25HSDNMSllYN0E9PQ==
Z0FBQUFBQm9IVGJCSXQ1ZVJMYXdWODZxcHB6OVBIWDRrNHVLN2syUEl6dHplSEpUV2I0VmxfdlRFeGpzbm9xdDdQSFQ3U2U3czN3VGZCZ0hRWnNNVUhfZmxIZW52QXE5SGlocl9lOHJzNXhDTWx6NHA4SU1UYXdsOWFrMlNacmJBdFU2RnVNVUpRVmktNC1QazQzdF9Hc2U4QThpRzh0U3Z4SG40Q2lGV3E5aG5ocmVLRWdqNEl4MHphTmI2MWhFcGJ6Ykk2SUM0bUlBSjlkdk9wdjJRMk5Obk9JZ1RkSXpEUT09
This is the best tl;dr I could make, [original](https://blog.singularitynet.io/data-as-labour-cfed2e2dc0d4) reduced by 92%. (I'm a bot) ***** > If our personal data can help a political party win an election, should we not have more bargaining power when exchanging that data, or at least have a say? Is our personal data on Facebook really only worth free access to Facebook? The artist, Jennifer Lyn Morone, more crudely calls it 'Data slavery'. > Data as labor also requires public institutions to 'check the ability of data platforms to exploit monopsony power over data providers and ensure a fair and vibrant market for data labor. > Realising the status that Personal Data holds legally, normalising extended data rights, acknowledging the full extent of PD's impact on social stratification, and capturing the ways that it can truly become 'personal' data again, are all necessary steps still missing from the education of today's "Onlife" worker -seamlessly living online and offline concurrently. ***** [**Extended Summary**](http://np.reddit.com/r/autotldr/comments/9zuubr/data_as_labour/) | [FAQ](http://np.reddit.com/r/autotldr/comments/31b9fm/faq_autotldr_bot/ "Version 2.02, ~366293 tl;drs so far.") | [Feedback](http://np.reddit.com/message/compose?to=%23autotldr "PM's and comments are monitored, constructive feedback is welcome.") | *Top* *keywords*: **Data**^#1 **Labor**^#2 **Free**^#3 **human**^#4 **work**^#5
r/aiethics
comment
r/AIethics
2018-11-24
Z0FBQUFBQm9IVGJBUWtzN3VzN0JrbzJlcW10eFdZRFBMTmFVWUlLUHNrNWVYWVRPTERmSTdhR0h1QTh5Mnl0R2h2aVNJaGZsNWN3VTlIV21aYVVoODdUc0lkX0JTVFM2QXc9PQ==
Z0FBQUFBQm9IVGJCOVRwTjl2QW1TN1dSOXBVQms0VTlrTHdsZHZRVGlDOGFPcHMzb0N0S0x4M05UMGk2T3NrZUY5aFhmd1F0WXdGMHdJc1BQeU9hc215ZTJBOTdkTzBXX2UxcUpWRUNvSnJhNnd3MFY0UnpuNXF3STFjWGwxN08yR3FuaUVDczhqNlYwSDFxeEM0Mkl2cDVOTzNiZ1BRSmR2azAtOEpLMVlwX2NoSDVWdXlGUmZnPQ==