Episodi

  • This and all episodes at: https://aiandyou.net/ .

    Artificial General Intelligence - AGI - an AI system that’s as intelligent as an average human being in all the ways that human beings are usually intelligent. Helping us understand what it means and how we might get there is Craig A. Kaplan, founder of iQ Company, where he invents advanced intelligence systems.He also founded and ran PredictWallStreet, a financial services firm whose clients included NASDAQ, TD Ameritrade, Schwab, and other well-known financial institutions. In 2018, PredictWallStreet harnessed the collective intelligence of millions of retail investors to power a top 10 hedge fund performance, and we talk about it in this episode.

    Craig is a visiting professor in computer science at the University of California, and earned master’s and doctoral degrees from famed robotics hub Carnegie Mellon University, where he co-authored research with the Nobel-Prize-winning economist and AI pioneer Dr. Herbert A. Simon.

    In the conclusion of the interview, we talk about the details of the collective intelligence architecture of agents, why Craig says it’s safe, morality of superintelligence, the risks of bad actors, and leading indicators of AGI.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    Artificial General Intelligence - AGI - an AI system that’s as intelligent as an average human being in all the ways that human beings are usually intelligent. Helping us understand what it means and how we might get there is Craig A. Kaplan, founder of iQ Company, where he invents advanced intelligence systems.He also founded and ran PredictWallStreet, a financial services firm whose clients included NASDAQ, TD Ameritrade, Schwab, and other well-known financial institutions. In 2018, PredictWallStreet harnessed the collective intelligence of millions of retail investors to power a top 10 hedge fund performance, and we talk about it in this episode.

    Craig is a visiting professor in computer science at the University of California, and earned master’s and doctoral degrees from famed robotics hub Carnegie Mellon University, where he co-authored research with the Nobel-Prize-winning economist and AI pioneer Dr. Herbert A. Simon.

    We talk about his work with Herb Simon, bounded rationality, connectionist vs symbolic architectures, jailbreaking large language models, collective intelligence architectures for AI, and a lot more!

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • Episodi mancanti?

    Fai clic qui per aggiornare il feed.

  • This and all episodes at: https://aiandyou.net/ .

    We are talking about international governance of AI again today, a field that is just growing and growing as governments across the globe grapple with the seemingly intractable idea of regulating something they don’t understand. Helping them understand that is Markus Anderljung, Director of Policy and Research at the Centre for the Governance of AI in the UK. He aims to produce rigorous recommendations for governments and AI companies, researching frontier AI regulation, responsible cutting-edge development, national security implications of AI, and compute governance. He is an Adjunct Fellow at the Center for a New American Security, and a member of the OECD AI Policy Observatory’s Expert Group on AI Futures. He was previously seconded to the UK Cabinet Office as a Senior Policy Specialist.

    I know “governance” sounds really dry and a million miles away from the drama of existential threats, and jobs going away, and loss of privacy on a global scale; but governance is exactly the mechanism by which we can hope to do something about all of those things. Whenever you say, or you hear someone say, “Someone ought to do something about that,” governance is what answers that call.

    In the conclusion, we talk about verification processes, ingenious schemes to verify hardware platforms, the frontier AI safety commitments, and who should set safety standards for the industry.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    We are talking about international governance of AI again today, a field that is just growing and growing as governments across the globe grapple with the seemingly intractable idea of regulating something they don’t understand. Helping them understand that is Markus Anderljung, Director of Policy and Research at the Centre for the Governance of AI in the UK. He aims to produce rigorous recommendations for governments and AI companies, researching frontier AI regulation, responsible cutting-edge development, national security implications of AI, and compute governance. He is an Adjunct Fellow at the Center for a New American Security, and a member of the OECD AI Policy Observatory’s Expert Group on AI Futures. He was previously seconded to the UK Cabinet Office as a Senior Policy Specialist.

    I know “governance” sounds really dry and a million miles away from the drama of existential threats, and jobs going away, and loss of privacy on a global scale; but governance is exactly the mechanism by which we can hope to do something about all of those things. Whenever you say, or you hear someone say, “Someone ought to do something about that,” governance is what answers that call.

    We talk about just what the Centre is, what it does and how it does it, and definitions of artificial general intelligence insofar as they affect governance – just what is the difference between training a system with 1025 and 1026 flops, for instance? And also in this part Markus will talk about how monitoring and verification might specifically work.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    Virtually everything that’s difficult about getting computers to do work for us is in getting them to understand our question or request and in our understanding their answer. How we interact with them is the problem.

    And that's where Sophie Kleber comes in. She is the UX – that’s User Experience – Director for the Future of Work at Google and an expert in ethical AI and future human-machine interaction. She deeply understands the emotional development of automated assistants, artificial intelligence, and physical spaces. Sophie develops technology that enables individuals to be their best selves. Before joining Google, Sophie held the Global Executive Creative Director role at Huge, collaborating with brands like IKEA and Thomson Reuters. She holds an MA in Communication Design and an MBA in Product Design, and is a Fulbright fellow.

    In the conclusion of our interview, we talk about about how she got into the user experience field, the emergence of a third paradigm of user interfaces, the future of smart homes, privacy, large language models coming to consumer devices, and brain-computer interfaces.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    Virtually everything that’s difficult about getting computers to do work for us is in getting them to understand our question or request and in our understanding their answer. How we interact with them is the problem.

    And that's where Sophie Kleber comes in. She is the UX – that’s User Experience – Director for the Future of Work at Google and an expert in ethical AI and future human-machine interaction. She deeply understands the emotional development of automated assistants, artificial intelligence, and physical spaces. Sophie develops technology that enables individuals to be their best selves. Before joining Google, Sophie held the Global Executive Creative Director role at Huge, collaborating with brands like IKEA and Thomson Reuters. She holds an MA in Communication Design and an MBA in Product Design, and is a Fulbright fellow.

    We talk about the Uncanny Valley and how we relate to computers as though they were human or inhuman, and what if they looked like Bugs Bunny. We talk about the environments and situations where some people have intimate relationships with AIs, gender stereotyping in large language models, and where emotional interactions with computers help or hinder.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    Teachers all over the world right now are having similar thoughts: Is AI going to take my job? How do I deal with homework that might have been done by ChatGPT? I know, because I've talked with many teachers, and these are universal concerns.

    So I'm visiting the topic of AI in education - not for the first time, not for the last. There are important and urgent issues to tackle; they become most acute at the high school level, but this episode will be useful for all levels.

    The reason it's so important for me to work with schools so much as an AI change management consultant is that there's no need for teachers to fear for their jobs. They are doing the most important job on the planet right now because they are literally educating the generation that is going to save the world. And generative AI has not created a learning problem: it's created learning opportunities. It's not created a teaching problem; it's created teaching opportunities. It has, however, created an assessment problem, and I'll talk about that.

    Kids need their human teachers more than ever before to model for them how to deal with disruption from technology, because change will never again happen as slowly as it does today, and all of their careers will be disrupted far more than anyone's is today. No student is going to remember something ChatGPT said for the rest of their life. The teacher’s job is to focus on the qualities that the AI cannot embody—the personal interactions that occur face to face when the teacher makes that lasting impression that inspires the student.

    Let's have honest, deep, and productive conversations about these issues now. A new school year is approaching and this is the time.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    Is work heading for utopia? My guest today is John Danaher, senior lecturer in law at the University of Galway and author of the 2019 book, Automation and Utopia: Human Flourishing in a World without Work, which is an amazingly broad discourse on the future of work ranging from today’s immediate issues to the different kinds of utopia – or dystopia, depending on your viewpoint – ultimately possible when automation becomes capable of replicating everything that humans do.

    John has published over 40 papers on topics including the risks of advanced AI, the meaning of life in the future of work, the ethics of human enhancement, the intersection of law and neuroscience, the utility of brain-based lie detection, and the philosophy of religion. He is co-editor of Robot Sex: Social And Ethical Implications from MIT Press, and his work has appeared in The Guardian, Aeon, and The Philosopher’s Magazine.

    In the conclusion of the interview we talk about generative AI extending our minds, the Luddite Fallacy and why this time things will be different, the effects of automation on class structure, and
 Taylor Swift.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    Is work heading for utopia? My guest today is John Danaher, senior lecturer in law at the University of Galway and author of the 2019 book, Automation and Utopia: Human Flourishing in a World without Work, which is an amazingly broad discourse on the future of work ranging from today’s immediate issues to the different kinds of utopia – or dystopia, depending on your viewpoint – ultimately possible when automation becomes capable of replicating everything that humans do.

    John has published over 40 papers on topics including the risks of advanced AI, the meaning of life in the future of work, the ethics of human enhancement, the intersection of law and neuroscience, the utility of brain-based lie detection, and the philosophy of religion. He is co-editor of Robot Sex: Social And Ethical Implications from MIT Press, and his work has appeared in The Guardian, Aeon, and The Philosopher’s Magazine.

    In the first part of the interview we talk about how much jobs may be automated and the methodology behind studies of that, the impact of automation on job satisfaction, what’s happening in academia, and much more.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    Helping the British Government understand AI since 2016 is our guest, Lord Tim Clement-Jones, co-founder and co-chair of Britain's All-Party Parliamentary Group on Artificial Intelligence since 2016. He is also former Liberal Democrat House of Lords spokesperson for Science, Innovation and Technology and former Chair of the House of Lords Select Committee on Artificial Intelligence which reported in 2018 with “AI in the UK: Ready Willing and Able?” and its follow-up report in 2020 “AI in the UK: No Room for Complacency.” His new book, "Living with the Algorithm: Servant or Master?: AI Governance and Policy for the Future" came out in the UK in March, with a North American release date of July 18.

    In the second half, we talk about elections, including the one just held in the UK, and disinformation, what AI and robots do to the flow of capital, the effects of AI upon education and enterprise culture, privacy and making AI accountable and trustworthy.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    Helping the British Government understand AI since 2016 is our guest, Lord Tim Clement-Jones, co-founder and co-chair of Britain's All-Party Parliamentary Group on Artificial Intelligence since 2016. He is also former Liberal Democrat House of Lords spokesperson for Science, Innovation and Technology and former Chair of the House of Lords Select Committee on Artificial Intelligence which reported in 2018 with “AI in the UK: Ready Willing and Able?” and its follow-up report in 2020 “AI in the UK: No Room for Complacency.” His new book, "Living with the Algorithm: Servant or Master?: AI Governance and Policy for the Future" came out in the UK in March, with a North American release date of July 18.

    In this first part, Tim gives a big picture of how #AI regulation has been proceeding on the global stage since before large language models were a thing, giving us the context that took us from the Asilomar Principles to today’s Hiroshima principles and the EU AI Act and the new ISO standard 42001 for AI. And we talk about long-term planning, intellectual property rights, the effects of the open letters that called for a pause or moratorium on model training, and much more.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    As the European Union AI Act rolls out, there are so many questions about what it will mean to businesses trying to navigate the incredibly volatile and complex field of AI regulation. Here to answer those questions is Antonina Burlachenko, Head of Quality and Regulatory Consulting at Star Global Consulting, calling from Poland. She explains what the Act really means for businesses and consumers, comparing it with GDPR, and providing some technical information around standards and regulations and other aspects of what it’s like for businesses to engage with the Act at a practical level.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    To help us get new and valuable insights into the future of work is Matt Beane, Assistant Professor in the Technology Management Program at the University of California, Santa Barbara. He has spent over a decade doing extensive field research on how workers, organizations and even AI defy norms and rules in the 21st century.

    His new book: The Skill Code: How to Save Human Ability in an Age of Intelligent Machines, was just published by Harper Business, and he has given you a special deal as a listener, to get a free copy of the first chapter, by going to http://aiandyou.theskillcodebook.com. The book lays out a plan for us to protect our skills and by extension the human connection between experts and novices (which is the foundation of skill-building) even as AI continues to take hold in our lives.

    In the conclusion, we talk more about what AIs do to the mentoring and learning pipelines in the workplace, and how education should pivot to deal with the changes to the future of work.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    To help us get new and valuable insights into the future of work is Matt Beane, Assistant Professor in the Technology Management Program at the University of California, Santa Barbara. He has spent over a decade doing extensive field research on how workers, organizations and even AI defy norms and rules in the 21st century.

    His new book: The Skill Code: How to Save Human Ability in an Age of Intelligent Machines, was just published by Harper Business, and he has given you a special deal as a listener, to get a free copy of the first chapter, by going to http://aiandyou.theskillcodebook.com. The book lays out a plan for us to protect our skills and by extension the human connection between experts and novices (which is the foundation of skill-building) even as AI continues to take hold in our lives.

    In this first part, we talk about how Matt studied surgeons in operating rooms for his PhD thesis and saw the effects that the introduction of a robot surgical system had in stifling the time-honored process of mentoring new surgeons, and generalized this to other fields, and observed the rise of “shadow learning,” where people bend or break the rules to get the learning they need.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    My guest is William A. Adams, technologist, philanthropist, and recorded by the Computer History Museum as one of the first Black entrepreneurs in Silicon Valley. He was the first technical advisor to Microsoft’s CTO Kevin Scott and has founded and overseen global initiatives at Microsoft from XML technologies as early as 1998, to DE&I initiatives in 2015. The Leap program, with a focus on diverse hiring, was named Microsoft’s D&I Program of the year in 2020.

    We talk about William’s experience creating the Leap program, its impact, the relationship between AI and diversity, equity, and inclusion programs like Leap, and creating personalized chatbots.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    Our relationship with time is dysfunctional. Here to help us explore possibly the most critical effect of AI on the pace of life is Oliver Burkeman, author of the best-selling self-help book Four Thousand Weeks: Time Management for Mortals and former author of the psychology column “This Column Will Change Your Life” in The Guardian.

    Most of us can attest to being severely overworked and with a shrinking amount of personal time left over. This is true despite the introduction into our lives of a huge amount of technology from the PC to the Internet. Why have tools like email, Google, and instant messaging not reduced our workload and stress? In fact, it’s not hard to believe that they are responsible for making those things worse. In which case, we must ask, what effect will unleashing AI – which accelerates everything it touches - have on our work life?

    This is exactly the thought space that Oliver inhabits, and his work has made a major difference in my own life. Read Oliver's posts and subscribe to his newsletter at OliverBurkeman.com.

    In the conclusion of the interview, we talk about whether this is Luddism, the influence of the Silicon Valley billionaires’ pursuit of immortality, the appropriate use of AI to save us time, and what will remain constant throughout any amount of technological evolution.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    Our relationship with time is dysfunctional. Here to help us explore possibly the most critical effect of AI on the pace of life is Oliver Burkeman, author of the best-selling self-help book Four Thousand Weeks: Time Management for Mortals and former author of the psychology column “This Column Will Change Your Life” in The Guardian.

    Most of us can attest to being severely overworked and with a shrinking amount of personal time left over. This is true despite the introduction into our lives of a huge amount of technology from the PC to the Internet. Why have tools like email, Google, and instant messaging not reduced our workload and stress? In fact, it’s not hard to believe that they are responsible for making those things worse. In which case, we must ask, what effect will unleashing AI – which accelerates everything it touches - have on our work life?

    This is exactly the thought space that Oliver inhabits, and his work has made a major difference in my own life. Read Oliver's posts and subscribe to his newsletter at OliverBurkeman.com.

    In this first half of the interview we talk about the parable of the rocks in the jar and how it’s a pernicious lie, the psychology of perceiving life as finite, and how technology has not changed our work stress and may be making it worse through induced demand.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    Mounir Shita, CEO of Kimera Systems, is author of the upcoming book The Science of Intelligence, which contains some interesting and thought-provoking explorations of intelligence that had me thinking about Pedro Domingos’ book The Master Algorithm. We talk about theories of AGI, free will, egg smashing, and Mounir's prototype smartphone app that learned how to silence itself in a movie theater!

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    There is, perhaps, no more burning topic at the moment than the future of work, and so I am particularly grateful to welcome to the show Gary Bolles, author of The Next Rules of Work and a co-founder of eParachute.com, helping job-hunters & career changers with programs inspired by the evergreen book “What Color Is Your Parachute?” written by his father. Gary's courses on LinkedIn Learning have over 1 million learners and he is a former Silicon Valley executive and a co-founder of SoCap, the world’s largest gathering of impact entrepreneurs and investors.

    Gary is adjunct Chair for the Future of Work for Singularity University, and as a partner in the consulting agency Charrette, he helps organizations, communities, educators and governments develop strategies for “what’s next.”

    In the conclusion of the interview, we talk about unbossing and holacracies, how AI will impact organizational structures, fear, FOMO, and agency, and the Singularity University.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.

  • This and all episodes at: https://aiandyou.net/ .

    There is, perhaps, no more burning topic at the moment than the future of work, and so I am particularly grateful to welcome to the show Gary Bolles, author of The Next Rules of Work and a co-founder of eParachute.com, helping job-hunters & career changers with programs inspired by the evergreen book “What Color Is Your Parachute?” written by his father. Gary's courses on LinkedIn Learning have over 1 million learners and he is a former Silicon Valley executive and a co-founder of SoCap, the world’s largest gathering of impact entrepreneurs and investors.

    Gary is adjunct Chair for the Future of Work for Singularity University, and as a partner in the consulting agency Charrette, he helps organizations, communities, educators and governments develop strategies for “what’s next.”

    In the first half of the interview, we talk about the gig economy, the new rules of work, what ChatGPT did to the job market, and an interesting concept called the community operating system.

    All this plus our usual look at today's AI headlines.

    Transcript and URLs referenced at HumanCusp Blog.