Jim and Mike on the Potential and Limitations of ChatGPT
Published April 7, 2023
Courses Mentioned in this Post: Data Privacy and Technology
Series Mentioned in this Post: Harvard On Digital
What are the 5 love languages? Who escaped from Alcatraz? When is the next full moon? These are some of the most searched questions on Google in 2023.
Since its introduction in 1998, Google has revolutionized the way we consume information and has allowed people all over the world to access news and information faster than ever before.
Mike Smith and Jim Waldo
In 2022, a new information processing and delivery technology called ChatGPT was announced and made available for free. ChatGPT is the fastest-growing consumer application in history, growing from one million users just after its launch in late November 2022, to over 100 million users in January 2023. Just like Google, this new generative AI technology has the potential to alter the way we engage with information and create content on the internet.
“Tech companies continue to encourage all of us to act first and think of the implications later, if at all.” — Michael D. Smith
While more and more people log in to ChatGPT to get answers and experiment with the technology, questions around the potential impact on society and privacy have been raised for many data technology experts and advocates.
Leaders in the fields of computer science and data privacy, Michael D. Smith (Professor of Engineering and Applied Sciences) and Jim Waldo (Professor of the Practice of Computer Science), answer questions about ChatGPT and offer their thoughts on the risks and rewards that accompany widespread generative AI technology use.
Have you used ChatGPT?
Michael D. Smith (MDS): Jim uses it, but I’ve used it only through my students.
Why? Two reasons. First, I’m less interested in how I might use the tool and much more interested in how those in their thirties, twenties, and teens think about and use it. Second, my students and I are completing a technical paper about language bias in tools like ChatGPT, Google, Wikipedia, and YouTube.
By language bias, I mean how these tools, and others like them, use the language of your query to present cultural stereotypes tied to the language you use in your query. Despite being trained on the global internet, these tools too often turn us into the proverbial blind person touching a small portion of an elephant, ignorant of the existence of other perspectives.
Jim Waldo (JW): As Mike said, I’ve been using ChatGPT recently, but in a somewhat non-standard way—I use it as a second in doing pair programming. This means that when I’m writing some code, I describe what I’m wanting in the user interface and let ChatGPT generate a first version of the code. I may iterate over this multiple times, as the AI isn’t all that great as a programmer, but it does have a wide scope on the appropriate libraries to use for various components. One of the things that I’ve been impressed with is that it will tell me where in the code it generates I should worry that it might have it wrong.
That said, the final code is something I write, since the first sketch coming out of ChatGPT is generally pretty inadequate. It isn’t very good about security, or error handling. Which, I suppose, is also a comment on the general quality of the code on which it was trained.
I’ve also used ChatGPT as an alternative to Google search when I’m starting research on a topic, but I find that I have to insist on getting references for the claims that it makes. I’ve encountered a number of times where the information it gives me is simply wrong. Everything needs to be verified (which, to its credit, it tells you as part of the answer).
What interests or concerns do you have about the rise of generative AI technologies?
MDS: Deepfakes. Soon anyone with a smartphone will be able to create really good ones with very little work. Perhaps in combination with a tool like ChatGPT:
“Hey ChatGPT, take this video I just took of some idiot doing something dumb, put this other person’s face on it, make the voice sound like this person, and then post it on my social media channel with some witty caption.”
This is just harmless fun, right? Maybe this post will be promoted by the platform and turn me into an influencer. Not only are tech companies too often focused on what novel things you can do with their latest product, but they continue to encourage all of us to act first and think of the implications later, if at all. Then again, the trend will make our course, Data Privacy and Technology, even more important!
JW: I agree with Mike about the concern over Deepfakes. We are going to have to figure out how to deal with the provenance of information in ways we have not had to do before.
I also worry about the combination of generative AI (such as ChatGPT) and bots that can be used to flood public discussions of policy issues. When agencies put out a regulation for public comment, it is now possible to flood those comment servers with plausible comments, either pro or con, that are written by AI. This is an automated version of the tactic of flooding the zone with s**t, that is going to be very hard to deal with.
Interested in learning more about trending topics in data privacy from Mike and Jim? Stay tuned to the Harvard Online blog page for their take on ChatGPT and Generative AI, or apply to join the next cohort of their course Data Privacy and Technology.
Related ArticlesJim and Mike on Data Privacy and TikTok Image
Keeping Your Data Secure Image
|