Dear Friends and University Students,
Today, Andrew Ng, Chairman of Coursera, Founder & CEO of Landing AI, Founder of DeepLearning.AI, Managing General Partner at AI Fund, with a B.S. from Carnegie Mellon University, M.S. from MIT, and a Ph.D. from the University of California, Berkeley, as well as Director of the Stanford AI Lab, raised an important issue regarding the "Risk of regulations for companies developing foundational AI models in the name of National Security".
You might be wondering, what is a Foundational AI Model? These models are typically trained on large datasets to learn patterns and relationships in data and can be adapted or fine-tuned for specific tasks.
Some examples include
-
LeNet-5 (a classic convolutional neural
network for handwritten digit recognition)
-
AlexNet (for image classification),
-
LSTM (Long Short-Term Memory, used in speech
recognition and sentiment analysis)
-
GPT (Generative Pre-trained Transformer, for
language translation, chatbots, and content generation like ChatGPT)
-
BERT (Bidirectional Encoder Representations
from Transformers, for text classification and question answering)
-
DQN (Deep Q-Network, used in video games and
controlling robots)
-
YOLO (You Only Look Once, for detecting
objects in images and video streams).
Andrew argues that "adding burdens to foundational model development
unnecessarily slows down AI's progress" and suggests that regulation
should focus on the application layer (e.g., underwriting software, healthcare
applications, self-driving, and chat applications).
In my view, we need to strike a balance between innovation and safety. In 2023, the European Parliament passed the AI Act, which is expected to come into force in 2025. The South Korean government has also announced a $480 million investment over five years to develop AI foundation models.
In this context, companies often face confusion regarding what to allow and what not to do. While it's relatively easy to restrict a particular feature, application, or use case, restrictions at the foundational level can hinder overall progress.
However, certain precautions are essential. For instance, what if a medical dataset is fabricated or
hallucinated to appear authentic? As Stephen King, Senior Fellow at HEA and
Senior Lecturer in Media at Middlesex University Dubai, points out,
"Visual artists legitimately fear that new works created in their style
will compete with their copyrighted works or worse, that works will be
attributable to them that they did not create."
What is the solution for this?
- Enterprises
need to create their own in-house proprietary AI models and restrict
external platforms/models.
- Developing
ethical talent is crucial. While it's easy to recruit high-IQ competitive
coders, the challenge is to hire and certify highly ethical AI developers.
- Researcher/Faculty
recruitment and promotions should not be based solely on the number of
papers, as AI can now generate research papers. Ethical considerations
should also play a significant role.
- Design Thinking should be incorporated as a mainstream subject in colleges and schools. Philosophy and Ethics are no longer free electives; we need to devise methods to instill these values. This is a key challenge for universities and enterprises.
The question we face is whether we are
prepared to adopt AI without compromising on ethics. This is achievable by
including ethics education in the curriculum with the same level of importance
as core domain subjects
Dr. Ravi Saripalle
Director, Center for Innovation and Incubation, GVPCE
Founder, Inspire to Innovate Storytelling Movement (http://i2iTM.blogspot.com)
https://www.linkedin.com/in/ravisaripalle/
No comments:
Post a Comment