translation

This is an AI translated post.

Byungchae Ryan Son

The Paradox of Leading AI Models, Transparency

Select Language

  • English
  • 汉语
  • Español
  • Bahasa Indonesia
  • Português
  • Русский
  • 日本語
  • 한국어
  • Deutsch
  • Français
  • Italiano
  • Türkçe
  • Tiếng Việt
  • ไทย
  • Polski
  • Nederlands
  • हिन्दी
  • Magyar

Summarized by durumis AI

  • Researchers at Stanford University have analyzed 10 state-of-the-art AI systems, including GPT-4, and found that there is a lack of transparency in AI models, such as datasets and learning methods.
  • In particular, major AI companies such as OpenAI and Google are taking a passive approach to data disclosure as they pursue a profit-oriented business model, which could hinder the development of AI technology and lead to a monopolistic future.
  • Experts argue that the transparency of AI models should be increased to ensure reproducibility and strengthen social responsibility, urging discussions on social consensus and regulation along with the development of AI technology.

A study published by Stanford University researchers on the 18th reveals how deep and potentially dangerous the secrecy surrounding GPT-4 and other cutting-edge AI systems is.

Introducing The Foundation Model Transparency Index, Stanford University


They investigated a total of 10 different AI systems, most of which were large language models like those used in ChatGPT and other chatbots. This included widely used commercial models such as OpenAI's GPT-4, Google's PaLM 2, and Amazon's Titan Text, and they assessed openness based on 13 criteria, including how transparent developers were about the data used to train the models (data collection and annotation methods, whether copyrighted materials were included, etc.). They also investigated whether they disclosed the hardware used to train and run the models, the software frameworks used, and the project's energy consumption.


The results showed that none of the AI models achieved more than 54% on the transparency scale across all criteria. Overall, Amazon's Titan Text was ranked as the least transparent, while Meta's Llama 2 was chosen as the most open. Interestingly, Llama 2, which is a representative of the recent confrontation between open and closed models and is an open-source model, did not disclose the data used in training, the data collection and curation methods, etc. In other words, despite the increasing influence of AI on our society, opacity in the industry is a widespread and persistent phenomenon.


This means that the AI industry is at risk of becoming a profit-driven sector rather than a scientific advancement sector, and it could lead to a monopolistic future driven by specific companies..

Eric Lee/Bloomberg via Getty Images


OpenAI CEO Sam Altman has already met with policymakers around the world, openly stating his willingness to explain this unfamiliar and new intelligence to them and help them concretize related regulations. However, while he supports the idea of an international body to oversee AI in principle, he believes that some limited rules, such as banning all copyrighted material from data sets, could become unfair obstacles. This is a clear sign that the "openness" embedded in the company name OpenAI has deteriorated from the radical transparency it presented at its launch.


However, as the results of this Stanford report show, it is also worth noting that there is no need to keep each model so secret for the sake of competition. Because these results are also an indicator of the shortcomings of almost all companies. For example, there are reportedly no companies that provide statistics on how many users rely on their models or on the regions or market segments where their models are used.


Among organizations that adhere to open source principles, there is a saying, "Many eyes make all bugs shallow." (Linus's law) Crude numbers help to find problems that can be solved and fixed.


However, the practice of open source tends to gradually lose social standing and value recognition within and outside public companies., so there is no great significance in emphasizing it unconditionally. Therefore, rather than sticking to the framework of whether a model is open or closed, it is a better choice to focus the discussion on gradually expanding external access to the "data" that forms the basis of powerful AI models..


Scientific progress requires ensuring reproducibility, confirming whether specific research results reappear. If concrete plans are not implemented to ensure transparency regarding the key components of each model's creation, the industry will ultimately be more likely to remain in a closed and stagnant monopolistic situation. And this needs to be considered a high priority in the current and future situations where AI technology is rapidly permeating all industries.


It has become important for journalists and scientists to understand data, and transparency is a prerequisite for planned policy efforts for policymakers. Transparency is also important for the public, as end users of AI systems can be perpetrators or victims of potential problems related to intellectual property, energy consumption, and bias. Sam Altman argues that the risk of human extinction from AI should be a global priority, on par with social-scale risks such as pandemics or nuclear war. However, we must not forget that our society's survival, maintaining a healthy relationship with developing AI, is a prerequisite for reaching the dangerous situation he mentions.


*This article is the original content published in the signed column of the e-newspaper on October 23, 2023.


References


Byungchae Ryan Son
Byungchae Ryan Son
찾아가 관찰하고 경청하는 일을 합니다.
Byungchae Ryan Son
OpenAI's Governance Drama: The Ethical Future of AI The firing and subsequent reinstatement of OpenAI CEO Sam Altman by the board, within five days, highlights the unusual structure of OpenAI, which claims to be pursuing a mission for humanity, and raises concerns about the potential risks of large-scale t

May 14, 2024

The era of algorithmic branding is coming Elon Musk, Google CEO Sundar Pichai, and Sam Altman of OpenAI, AI experts, warn of the dangers of artificial intelligence and offer different solutions for the future of humanity. In an electronic newspaper column on April 24, 2023, check out their argume

May 10, 2024

The Evolving Relationship Between Us and Algorithms Recent advancements in generative AI technology have redefined the relationship between humans and algorithms. The author delves into the question of how humans should engage with algorithms in the age of generative AI like ChatGPT, particularly addressin

May 9, 2024

Building an AI Full Stack with Open Source New open source LLM (Large Language Model) models are emerging in the AI ecosystem. Powerful models with open licenses, such as Mistral, Llama, and phi-2, have been released, and various tools to use them are also being developed. From LLM frameworks such
RevFactory
RevFactory
RevFactory
RevFactory

February 5, 2024

Snowflake Launches 'Arctic,' an Enterprise-Grade LLM with Industry-Leading Openness Snowflake has launched 'Arctic,' an open-source large language model. It's free for commercial use under the Apache 2.0 license and boasts industry-leading performance and efficiency. It offers exceptional token efficiency and top-notch quality, supportin
스타트업 커뮤니티 씬디스 (SeenThis.kr)
스타트업 커뮤니티 씬디스 (SeenThis.kr)
스타트업 커뮤니티 씬디스 (SeenThis.kr)
스타트업 커뮤니티 씬디스 (SeenThis.kr)

April 25, 2024

Apple's OpenELM / MS's Phi-3 / Meta's Llama 3 Released Major tech companies such as Apple, Microsoft, and Meta are injecting new energy into the AI industry by recently releasing their own large language models. These newly released models are evolving in various directions, including size reduction, data opt
해리슨 블로그
해리슨 블로그
해리슨 블로그
해리슨 블로그

April 27, 2024

Mr. Know-All – 2023.7 The first issue of "Mr. Know-All," a monthly AI magazine in July 2023, introduces the latest AI technologies and trends, including Claude 2, Azure OpenAI, LangChain, and LlamaIndex. In particular, it provides a detailed explanation of LlamaIndex, which em
Pilot AISmrteasy
Pilot AISmrteasy
Pilot AISmrteasy
Pilot AISmrteasy

March 21, 2024

Frost & Sullivan Announces Top 10 AI Market Forecasts for 2024 The global AI market is expected to grow by approximately 10% in 2024, reaching $340 billion, with AI adoption expected to increase particularly in the manufacturing, financial, and healthcare sectors. Frost & Sullivan believes that generative AI will pla
스타트업 커뮤니티 씬디스 (SeenThis.kr)
스타트업 커뮤니티 씬디스 (SeenThis.kr)
스타트업 커뮤니티 씬디스 (SeenThis.kr)
스타트업 커뮤니티 씬디스 (SeenThis.kr)

May 10, 2024

Splunk Releases 2024 Security State Report According to Splunk's 2024 Security State Report, 93% of organizations are using generative AI, but 34% haven't established policies yet. Generative AI offers new opportunities for both cybersecurity professionals and attackers, and it is important to und
스타트업 커뮤니티 씬디스 (SeenThis.kr)
스타트업 커뮤니티 씬디스 (SeenThis.kr)
스타트업 커뮤니티 씬디스 (SeenThis.kr)
스타트업 커뮤니티 씬디스 (SeenThis.kr)

May 13, 2024