“Open source” Amnesty International is not really open – here is how researchers can restore the term

About 50 years ago this month, the Homebrew Computer Club-a group of computer and amateur-meeting lovers at Menlo Park, California, promotes a culture of cooperation, exchange of knowledge and open participation of programs. These values, which helped form an open source movement, are now sabotaged by some artificial intelligence companies (AI).

Several founding artificial intelligence models are classified as an “open source” because its structure, including the structure and design of the nerve networks available freely. However, a few information is revealed on how to train models. As the executive director of the Open source initiative (OSI), based in Palo Alto, California, my priority since 2022 shows what the term already means in the era of artificial intelligence.

Contracts of free access to non-refrigerated software-such as R Studio for Statistical Computing and OpenFOAm for fluid dynamics-hastened a scientific discovery. Open source programs protect the safety of research by ensuring cloning. It also enhances global cooperation, allowing scientists to share data and solutions freely.

Traditional open sources are built around the source code, which is easy to participate with full transparency, but artificial intelligence systems are different. They rely heavily on training data, often from property or protected sources under privacy laws, such as healthcare information.

Since artificial intelligence pushes discoveries in fields ranging from genome to climate modeling, the lack of a strong consensus as it is and not anxious open source. In the future, the scientific community can find its arrival limited to closed corporate systems and indispensable models.

In order for artificial intelligence systems to be in line with open source models, the freedom to use, study, modify and share their basic models must support their basic models. Although many of the artificial intelligence models that use the “open source” sign are free to use and participate, the inability to access training data and the source symbol strongly restricts a deeper study and amendment. For example, OSI analysis was found that many of the most famous large language models, such as Llama2 and Llama 3.x (developed by Meta), GROK (X), Phi-2 (Microsoft) and Mixtral (Mistral Ai), do not correspond to the principles of open source. In contrast, models such as Olmo, developed by the Allen Institute of the Antioral Organization, is a non-profit organization in Seattle in Washington and community-led projects such as Crystalcoder llm360-a language model designed to perform both programming tasks and natural language-a better vision than OSI from the open source.

The main reason why some companies may harm the use of the open source poster is to avoid the proposed regulations under the AI ​​Law of the European Union for the year 2024, which exempt free and open programs from strict scrutiny. This practice is called – companies that demand openness while restricting access to major components such as information about training data – OpenWooking.

Leave a Comment