In recent years, the field of artificial intelligence (AI) has made significant strides, sparking growing attention towards regulations governing its use. Among these, an important initiative is the EU AI Act, which introduces stringent requirements for providers of general-purpose AI models. Among the most notable provisions of the regulation is the obligation to provide detailed documentation regarding the training process of the models.
This requirement, far from being a mere bureaucratic formality, serves as an important step towards greater transparency in the functioning of AI models, which are often described as “black boxes.” The need to clarify how these algorithms are developed and trained is vital, especially in a context where automated decisions can have significant repercussions on people’s lives.
AI model providers are required to draft and maintain up-to-date technical documentation that must include crucial information. Among these, the sources of the data used for training stand out, which can include texts, images, and code, as well as the methods used for filtering and cleaning this data, not to mention the computational resources employed. This information not only helps to ensure greater clarity but is also essential for identifying and mitigating potential risks associated with the use of AI.
Indeed, one of the main objectives of this obligation is to assess and reduce the risks arising from the use of AI models. In particular, it is necessary to pay attention to the possibility that biases may be present in the training data, which, if uncorrected, can lead to discriminatory outcomes. Furthermore, there is a need to ensure that personal data is not used without a valid legal basis, thereby avoiding potential violations of privacy.
Another aspect of fundamental importance regarding the documentation required by the EU AI Act concerns compliance with copyright law. It is essential that the documentation includes a detailed summary of the copyrighted data used during the training phase, allowing rights holders to verify whether their works have been included and to exercise their rights, such as the possibility to request the exclusion of their creations (opt-out). This aspect of the regulation highlights the need for careful governance in the development processes of AI, as violations of copyright could incur significant penalties for the companies involved.
However, for companies engaged in the development of AI models, compliance with these requirements poses a considerable challenge. Robust systems will need to be adopted to trace the provenance of the data and to implement rigorous governance processes. These implementations will not only require significant investments in terms of resources but also a cultural shift within the organizations where the technologies are developed. Transparency and accountability will become essential aspects of work practice, thereby contributing to building a more ethical and responsible AI ecosystem.
On the other hand, compliance with these obligations represents a unique opportunity for society. The demand for greater transparency in the documentation of AI models marks a crucial step towards shared responsibility and a deeper understanding of how these sophisticated technologies are created and implemented. In a context where public trust in AI is essential, such practices will help to establish a stronger relationship between developers and users.
In conclusion, the requirement for detailed documentation imposed by the EU AI Act is not merely a legal formality, but represents a fundamental element to ensure that artificial intelligence is used safely and ethically. The challenge for companies will be to adapt to these new regulations by developing governance models and work practices that uphold transparency and accountability.
We invite all readers to follow our updates.

