Blog

Open source artificial intelligence (AI) is an opportunity to democratise innovation and avoid the concentration of power in the technology industry. However, their development is highly dependent on the availability of high quality datasets and the implementation of robust data governance frameworks. A recent report by Open Future and the Open Source Initiative (OSI) analyses the challenges and opportunities at this intersection, proposing solutions for equitable and accountable data governance.  You can read the full report here.

In this post, we will analyse the most relevant ideas of the document, as well as the advice it offers to ensure a correct and effective data governance in artificial intelligence open source and take advantage of all its benefits.

The challenges of data governance in AI

Despite the vast amount of data available on the web, accessing and using it to train AI models poses significant ethical, legal and technical challenges. For example:

  1.  Balancing openness and rights: In line with the Data Governance Regulation (DGA), broad access to data should be guaranteed without compromising intellectual property rights, privacy and fairness.
  2. Lack of transparency and openness standards: It is important that models labelled as "open" meet clear criteria for transparency in the use of data.
  3. Structural biases: Many datasets reflect linguistic, geographic and socio-economic biases that can perpetuate inequalities in AI systems.
  4. Environmental sustainability: the intensive use of resources to train AI models poses sustainability challenges that must be addressed with more efficient practices.
  5. Engage more stakeholders: Currently, developers and large corporations dominate the conversation on AI, leaving out affected communities and public organisations.

Having identified the challenges, the report proposes a strategy for achieving the main goal: adequate data governance in open source AI models. This approach is based on two fundamental pillars.

Towards a new paradigm of data governance

Currently, access to and management of data for training AI models is marked by increasing inequality. While some large corporations have exclusive access to vast data repositories, many open source initiatives and marginalised communities lack the resources to access quality, representative data. To address this imbalance, a new approach to data management and use in open source AI is needed. The report highlights two fundamental changes in the way data governance is conceived:

On the one hand, adopting a data commons approach  which is nothing more than an access model that ensures a balance between data openness and rights protection.. To this end, it would be important to use innovative licences that allow data sharing without undue exploitation. It is also relevant to create governance structures that regulate access to and use of data. And finally, implement compensation mechanisms for communities whose data is used in artificial intelligence.

On the other hand, it is necessary to transcend the vision focused on AI developers and include more actors in data governance, such as:

  • Data owners and content-generating communities.
  • Public institutions that can promote openness standards.
  • Civil society organisations that ensure fairness and responsible access to data.

By adopting these changes, the AI community will be able to establish a more inclusive system, in which the benefits of data access are distributed in a manner that is equitable and respectful of the rights of all stakeholders. According to the report, the implementation of these models will not only increase the amount of data available for open source AI, but will also encourage the creation of fairer and more sustainable tools for society as a whole.

Advice and strategy

To make robust data governance effective in open source AI, the report proposes six priority areas for action:

  1. Data preparation and traceability: Improve the quality and documentation of data sets.
  2. Licensing and consent mechanisms: allow data creators to clearly define their use.
  3. Data stewardship: strengthen the role of intermediaries who manage data ethically.
  4. Environmental sustainability: Reduce the impact of AI training with efficient practices.
  5. Compensation and reciprocity: ensure that the benefits of AI reach those who contribute data.
  6. Public policy interventions: promote regulations that encourage transparency and equitable access to data.

How to achieve good data governance in open source AI?  Improving data quality and traceability  Enable clear licensing and consent  Strengthen data stewardship of intermediaries  Reduce the environmental impact of AI training  Ensure proper compensation of AI benefits.  Promote public policies that encourage transparency and equitable access to data.  Source: ‘Data Governance in Open Source AI’. Open Source Initiative and Open Future. Available here: Reimagining data for Open Source AI: A call to action

Open source artificial intelligence can drive innovation and equity, but to achieve this requires a more inclusive and sustainable approach to data governance. Adopting common data models and broadening the ecosystem of actors will build AI systems that are fairer, more representative and accountable to the common good.

The report published by Open Future and Open Source Initiative calls for action from developers, policymakers and civil society to establish shared standards and solutions that balance open data with the protection of rights. With strong data governance, open source AI will be able to deliver on its promise to serve the public interest.

calendar icon
Blog

The   Big Data Test Infrastructure (BDTI) is a tool funded by the European Digital Agenda, which enables public administrations to perform analysis with open data and open source tools in order to drive innovation.

This free-to-use, cloud-based tool was created in 2019 to accelerate digital and social transformation. With this approach and also following the European Open Data Directive, the European Commission concluded that in order to achieve a digital and economic boost, the power of public administrations' data should be harnessed, i.e. its availability, quality and usability should be increased. This is how BDTI was born, with the purpose of encouraging the reuse of this information by providing a free analysis test environment that allows public administrations to prototype solutions in the cloud before implementing them in the production environment of their own facilities.

What tools does BDTI offer?

Big Data Test Infrastructure offers European public administrations a set of standard open source tools for storing, processing and analysing their data. The platform consists of virtual machines, analysis clusters, storage and network facilities. The tools it offers are:

  • Databases: to store data and perform queries on the stored data. The BDTI currently includes a relational database(PostgreSQL), a document-oriented database(MongoDB) and a graph database(Virtuoso).
  • Data lake: for storing large amounts of structured and unstructured data (MinIO). Unstructured raw data can be processed with deployed configurations of other building blocks (BDTI components) and stored in a more structured format within the data lake solution.
  • Development environments: provide the computing capabilities and tools necessary to perform standard data analysis activities on data from external sources, such as data lakes and databases.
    • JupyterLab, an interactive, online development environment for creating Jupyter notebooks, code and data.
    • Rstudio, an integrated development environment for R, a programming language for statistical computing and graphics.
    • KNIME, an open source data integration, reporting and analytics platform with machine learning and data mining components, can be used for the entire data science lifecycle.
    • H2O.ai, an open sourcemachine learning ( ML) and artificial intelligence (AI) platform designed to simplify and accelerate the creation, operation and innovation with ML and AI in any environment.
  • Advanced processing: clusters and tools can also be created to process large volumes of data and perform real-time search operations(Apache Spark, Elasticsearch and Kibana)
  • Display: BDTI also offers data visualisation applications such as Apache Superset, capable of handling petabyte-scale data, or Metabase.
  • Orchestration: for the automation of data-driven processes throughout their lifecycle, from preparing data to making data-driven decisions and taking actions based on those decisions, is offered:
    • Apache Airflow, an open source workflow management platform that allows complex data pipelines to be easily scheduled and executed.

Through these cloud-based tools, public workers in EU countries can create their own pilot projects to demonstrate the value that data can bring to innovation. Once the project is completed, users have the possibility to download the source code and data to continue the work themselves, using environments of their choice. In addition, civil society, academia and the private sector can participate in these pilot projects, as long as there is a public entity involved in the use case.

Success stories

These resources have enabled the creation of various projects in different EU countries. Some examples of use cases can be found on the BDTI website. For example, Eurostat carried out a pilot project using open data from internet job advertisements to map the situation of European labour markets. Other success stories included the optimisation of public procurement by the Norwegian Agency for Digitisation, data sharing efforts by the European Blood Alliance and work to facilitate understanding of the impact of COVID-19 on the city of Florence .

In Spain, BDTI enabled a data mining project atthe  Conselleria de Sanitat de la Comunidad Valenciana. Thanks to BDTI, knowledge could be extracted from the enormous amount of scientific clinical articles, a task that supported clinicians and managers in their clinical practices and daily work.

OVERVIEW OF BDTI SUCCESS STORIES   Conselleria de Sanitat (Generalitat Valenciana): Text Mining   Extract knowledge from the huge quantity ofd scientific clinical articles, spporting clinicians and managers in their clinical practices and day-to-day work.   Norwegian Digital Agency (Digdir): Optimisation   Optimise public procurement in Norway, by gathering and analysing big datasets on transactions in this area.   European Blood Alliance: Data sharing   A ready-to-use virtual enviroment in which data collected though a custom-built website are ingested and anonymized, to be then analysed with advanced data visualization and analytical tools.   City of Florence: Mobility data   Predictive, descriptive and time-series analysis on multiple datasets collected before, aduring and after covid-19 pandemic such are public WiFi sensors, sharing and geo-referenced data of people  movement.   Eurostat, European Centre for Development of Vocational Training National Statistical Institutes: Labour market intelligence   Use online job advertisement data to provide timely information about the EU labour markets, application of Artificial Intelligence, Natural Language Processing and Machine Learning to clean text and extract the relevant data.

Courses, newsletter and other resources

 In addition to publishing use cases, theBig Data Test Infrastructure website offers an free online course to learn how to get the most out of BDTI. This course focuses on a highly practical use case: analysing the financing of green projects and initiatives in polluted regions of the EU, using open data from data.europa.eu and other open sources.

In addition, a monthly  newsletter on the latest BDTI news, best practices and data analytics opportunities for the public sector has recently been launched .

In short, the re-use of public sector data (RISP) is a priority for the European Commission and BDTI(Big Data Test Infrastructure) is one of the tools contributing to its development. If you work in the public administration and you are interested in using BDTI register here.

calendar icon
Evento

On June 1, Madrid will host the fourth edition of Conference on FLOSS (software and free code) and Open Economy. The aim of the Open Expo is bringing together leading companies and institutions, developers, hackers, experts, suppliers and users to learn about technology solutions and trends on open source, open source, open data and innovation.

Since 2012, each of the events has sought to promote the use and development of free and open software to boost the collaborative philosophy and democratize the access to information technologies. For that purpose, there have been several events dedicated to specific topics such as e-commerce, business intelligence, content managers or elearning, among others.

On this occasion, the Open Expo is focused on addressing the latest challenges related to open source and digital transformation. An opportunity to discover how this technology can modernize businesses and help companies on their way to innovation and the digital transformation of their corporate operations.

In this regard, the organisers of the event has launched a call to find speakers who participate in the congress sharing their success stories and experiences in the field of open technologies, showing how open source and free software have helped improve their business activities or presenting their open source projects.

To participate, it is necessary to apply before March 2 through the official page of the event; a jury will analyze the proposals and select the most relevant ideas in the field to be discussed.

In addition, among the activities organized this year, apart from the showroom where the main companies of the industry show their services and products, there will be an investment forum, Open StartUp Connector, where a twelve start-ups will present to potential investors their ICT projects based onpen code/data, or developed through free tools or software.

At the same time, networking activities with experts are also planned to discuss on cybersecurity, big data and IoT solutions and, moreover, the Open Awards Spain 2017 will take place, which awardsthe best solutions developed with open source technology at the national level.

 

 

calendar icon