HORIZON-CL4-2025-04-DIGITAL-EMERGING-04
Assessment methodologies for General Purpose AI capabilities and risks (RIA) (AI/Data/Robotics Partnership) -
About the connections
The graph above was generated based on the following links
Call text (as on F&T portal)
View on F&T portalProject results are expected to contribute to some of the following expected outcomes:
- New assessment and validations methodologies developed allowing to evaluate General Purpose AI (GPAI) models, including multimodal systems, and systems’ capabilities and risks.
- Use of the research outcomes by GPAI providers, policymakers, public institutions, and other relevant stakeholders to evaluate GPAI models and systems’ capabilities and risks.
- Support to the AI Office in its function to conduct evaluations of general purpose AI models with a view to enforce the AI Act' rules for general purpose AI models and facilitate self-evaluation for GPAI model developers to ensure compliance with AI Act requirements.
The rapid advancement of artificial intelligence (AI) has led to the development of increasingly sophisticated general-purpose AI (GPAI) models and systems. These models, such as large language models and multimodal AI systems, demonstrate remarkable capabilities across a wide range of tasks. However, assessing the capabilities of these models remains a significant challenge. Traditional evaluation methods often fail to capture the full spectrum of abilities exhibited by GPAI models and systems. Therefore, there is a pressing need for the development of new assessment frameworks, methodologies and tools that can comprehensively evaluate these models in terms of their trustworthy and ethical behaviour and operation, ensuring their reliability, fairness, and alignment with human values.
This topic aims to develop robust assessment tools, techniques, and benchmarks specifically designed to rigorously evaluate GPAI models and systems, including multimodal systems. Proposals should cover one or more of the following research areas:
- Innovative methods for proactively identifying and forecasting emergent capabilities in GPAI models and systems. This encompasses the identification of capabilities with both beneficial and potentially detrimental uses.
- Assessment of GPAI capabilities with a significant economic impact or potential for misuse. This includes assessing capabilities that drive beneficial innovation and societal good, as well as evaluating potential risks in areas such as chemical, biological, radiological, and nuclear (CBRN) hazards or cybersecurity threats.
- Developing assessment techniques that illuminate the underlying mechanisms of emergent capabilities in AI systems, emphasising interpretability and explainability.
Projects should generate example benchmark tests to examine trained AI models, systematically uncovering latent capabilities. These benchmarks will be made available to GPAI providers, policymakers, and other relevant stakeholders to implement robust evaluation tools.
This topic strongly encourages the formation of interdisciplinary teams combining the necessary technical expertise. Such a collaborative approach will ensure that assessments accurately capture real-world use cases, including capabilities elicitation techniques, and that the developed frameworks, methodologies and tools are responsive to the concerns of all relevant stakeholders.
This topic requires the effective contribution of SSH disciplines and the involvement of SSH experts, institutions as well as the inclusion of relevant SSH expertise, in order to produce meaningful and significant effects enhancing the societal impact of the related research activities.
Proposals must adhere to Horizon Europe's requirements regarding Open Science. Open access to research outputs should be provided unless there is a legitimate reason or constraint; in such cases, the proposal should detail how GPAI providers, policymakers, and other stakeholders will access the research outcomes.
All proposals are expected to incorporate mechanisms for assessing and demonstrating progress, including qualitative and quantitative KPIs, benchmarking, and progress monitoring. This should include participation in international evaluation contests and the presentation of illustrative application use-cases that demonstrate concrete potential added value. Communicable results should be shared with the European R&D community through the AI-on-demand platform, and if necessary, other relevant digital resource platforms to bolster the European AI, Data, and Robotics ecosystem by disseminating results and best practices.
This topic implements the co-programmed European Partnership on AI, data and robotics (ADRA), and all proposals are expected to allocate tasks for cohesion activities with ADRA and the CSA HORIZON-CL4-2025-03-HUMAN-18: GenAI4EU central Hub.
Proposals should also build on or seek collaboration with existing projects and develop synergies with other relevant International, European, national or regional initiatives. Regarding European programmes, projects are expected to develop synergies and complementarities with relevant projects funded under Horizon Europe but also under the Digital Europe Programme (DEP).
News flashes
Please note that due to a technical issue, during the first days of publication of this call, the topic page did not display the description of the corresponding destination. This problem is now solved.
In addition to the information published in the topic page, you can always find a full description of the Destination 4 ("Achieving open strategic autonomy in digital and emerging enabling technologies") that is relevant for the call in the Work Programme 2025 part for "Digital, Industry and Space". Please select from the work programme the destination relevant to your topic and take into account the description and expected impacts of that destination for the preparation of your proposal.
Publication date: 2025-05-14 (1 month ago)
Opening date: 2025-06-10 (4 days ago)
Closing date: 2025-10-02 (3 months from now)
Procedure: single-stage
Budget: 7000000
Expected grants: 2
Contribution: 3000000 - 4000000
This call topic has been appended 2 times by the EC with news.
-
2025-06-14
please note that due to a technical issu... -
2025-06-14
the submission session is now available...
HORIZON-CL4-2025-04
Call topics are often grouped together in a call. Sometimes this is for a thematic reason, but often it is also for practical reasons.
There are 6 other topics in this call:
Showing the latest information. Found 4 versions of this call topic in the F&T portal.
Information from
- 2025-06-13_03-30-09
- 2025-06-10_03-30-13
- 2025-06-03_03-30-15
- 2025-05-15_03-30-17
Check the differences between the versions.
Annotations (will be publicly visible when approved)
Events
Events are added by the ideal-ist NCP community and are hand-picked. If you would like to suggest an event, please contact idealist@ffg.at.
Call topic timeline
-
Work programme available
- 3 months agoThe call topics are published first in the Work Programme, which is available a while before the call opens. By following up the Work Programme publications, you can get a headstart.
-
Publication date
- 1 month agoThe call was published on the Funding & Tenders Portal.
-
Opening date
- 4 days agoThe call opened for submissions.
-
Today
-
Closing date
- 3 months from nowDeadline for submitting a project.
-
Time to inform applicants Estimate
- 8 months from nowThe maximum time to inform applicants (TTI) of the outcome of the evaluation is five months from the call closure date.
-
Sign grant agreement Estimate
- 11 months from nowThe maximum time to sign grant agreements (TTG) is three months from the date of informing applicants.
Funded Projects
Loading...
Project information comes from CORDIS (for Horizon 2020 and Horizon Europe) and will be sourced from F&T Portal (for Digital Europe projects)