IEEE Logo
Trismik Logo Vela Logo

Thanks to the AI4Finance Foundation open source community for their support.

Introduction

As AI continues to advance at a fast pace, more FinAI agents are being developed for the finance sector, such as FinRL trading agents [1,2,3], FinGPT agents [4,5] with multimodal capabilities [6], and regulatory reporting agents [7]. The Secure FinAI Contest 2026 encourages the development of FinAI agents based on the frameworks FinRL [2,3] and FinGPT [4].

We design four tasks. These challenges allow contestants to participate in various financial tasks and contribute to secure finance using state-of-the-art technologies with privacy-preserving and verifiable computation frameworks. We welcome students, researchers, and engineers who are passionate about finance, machine learning, and security to partake in the contest.

Tasks

Each team can choose to participate in one or more tasks. The prizes will be awarded for each task.

Task I: Adaptive Evaluation and Benchmarking Suite for Financial LLMs and Agents

Why this matters: Adaptive testing enables accurate model evaluation using only a dynamic subset of test items, significantly reducing evaluation overhead and cost compared to static benchmarks. Furthermore, we encourage comparison against and defeating BloombergGPT to showcase superior financial reasoning capabilities.

This task focuses on benchmarking Financial Large Language Models (FinLLMs) and agents using an adaptive testing pipeline. Unlike traditional benchmarks, the adaptive pipeline partitions the test set into difficulty levels and dynamically selects test items based on model performance. This enables more efficient evaluation while preserving rigour. Participants are expected to submit models that can handle a diverse range of financial reasoning and comprehension tasks, optimised for both accuracy and inference efficiency.

Datasets: We utilize the standard evaluation suite referenced in BloombergGPT, integrated into the adaptive framework. This includes FPB (sentiment analysis), FiQA-SA (aspect-based sentiment), Headlines (market news classification), NER (entity recognition), and ConvFinQA (conversational financial Q&A).

Performance Comparison:

  BloombergGPT GPT-NeoX OPT 66B BLOOM 176B
ConvFinQA 43.41 30.06 27.88 36.31
FiQA SA 75.07 50.59 51.60 53.12
FPB 51.07 44.64 48.67 50.25
Headline 82.20 73.22 79.41 76.51
NER 60.82 60.98 57.49 55.56
All Tasks (avg) 62.51 51.90 53.01 54.35
All Tasks (WR) 0.93 0.27 0.33 0.47

Table 8: Results on financial domain tasks.

Task II: Reliable Agentic FinSearch

Why this matters: Accurate, hallucination-free retrieval of real-time financial data is fundamental for trustworthy financial agents.

Task Description: Numerical accuracy is critical for financial tasks. On a small curated question set, we manually compared Agentic FinSearch (a google chrome extension) with Perplexity (with financial support, Link). Our Agentic FinSearch outperforms Perplexity substantially!

This task invites teams to try our Agentic FinSearch demo and test your LLMs or agents on a large question set. Accurate retrieval of real-time financial data is a core infrastructure for building FinAgents. Higher numerical accuracy, lower hallucination & misinformation, and air-gapped deployment are three features. We encourage you to build powerful Agentic FinSearch agents.

Fig. 1. Agentic FinSearch on Yahoo Finance

Fig. 1. Agentic FinSearch on Yahoo Finance.

Question sets for benchmarking:

  • Real-time retrieval of market data
  • Simple lookup of historical data
  • Complex computations over historical data
  • Hallucinations and misinformation

Task III: Prediction Market Arbitrage

Why this matters: Exploiting inefficiencies in prediction markets demonstrates an agent’s ability to synthesize information and react faster than human traders.

This task focuses on developing trading agents that identify and execute arbitrage opportunities across two prediction markets, Kalshi and Polymarket, for a series of sports events with binary options. Models may incorporate sentiment signals in addition to market data to anticipate market moves when new information changes expectations during a game. Evaluation will be conducted via paper trading, where agents perform simulated trading on real market data without real capital.

Datasets:

  • Kalshi Data: Market metadata and public market snapshots collected via the Kalshi API, with optional WebSocket streams for real-time updates including order book changes, public trades, and price/ticker movements.

Kalshi API Docs — https://docs.kalshi.com/

  • Polymarket Data: Market metadata from Polymarket’s public Gamma API (series, events, markets, and outcome identifiers), paired with real-time updates from the CLOB WebSocket market feed to track live pricing and liquidity.

Polymarket Developer Docs — https://docs.polymarket.com/quickstart/overview

  • Sentiment Data: Sports news collected from RSS feeds across multiple sources, including breaking headlines and alerts, injury and lineup updates, and other game-related developments that can be used as sentiment inputs.

Task IV: AI for Venture Capital - Prediction of Startup Success

Why this matters: Venture capital investment is high-risk; AI can help identify promising startups by analyzing patterns in founder profiles that humans might miss.

This task tests the ability of Large Language Models to act as Venture Capitalists by predicting the potential success of early-stage startups. Using the VCBench dataset, which consists of anonymized founder profiles, participants must predict whether a startup will achieve a significant liquidity event (IPO, M&A >$500M, or high-tier funding).

Goal & Constraints:

  • Objective: Predict the binary “Success” label for given founder profiles.
  • Optimization: Participants are encouraged to optimize input templates and output extraction methods alongside model fine-tuning.
  • Metric: F1-Score.

Task V: Agentic Trading Using OpenClaw

Why this matters: AI agents have seen rapid development and have been applied to financial trading tasks. Recent work such as Agent Market Arena (AMA) by FinAI and Alpha Arena by NoF1 introduces arena-style benchmarks that demonstrate the capability of AI agents to trade in market environments under standardized settings. While these benchmarks focus on comparing agent performance within fixed interfaces, this task emphasizes flexibility and agent design. Participants are encouraged to build their own trading agents, including data, agent architecture, and trading strategy, and evaluate them through live paper trading on Alpaca for stock or cryptocurrency markets.

This task focuses on developing LLM agents using OpenClaw for quantitative trading in a real-time paper trading setting. It formulates trading as a reasoning-to-action problem, where OpenClaw agents integrate real-time signals such as prices, newsflow, and asset-specific fundamentals and make trading actions. Evaluation will be conducted via paper trading on an Alpaca paper trading account, with metrics emphasizing profitability and risk.

Datasets Participants may fetch historical market data from Alpaca Market Data API:

Evaluation Participants will create an Alpaca paper trading account, run their agent to trade during the evaluation period, and submit required files at the end of the evaluation period.

  • Time Period: April 20 – May 1.
  • Initial Capital: Each account starts with a fixed capital of $100,000.

What to Submit

  • an orders JSONL file (time-ordered trading actions),
  • a daily equity CSV file,
  • a snapshot of the Alpaca portfolio value at the end of the evaluation period.

Metrics

  • Primary: Cumulative Return (CR)
  • Secondary: Sharpe Ratio (SR), Maximum Drawdown (MD), Daily Volatility (DV), Annualized Volatility (AV)

A single paper trading account trade asset from only one market type: stock/crypto. Create two accounts if you are participating in both the stock and crypto tracks. Stock and crypto tracks are evaluated separately.

Task VI: Financial Auditing Model Training and Evaluation over XBRL Filings (Taxonomy-Grounded)

Why this matters: Automating financial auditing with high precision ensures compliance and trust in financial reporting, reducing manual effort and errors.

This task focuses on training, developing, and evaluating specialized auditing models for professional-grade financial auditing over structured, taxonomy-driven financial disclosures. Participants are required to build an auditing-oriented model that reasons jointly over multi-document XBRL filings and the US-GAAP taxonomy, with the goal of detecting semantic, structural, and numerical inconsistencies in real-world financial reports. Unlike traditional financial NLP tasks that treat documents as unstructured text, this task emphasizes model learning under explicit accounting and taxonomy constraints. Models must internalize the hierarchical structure of XBRL filings, cross-document dependencies, and formal accounting logic, closely reflecting real-world auditing workflows. The task is grounded in real SEC XBRL filings and Data Quality Committee (DQC) error cases, and is designed to evaluate whether trained models can perform reliable, regulation-aligned financial auditing, rather than surface-level pattern matching.

Datasets Participants must train their auditing models using datasets constructed from real US-GAAP-compliant XBRL filings collected from SEC disclosures between 2022 and 2024: https://www.sec.gov/edgar/search/. Each training and evaluation instance consists of multiple interconnected XBRL documents, including:

  • Instance documents containing reported financial facts
  • Schema documents defining financial elements
  • Presentation, calculation, definition, and label linkbases capturing hierarchical and computational relations
  • Reference US-GAAP taxonomy, including concepts and their semantic and structural relationships

These datasets are intended to support end-to-end auditing model training, including pretraining, fine-tuning, and task-specific adaptation. Important (Contest Dataset Note): This repo may reference the original auditing datasets (e.g., TheFinAI/FinMR) released for the auditing paper [5]. For this contest, the official evaluation should use the curated subset datasets: (1) FinSM_Sub: https://huggingface.co/datasets/TheFinAI/FinSM_Sub (2) FinRE_Sub: https://huggingface.co/datasets/TheFinAI/FinRE_Sub (3) FinMR_Sub: https://huggingface.co/datasets/TheFinAI/FinMR_Sub These subsets were specifically prepared for the contest and are recommended for final reporting/leaderboard evaluation.

Auditing Subtasks The auditing model is evaluated on three complementary subtasks, each corresponding to a core auditing capability. Participants are encouraged to design a unified model or multi-head architecture that can handle all three tasks.

(1) FinSM – Financial Semantic Matching

  • Goal: Train a model to identify semantically incorrect US-GAAP concept tags used in XBRL filings.
  • Input: A financial query, an XBRL filing segment, and the relevant portion of the US-GAAP taxonomy.
  • Output: A set of US-GAAP concepts that are semantically mismatched or inappropriate for the reported financial fact.
  • Output Example: ["us-gaap:Revenues", "us-gaap:OperatingIncomeLoss"]

(2) FinRE – Financial Relationship Extraction

  • Goal: Train a model to detect erroneous hierarchical or compositional relationships between financial concepts in XBRL filings.
  • Input: Two financial concepts, the XBRL filing context, and corresponding taxonomy definitions.
  • Output: A single error-type label describing the incorrect relationship: Reversal, Inappropriateness, or CombinationErr.

(3) FinMR – Financial Mathematical Reasoning

  • Goal: Train a model to verify numerical consistency by reasoning over accounting logic and calculation linkbases.
  • Input: A question about a reported financial value, together with its implied calculated value, the XBRL filing context, and the relevant taxonomy information.
  • Output: A JSON object containing:
    {
    "extracted_value": <reported value>,
    "calculated_value": <computed value>
    }
    

Objective & Constraints (1) Training Objective Participants must train an auditing-oriented model that maximizes performance across all three subtasks by learning to reason over:

  • Taxonomy semantics (concept meaning and usage)
  • Hierarchical and relational structures in XBRL filings
  • Multi-step numerical calculations grounded in accounting rules

The ultimate objective is to systematically improve auditing performance on FinSM, FinRE, and FinMR, rather than optimizing for a single isolated task.

(2) Output Constraints Model outputs must strictly follow the specified formats:

  • JSON arrays for FinSM
  • Single-label classification for FinRE
  • JSON objects for FinMR

No explanations, chain-of-thought, or free-form text are allowed unless explicitly required by the task specification.

(3) Evaluation Setting

  • All evaluations are conducted in a static, reproducible offline setting using held-out SEC XBRL filings.
  • Input contexts may be long (often exceeding 30k tokens), reflecting realistic auditing complexity.
  • Models must rely solely on the provided SEC XBRL data and taxonomy resources. External financial knowledge sources are not permitted during evaluation. Repository Scope (Inference + Evaluation Only): This repository provides inference pipelines and evaluation scripts only.
  • pipeline-vllm: inference using vLLM
  • pipeline-hf: inference using Hugging Face Transformers After generating predictions, use the evaluation scripts for FinSM/FinRE/FinMR to compute metrics. Training code is NOT provided; participants are expected to train their own models.

Metrics Model performance is evaluated separately and jointly across the three subtasks using auditing-oriented metrics:

  • FinSM (Semantic Matching)
    • Hit Rate@k
    • Recall@k
    • Macro-F1@k
  • FinRE (Relationship Extraction)
    • Accuracy
    • Macro Precision, Recall, and F1
  • FinMR (Mathematical Reasoning)
    • Overall Accuracy (LLM-as-a-judge)
    • Structural Error Rate (SER)
    • Extraction Error Rate (EER)
    • Calculation Error Rate (CER)

These metrics jointly assess semantic correctness, structural grounding, numerical reliability, and error localization, providing a comprehensive evaluation of trained auditing models.

[1] Wang, Keyi, et al. "FinRL Contests: Data‐Driven Financial Reinforcement Learning Agents for Stock and Crypto Trading." Artificial Intelligence for Engineering (2025). [IET] [arXiv]

[2] Liu, Xiao-Yang, et al. "Finrl-meta: Market environments and benchmarks for data-driven financial reinforcement learning." Advances in Neural Information Processing Systems 35 (2022): 1835-1849. [NeurIPS]

[3] Liu, Xiao-Yang, et al. "Fingpt: Democratizing internet-scale data for financial large language models." arXiv preprint arXiv:2307.10485 (2023). [arXiv]

[4] Lin, Shengyuan, et al. "Evaluation and Benchmarking Suite for Financial Large Language Models and Agents." NeurIPS 2025 Workshop on Evaluating the Evolving LLM Lifecycle: Benchmarks, Emergent Abilities, and Scaling. [NeurIPS 2025 Workshop]

[5] Yan Wang, Keyi Wang, Shanshan Yang, Jaisal Patel, Jeff Zhao, Fengran Mo, Xueqing Peng, Lingfei Qian, Jimin Huang, Guojun Xiong, Xiao-Yang Liu, Jian-Yun Nie. "FinAuditing: A Financial Taxonomy-Structured Multi-Document Benchmark for Evaluating LLMs." arXiv preprint arXiv:2510.08886 (2025). [arXiv]

[6] Yan Wang, Lingfei Qian, Xueqing Peng, Yang Ren, Keyi Wang, Yi Han, Dongji Feng, Fengran Mo, Shengyuan Lin, Qinchuan Zhang, Kaiwen He, Chenri Luo, Jianxing Chen, Junwei Wu, Chen Xu, Ziyang Xu, Jimin Huang, Guojun Xiong, Xiao-Yang Liu, Qianqian Xie, Jian-Yun Nie. "FinTagging: Benchmarking LLMs for Extracting and Structuring Financial Information." arXiv preprint arXiv:2505.20650 (2025). [arXiv] ## Contact Contact email: [finrlcontest@gmail.com](mailto:finrlcontest@gmail.com) Contestants can communicate any questions on * [Discord](https://discord.gg/dJY5cKzmkv).