In Lesson 1, we discussed why AI is becoming the new infrastructure for crypto trading. The key question that follows is: no matter how powerful AI is, it can only operate within the boundaries of the data you provide.
Many strategies fail not because the model is too simple, but because there are directional errors at the data layer: either data quality is insufficient, feature design is distorted, or validation methods are biased.
Therefore, real AI trading often doesn’t start with “choosing a model,” but with “building the data foundation.” What you feed the model determines what it can see; what it can see determines what judgments it can make.
Traders new to AI often fall into the “data hoarding” mindset: grabbing every piece of data possible, believing more features make it easier to find alpha.
In reality, low-quality, noisy, weakly correlated data actually reduces model stability. The reason is simple:
So, the first principle of building a data system is:
Select data around trading problems—not hunt for problems around data itself.
If you’re solving “short-term direction prediction,” prioritize microstructure and sentiment shocks; if you’re working on “medium-term position management,” focus more on liquidity, volatility structure, and macro factors.

In crypto markets, the most valuable data usually comes from four layers: market data, derivatives, on-chain, and external information.
This is the foundational layer for all strategies, including:
It answers: how prices change, how liquidity changes, how trading behavior changes.
Many basic strategies can be built with just market data, but its limitation is: it’s more like a “result variable,” with limited explanatory power for “why things change.”
Especially crucial in crypto markets, including:
This data reflects market leverage crowding and position vulnerability.
For example, “price rising + OI rising + high funding rate” vs. “price rising + OI falling” mean completely different things. The former may signal trend strengthening or leverage crowding; the latter is more likely driven by short covering.
Without derivatives dimension, it’s hard to judge position structure behind market moves.
A key advantage distinguishing crypto markets from traditional ones, including:
The value of on-chain data lies in observing “capital and behavioral trajectories,” but the challenge is delayed interpretation and noise filtering.
For example, increased exchange inflow could mean preparing to sell or preparing to hedge. On-chain data must be interpreted together with price structure and derivatives data—using it alone easily leads to misjudgment.
Includes news, social media discussion heat, policy events, macro data release timings.
These are more like “shock source data”: explaining why volatility suddenly spikes or trends shift briefly.
But this type of data has obvious issues: highly subjective, noisy, mixed true/false information.
Therefore, external text is better used as “risk alert factors” and “event filters,” not recommended as sole entry signals.
AI doesn’t directly understand “market narratives”; it only recognizes feature patterns.
So the second step isn’t rushing to train models but transforming raw data into learnable, verifiable, tradable features.
Common useful features can be grouped into four categories:
The key isn’t “flashy features,” but three standards:
Many people default to having the model predict “the next K-line up/down,” but that’s not necessarily optimal.
Trading objectives can have various label forms:
If your strategy goal is to “avoid large drawdowns” but you use “short-term price direction” as a label, no matter how accurate the model is, it may not be useful.
So labels should match strategy goals: whatever profit you seek in trading, have the model learn that target.
In typical machine learning tasks, randomly shuffling training and test sets is common and reasonable; but in trading this causes severe distortion.
Because markets have time-dependent structure—future information must never “leak” into the past.
AI trading should adhere to three validation rules at minimum:
Many “backtest miracle strategies” collapse not because markets worsen but because testing methods were optimistically biased from the start.
Using unavailable-at-the-time data leads to inflated results.
Training only on surviving coins or platforms—ignoring failed samples.
Deleting real noise as dirty data—model loses adaptability to extreme markets.
Features implicitly contain label information—making model appear overly accurate.
Forcing low-frequency on-chain features into high-frequency trading tasks—causing false signals.
These issues don’t trigger alarms during backtesting but will quickly magnify in live trading.
For course learners, the safest approach isn’t to start with a “full-market all-factor mega-model,” but begin with a minimal viable data framework:
This approach keeps problem localization clear, iteration costs low, and deployment path short.
Complex systems aren’t built all at once—they expand layer by layer from interpretable small systems.
In actual implementation, the data stage is often the most time-consuming part: multi-source collection, format cleaning, time alignment, feature pipelines, strategy integration.
That’s why platform-based AI tools are increasingly important. Taking Gate for AI as an example of such infrastructure—the value isn’t in “generating a universal strategy for you,” but in helping traders efficiently complete the engineering loop from data to strategy and reducing friction between research and execution. Traders still need to define problems, set constraints, manage risks—but underlying workflows can be more standardized and reusable.