Technology & Product9 min read

Which Features Actually Predict Ad Clicks in HypeLab's 25-Feature PCTR Model

HypeLab's PCTR model uses 25 features to predict ad clicks. Learn which signals matter most and how feature weight homogeneity prevents overfitting.

Joe Kim
Joe Kim
Founder @ HypeLab ·
Share

HypeLab's PCTR model uses 25 carefully selected features to predict ad clicks, including placement quality indicators, viewport geometry, user engagement history, category matching between advertisers and publishers, and wallet-native signals like Phantom and MetaMask preferences. The key innovation is not just which features we use, but how we enforce balanced feature weights to prevent overfitting and ensure consistent performance across our network of 200+ premium Web3 publishers.

Quick Answers

What features predict crypto ad clicks? Placement slug (quality indicator), viewport geometry (position on page), user history (session length, past clicks), category matching (advertiser-publisher alignment), and wallet preferences (Phantom, MetaMask, Rainbow) are the strongest signals.

How does HypeLab prevent overfitting? We train 50 models per run and reject any model that puts excessive weight on a small subset of features. The winning model must demonstrate balanced reliance across all 25 features.

Why does feature balance matter? Models that over-rely on a single signal fail catastrophically when that signal shifts. Balanced models degrade gracefully and maintain performance as the Web3 advertising ecosystem evolves.

Why Does Feature Selection Matter in Crypto Ad Prediction?

A machine learning model is only as good as the data it learns from. Feed it irrelevant features and it will find spurious correlations. Feed it too few features and it will miss real patterns. Feed it features with wildly different scales or cardinalities and it will struggle to weight them appropriately.

The challenge in ad tech is that you have access to hundreds of potential signals: user demographics, device specifications, time of day, page content, historical behavior, and more. Most crypto ad networks throw everything into the model and hope the algorithm figures it out. That approach creates two problems.

First, models trained on too many features tend to overfit. They memorize the training data instead of learning generalizable patterns. A model might learn that users on a specific obscure device model always click, when in reality it just saw a few bot-heavy sessions from that device in training data.

Second, models trained without feature discipline become uninterpretable. When you cannot explain why the model made a prediction, you cannot debug it when it breaks. In Web3 advertising, models break constantly as user behavior shifts, publishers change, and advertisers rotate creative.

What Are the 25 Features in HypeLab's PCTR Model?

After extensive experimentation with our Web3 ad platform, HypeLab's PCTR model converged on approximately 25 features organized into several categories:

Placement Features: The placement slug acts as a quality indicator. Different publishers and placements have dramatically different baseline CTRs. A wallet's home screen (like Phantom or MetaMask) converts differently than a news site's sidebar. The model learns these baseline rates through target encoding (more on this in a related post).

Geometry Features: Viewport area and placement angle tell the model where on the page the ad sits. These are combined into a composite feature representing the relationship of the ad space to the overall viewport. Ads at the bottom of long pages where nobody scrolls get lower predicted CTR. Ads above the fold in prominent positions get higher predictions.

User History Features: How long have we been seeing this user? How many impressions have they had? How many clicks? Users with longer session histories and demonstrated engagement patterns behave differently than brand new users we know nothing about.

Category Matching: Does the advertiser's category align with the publisher's category? A DeFi protocol (like Aave or Compound) advertising on a DeFi dashboard (like Zapper or DeBank) converts better than the same ad on a gaming site. Category matching is one of the strongest signals in the model.

Device and Context Features: Operating system, device category, and preferred wallet. These individually carry less weight but contribute to the overall prediction. An iOS user on Safari behaves differently than an Android user on Chrome. A user with a Phantom wallet prefers Solana content, while MetaMask users gravitate toward Ethereum and L2 ecosystems like Arbitrum and Optimism.

How Does HypeLab Train 50 Models and Pick the Winner?

HypeLab does not train a single model and deploy it. We train approximately 50 models simultaneously per training run, each with slightly different hyperparameters, random seeds, and tree configurations. Then we evaluate them not just on accuracy metrics, but on a property we call feature weight homogeneity.

Here is what that means. After a model is trained, we examine how much weight it assigns to each feature. In gradient boosted tree models, this corresponds to how often a feature is used in splits and how much information gain those splits provide. A well-balanced model distributes this weight across all 25 features. A poorly balanced model concentrates weight on just a handful.

The homogeneity test: If a model puts excessive weight on a small subset of features, we reject it regardless of its accuracy score. That model has learned shortcuts, not patterns. It will fail when those features shift even slightly.

Why this matters for advertisers: This discipline is what separates production-grade ad prediction from academic exercises. A model that achieves 0.1% higher AUC by overfitting to placement slug will catastrophically fail when a major publisher changes their integration. A model with balanced weights degrades gracefully, meaning your crypto advertising campaigns maintain consistent performance.

Why Is Feature Dominance Dangerous in Ad Models?

Consider what happens when a model over-relies on a single feature. Say it learns that placement X has a 5% CTR while all other placements average 0.5%. If the model just predicts high for placement X and low for everything else, it will score well on historical data. But placement X might change their ad position. They might get hit with bot traffic. Their user base might shift.

When that happens, a feature-dominant model collapses. It has no fallback because it never learned to use the other 24 features properly. A balanced model continues to function because it learned from all signals. Placement X's predictions degrade, but the model still uses geometry, user history, and category matching to make reasonable predictions.

This is why we explicitly test for and reject feature dominance. It is not a bug to have a model that is slightly less accurate in aggregate but much more robust to distribution shift. In production ad systems, robustness matters more than peak accuracy.

How Does Category Matching Improve Ad Performance?

One of the strongest signals in our model is category matching: the alignment between what the advertiser is promoting and what the publisher's audience expects. This is intuitive. A user reading about DeFi yield strategies is more likely to click an ad for a yield aggregator than an ad for an NFT marketplace.

But category matching is also dangerous if over-relied upon. Not every impression has clear category signals. Some publishers serve diverse content. Some advertisers run campaigns that could fit multiple categories. A model that depends too heavily on category matching will produce poor predictions whenever the signal is ambiguous.

Our approach is to let category matching contribute meaningfully but not dominantly. The model uses it when the signal is strong and falls back on other features (geometry, user history, device context) when it is weak. This is only possible because we enforce homogeneity across the full feature set.

Why Does Ad Position on the Page Matter So Much?

A detail that many Web3 ad platforms ignore is the physical position of the ad on the page. An ad above the fold, visible immediately when the page loads, converts at dramatically higher rates than an ad below three screens worth of content where users rarely scroll.

HypeLab captures this through geometry features: viewport area, placement angle, and a combined metric we call the directional relationship to the top-left corner of the viewport. These features let the model learn that bottom-of-page placements convert poorly even on high-quality publishers, while top-of-page placements convert well even on lower-tier sites.

The insight is that placement quality and position quality are independent signals. A premium publisher with a terrible ad position might underperform a mid-tier publisher with perfect ad placement. The model learns both dimensions.

How Does HypeLab Handle the Cold Start Problem?

User history features (session length, historical impressions, historical clicks) create a cold start problem. For brand new users, we have no history. How does the model predict?

Tree-based models handle this gracefully. When user history features are null or empty, the trees that rely on those features simply contribute less to the overall prediction. Other features (placement, geometry, device, category) carry the load. As we accumulate history for a user, the user history features gradually take on more weight.

This is architecturally different from systems that require complete feature vectors. We never have to impute fake history for new users. The model structure naturally adapts to whatever information is available.

What Features Did HypeLab Deliberately Exclude?

As important as what is in the model is what we leave out. Features we experimented with but ultimately excluded:

  • Time of day: Too noisy. While there are aggregate patterns (more activity during US business hours), the signal-to-noise ratio at the individual prediction level was poor.
  • Creative content analysis: We experimented with features derived from ad creative (image analysis, text sentiment). The computation cost did not justify the marginal accuracy gain.
  • Detailed device models: We reduced over 5,000 device model strings to the top 500 that cover 90%+ of traffic. The long tail added sparsity without predictive value.
  • IP-based signals: Too privacy-invasive and too easily manipulated by bots. We rely on first-party signals from publishers instead.

How Does HypeLab Compare to Other Crypto Ad Networks?

Most crypto ad networks do not disclose their model architectures, but based on public information and industry patterns, HypeLab's approach differs from competitors like Coinzilla, Bitmedia, and A-Ads in several key ways:

CapabilityHypeLabTraditional Crypto Ad Networks
Feature homogeneity optimizationYes - explicit testing and rejection of dominant featuresRarely - most chase pure accuracy metrics
Wallet-native signalsYes - Phantom, MetaMask, Rainbow preferencesLimited or none - rely on cookies/fingerprinting
Prediction latencyMilliseconds per predictionVaries widely, often much slower
Real-time biddingYes - programmatic RTBOften static or delayed
Model interpretabilityHigh - balanced feature weightsOften black-box

A user's wallet preference provides information about their ecosystem allegiance (Ethereum vs. Solana vs. Base vs. Arbitrum) that no amount of cookie tracking can replicate. This is why Web3-native advertising outperforms generic crypto ad placement.

Why Does This Matter for Crypto Advertisers?

For crypto advertisers running campaigns for DeFi protocols, NFT projects, Web3 games, and blockchain infrastructure, this technical foundation translates into practical benefits:

  • Consistent performance: Because the model does not over-rely on any single signal, your campaign performance is stable even as the ad ecosystem changes.
  • Fair distribution: Category matching ensures your DeFi ads appear to DeFi-interested users on platforms like Zapper and DeBank, not wasted on unrelated audiences.
  • Quality placement: Geometry features ensure your ads appear in visible positions, not buried below the fold where nobody sees them.
  • New user reach: The cold start handling means you can reach new users effectively, not just retarget existing ones.

The result: HypeLab advertisers see 2-3x higher engagement rates compared to generic crypto ad placements, with consistent performance across market cycles. Leading protocols like Aave, Lido, and Uniswap trust HypeLab's prediction system to deliver quality traffic.

Machine learning in ad tech is not magic. It is engineering discipline applied to pattern recognition. HypeLab's 25-feature PCTR model represents years of experimentation, thousands of A/B tests, and a deliberate philosophy of robustness over raw accuracy. The result is a prediction system that actually works in production, not just in backtests.

Ready to Launch a Campaign with Intelligent Targeting?

HypeLab is the Web3 ad platform with machine learning-powered targeting that actually works. Our PCTR model ensures your ads reach the right users at the right time on the right placements across Phantom, MetaMask, StepN, and 200+ other premium Web3 publishers. Q1 2026 inventory is filling fast as crypto advertisers increase spend ahead of market momentum.

  • 25-feature prediction model: Balanced, robust, and resistant to overfitting.
  • Wallet-native signals: Targeting based on actual crypto behavior, not inferred interest.
  • 200+ premium publishers: From wallets like Phantom and Rainbow to DeFi dashboards like Zapper and DeBank to crypto media.
  • Dual payment rails: Pay with crypto (USDC, ETH) or credit card.
  • No minimum budget: Launch your campaign in minutes with our self-serve platform.
  • Trusted by top protocols: Aave, Uniswap, Lido, and 500+ crypto advertisers use HypeLab.

Key Takeaways

  • HypeLab's PCTR model uses 25 features including placement quality, viewport geometry, user history, category matching, and wallet preferences
  • We train 50 models per run and reject any that over-rely on a small subset of features
  • Feature weight homogeneity prevents overfitting and ensures consistent campaign performance
  • Wallet-native signals (Phantom, MetaMask, Rainbow) provide targeting precision that cookie-based systems cannot match
  • The model operates under millisecond latency constraints for real-time bidding

Frequently Asked Questions

HypeLab's PCTR model uses approximately 25 features including: placement slug (quality indicator), viewport geometry (area and placement angle), user history features (session length, historical clicks), category matching (advertiser vs publisher category alignment), operating system, device category, and preferred wallet. The model is designed to distribute weight evenly across features to prevent overfitting.
HypeLab trains approximately 50 models simultaneously per run and evaluates them for feature weight homogeneity. A model that assigns excessive weight to a small subset of features is rejected for overfitting. The winning model must demonstrate balanced reliance across the full feature set.
No single feature dominates by design. However, placement slug (a quality indicator for where the ad appears), viewport geometry (position on page), user history, and category matching (alignment between advertiser and publisher categories) tend to be the strongest signals. The model deliberately prevents any feature from becoming too dominant.

Continue Reading

Contact our sales team.

Got questions or ready to get started? Our sales team is here to help. Whether you want to learn more about our Web3 ad network, explore partnership opportunities, or see how HypeLab can support your goals, just reach out - we'd love to chat.