XGBoost 2.0 is Here

Share via:

The best solution for making sense of tabular data, XGBoost, has just been upgraded. XGBoost 2.0 brings forth a plethora of new features and enhancements aimed at revolutionising the machine learning landscape. 

Click here to check out the full release.

XGBoost 2.0 introduces a novel feature under development, focusing on vector-leaf tree models for multi-target regression, multi-label classification, and multi-class classification. Unlike the previous approach of building separate models for each target, this feature allows XGBoost to construct a single tree for all targets, offering several advantages, including prevention of overfitting, smaller model sizes, and the ability to consider correlations between targets. 

Users can combine vector leaf and scalar leaf trees during training through a callback. It’s important to note that this feature is a work in progress, and some aspects are still under development.

You thought XGBoost was powerful enough?

Introducing: XGBoost 2.

– Multi-target trees (vector-leafs)
– Finally: 𝚍𝚎𝚟𝚒𝚌𝚎 parameter.
– 𝚊𝚙𝚙𝚛𝚘𝚡 for GPU devices.
– Controllable histogram bound (to optimize RAM)
– Batch by batch training (External Memory)
– Learning to…

— Yam Peleg (@Yampeleg) October 3, 2023

Read: XGBoost is All You Need 

New Device Parameter 

A significant change is the introduction of a new ‘device’ parameter, replacing existing parameters like ‘gpu_id,’ ‘gpu_hist,’ ‘gpu_predictor,’ ‘cpu_predictor,’ ‘gpu_coord_descent,’ and the PySpark-specific ‘use_gpu.’ Users can now use the ‘device’ parameter to select their preferred device for computation, simplifying the configuration process.

Default Tree Method

Starting from XGBoost 2.0, the ‘hist’ tree method is set as the default. In previous versions, XGBoost would automatically choose between ‘approx’ and ‘exact’ based on input data and the training environment. The new default method aims to improve model training efficiency and consistency.

GPU-Based Approximate Tree Method 

XGBoost 2.0 offers initial support for the ‘approx’ tree method on GPU. While performance optimisation is ongoing, the feature is considered feature-complete, except for the JVM packages. 

Users can access this capability by specifying ‘device=”cuda”‘ and ‘tree_method=”approx”.’ It’s important to note that the Scala-based Spark interface is not yet supported.

Memory Footprint Optimization 

This release also introduces a new parameter, ‘max_cached_hist_node,’ allowing users to limit CPU cache size for histograms. This helps prevent aggressive caching of histograms, especially in deep trees. Additionally, memory usage for ‘hist’ and ‘approx’ tree methods on distributed systems is reduced by half.

Improved External Memory Support 

External memory support receives a significant boost in XGBoost 2.0. The default ‘hist’ tree method now utilises memory mapping, enhancing performance and reducing CPU memory usage. Users are encouraged to try this feature, particularly when memory savings are required.

Learning-to-Rank Enhancements XGBoost 2.0 introduces a new implementation for learning-to-rank tasks, offering a range of new features and parameters to improve ranking performance. 

Notable additions include parameters for pair construction strategy, control over the number of samples per group, experimental unbiased learning-to-rank support, and custom gain functions with NDCG.

Column-Based Split and Federated Learning Significant progress has been made in column-based split for federated learning, with support for various tree methods and vertical federated learning. GPU support for this feature is still in development.

PySpark Enhancements 

The PySpark interface in XGBoost 2.0 has received numerous new features and optimisations, including GPU-based prediction, data initialisation improvements, support for predicting feature contributions, Python typing support, and improved logs for training.

The post XGBoost 2.0 is Here appeared first on Analytics India Magazine.

Disclaimer

We strive to uphold the highest ethical standards in all of our reporting and coverage. We StartupNews.fyi want to be transparent with our readers about any potential conflicts of interest that may arise in our work. It’s possible that some of the investors we feature may have connections to other businesses, including competitors or companies we write about. However, we want to assure our readers that this will not have any impact on the integrity or impartiality of our reporting. We are committed to delivering accurate, unbiased news and information to our audience, and we will continue to uphold our ethics and principles in all of our work. Thank you for your trust and support.

Popular

More Like this

XGBoost 2.0 is Here

The best solution for making sense of tabular data, XGBoost, has just been upgraded. XGBoost 2.0 brings forth a plethora of new features and enhancements aimed at revolutionising the machine learning landscape. 

Click here to check out the full release.

XGBoost 2.0 introduces a novel feature under development, focusing on vector-leaf tree models for multi-target regression, multi-label classification, and multi-class classification. Unlike the previous approach of building separate models for each target, this feature allows XGBoost to construct a single tree for all targets, offering several advantages, including prevention of overfitting, smaller model sizes, and the ability to consider correlations between targets. 

Users can combine vector leaf and scalar leaf trees during training through a callback. It’s important to note that this feature is a work in progress, and some aspects are still under development.

You thought XGBoost was powerful enough?

Introducing: XGBoost 2.

– Multi-target trees (vector-leafs)
– Finally: 𝚍𝚎𝚟𝚒𝚌𝚎 parameter.
– 𝚊𝚙𝚙𝚛𝚘𝚡 for GPU devices.
– Controllable histogram bound (to optimize RAM)
– Batch by batch training (External Memory)
– Learning to…

— Yam Peleg (@Yampeleg) October 3, 2023

Read: XGBoost is All You Need 

New Device Parameter 

A significant change is the introduction of a new ‘device’ parameter, replacing existing parameters like ‘gpu_id,’ ‘gpu_hist,’ ‘gpu_predictor,’ ‘cpu_predictor,’ ‘gpu_coord_descent,’ and the PySpark-specific ‘use_gpu.’ Users can now use the ‘device’ parameter to select their preferred device for computation, simplifying the configuration process.

Default Tree Method

Starting from XGBoost 2.0, the ‘hist’ tree method is set as the default. In previous versions, XGBoost would automatically choose between ‘approx’ and ‘exact’ based on input data and the training environment. The new default method aims to improve model training efficiency and consistency.

GPU-Based Approximate Tree Method 

XGBoost 2.0 offers initial support for the ‘approx’ tree method on GPU. While performance optimisation is ongoing, the feature is considered feature-complete, except for the JVM packages. 

Users can access this capability by specifying ‘device=”cuda”‘ and ‘tree_method=”approx”.’ It’s important to note that the Scala-based Spark interface is not yet supported.

Memory Footprint Optimization 

This release also introduces a new parameter, ‘max_cached_hist_node,’ allowing users to limit CPU cache size for histograms. This helps prevent aggressive caching of histograms, especially in deep trees. Additionally, memory usage for ‘hist’ and ‘approx’ tree methods on distributed systems is reduced by half.

Improved External Memory Support 

External memory support receives a significant boost in XGBoost 2.0. The default ‘hist’ tree method now utilises memory mapping, enhancing performance and reducing CPU memory usage. Users are encouraged to try this feature, particularly when memory savings are required.

Learning-to-Rank Enhancements XGBoost 2.0 introduces a new implementation for learning-to-rank tasks, offering a range of new features and parameters to improve ranking performance. 

Notable additions include parameters for pair construction strategy, control over the number of samples per group, experimental unbiased learning-to-rank support, and custom gain functions with NDCG.

Column-Based Split and Federated Learning Significant progress has been made in column-based split for federated learning, with support for various tree methods and vertical federated learning. GPU support for this feature is still in development.

PySpark Enhancements 

The PySpark interface in XGBoost 2.0 has received numerous new features and optimisations, including GPU-based prediction, data initialisation improvements, support for predicting feature contributions, Python typing support, and improved logs for training.

The post XGBoost 2.0 is Here appeared first on Analytics India Magazine.

Disclaimer

We strive to uphold the highest ethical standards in all of our reporting and coverage. We StartupNews.fyi want to be transparent with our readers about any potential conflicts of interest that may arise in our work. It’s possible that some of the investors we feature may have connections to other businesses, including competitors or companies we write about. However, we want to assure our readers that this will not have any impact on the integrity or impartiality of our reporting. We are committed to delivering accurate, unbiased news and information to our audience, and we will continue to uphold our ethics and principles in all of our work. Thank you for your trust and support.

Website Upgradation is going on for any glitch kindly connect at office@startupnews.fyi

More like this

Anthropic Sparks SaaSpocalypse, Boost For Deeptechs & More

Anthropic Rattles Markets Anthropic’s launch of Cowork plugins on...

Open source USearch library jumpstarts ScyllaDB vector search

ScyllaDB, an open source columnar database, recently added...

‘Clearly dishonest’: Sam Altman slams Anthropic’s Super Bowl ad,...

Anthropic took a shot at OpenAI with a...

Popular

iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv