Recent from talks
Contribute something
Nothing was collected or created yet.
RelieF
View on Wikipedia| RelieF | |
|---|---|
| Presented by | Gisèle Quenneville |
| Country of origin | Canada |
| Original release | |
| Network | TFO |
Relief, formerly known as Panorama, was a public affairs newsmagazine series in Canada, airing nightly in Ontario on TFO, the Franco-Ontarian public television channel.
The series was hosted by Gisèle Quenneville. Associated reports include Melanie Routhier-Boudreau, Isabelle Brunet, Marie Duchesneau, Luce Gauthier, Frédéric Projean and Chantal Racine. Longtime host Pierre Granger retired in 2009.[1] The series was renamed RelieF in autumn 2010.
RelieF aired seven nights a week at 7 p.m. From Monday to Thursday, it focused on news and public affairs. On Fridays, it showed documentaries. Saturday had a "week in review" edition and Sunday an arts and culture magazine.
References
[edit]- ^ "Pierre Granger quitte l'antenne". L'Express de Toronto, December 21, 2009.
RelieF
View on GrokipediaIntroduction and Background
Definition and Core Principles
ReliefF is a filter-based feature selection algorithm proposed by Igor Kononenko and colleagues as an extension of the original Relief method (developed by Kenji Kira and Larry Rendell) to address limitations in handling multi-class problems, noisy data, and incomplete datasets. Introduced as an improvement over Relief, which was limited to binary classification and single nearest neighbors, ReliefF estimates the relevance of features by iteratively evaluating their ability to distinguish a randomly selected target instance from its k nearest neighbors of the same class (hits) and from different classes (misses). The algorithm assigns weights to features ranging from -1 (indicating irrelevance or redundancy) to 1 (indicating high relevance), with weights updated based on observed differences in feature values among these neighbors.[3][4] The core principle underlying ReliefF is the contextual estimation of feature quality through local instance-based comparisons, which captures both univariate relevance and interactions with other features without explicitly searching subsets. For each target instance, ReliefF identifies k nearest hits using a distance metric (typically Euclidean for continuous features or Hamming for discrete), then finds k nearest misses from each differing class, weighting contributions by the prior probability of that class conditional on it differing from the target's class. Feature weights are updated by subtracting the average difference to hits (penalizing features that vary within classes) and adding the average difference to misses (rewarding features that separate classes), normalized across iterations. For continuous features, differences are scaled by the feature's range; for discrete, they are binary (0 if matching, 1 otherwise). This process repeats for all or a subset of instances, yielding a relevance score interpretable as the difference between the probability of differing feature values given a same-class neighbor versus a different-class neighbor.[3] ReliefF's robustness stems from its use of multiple (k, often 10) neighbors to mitigate noise—unlike Relief's single-neighbor approach, which is sensitive to outliers—and its probabilistic handling of multi-class scenarios, where miss contributions are averaged across classes proportional to their prevalence. It also accommodates missing values by estimating differences via class-conditional probabilities rather than direct comparisons. These principles enable ReliefF to detect conditional dependencies and epistatic interactions, performing in linear time relative to dataset size and feature count, making it suitable for high-dimensional data while avoiding parametric assumptions or global statistics. Theoretical analysis confirms that under independence, irrelevant features converge to zero weight, while relevant ones yield positive scores proportional to their discriminative power.[3][5]Historical Context and Development
The Relief algorithm originated in 1992, developed by Kenji Kira and Larry A. Rendell at the University of Western Ontario. Presented at the Tenth National Conference on Artificial Intelligence (AAAI-92), it introduced a filter-based feature selection method that estimates attribute relevance by measuring how well features distinguish nearest-neighbor instances across different classes, particularly excelling at capturing feature interactions and conditional dependencies ignored by independence-assuming techniques like information gain. This approach drew inspiration from instance-based learning paradigms, iteratively sampling random instances and updating feature weights based on the difference in attribute values between a given instance and its nearest "hit" (same-class neighbor) and "miss" (opposite-class neighbor). ReliefF emerged as a key extension in the mid-1990s, primarily by Igor Kononenko and collaborators at the University of Ljubljana, addressing Relief's sensitivities to noisy data and its limitation to binary classification with exact nearest neighbors. First described around 1994, ReliefF mitigated these issues by averaging contributions over k nearest hits and misses (typically k=10), enhancing robustness to incomplete or erroneous data while extending applicability to multiclass problems through class-specific miss calculations.[6] A seminal 1997 publication formalized its integration with inductive learners, demonstrating how ReliefF's heuristic guidance improved attribute estimation in domains with irrelevant or redundant features, thus overcoming "myopia" in algorithms prone to overfitting noise.[4] Subsequent refinements in the late 1990s and early 2000s, including theoretical analyses by Marko Robnik-Šikonja and Kononenko, clarified ReliefF's probabilistic foundations and parameter sensitivities, confirming its ability to detect epistatic interactions without assuming feature independence.[5] These developments established Relief-based methods as a distinct family of nearest-neighbor-driven filters, influencing broader machine learning applications in high-dimensional data like genomics, where traditional selectors faltered. By the early 2000s, adaptations such as RReliefF for regression tasks further expanded the framework, building directly on ReliefF's core sampling and weighting mechanics.[7]Core Algorithms
Original Relief Algorithm
The original Relief algorithm, introduced by Kenji Kira and Larry A. Rendell in 1992, is a filter-based method for feature selection in binary classification tasks. It evaluates feature relevance by assessing how well individual features distinguish instances of the same class from those of the opposite class, using nearest-neighbor comparisons to estimate quality scores. Unlike traditional methods that rely on statistical independence or exhaustive search, Relief incorporates contextual information from instance similarities, enabling it to detect interacting features and tolerate noise in the data.[8] The algorithm operates on a training set of instances, each described by features (numerical or nominal), divided into positive () and negative () classes. It initializes a weight vector of zeros for the features and performs random samplings (where is typically set to a constant like 50–100 for efficiency). For each sampled instance , Relief identifies the nearest hit (closest instance of the same class to ) and the nearest miss (closest instance of the opposite class to ), using Euclidean distance across all features. Feature weights are then updated to penalize values that fail to distinguish from its hit (subtracting the squared difference) while rewarding those that distinguish it from its miss (adding the squared difference), with differences normalized to [0,1] for comparability. The final relevance score for each feature is the average weight update over all iterations, and features exceeding a statistical threshold (e.g., derived from Chebyshev's inequality to bound false positives) are selected as relevant. This process yields a time complexity of , which simplifies to for fixed , making it scalable for high-dimensional data.[8] Key to Relief's effectiveness is its ability to capture feature dependencies without assuming independence; relevant features, especially those interacting with others, receive positive scores because hits cluster closely while misses differ substantially, whereas irrelevant features average to near-zero or negative scores under random sampling. Empirical tests in the original work on noisy domains like parity problems and real-world datasets (e.g., from the Promoter Gene database) demonstrated that Relief selected minimal relevant feature subsets, outperforming methods like FOCUS in accuracy and efficiency, with noise tolerance stemming from aggregation over multiple neighbors. However, it assumes a two-class problem and complete label accuracy, limiting direct applicability to multi-class or regression settings without extensions.[8] The pseudocode for Relief, as formalized in the 1992 paper, is as follows:Relief(S, m, τ)
1. Separate S into S+ = {positive instances} and S- = {negative instances}
2. Initialize weight vector W = (0, 0, ..., 0)
3. For i = 1 to m:
a. Pick at random an instance X ∈ S
b. From X, pick at random:
Z+ from S+ and Z- from S- that are closest to X
c. If (X is a positive instance) then
Near-hit = Z+; Near-miss = Z-
else
Near-hit = Z-; Near-miss = Z+
d. update-weight(W, X, Near-hit, Near-miss)
4. Compute relevance = (1/m) W
5. For i = 1 to p:
a. If (relevance_i ≥ τ) then f_i is a relevant feature
b. Else f_i is an irrelevant feature
update-weight(W, X, Near-hit, Near-miss)
For i = 1 to p:
W_i = W_i - diff(x_i, near-hit_i)^2 + diff(x_i, near-miss_i)^2
Relief(S, m, τ)
1. Separate S into S+ = {positive instances} and S- = {negative instances}
2. Initialize weight vector W = (0, 0, ..., 0)
3. For i = 1 to m:
a. Pick at random an instance X ∈ S
b. From X, pick at random:
Z+ from S+ and Z- from S- that are closest to X
c. If (X is a positive instance) then
Near-hit = Z+; Near-miss = Z-
else
Near-hit = Z-; Near-miss = Z+
d. update-weight(W, X, Near-hit, Near-miss)
4. Compute relevance = (1/m) W
5. For i = 1 to p:
a. If (relevance_i ≥ τ) then f_i is a relevant feature
b. Else f_i is an irrelevant feature
update-weight(W, X, Near-hit, Near-miss)
For i = 1 to p:
W_i = W_i - diff(x_i, near-hit_i)^2 + diff(x_i, near-miss_i)^2
diff computes the normalized absolute difference between feature values.[8]
ReliefF Algorithm Mechanics
The ReliefF algorithm extends the original Relief method to address multi-class classification problems, incomplete data, and noisy environments by iteratively estimating feature weights based on the ability of features to differentiate nearest neighbors from the same class (hits) versus different classes (misses). It initializes all feature weights to 0 and performs iterations, where in each iteration a random training instance is selected without replacement. For the selected , the algorithm identifies nearest hits from the same class using a distance metric, typically Euclidean for continuous features or Hamming for discrete. To handle multiple classes, it then finds, for each distinct class class(), the nearest misses from that class, enabling evaluation across all class boundaries rather than pairwise comparisons. Feature weights are updated for each attribute by contrasting differences to hits and misses, formalized as: where is the prior probability of class , ensuring misses from rarer classes contribute proportionally less to avoid bias toward majority classes. The negative term penalizes features that vary among same-class instances, indicating poor relevance or noise, while the positive term rewards features that differ between and opposite-class instances, highlighting discriminative power. This contextual scoring captures feature interactions by leveraging local neighborhoods, approximating . The difference function adapts to feature types: for discrete attributes, it equals 0 if values match and 1 otherwise; for continuous, it applies a ramp function between thresholds (e.g., 5-10% of range) to mitigate underestimation of relevance in scaled differences. Incomplete data is handled probabilistically: if one instance lacks a value, ; if both lack values, it averages over possible value probabilities conditioned on classes. Parameters (typically 10 for noise tolerance) and (often dataset size for coverage) tune locality and sampling; higher averages more neighbors for robustness but increases computation. Pseudocode for ReliefF is as follows:Input: Training instances with feature values and class labels
Output: Weight vector W for each feature A
1. Set all W[A] := 0.0
2. For i = 1 to m:
a. Randomly select instance R_i
b. Find k nearest hits H_j from same class as R_i
c. For each class C ≠ class(R_i):
- Find k nearest misses M_j(C) from C
d. For each feature A:
- W[A] := W[A] - (sum_{j=1 to k} diff(A, R_i, H_j)) / (m * k)
+ sum_{C ≠ class(R_i)} [P(C) / (1 - P(class(R_i))) * (sum_{j=1 to k} diff(A, R_i, M_j(C))) / (m * k)]
Input: Training instances with feature values and class labels
Output: Weight vector W for each feature A
1. Set all W[A] := 0.0
2. For i = 1 to m:
a. Randomly select instance R_i
b. Find k nearest hits H_j from same class as R_i
c. For each class C ≠ class(R_i):
- Find k nearest misses M_j(C) from C
d. For each feature A:
- W[A] := W[A] - (sum_{j=1 to k} diff(A, R_i, H_j)) / (m * k)
+ sum_{C ≠ class(R_i)} [P(C) / (1 - P(class(R_i))) * (sum_{j=1 to k} diff(A, R_i, M_j(C))) / (m * k)]
