Distributed Feature Selection for Efficient Economic Big Data Analysis
read more
Citations
An Adaptive Dropout Deep Computation Model for Industrial IoT Big Data Learning With Crowdsourcing to Cloud Computing
Energy-Efficient Scheduling for Real-Time Systems Based on Deep Q-Learning Model
Privacy-Preserving Double-Projection Deep Computation Model With Crowdsourcing on Cloud for Big Data Feature Learning
Incomplete multi-view clustering via deep semantic mapping
References
Data Mining: Practical Machine Learning Tools and Techniques
Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy
Feature selection based on mutual information: criteria ofmax-dependency, max-relevance, and min-redundancy
Correlation-based Feature Selection for Machine Learning
Related Papers (5)
Frequently Asked Questions (14)
Q2. What are the future works in "Distributed feature selection for efficient economic big data analysis" ?
In the future work, the authors plan to establish a platform of algorithm library based on the proposed framework.
Q3. What have the authors contributed in "Distributed feature selection for efficient economic big data analysis" ?
To address the challenges, this paper presents a new framework for efficient analysis of high-dimensional economic big data based on innovative distributed feature selection.
Q4. What are the three main groups of feature selection algorithms?
With respect to different selection strategies, feature selection algorithms can be categorized into four groups, namely the filter, wrapper, embedded, and hybrid methods.
Q5. How can the authors estimate the effects of inflation on economic growth?
By approaching data on developing economies, the semi-parametric method can estimate the potentially nonlinear effects of inflation on economic growth [9].
Q6. What is the density of the data point with the highest remaining density?
After the reduction, the data point with the highest remaining density is selected as the second cluster center and the density of each data point is further reduced according to its distance to the second cluster center.
Q7. What is the importance of the a-th attribute to select the k-th representative?
the importance of the a-th attribute to select the k-th representative economic record can be defined asI(a)k = n∑i=1I(i, a)k. (7)The attributes with the higher-ranking value contain more information of clusters than others, namely they have powerful impacts on typical economic phenomena analysis.
Q8. What is the key to capturing valuable information, meanings, and insights hidden in big data?
to support social and economic development, the key is to capture valuable information, meanings, and insights hidden in big data.
Q9. What is the main weakness of traditional methods for constructing models?
a weakness of most traditional econometric methods for constructing models is that they take no consideration of the indirect relations between response indicators and economic factors.
Q10. How are the two components of the economic data preprocessing and economic feature selection deployed?
1. Specially, to speed up the process of data analysis, the Economic Data Preprocess and EconomicFeature Selection are deployed in distributed platform [36].•
Q11. How is the density of the data set selected?
in order to avoid the points near the first cluster center being selected as other centers of clusters, an amount of density proportional is subtracted from each point to its distance from the first cluster center.
Q12. What is the effect of migratory distance on economic development?
in [10], the hypothesis is established that variation in migratory distance has a long-lasting effect on genetic diversity and the pattern of economic development.
Q13. What is the purpose of this paper?
This paper aims to reduce the potentially huge set of candidate attributes produced by the preprocess layer to a small set of possible attributes, which are diverse and similar to the attributes in the original data set.
Q14. What are the main issues of the big data analysis?
While all of these provide sufficient information for economic analysis, the issues of dimension and volume overload pose great challenges: (1) The collected huge volume data usually contains incomplete, incorrect and nonstandard items, which are difficult for processing.