Descripción
|
|
---|---|
Dealing with the whole dataset in big data estimation problems is usually unfeasible. A common solution then consists of dividing the data into several smaller sets, performing distributed Bayesian estimation and combining these partial estimates to obtain a global estimate. A major problem of this approach is the presence of a non-negligible bias in the partial estimators due to the mismatch between the unknown true prior and the prior assumed in the estimation. A simple method to mitigate the effect of this bias is proposed in this paper. Essentially, the approach is based on using a reference data set to obtain a rough estimation of the parameter of interest, i.e., a reference parameter. This information is then communicated to the partial filters that handle the smaller data sets, which can thus use a refined prior centered around this parameter. Simulation results confirm the good performance of this scheme. | |
Internacional
|
Si |
Nombre congreso
|
IEEE International Workshop on Computational Advances in Multi-Sensor Adaptive Processing (CAMSAP) |
Tipo de participación
|
970 |
Lugar del congreso
|
Cancún (México) |
Revisores
|
Si |
ISBN o ISSN
|
978-1-4799-1963-5 |
DOI
|
|
Fecha inicio congreso
|
13/12/2015 |
Fecha fin congreso
|
16/12/2015 |
Desde la página
|
257 |
Hasta la página
|
260 |
Título de las actas
|
Proceedings of the 6th IEEE International Workshop on Computational Advances in Multi-Sensor Adaptive Processing (CAMSAP) |