DATAI Seminars. Course 2024-2025
Complex survey data are derived by sampling the population of interest for the study based on a complex sampling design, implementing techniques such as stratification or clustering. In order to compensate for the unequal selection probabilities derived from the implementation of complex sampling design techniques, a sampling weight is assigned to each individual, which indicates the number of units this represents in the population. This kind of data is usually used for the same purposes as traditional simple random samples, including the development of prediction models, among others. However, due to the special characteristics of the data collection process, traditional statistical techniques are usually not valid in this context, and complex survey data need to be treated in particular ways. The goal of this talk is to present new methodological proposals that account for complex sampling designs in the development process of prediction models. In particular, a new design-based variable selection technique that accounts for the complex sampling structure has been proposed based on LASSO regression models. In addition, new design-based estimators have been proposed for estimating the discrimination ability of logistic regression models for dichotomous outcomes fitted to complex survey data. The validity of both proposals has been evaluated through simulation studies. Finally, these proposals have been implemented into the svyVarSel and svyROC R packages, available on CRAN.
Ensuring fairness in AI is a complex challenge that requires bridging ethical principles with practical implementation. This talk explores key concepts in algorithmic fairness, from defining bias and fairness in predictive machine learning to assessing stigmas in large language models (LLMs). Using real-world examples, we illustrate the sociotechnical nature of fairness and the need for case-specific approaches.
As LLMs reshape the landscape, we examine new challenges, including how bias manifests differently across languages, methods for measuring these disparities, and strategies for mitigation. Finally, we advocate for an ethics-by-design mindset, emphasizing continuous monitoring, multidisciplinary collaboration, and proactive governance to ensure fairness remains at the core of AI development.
Load forecasting is essential for the efficient, reliable, and cost-effective management of power systems. Load forecasting performance can be improved by learning the similarities among multiple entities (e.g., regions, buildings). Techniques based on multi-task learning obtain predictions by leveraging consumption patterns from the historical load demand of multiple entities and their relationships. However, existing techniques cannot effectively assess inherent uncertainties in load demand or cannot account for dynamic changes in consumption patterns. This talk proposes a multi-task learning technique for online and probabilistic load forecasting. This technique provides accurate probabilistic predictions for loads of multiple entities by leveraging their dynamic similarities. The method's performance is evaluated using datasets that register the load demand of multiple entities and contain diverse and dynamic consumption patterns. The experimental results show that the proposed method can significantly enhance the effectiveness of current multi-task learning approaches across a wide variety of load consumption scenarios.
Insights in Quantum Information Theory
27/11/2024 / Pedro Crespo Bofill, Tecnun School of Engineering
In this talk we will look at the interplay between Shannon Classical Information Theory and Quantum Information.
The concepts of Classical Compression and Channel Coding will be related with their Quantum counterparts.
The use of artificial intelligence (AI) tools in all fields of human activity raises crucial questions about professional ethical responsibility. To what extent do AI techniques represent an additional ethical challenge compared to traditional computational tools? The talk focuses on the problem of explainability and understanding of work with AI by means of "black box" processes: how can an engineer, a doctor, a lawyer, assume ethical responsibility if he cannot rationally explain his work, both to himself and to others? A new understanding of ethics in the design and use of artificial intelligence technologies thus proves necessary.