Academic literature on the topic 'Data-dependent complexity'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Data-dependent complexity.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Data-dependent complexity"

1

Synnevåg, J.-F., A. Austeng, and S. Holm. "A low-complexity data-dependent beamformer." IEEE Transactions on Ultrasonics, Ferroelectrics and Frequency Control 58, no. 2 (February 2011): 281–89. http://dx.doi.org/10.1109/tuffc.2011.1805.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Fritzsche, Paula, Dolores Rexachs, and Emilio Luque. "Defining Asymptotic Parallel Time Complexity of Data-dependent Algorithms." New Generation Computing 32, no. 2 (April 2014): 123–44. http://dx.doi.org/10.1007/s00354-014-0202-2.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Di, J., and E. Kolaczyk. "Complexity-penalized estimation of minimum volume sets for dependent data." Journal of Multivariate Analysis 101, no. 9 (October 2010): 1910–26. http://dx.doi.org/10.1016/j.jmva.2010.04.009.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Bazin, Jérémie, Pilar Bustos-Sanmamed, Caroline Hartmann, Christine Lelandais-Brière, and Martin Crespi. "Complexity of miRNA-dependent regulation in root symbiosis." Philosophical Transactions of the Royal Society B: Biological Sciences 367, no. 1595 (June 5, 2012): 1570–79. http://dx.doi.org/10.1098/rstb.2011.0228.

Full text
Abstract:
The development of root systems may be strongly affected by the symbiotic interactions that plants establish with soil organisms. Legumes are able to develop symbiotic relationships with both rhizobial bacteria and arbuscular mycorrhizal fungi leading to the formation of nitrogen-fixing nodules and mycorrhizal arbuscules, respectively. Both of these symbiotic interactions involve complex cellular reprogramming and profound morphological and physiological changes in specific root cells. In addition, the repression of pathogenic defence responses seems to be required for successful symbiotic interactions. Apart from typical regulatory genes, such as transcription factors, microRNAs (miRNAs) are emerging as riboregulators that control gene networks in eukaryotic cells through interactions with specific target mRNAs. In recent years, the availability of deep-sequencing technologies and the development of in silico approaches have allowed for the identification of large sets of miRNAs and their targets in legumes . A number of conserved and legume-specific miRNAs were found to be associated with symbiotic interactions as shown by their expression patterns or actions on symbiosis-related targets. In this review, we combine data from recent literature and genomic and deep-sequencing data on miRNAs controlling nodule development or restricting defence reactions to address the diversity and specificity of miRNA-dependent regulation in legume root symbiosis. Phylogenetic analysis of miRNA isoforms and their potential targets suggests a role for miRNAs in the repression of plant defence during symbiosis and revealed the evolution of miRNA-dependent regulation in legumes to allow for the modification of root cell specification, such as the formation of mycorrhized roots and nitrogen-fixing nodules.
APA, Harvard, Vancouver, ISO, and other styles
5

Yuen, Hok Pan, and Andrew Mackinnon. "Performance of joint modelling of time-to-event data with time-dependent predictors: an assessment based on transition to psychosis data." PeerJ 4 (October 19, 2016): e2582. http://dx.doi.org/10.7717/peerj.2582.

Full text
Abstract:
Joint modelling has emerged to be a potential tool to analyse data with a time-to-event outcome and longitudinal measurements collected over a series of time points. Joint modelling involves the simultaneous modelling of the two components, namely the time-to-event component and the longitudinal component. The main challenges of joint modelling are the mathematical and computational complexity. Recent advances in joint modelling have seen the emergence of several software packages which have implemented some of the computational requirements to run joint models. These packages have opened the door for more routine use of joint modelling. Through simulations and real data based on transition to psychosis research, we compared joint model analysis of time-to-event outcome with the conventional Cox regression analysis. We also compared a number of packages for fitting joint models. Our results suggest that joint modelling do have advantages over conventional analysis despite its potential complexity. Our results also suggest that the results of analyses may depend on how the methodology is implemented.
APA, Harvard, Vancouver, ISO, and other styles
6

El-Yaniv, R., and D. Pechyony. "Transductive Rademacher Complexity and its Applications." Journal of Artificial Intelligence Research 35 (June 22, 2009): 193–234. http://dx.doi.org/10.1613/jair.2587.

Full text
Abstract:
We develop a technique for deriving data-dependent error bounds for transductive learning algorithms based on transductive Rademacher complexity. Our technique is based on a novel general error bound for transduction in terms of transductive Rademacher complexity, together with a novel bounding technique for Rademacher averages for particular algorithms, in terms of their "unlabeled-labeled" representation. This technique is relevant to many advanced graph-based transductive algorithms and we demonstrate its effectiveness by deriving error bounds to three well known algorithms. Finally, we present a new PAC-Bayesian bound for mixtures of transductive algorithms based on our Rademacher bounds.
APA, Harvard, Vancouver, ISO, and other styles
7

Chen, Bo, Yuanyuan Guo, Xiue Gao, and Yunming Wang. "A Novel Multi-Attribute Decision Making Approach: Addressing the Complexity of Time Dependent and Interdependent Data." IEEE Access 6 (2018): 55838–49. http://dx.doi.org/10.1109/access.2018.2872636.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Nasution, Mahyuddin K. M., Rahmad Syah, and Marischa Elveny. "Social Network Analysis: Towards Complexity Problem." Webology 18, no. 2 (December 23, 2021): 449–61. http://dx.doi.org/10.14704/web/v18i2/web18332.

Full text
Abstract:
Social network analysis is a advances from field of social networks. The structuring of social actors, with data models and involving intelligence abstracted in mathematics, and without analysis it will not present the function of social networks. However, graph theory inherits process and computational procedures for social network analysis, and it proves that social network analysis is mathematical and computational dependent on the degree of nodes in the graph or the degree of social actors in social networks. Of course, the process of acquiring social networks bequeathed the same complexity toward the social network analysis, where the approach has used the social network extraction and formulated its consequences in computing.
APA, Harvard, Vancouver, ISO, and other styles
9

Romero, E., and R. Sullivan. "Complexity of the outward K+ current of the rat megakaryocyte." American Journal of Physiology-Cell Physiology 272, no. 5 (May 1, 1997): C1525—C1531. http://dx.doi.org/10.1152/ajpcell.1997.272.5.c1525.

Full text
Abstract:
Megakaryocytes isolated from rat bone marrow express a voltage-dependent, outward K+ current with complex kinetics of activation and inactivation. We found that this current could be separated into at least two components based on differential responses to K+ channel blockers. One component, which exhibited features of the "transient" or "A-type" K+ current of excitable cells, was more strongly blocked by 4-aminopyridine (4-AP) than by tetrabutylammonium (TBA). This current, which we designated as "4-AP-sensitive" current, activated rapidly at potentials more positive than -40 mV and subsequently underwent rapid voltage-dependent inactivation. A separate current that activated slowly was blocked much more effectively by TBA than by 4-AP. This "TBA-sensitive" component, which resembled a typical delayed rectifier current, was much more resistant to voltage-dependent inactivation. The relative contribution of each of these components varied from cell to cell. The effect of charybdotoxin was similar to that of 4-AP. Our data indicate that the voltage-dependent K+ current of resting megakaryocytes is more complex than heretofore believed and support the emerging concept that megakaryocytes possess intricate electrophysiological properties.
APA, Harvard, Vancouver, ISO, and other styles
10

Tompkins, Catherine J. "SURVIVING THE COMPLEXITY: USING GROUNDED THEORY TO UNDERSTAND KINSHIP CAREGIVING." Innovation in Aging 3, Supplement_1 (November 2019): S803—S804. http://dx.doi.org/10.1093/geroni/igz038.2956.

Full text
Abstract:
Abstract The role of a caregiver often goes beyond the task of caring for someone who is dependent in managing activities of daily living. Children are dependent on others to care for them due solely to their age and maturity; others are dependent due to chronic ailments or short-term disabilities. Regardless of why someone is dependent, the caregiving relationship is complex. This paper focuses on a grounded theory, developed and applied to understand the complexities of kinship caregiving. The literature continues to support the identified needs of kinship caregivers (Tompkins, 2015; Lee, Clarkson-Hendrix, & Lee, 2016). To understand the unique needs of kinship families, the following grand tour question was asked: What is it like for you to live within a kinship caregiving household? The theory was developed over several years based on observational data and 15 interviews with grandparent caregivers and at least one of the grandchildren they were raising. The theory, Surviving the Complexity, is a survival process of taking on the caregiving role and doing one’s best in spite of multiple obstacles. Surviving the complexity consists of three stages: rescuing, taking-on and role reversal. The theory identifies and explains emotional, relationship and situational complexity within kinship families. Hope and denial are factors of emotional complexity: “It’s not that she (my daughter) does not love him (the child), she is just unable to right now. She will get better.” Theory development and further application of the theory will be discussed.
APA, Harvard, Vancouver, ISO, and other styles

Dissertations / Theses on the topic "Data-dependent complexity"

1

Philips, Petra Camilla, and petra philips@gmail com. "Data-Dependent Analysis of Learning Algorithms." The Australian National University. Research School of Information Sciences and Engineering, 2005. http://thesis.anu.edu.au./public/adt-ANU20050901.204523.

Full text
Abstract:
This thesis studies the generalization ability of machine learning algorithms in a statistical setting. It focuses on the data-dependent analysis of the generalization performance of learning algorithms in order to make full use of the potential of the actual training sample from which these algorithms learn.¶ First, we propose an extension of the standard framework for the derivation of generalization bounds for algorithms taking their hypotheses from random classes of functions. This approach is motivated by the fact that the function produced by a learning algorithm based on a random sample of data depends on this sample and is therefore a random function. Such an approach avoids the detour of the worst-case uniform bounds as done in the standard approach. We show that the mechanism which allows one to obtain generalization bounds for random classes in our framework is based on a “small complexity” of certain random coordinate projections. We demonstrate how this notion of complexity relates to learnability and how one can explore geometric properties of these projections in order to derive estimates of rates of convergence and good confidence interval estimates for the expected risk. We then demonstrate the generality of our new approach by presenting a range of examples, among them the algorithm-dependent compression schemes and the data-dependent luckiness frameworks, which fall into our random subclass framework.¶ Second, we study in more detail generalization bounds for a specific algorithm which is of central importance in learning theory, namely the Empirical Risk Minimization algorithm (ERM). Recent results show that one can significantly improve the high-probability estimates for the convergence rates for empirical minimizers by a direct analysis of the ERM algorithm. These results are based on a new localized notion of complexity of subsets of hypothesis functions with identical expected errors and are therefore dependent on the underlying unknown distribution. We investigate the extent to which one can estimate these high-probability convergence rates in a data-dependent manner. We provide an algorithm which computes a data-dependent upper bound for the expected error of empirical minimizers in terms of the “complexity” of data-dependent local subsets. These subsets are sets of functions of empirical errors of a given range and can be determined based solely on empirical data. We then show that recent direct estimates, which are essentially sharp estimates on the high-probability convergence rate for the ERM algorithm, can not be recovered universally from empirical data.
APA, Harvard, Vancouver, ISO, and other styles
2

Mehta, Nishant A. "On sparse representations and new meta-learning paradigms for representation learning." Diss., Georgia Institute of Technology, 2013. http://hdl.handle.net/1853/52159.

Full text
Abstract:
Given the "right" representation, learning is easy. This thesis studies representation learning and meta-learning, with a special focus on sparse representations. Meta-learning is fundamental to machine learning, and it translates to learning to learn itself. The presentation unfolds in two parts. In the first part, we establish learning theoretic results for learning sparse representations. The second part introduces new multi-task and meta-learning paradigms for representation learning. On the sparse representations front, our main pursuits are generalization error bounds to support a supervised dictionary learning model for Lasso-style sparse coding. Such predictive sparse coding algorithms have been applied with much success in the literature; even more common have been applications of unsupervised sparse coding followed by supervised linear hypothesis learning. We present two generalization error bounds for predictive sparse coding, handling the overcomplete setting (more original dimensions than learned features) and the infinite-dimensional setting. Our analysis led to a fundamental stability result for the Lasso that shows the stability of the solution vector to design matrix perturbations. We also introduce and analyze new multi-task models for (unsupervised) sparse coding and predictive sparse coding, allowing for one dictionary per task but with sharing between the tasks' dictionaries. The second part introduces new meta-learning paradigms to realize unprecedented types of learning guarantees for meta-learning. Specifically sought are guarantees on a meta-learner's performance on new tasks encountered in an environment of tasks. Nearly all previous work produced bounds on the expected risk, whereas we produce tail bounds on the risk, thereby providing performance guarantees on the risk for a single new task drawn from the environment. The new paradigms include minimax multi-task learning (minimax MTL) and sample variance penalized meta-learning (SVP-ML). Regarding minimax MTL, we provide a high probability learning guarantee on its performance on individual tasks encountered in the future, the first of its kind. We also present two continua of meta-learning formulations, each interpolating between classical multi-task learning and minimax multi-task learning. The idea of SVP-ML is to minimize the task average of the training tasks' empirical risks plus a penalty on their sample variance. Controlling this sample variance can potentially yield a faster rate of decrease for upper bounds on the expected risk of new tasks, while also yielding high probability guarantees on the meta-learner's average performance over a draw of new test tasks. An algorithm is presented for SVP-ML with feature selection representations, as well as a quite natural convex relaxation of the SVP-ML objective.
APA, Harvard, Vancouver, ISO, and other styles
3

Philips, Petra. "Data-Dependent Analysis of Learning Algorithms." Phd thesis, 2005. http://hdl.handle.net/1885/47998.

Full text
Abstract:
This thesis studies the generalization ability of machine learning algorithms in a statistical setting. It focuses on the data-dependent analysis of the generalization performance of learning algorithms in order to make full use of the potential of the actual training sample from which these algorithms learn.¶ First, we propose an extension of the standard framework for the derivation of generalization bounds for algorithms taking their hypotheses from random classes of functions. ... ¶ Second, we study in more detail generalization bounds for a specific algorithm which is of central importance in learning theory, namely the Empirical Risk Minimization algorithm (ERM). ...
APA, Harvard, Vancouver, ISO, and other styles
4

Abd-Rabbo, Diala. "Beyond hairballs: depicting complexity of a kinase-phosphatase network in the budding yeast." Thèse, 2017. http://hdl.handle.net/1866/19318.

Full text
APA, Harvard, Vancouver, ISO, and other styles

Books on the topic "Data-dependent complexity"

1

Wikle, Christopher K. Spatial Statistics. Oxford University Press, 2018. http://dx.doi.org/10.1093/acrefore/9780190228620.013.710.

Full text
Abstract:
The climate system consists of interactions between physical, biological, chemical, and human processes across a wide range of spatial and temporal scales. Characterizing the behavior of components of this system is crucial for scientists and decision makers. There is substantial uncertainty associated with observations of this system as well as our understanding of various system components and their interaction. Thus, inference and prediction in climate science should accommodate uncertainty in order to facilitate the decision-making process. Statistical science is designed to provide the tools to perform inference and prediction in the presence of uncertainty. In particular, the field of spatial statistics considers inference and prediction for uncertain processes that exhibit dependence in space and/or time. Traditionally, this is done descriptively through the characterization of the first two moments of the process, one expressing the mean structure and one accounting for dependence through covariability.Historically, there are three primary areas of methodological development in spatial statistics: geostatistics, which considers processes that vary continuously over space; areal or lattice processes, which considers processes that are defined on a countable discrete domain (e.g., political units); and, spatial point patterns (or point processes), which consider the locations of events in space to be a random process. All of these methods have been used in the climate sciences, but the most prominent has been the geostatistical methodology. This methodology was simultaneously discovered in geology and in meteorology and provides a way to do optimal prediction (interpolation) in space and can facilitate parameter inference for spatial data. These methods rely strongly on Gaussian process theory, which is increasingly of interest in machine learning. These methods are common in the spatial statistics literature, but much development is still being done in the area to accommodate more complex processes and “big data” applications. Newer approaches are based on restricting models to neighbor-based representations or reformulating the random spatial process in terms of a basis expansion. There are many computational and flexibility advantages to these approaches, depending on the specific implementation. Complexity is also increasingly being accommodated through the use of the hierarchical modeling paradigm, which provides a probabilistically consistent way to decompose the data, process, and parameters corresponding to the spatial or spatio-temporal process.Perhaps the biggest challenge in modern applications of spatial and spatio-temporal statistics is to develop methods that are flexible yet can account for the complex dependencies between and across processes, account for uncertainty in all aspects of the problem, and still be computationally tractable. These are daunting challenges, yet it is a very active area of research, and new solutions are constantly being developed. New methods are also being rapidly developed in the machine learning community, and these methods are increasingly more applicable to dependent processes. The interaction and cross-fertilization between the machine learning and spatial statistics community is growing, which will likely lead to a new generation of spatial statistical methods that are applicable to climate science.
APA, Harvard, Vancouver, ISO, and other styles

Book chapters on the topic "Data-dependent complexity"

1

Peinke, J., Ch Renner, and R. Friedrich. "Turbulence and Financial Market Data Analyzed with Respect to Their Scale Dependent Complexity." In Complexity from Microscopic to Macroscopic Scales: Coherence and Large Deviations, 151–69. Dordrecht: Springer Netherlands, 2002. http://dx.doi.org/10.1007/978-94-010-0419-0_9.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Blocki, Jeremiah, and Blake Holman. "Sustained Space and Cumulative Complexity Trade-Offs for Data-Dependent Memory-Hard Functions." In Advances in Cryptology – CRYPTO 2022, 222–51. Cham: Springer Nature Switzerland, 2022. http://dx.doi.org/10.1007/978-3-031-15982-4_8.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Cohen, Albert, Wolfgang Dahmen, and Ron DeVore. "State Estimation—The Role of Reduced Models." In SEMA SIMAI Springer Series, 57–77. Cham: Springer International Publishing, 2022. http://dx.doi.org/10.1007/978-3-030-86236-7_4.

Full text
Abstract:
AbstractThe exploration of complex physical or technological processes usually requires exploiting available information from different sources: (i) physical laws often represented as a family of parameter dependent partial differential equations and (ii) data provided by measurement devices or sensors. The amount of sensors is typically limited and data acquisition may be expensive and in some cases even harmful. This article reviews some recent developments for this “small-data” scenario where inversion is strongly aggravated by the typically large parametric dimensionality. The proposed concepts may be viewed as exploring alternatives to Bayesian inversion in favor of more deterministic accuracy quantification related to the required computational complexity. We discuss optimality criteria which delineate intrinsic information limits, and highlight the role of reduced models for developing efficient computational strategies. In particular, the need to adapt the reduced models—not to a specific (possibly noisy) data set but rather to the sensor system—is a central theme. This, in turn, is facilitated by exploiting geometric perspectives based on proper stable variational formulations of the continuous model.
APA, Harvard, Vancouver, ISO, and other styles
4

Stojanovski, David, Uxio Hermida, Marica Muffoletto, Pablo Lamata, Arian Beqiri, and Alberto Gomez. "Efficient Pix2Vox++ for 3D Cardiac Reconstruction from 2D Echo Views." In Simplifying Medical Ultrasound, 86–95. Cham: Springer International Publishing, 2022. http://dx.doi.org/10.1007/978-3-031-16902-1_9.

Full text
Abstract:
AbstractAccurate geometric quantification of the human heart is a key step in the diagnosis of numerous cardiac diseases, and in the management of cardiac patients. Ultrasound imaging is the primary modality for cardiac imaging, however acquisition requires high operator skill, and its interpretation and analysis is difficult due to artifacts. Reconstructing cardiac anatomy in 3D can enable discovery of new biomarkers and make imaging less dependent on operator expertise, however most ultrasound systems only have 2D imaging capabilities. We propose both a simple alteration to the Pix2Vox++ networks for a sizeable reduction in memory usage and computational complexity, and a pipeline to perform reconstruction of 3D anatomy from 2D standard cardiac views, effectively enabling 3D anatomical reconstruction from limited 2D data. We evaluate our pipeline using synthetically generated data achieving accurate 3D whole-heart reconstructions (peak intersection over union score $$> 0.88$$ > 0.88 ) from just two standard anatomical 2D views of the heart. We also show preliminary results using real echo images.
APA, Harvard, Vancouver, ISO, and other styles
5

Roden, Dylan F., Jennifer M. Johnson, Petr Szturz, Paolo Bossi, and Athanassios Argiris. "New and Promising Targeted Therapies in First and Second-Line Settings." In Critical Issues in Head and Neck Oncology, 277–96. Cham: Springer International Publishing, 2021. http://dx.doi.org/10.1007/978-3-030-63234-2_18.

Full text
Abstract:
AbstractDeeper understanding of the molecular pathogenesis of malignancies, including head and neck squamous cell carcinoma (HNSCC), has led to the investigation of several novel targeted therapies. These therapeutic approaches may eventually replace or complement existing treatment modalities, such as surgery, radiation therapy, and traditional cytotoxic chemotherapy. Epidermal growth factor receptor (EGFR) inhibitors, and specifically cetuximab, are as of now the only class of targeted agents, excluding immune checkpoint inhibitors, with approval in the treatment of HNSCC. Beyond EGFR inhibition, novel therapies under evaluation are directed against vascular endothelial growth factor (VEGF) and VEGF receptor (VEGFR), PI3K/AKT/mTOR pathway, cell cycle regulation (for example, cyclin dependent kinases 4 and 6), HRAS, DNA repair mechanisms, and others. Development of new therapies has to take into consideration the complexity of solid tumors and their heterogeneity. Multitargeted combination therapy approaches may be required in certain cases in order to maximize antitumor effect. Ways to individualize treatment using validated biomarkers are likely to improve outcomes. We review the most relevant molecular targets in HNSCC and provide updates on clinical trial data with promising new targeted agents.
APA, Harvard, Vancouver, ISO, and other styles
6

Sarkar, Saswati, Anirban Kundu, and Ayan Banerjee. "Evaluation of Reliable Data Storage in Cloud Using an Efficient Encryption Technique." In Research Anthology on Privatizing and Securing Data, 758–72. IGI Global, 2021. http://dx.doi.org/10.4018/978-1-7998-8954-0.ch034.

Full text
Abstract:
Cloud-based reliable and protected data storage technique is proposed in this chapter. The proposed technique encrypts and protects data with less time consumption. Power consumption of storage is dependent upon capacity of storage and physical size of storage. Time analysis is presented graphically in this chapter. Reliable data storage is represented in cloud based proposed approach. Data is encrypted with minimum time complexity due to usage of proposed cloud-based reliable data storage. The competent ratio of time complexity is graphically observed in proposed data storage technique. Power consumption of storage has been typically dependent on the basis of capacity of storage and amount of storage. A ratio of power consumption and capacity of storage is presented in cloud-based approach. An efficient usage of energy is shown depending on current consumption and voltage in proposed reliable approach.
APA, Harvard, Vancouver, ISO, and other styles
7

Sarkar, Saswati, Anirban Kundu, and Ayan Banerjee. "Evaluation of Reliable Data Storage in Cloud Using an Efficient Encryption Technique." In Handbook of Research on Cloud Computing and Big Data Applications in IoT, 229–42. IGI Global, 2019. http://dx.doi.org/10.4018/978-1-5225-8407-0.ch012.

Full text
Abstract:
Cloud-based reliable and protected data storage technique is proposed in this chapter. The proposed technique encrypts and protects data with less time consumption. Power consumption of storage is dependent upon capacity of storage and physical size of storage. Time analysis is presented graphically in this chapter. Reliable data storage is represented in cloud based proposed approach. Data is encrypted with minimum time complexity due to usage of proposed cloud-based reliable data storage. The competent ratio of time complexity is graphically observed in proposed data storage technique. Power consumption of storage has been typically dependent on the basis of capacity of storage and amount of storage. A ratio of power consumption and capacity of storage is presented in cloud-based approach. An efficient usage of energy is shown depending on current consumption and voltage in proposed reliable approach.
APA, Harvard, Vancouver, ISO, and other styles
8

Jamalmohammed, Saira Banu, Lavanya K., Sumaiya Thaseen I., and Biju V. "Review on Sparse Matrix Storage Formats With Space Complexity Analysis." In Applications of Artificial Intelligence for Smart Technology, 122–45. IGI Global, 2021. http://dx.doi.org/10.4018/978-1-7998-3335-2.ch009.

Full text
Abstract:
Sparse matrix-vector multiplication (SpMV) is a challenging computational kernel in linear algebra applications, like data mining, image processing, and machine learning. The performance of this kernel is greatly dependent on the size of the input matrix and the underlying hardware features. Various sparse matrix storage formats referred to commonly as sparse formats have been proposed in the literature to reduce the size of the matrix. In modern multi-core and many-core architectures, the performance of the kernel is mainly dependent on memory wall and power wall problem. Normally review on sparse formats is done with specific architecture or with specific application. This chapter presents a comparative study on various sparse formats in cross platform architecture like CPU, graphics processor unit (GPU), and single instruction multiple data stream (SIMD) registers. Space complexity analysis of various formats with its representation is discussed. Finally, the merits and demerits of each format have been summarized into a table.
APA, Harvard, Vancouver, ISO, and other styles
9

Mic, Vladimir, and Pavel Zezula. "On the Similarity Search With Hamming Space Sketches." In Advances in Data Mining and Database Management, 97–127. IGI Global, 2021. http://dx.doi.org/10.4018/978-1-7998-4963-6.ch005.

Full text
Abstract:
This chapter focuses on data searching, which is nowadays mostly based on similarity. The similarity search is challenging due to its computational complexity, and also the fact that similarity is subjective and context dependent. The authors assume the metric space model of similarity, defined by the domain of objects and the metric function that measures the dissimilarity of object pairs. The volume of contemporary data is large, and the time efficiency of similarity query executions is essential. This chapter investigates transformations of metric space to Hamming space to decrease the memory and computational complexity of the search. Various challenges of the similarity search with sketches in the Hamming space are addressed, including the definition of sketching transformation and efficient search algorithms that exploit sketches to speed-up searching. The indexing of Hamming space and a heuristic to facilitate the selection of a suitable sketching technique for any given application are also considered.
APA, Harvard, Vancouver, ISO, and other styles
10

Castellano, Leonardo, Nicoletta Sala, Angelo Rolla, and Walter Ambrosetti. "The Residence Time of the Water in Lake MAGGIORE. Through an Eulerian-Lagrangian Approach." In Complexity Science, Living Systems, and Reflexing Interfaces, 218–34. IGI Global, 2013. http://dx.doi.org/10.4018/978-1-4666-2077-3.ch011.

Full text
Abstract:
This chapter describes a study designed to evaluate the spectrum of the residence time of the water at different depths of a deep lake, and to examine the mechanisms governing the seasonal cycle of thermal stratification and destratification, with the ultimate aim of assessing the actual exchange time of the lake water. The study was performed on Lake Maggiore (depth 370m) using a multidimensional mathematical model and computer codes for the heat and mass transfer in very large natural water bodies. A 3D Eulerian time-dependent CFD (Computational Fluid Dynamics) code was applied under real conditions, taking into account the effects of the monthly mean values of the mass flow rates and temperatures of all the tributaries, mass flow rate of the Ticino effluent and meteorological, hydrological, and limnological parameters available from the rich data-base of the CNR-ISE (Pallanza). The velocity distributions from these simulations were used to compute the paths of a large number of massless markers with different initial positions and evaluate their residence times in the lake.
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Data-dependent complexity"

1

Synnevag, Johan-Fredrik, Sverre Holm, and Andreas Austeng. "A low complexity data-dependent beamformer." In 2008 IEEE Ultrasonics Symposium (IUS). IEEE, 2008. http://dx.doi.org/10.1109/ultsym.2008.0261.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Li, Peng, and Rodrigo C. de Lamare. "Low-complexity robust data-dependent dimensionality reduction based on joint iterative optimization of parameters." In 2013 IEEE 5th International Workshop on Computational Advances in Multi-Sensor Adaptive Processing (CAMSAP). IEEE, 2013. http://dx.doi.org/10.1109/camsap.2013.6714004.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Maximov, Yury, Massih-Reza Amini, and Zaid Harchaoui. "Rademacher Complexity Bounds for a Penalized Multi-class Semi-supervised Algorithm (Extended Abstract)." In Twenty-Seventh International Joint Conference on Artificial Intelligence {IJCAI-18}. California: International Joint Conferences on Artificial Intelligence Organization, 2018. http://dx.doi.org/10.24963/ijcai.2018/800.

Full text
Abstract:
We propose Rademacher complexity bounds for multi-class classifiers trained with a two-step semi-supervised model. In the first step, the algorithm partitions the partially labeled data and then identifies dense clusters containing k predominant classes using the labeled training examples such that the proportion of their non-predominant classes is below a fixed threshold stands for clustering consistency. In the second step, a classifier is trained by minimizing a margin empirical loss over the labeled training set and a penalization term measuring the disability of the learner to predict the k predominant classes of the identified clusters. The resulting data-dependent generalization error bound involves the margin distribution of the classifier, the stability of the clustering technique used in the first step and Rademacher complexity terms corresponding to partially labeled training data. Our theoretical result exhibit convergence rates extending those proposed in the literature for the binary case, and experimental results on different multi-class classification problems show empirical evidence that supports the theory.
APA, Harvard, Vancouver, ISO, and other styles
4

Ding, Jingtao, Guanghui Yu, Xiangnan He, Yuhan Quan, Yong Li, Tat-Seng Chua, Depeng Jin, and Jiajie Yu. "Improving Implicit Recommender Systems with View Data." In Twenty-Seventh International Joint Conference on Artificial Intelligence {IJCAI-18}. California: International Joint Conferences on Artificial Intelligence Organization, 2018. http://dx.doi.org/10.24963/ijcai.2018/464.

Full text
Abstract:
Most existing recommender systems leverage the primary feedback data only, such as the purchase records in E-commerce. In this work, we additionally integrate view data into implicit feedback based recommender systems (dubbed as Implicit Recommender Systems). We propose to model the pairwise ranking relations among purchased, viewed, and non-viewed interactions, being more effective and flexible than typical pointwise matrix factorization (MF) methods. However, such a pairwise formulation poses efficiency challenges in learning the model. To address this problem, we design a new learning algorithm based on the element-wise Alternating Least Squares (eALS) learner. Notably, our algorithm can efficiently learn model parameters from the whole user-item matrix (including all missing data), with a rather low time complexity that is dependent on the observed data only. Extensive experiments on two real-world datasets demonstrate that our method outperforms several state-of-the-art MF methods by 10% ∼ 28.4%. Our implementation is available at: https://github.com/ dingjingtao/View_enhanced_ALS.
APA, Harvard, Vancouver, ISO, and other styles
5

Gao, Jianbo, Jing Hu, and Wen-wen Tung. "Multiscale Analysis of Biological Signals." In ASME 2011 Dynamic Systems and Control Conference and Bath/ASME Symposium on Fluid Power and Motion Control. ASMEDC, 2011. http://dx.doi.org/10.1115/dscc2011-6084.

Full text
Abstract:
Complex systems often generate highly nonstationary and multiscale signals, due to nonlinear and stochastic interactions among their component systems and hierarchical regulations imposed by the operating environments. The further advances in the fields of life sciences, systems biology, nano-sciences, information systems, and physical sciences, have made it increasingly important to develop complexity measures that incorporate the concept of scale explicitly, so that different behaviors of the signals on varying scales can be simultaneously characterized by the same scale-dependent measure. Here, we propose such a measure, the scale-dependent Lyapunov exponent (SDLE), and develop a unified theory of multiscale analysis of complex data. We show that the SDLE can readily characterize low-dimensional chaos and random 1/fα processes, as well as accurately detect epileptic seizures from EEG data and distinguish healthy subjects from patients with congestive heart failure from heart rate variability (HRV) data. More importantly, our analyses of EEG and HRV data illustrate that commonly used complexity measures from information theory, chaos theory, and random fractal theory can be related to the values of the SDLE at specific scales, and useful information on the structured components of the data is also embodied by the SDLE.
APA, Harvard, Vancouver, ISO, and other styles
6

Cotrim, Lucas Pereira, Henrique Barros Oliveira, Asdrubal N. Queiroz Filho, Ismael H. F. Santos, Rodrigo Augusto Barreira, Eduardo Aoun Tannuri, Anna Helena Reali Costa, and Edson Satoshi Gomi. "Neural Network Meta-Models for FPSO Motion Prediction From Environmental Data." In ASME 2021 40th International Conference on Ocean, Offshore and Arctic Engineering. American Society of Mechanical Engineers, 2021. http://dx.doi.org/10.1115/omae2021-62674.

Full text
Abstract:
Abstract The current design process of mooring systems for FPSOs is highly dependent on the availability of the platform’s mathematical model and accuracy of dynamic simulations, through which resulting time series motion is evaluated according to design constraints. This process can be time-consuming and present inaccurate results due to the mathematical model’s limitations and overall complexity of the vessel’s dynamics. We propose a Neural Simulator, a set of data-based surrogate models with environmental data as input, each specialized in the prediction of different motion statistics relevant to mooring system design: Maximum Roll, Platform Offset and Fairlead Displacements. The meta-models are trained by real current, wind and wave data measured in 3h periods at the Campos Basin (Brazil) from 2003 to 2010 and the associated dynamic response of a spread-moored FPSO obtained through time-domain simulations using the Dynasim software. A comparative analysis of different model architectures is conducted and the proposed models are shown to correctly capture platform dynamics, providing good results when compared to the statistical analysis of time series motion obtained from Dynasim.
APA, Harvard, Vancouver, ISO, and other styles
7

Mendoza, Alberto, Çağrı Cerrahoğlu, Alessandro Delfino, and Martin Sundin. "Signal Processing and Machine Learning for Effective Integration of Distributed Fiber Optic Sensing Data in Production Petrophysics." In 2022 SPWLA 63rd Annual Symposium. Society of Petrophysicists and Well Log Analysts, 2022. http://dx.doi.org/10.30632/spwla-2022-0016.

Full text
Abstract:
Distributed fibre optic sensing (DFOS) is progressively being considered in the mix of customary surveillance tools for oil and gas producing assets. Its applications are beyond monitoring of wells for production and reservoir optimization, including detection of well integrity risks and other well completion failures. However, while DFOS can uniquely yield time-dependent spatially distributed measurements, these are yet to be routinely used in formation evaluation and production logging workflows. The large volumes and complexity of time- and depth-dependent data produced by DFOS often require the usage of Digital Signal Processing (DSP) to reduce the amount of stored data and data-driven techniques such as machine learning (ML) for analysis. Distributed sensing data is sampled at high rates; up to 10,000 samples per second and depth for Distributed Acoustic Sensing (DAS), and one sample per minute and depth for distributed temperature sensing (DTS). The high sampling rate in time, across hundreds or thousands of meters, creates a big data problem. Consequently, managing and transferring data acquired in the field to an expert analyst is extremely challenging. Even when these data management challenges are overcome, the amount of data itself is still not suitable for manual analysis. Starting from edge computing for feature extraction, we illustrate the principles of using DSP and ML to effectively handle the challenges of analyzing time-dependent distributed data from DFOS. Results enable integration of DFOS with customary formation evaluation and production surveillance workflows. Feature extraction, a crucial DSP step used to generate inputs to ML, reduces data size by orders of magnitude while ML models analyse continuous data streams from the field. We derive thermal features from DTS data effectively characterizing Joule Thomson effects. Moreover, we combine DTS thermal features with acoustic features from DAS in supervised ML for multiphase downhole inflow predictions. In so doing, we have successfully applied ML on DFOS for real-time detection of sand production, production and injection profiling, and well integrity surveillance. With use cases in a range of well completion types and well operating conditions, we demonstrate an endto- end system of DFOS that effectively integrates DAS and DTS into routine analysis techniques for Formation Evaluation Specialists and Production Petrophysicists.
APA, Harvard, Vancouver, ISO, and other styles
8

Weiss, Benjamin M., Joshua M. Hamel, Mark A. Ganter, and Duane W. Storti. "Data-Driven Additive Manufacturing Constraints for Topology Optimization." In ASME 2018 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference. American Society of Mechanical Engineers, 2018. http://dx.doi.org/10.1115/detc2018-85391.

Full text
Abstract:
The topology optimization (TO) of structures to be produced using additive manufacturing (AM) is explored using a data-driven constraint function that predicts the minimum producible size of small features in different shapes and orientations. This shape- and orientation-dependent manufacturing constraint, derived from experimental data, is implemented within a TO framework using a modified version of the Moving Morphable Components (MMC) approach. Because the analytic constraint function is fully differentiable, gradient-based optimization can be used. The MMC approach is extended in this work to include a “bootstrapping” step, which provides initial component layouts to the MMC algorithm based on intermediate Solid Isotropic Material with Penalization (SIMP) topology optimization results. This “bootstrapping” approach improves convergence compared to reference MMC implementations. Results from two compliance design optimization example problems demonstrate the successful integration of the manufacturability constraint in the MMC approach, and the optimal designs produced show minor changes in topology and shape compared to designs produced using fixed-radius filters in the traditional SIMP approach. The use of this data-driven manufacturability constraint makes it possible to take better advantage of the achievable complexity in additive manufacturing processes, while resulting in typical penalties to the design objective function of around only 2% when compared to the unconstrained case.
APA, Harvard, Vancouver, ISO, and other styles
9

Leenders, Arne, Hamed Vahdati Zadeh, and Matthias Wangenheim. "Modelling the Time-Dependent Behavior of Elastomers Using Fractional Viscoelastic Material Formulations." In ASME 2021 International Mechanical Engineering Congress and Exposition. American Society of Mechanical Engineers, 2021. http://dx.doi.org/10.1115/imece2021-71178.

Full text
Abstract:
Abstract Elastomer materials are often used for components such as tire treads or hydraulic sealings, when deformable and damping behavior of components are desired and high dynamic loads appear. Such elastomers show time- and frequency-dependent characteristics, called viscoelasticity. The modelling of viscoelastic material is mainly implemented in simulations by rheological models, which often consists of elastic and damping elements. A viscoelastic model can be parametrized to experimental data to describe a specific elastomer with high accuracy. The most common model is the Prony-series. This model uses several Maxwell-branches (connection of one elastic and one damping element in series). Every branch is only able to fit the experimental behavior at one single excitation frequency. This fact makes it necessary to use a lot of parameters for adapting the frequency- and temperature-dependent characteristics over decades of the excitation frequency. To overcome this need for a huge amount of parameters we formulate a fractional viscoelastic model approach that gets along with a much smaller set of parameters, using finite elements. In order to reduce the numerical effort, a similarly formulated model is set up on force-displacement level additionally. In this way, the complexity of the simulation can be reduced with mapping of the material behavior.
APA, Harvard, Vancouver, ISO, and other styles
10

Aktosun, Erdem, Nikolaos I. Xiros, and Jason M. Dahl. "Error Analysis of Models for the Forces on a Cylinder Undergoing 2-DOF Prescribed Motion in a Stream." In ASME 2021 International Mechanical Engineering Congress and Exposition. American Society of Mechanical Engineers, 2021. http://dx.doi.org/10.1115/imece2021-70102.

Full text
Abstract:
Abstract Based on experimental data, a data error estimation model is developed for an existing neural network time dependent hydrodynamic force model. The force model, based on forces measured in forced motion experiments, is used to approximate the time dependent forces on a cylinder that may undergo combined in-line and cross-flow vortex-induced vibrations. This model will be used to develop future control models to improve VIV-based energy harvesting systems. Previously, a dynamic model of appropriate complexity was created in order to approximate time dependent lift and drag forces based on force time histories measured in an expansive set of forced motion experiments. Position and velocity were used as input to the dynamic model. A feed forward neural network was trained using the force database in order to develop time dependent models of forces on the cylinder for prescribed sinusoidal motion. The time series error between the measured and feed forward Artificial Neural Network (ANN) model was found for the lift and drag force time histories. In the present study, an autoregressive error predictor is developed from the existing neural network time dependent model of forces on the cylinder for given kinematic conditions. This autoregressive (AR) error predictor is developed based on the error between the measured signal and artificial neural network model and can be used to improve predictions from the model.
APA, Harvard, Vancouver, ISO, and other styles

Reports on the topic "Data-dependent complexity"

1

McKenna, Patrick, and Mark Evans. Emergency Relief and complex service delivery: Towards better outcomes. Queensland University of Technology, June 2021. http://dx.doi.org/10.5204/rep.eprints.211133.

Full text
Abstract:
Emergency Relief (ER) is a Department of Social Services (DSS) funded program, delivered by 197 community organisations (ER Providers) across Australia, to assist people facing a financial crisis with financial/material aid and referrals to other support programs. ER has been playing this important role in Australian communities since 1979. Without ER, more people living in Australia who experience a financial crisis might face further harm such as crippling debt or homelessness. The Emergency Relief National Coordination Group (NCG) was established in April 2020 at the start of the COVID-19 pandemic to advise the Minister for Families and Social Services on the implementation of ER. To inform its advice to the Minister, the NCG partnered with the Institute for Governance at the University of Canberra to conduct research to understand the issues and challenges faced by ER Providers and Service Users in local contexts across Australia. The research involved a desktop review of the existing literature on ER service provision, a large survey which all Commonwealth ER Providers were invited to participate in (and 122 responses were received), interviews with a purposive sample of 18 ER Providers, and the development of a program logic and theory of change for the Commonwealth ER program to assess progress. The surveys and interviews focussed on ER Provider perceptions of the strengths, weaknesses, future challenges, and areas of improvement for current ER provision. The trend of increasing case complexity, the effectiveness of ER service delivery models in achieving outcomes for Service Users, and the significance of volunteering in the sector were investigated. Separately, an evaluation of the performance of the NCG was conducted and a summary of the evaluation is provided as an appendix to this report. Several themes emerged from the review of the existing literature such as service delivery shortcomings in dealing with case complexity, the effectiveness of case management, and repeat requests for service. Interviews with ER workers and Service Users found that an uplift in workforce capability was required to deal with increasing case complexity, leading to recommendations for more training and service standards. Several service evaluations found that ER delivered with case management led to high Service User satisfaction, played an integral role in transforming the lives of people with complex needs, and lowered repeat requests for service. A large longitudinal quantitative study revealed that more time spent with participants substantially decreased the number of repeat requests for service; and, given that repeat requests for service can be an indicator of entrenched poverty, not accessing further services is likely to suggest improvement. The interviews identified the main strengths of ER to be the rapid response and flexible use of funds to stabilise crisis situations and connect people to other supports through strong local networks. Service Users trusted the system because of these strengths, and ER was often an access point to holistic support. There were three main weaknesses identified. First, funding contracts were too short and did not cover the full costs of the program—in particular, case management for complex cases. Second, many Service Users were dependent on ER which was inconsistent with the definition and intent of the program. Third, there was inconsistency in the level of service received by Service Users in different geographic locations. These weaknesses can be improved upon with a joined-up approach featuring co-design and collaborative governance, leading to the successful commissioning of social services. The survey confirmed that volunteers were significant for ER, making up 92% of all workers and 51% of all hours worked in respondent ER programs. Of the 122 respondents, volunteers amounted to 554 full-time equivalents, a contribution valued at $39.4 million. In total there were 8,316 volunteers working in the 122 respondent ER programs. The sector can support and upskill these volunteers (and employees in addition) by developing scalable training solutions such as online training modules, updating ER service standards, and engaging in collaborative learning arrangements where large and small ER Providers share resources. More engagement with peak bodies such as Volunteering Australia might also assist the sector to improve the focus on volunteer engagement. Integrated services achieve better outcomes for complex ER cases—97% of survey respondents either agreed or strongly agreed this was the case. The research identified the dimensions of service integration most relevant to ER Providers to be case management, referrals, the breadth of services offered internally, co-location with interrelated service providers, an established network of support, workforce capability, and Service User engagement. Providers can individually focus on increasing the level of service integration for their ER program to improve their ability to deal with complex cases, which are clearly on the rise. At the system level, a more joined-up approach can also improve service integration across Australia. The key dimensions of this finding are discussed next in more detail. Case management is key for achieving Service User outcomes for complex cases—89% of survey respondents either agreed or strongly agreed this was the case. Interviewees most frequently said they would provide more case management if they could change their service model. Case management allows for more time spent with the Service User, follow up with referral partners, and a higher level of expertise in service delivery to support complex cases. Of course, it is a costly model and not currently funded for all Service Users through ER. Where case management is not available as part of ER, it might be available through a related service that is part of a network of support. Where possible, ER Providers should facilitate access to case management for Service Users who would benefit. At a system level, ER models with a greater component of case management could be implemented as test cases. Referral systems are also key for achieving Service User outcomes, which is reflected in the ER Program Logic presented on page 31. The survey and interview data show that referrals within an integrated service (internal) or in a service hub (co-located) are most effective. Where this is not possible, warm referrals within a trusted network of support are more effective than cold referrals leading to higher take-up and beneficial Service User outcomes. However, cold referrals are most common, pointing to a weakness in ER referral systems. This is because ER Providers do not operate or co-locate with interrelated services in many cases, nor do they have the case management capacity to provide warm referrals in many other cases. For mental illness support, which interviewees identified as one of the most difficult issues to deal with, ER Providers offer an integrated service only 23% of the time, warm referrals 34% of the time, and cold referrals 43% of the time. A focus on referral systems at the individual ER Provider level, and system level through a joined-up approach, might lead to better outcomes for Service Users. The program logic and theory of change for ER have been documented with input from the research findings and included in Section 4.3 on page 31. These show that ER helps people facing a financial crisis to meet their immediate needs, avoid further harm, and access a path to recovery. The research demonstrates that ER is fundamental to supporting vulnerable people in Australia and should therefore continue to be funded by government.
APA, Harvard, Vancouver, ISO, and other styles
2

African Open Science Platform Part 1: Landscape Study. Academy of Science of South Africa (ASSAf), 2019. http://dx.doi.org/10.17159/assaf.2019/0047.

Full text
Abstract:
This report maps the African landscape of Open Science – with a focus on Open Data as a sub-set of Open Science. Data to inform the landscape study were collected through a variety of methods, including surveys, desk research, engagement with a community of practice, networking with stakeholders, participation in conferences, case study presentations, and workshops hosted. Although the majority of African countries (35 of 54) demonstrates commitment to science through its investment in research and development (R&D), academies of science, ministries of science and technology, policies, recognition of research, and participation in the Science Granting Councils Initiative (SGCI), the following countries demonstrate the highest commitment and political willingness to invest in science: Botswana, Ethiopia, Kenya, Senegal, South Africa, Tanzania, and Uganda. In addition to existing policies in Science, Technology and Innovation (STI), the following countries have made progress towards Open Data policies: Botswana, Kenya, Madagascar, Mauritius, South Africa and Uganda. Only two African countries (Kenya and South Africa) at this stage contribute 0.8% of its GDP (Gross Domestic Product) to R&D (Research and Development), which is the closest to the AU’s (African Union’s) suggested 1%. Countries such as Lesotho and Madagascar ranked as 0%, while the R&D expenditure for 24 African countries is unknown. In addition to this, science globally has become fully dependent on stable ICT (Information and Communication Technologies) infrastructure, which includes connectivity/bandwidth, high performance computing facilities and data services. This is especially applicable since countries globally are finding themselves in the midst of the 4th Industrial Revolution (4IR), which is not only “about” data, but which “is” data. According to an article1 by Alan Marcus (2015) (Senior Director, Head of Information Technology and Telecommunications Industries, World Economic Forum), “At its core, data represents a post-industrial opportunity. Its uses have unprecedented complexity, velocity and global reach. As digital communications become ubiquitous, data will rule in a world where nearly everyone and everything is connected in real time. That will require a highly reliable, secure and available infrastructure at its core, and innovation at the edge.” Every industry is affected as part of this revolution – also science. An important component of the digital transformation is “trust” – people must be able to trust that governments and all other industries (including the science sector), adequately handle and protect their data. This requires accountability on a global level, and digital industries must embrace the change and go for a higher standard of protection. “This will reassure consumers and citizens, benefitting the whole digital economy”, says Marcus. A stable and secure information and communication technologies (ICT) infrastructure – currently provided by the National Research and Education Networks (NRENs) – is key to advance collaboration in science. The AfricaConnect2 project (AfricaConnect (2012–2014) and AfricaConnect2 (2016–2018)) through establishing connectivity between National Research and Education Networks (NRENs), is planning to roll out AfricaConnect3 by the end of 2019. The concern however is that selected African governments (with the exception of a few countries such as South Africa, Mozambique, Ethiopia and others) have low awareness of the impact the Internet has today on all societal levels, how much ICT (and the 4th Industrial Revolution) have affected research, and the added value an NREN can bring to higher education and research in addressing the respective needs, which is far more complex than simply providing connectivity. Apart from more commitment and investment in R&D, African governments – to become and remain part of the 4th Industrial Revolution – have no option other than to acknowledge and commit to the role NRENs play in advancing science towards addressing the SDG (Sustainable Development Goals). For successful collaboration and direction, it is fundamental that policies within one country are aligned with one another. Alignment on continental level is crucial for the future Pan-African African Open Science Platform to be successful. Both the HIPSSA ((Harmonization of ICT Policies in Sub-Saharan Africa)3 project and WATRA (the West Africa Telecommunications Regulators Assembly)4, have made progress towards the regulation of the telecom sector, and in particular of bottlenecks which curb the development of competition among ISPs. A study under HIPSSA identified potential bottlenecks in access at an affordable price to the international capacity of submarine cables and suggested means and tools used by regulators to remedy them. Work on the recommended measures and making them operational continues in collaboration with WATRA. In addition to sufficient bandwidth and connectivity, high-performance computing facilities and services in support of data sharing are also required. The South African National Integrated Cyberinfrastructure System5 (NICIS) has made great progress in planning and setting up a cyberinfrastructure ecosystem in support of collaborative science and data sharing. The regional Southern African Development Community6 (SADC) Cyber-infrastructure Framework provides a valuable roadmap towards high-speed Internet, developing human capacity and skills in ICT technologies, high- performance computing and more. The following countries have been identified as having high-performance computing facilities, some as a result of the Square Kilometre Array7 (SKA) partnership: Botswana, Ghana, Kenya, Madagascar, Mozambique, Mauritius, Namibia, South Africa, Tunisia, and Zambia. More and more NRENs – especially the Level 6 NRENs 8 (Algeria, Egypt, Kenya, South Africa, and recently Zambia) – are exploring offering additional services; also in support of data sharing and transfer. The following NRENs already allow for running data-intensive applications and sharing of high-end computing assets, bio-modelling and computation on high-performance/ supercomputers: KENET (Kenya), TENET (South Africa), RENU (Uganda), ZAMREN (Zambia), EUN (Egypt) and ARN (Algeria). Fifteen higher education training institutions from eight African countries (Botswana, Benin, Kenya, Nigeria, Rwanda, South Africa, Sudan, and Tanzania) have been identified as offering formal courses on data science. In addition to formal degrees, a number of international short courses have been developed and free international online courses are also available as an option to build capacity and integrate as part of curricula. The small number of higher education or research intensive institutions offering data science is however insufficient, and there is a desperate need for more training in data science. The CODATA-RDA Schools of Research Data Science aim at addressing the continental need for foundational data skills across all disciplines, along with training conducted by The Carpentries 9 programme (specifically Data Carpentry 10 ). Thus far, CODATA-RDA schools in collaboration with AOSP, integrating content from Data Carpentry, were presented in Rwanda (in 2018), and during17-29 June 2019, in Ethiopia. Awareness regarding Open Science (including Open Data) is evident through the 12 Open Science-related Open Access/Open Data/Open Science declarations and agreements endorsed or signed by African governments; 200 Open Access journals from Africa registered on the Directory of Open Access Journals (DOAJ); 174 Open Access institutional research repositories registered on openDOAR (Directory of Open Access Repositories); 33 Open Access/Open Science policies registered on ROARMAP (Registry of Open Access Repository Mandates and Policies); 24 data repositories registered with the Registry of Data Repositories (re3data.org) (although the pilot project identified 66 research data repositories); and one data repository assigned the CoreTrustSeal. Although this is a start, far more needs to be done to align African data curation and research practices with global standards. Funding to conduct research remains a challenge. African researchers mostly fund their own research, and there are little incentives for them to make their research and accompanying data sets openly accessible. Funding and peer recognition, along with an enabling research environment conducive for research, are regarded as major incentives. The landscape report concludes with a number of concerns towards sharing research data openly, as well as challenges in terms of Open Data policy, ICT infrastructure supportive of data sharing, capacity building, lack of skills, and the need for incentives. Although great progress has been made in terms of Open Science and Open Data practices, more awareness needs to be created and further advocacy efforts are required for buy-in from African governments. A federated African Open Science Platform (AOSP) will not only encourage more collaboration among researchers in addressing the SDGs, but it will also benefit the many stakeholders identified as part of the pilot phase. The time is now, for governments in Africa, to acknowledge the important role of science in general, but specifically Open Science and Open Data, through developing and aligning the relevant policies, investing in an ICT infrastructure conducive for data sharing through committing funding to making NRENs financially sustainable, incentivising open research practices by scientists, and creating opportunities for more scientists and stakeholders across all disciplines to be trained in data management.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography