Dissertations / Theses on the topic 'Distance-based measure'

To see the other types of publications on this topic, follow the link: Distance-based measure.

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 16 dissertations / theses for your research on the topic 'Distance-based measure.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Rogers, Wendy Laurel. "A Mahalanobis-distance-based image segmentation error measure with applications in automated microscopy /." Thesis, McGill University, 1985. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=66025.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Nordström, Markus. "Automatic Source Code Classification : Classifying Source Code for a Case-Based Reasoning System." Thesis, Mittuniversitetet, Avdelningen för informations- och kommunikationssystem, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:miun:diva-25519.

Full text
Abstract:
This work has investigated the possibility of classifying Java source code into cases for a case-based reasoning system. A Case-Based Reasoning system is a problem solving method in Artificial Intelligence that uses knowledge of previously solved problems to solve new problems. A case in case-based reasoning consists of two parts: the problem part and solution part. The problem part describes a problem that needs to be solved and the solution part describes how this problem was solved. In this work, the problem is described as a Java source file using words that describes the content in the source file and the solution is a classification of the source file along with the source code. To classify Java source code, a classification system was developed. It consists of four analyzers: type filter, documentation analyzer, syntactic analyzer and semantic analyzer. The type filter determines if a Java source file contains a class or interface. The documentation analyzer determines the level of documentation in asource file to see the usefulness of a file. The syntactic analyzer extracts statistics from the source code to be used for similarity, and the semantic analyzer extracts semantics from the source code. The finished classification system is formed as a kd-tree, where the leaf nodes contains the classified source files i.e. the cases. Furthermore, a vocabulary was developed to contain the domain knowledge about the Java language. The resulting kd-tree was found to be imbalanced when tested, as the majority of source files analyzed were placed inthe left-most leaf nodes. The conclusion from this was that using documentation as a part of the classification made the tree imbalanced and thus another way has to be found. This is due to the fact that source code is not documented to such an extent that it would be useful for this purpose.
APA, Harvard, Vancouver, ISO, and other styles
3

Gashayija, Jean Marie. "Image classification, storage and retrieval system for a 3 u cubesat." Thesis, Cape Peninsula University of Technology, 2014. http://hdl.handle.net/20.500.11838/1189.

Full text
Abstract:
Thesis submitted in fulfillment of the requirements for the degree Master of Technology: Electrical Engineering in the Faculty of Engineering at the Cape Peninsula University of Technology
Small satellites, such as CubeSats are mainly utilized for space and earth imaging missions. Imaging CubeSats are equipped with high resolution cameras for the capturing of digital images, as well as mass storage devices for storing the images. The captured images are transmitted to the ground station and subsequently stored in a database. The main problem with stored images in a large image database, identified by researchers and developers within the last number of years, is the retrieval of precise, clear images and overcoming the semantic gap. The semantic gap relates to the lack of correlation between the semantic categories the user requires and the low level features that a content-based image retrieval system offers. Clear images are needed to be usable for applications such as mapping, disaster monitoring and town planning. The main objective of this thesis is the design and development of an image classification, storage and retrieval system for a CubeSat. This system enables efficient classification, storing and retrieval of images that are received on a daily basis from an in-orbit CubeSat. In order to propose such a system, a specific research methodology was chosen and adopted. This entails extensive literature reviews on image classification techniques and image feature extraction techniques, to extract content embedded within an image, and include studies on image database systems, data mining techniques and image retrieval techniques. The literature study led to a requirement analysis followed by the analyses of software development models in order to design the system. The proposed design entails classifying images using content embedded in the image and also extracting image metadata such as date and time. Specific features extraction techniques are needed to extract required content and metadata. In order to achieve extraction of information embedded in the image, colour feature (colour histogram), shape feature (Mathematical Morphology) and texture feature (GLCM) techniques were used. Other major contributions of this project include a graphical user interface which enables users to search for similar images against those stored in the database. An automatic image extractor algorithm was also designed to classify images according to date and time, and colour, texture and shape features extractor techniques were proposed. These ensured that when a user wishes to query the database, the shape objects, colour quantities and contrast contained in an image are extracted and compared to those stored in the database. Implementation and test results concluded that the designed system is able to categorize images automatically and at the same time provide efficient and accurate results. The features extracted for each image depend on colour, shape and texture methods. Optimal values were also incorporated in order to reduce retrieval times. The mathematical morphological technique was used to compute shape objects using erosion and dilation operators, and the co-occurrence matrix was used to compute the texture feature of the image.
APA, Harvard, Vancouver, ISO, and other styles
4

Suñé, Socias Víctor Manuel. "Failure distance based bounds of dependability measures." Doctoral thesis, Universitat Politècnica de Catalunya, 2000. http://hdl.handle.net/10803/6375.

Full text
Abstract:
El tema d'aquesta tesi és el desenvolupament de mètodes de fitació per a una classe de models de confiabilitat basats en cadenes de Markov de temps continu (CMTC) de sistemes tolerants a fallades.

Els sistemes considerats a la tesi es conceptualitzen com formats per components (hardware o software) que fallen i, en el cas de sistemes reparables, són reparats. Els components s'agrupen en classes de forma que els components d'una mateixa classe són indistingibles. Per tant, un component és considerat com a una instància d'una classe de components i el sistema inclou un bag de classes de components definit sobre un cert domini. L'estat no fallada/fallada del sistema es determina a partir de l'estat no fallada/fallada dels components mitjançant una funció d'estructura coherent que s'especifica amb un arbre de fallades amb classes d'esdeveniments bàsics. (Una classe d'esdeveniment bàsic és la fallada d'un component d'una classe de components.)

La classe de models basats en CMTC considerada a la tesi és força àmplia i permet, per exemple, de modelar el fet que un component pot tenir diversos modes de fallada. També permet de modelar fallades de cobertura mitjançant la introducció de components ficticis que no fallen per ells mateixos i als quals es propaguen les fallades d'altres components. En el cas de sistemes reparables, la classe de models considerada admet polítiques de reparació complexes (per exemple, nombre limitat de reparadors, prioritats, inhibició de reparació) així com reparació en grup (reparació simultània de diversos components). Tanmateix, no és possible de modelar la reparació diferida (és a dir, el fet de diferir la reparació d'un component fins que una certa condició es compleixi).

A la tesi es consideren dues mesures de confiabilitat: la no fiabilitat en un instant de temps donat en el cas de sistemes no reparables i la no disponibilitat en règim estacionari en el cas sistemes reparables.

Els mètodes de fitació desenvolupats a la tesi es basen en el concepte de "distància a la fallada", que es defineix com el nombre mínim de components que han de fallar a més dels que ja han fallat per fer que el sistema falli.

A la tesi es desenvolupen quatre mètodes de fitació. El primer mètode dóna fites per a la no fiabilitat de sistemes no reparables emprant distàncies a la fallada exactes. Aquestes distàncies es calculen usant el conjunt de talls mínims de la funció d'estructura del sistema. El conjunt de talls mínims s'obté amb un algorisme desenvolupat a la tesi que obté els talls mínims per a arbres de fallades amb classes d'esdeveniments bàsics. El segon mètode dóna fites per a la no fiabilitat usant fites inferiors per a les distàncies a la fallada. Aquestes fites inferiors s'obtenen analitzant l'arbre de fallades del sistema, no requereixen de conèixer el conjunt de talls mínims i el seu càlcul és poc costós. El tercer mètode dóna fites per a la no disponibilitat en règim estacionari de sistemes reparables emprant distàncies a la fallada exactes. El quart mètode dóna fites per a la no disponibilitat en règim estacionari emprant les fites inferiors per a les distàncies a la fallada.

Finalment, s'il·lustren les prestacions de cada mètode usant diversos exemples. La conclusió és que cada un dels mètodes pot funcionar molt millor que altres mètodes prèviament existents i estendre de forma significativa la complexitat de sistemes tolerants a fallades per als quals és possible de calcular fites ajustades per a la no fiabilitat o la no disponibilitat en règim estacionari.
The subject of this dissertation is the development of bounding methods for a class of continuous-time Markov chain (CTMC) dependability models of fault-tolerant systems.

The systems considered in the dissertation are conceptualized as made up of components (hardware or software) that fail and, for repairable systems, are repaired. Components are grouped into classes, the components of the same class being indistinguishable. Thus, a component is regarded as an instance of some component class and the system includes a bag of component classes defined over a certain domain. The up/down state of the system is determined from the unfailed/failed state of the components through a coherent structure function specified by a fault tree with basic event classes. (A basic event class is the failure of a component of a component class.)

The class of CTMC models considered in the dissertation is quite wide and allows, for instance, to model the fact that a component may have different failure modes. It also allows to model coverage failures by means of introducing fictitious components that do not fail by themselves and to which uncovered failures of other components are propagated. In the case of repairable systems, the considered class of models supports very complex repair policies (e.g., limited repairpersons, priorities, repair preemption) as well as group repair (i.e., simultaneous repair of several components). However, deferred repair (i.e., the deferring of repair until some condition is met) is not allowed.

Two dependability measures are considered in the dissertation: the unreliability at a given time epoch for non-repairable systems and the steady-state unavailability for repairable systems.

The bounding methods developed in the dissertation are based on the concept of "failure distance from a state," which is defined as the minimum number of components that have to fail in addition to those already failed to take the system down.

We develop four bounding methods. The first method gives bounds for the unreliability of non-repairable fault-tolerant systems using (exact) failure distances. Those distances are computed using the set of minimal cuts of the structure function of the system. The set of minimal cuts is obtained using an algorithm developed in the dissertation that obtains the minimal cuts for fault trees with basic event classes. The second method gives bounds for the unreliability using easily computable lower bounds for failure distances. Those lower bounds are obtained analyzing the fault tree of the system and do not require the knowledge of the set of minimal cuts. The third method gives bounds for the steady-state unavailability using (exact) failure distances. The fourth method gives bounds for the steady-state unavailability using the lower bounds for failure distances.

Finally, the performance of each method is illustrated by means of several large examples. We conclude that the methods can outperform significantly previously existing methods and extend significantly the complexity of the fault-tolerant systems for which tight bounds for the unreliability or steady-state unavailability can be computed.
APA, Harvard, Vancouver, ISO, and other styles
5

Goussakov, Roma. "Hellinger Distance-based Similarity Measures for Recommender Systems." Thesis, Umeå universitet, Statistik, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-172385.

Full text
Abstract:
Recommender systems are used in online sales and e-commerce for recommending potential items/products for customers to buy based on their previous buying preferences and related behaviours. Collaborative filtering is a popular computational technique that has been used worldwide for such personalized recommendations. Among two forms of collaborative filtering, neighbourhood and model-based, the neighbourhood-based collaborative filtering is more popular yet relatively simple. It relies on the concept that a certain item might be of interest to a given customer (active user) if, either he appreciated similar items in the buying space, or if the item is appreciated by similar users (neighbours). To implement this concept different kinds of similarity measures are used. This thesis is set to compare different user-based similarity measures along with defining meaningful measures based on Hellinger distance that is a metric in the space of probability distributions. Data from a popular database MovieLens will be used to show the effectiveness of dierent Hellinger distance-based measures compared to other popular measures such as Pearson correlation (PC), cosine similarity, constrained PC and JMSD. The performance of dierent similarity measures will then be evaluated with the help of mean absolute error, root mean squared error and F-score. From the results, no evidence were found to claim that Hellinger distance-based measures performed better than more popular similarity measures for the given dataset.
APA, Harvard, Vancouver, ISO, and other styles
6

Dey, Rajarshi. "Inference for the K-sample problem based on precedence probabilities." Diss., Kansas State University, 2011. http://hdl.handle.net/2097/12000.

Full text
Abstract:
Doctor of Philosophy
Department of Statistics
Paul I. Nelson
Rank based inference using independent random samples to compare K>1 continuous distributions, called the K-sample problem, based on precedence probabilities is developed and explored. There are many parametric and nonparametric approaches, most dealing with hypothesis testing, to this important, classical problem. Most existing tests are designed to detect differences among the location parameters of different distributions. Best known and most widely used of these is the F- test, which assumes normality. A comparable nonparametric test was developed by Kruskal and Wallis (1952). When dealing with location-scale families of distributions, both of these tests can perform poorly if the differences among the distributions are among their scale parameters and not in their location parameters. Overall, existing tests are not effective in detecting changes in both location and scale. In this dissertation, I propose a new class of rank-based, asymptotically distribution- free tests that are effective in detecting changes in both location and scale based on precedence probabilities. Let X_{i} be a random variable with distribution function F_{i} ; Also, let _pi_ be the set of all permutations of the numbers (1,2,...,K) . Then P(X_{i_{1}}<...
APA, Harvard, Vancouver, ISO, and other styles
7

Sooful, Jayren Jugpal. "Automated phoneme mapping for cross-language speech recognition." Diss., Pretoria [s.n.], 2004. http://upetd.up.ac.za/thesis/available/etd-01112005-131128.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Lin, Chia-Wen, and 林嘉文. "A Novel Content-Based Image Retrieval System Based on Distance Measure Approach." Thesis, 2013. http://ndltd.ncl.edu.tw/handle/4m2xy9.

Full text
Abstract:
碩士
國立臺中科技大學
資訊工程系碩士班
101
For the last two dacades, content-based image retrieval (CBIR) is a very popular topic in image processing area. Researchers aim to extract features of color, texture, shape or position information from an image and then via a distance measure method, they retrieve images most similar to the query image from a particular image database. A lot of content-based image retrieval methods have been proposed and the performance of each mothod has been demonstrated in each proposed paper. In this paper, we proposed a novel distance meature approach (re-rank,shorted as RRK) for CBIR system. First, we get the retrieved images from our simple CBIR system. Then we used these images to set reference feature. Finally, we retrieved again using the referenced feature and obtained more precise result. According to the experimental results, the proposed method is simple and efficient. Furthermore, the proposed method can be applied in other CBIR system easily.
APA, Harvard, Vancouver, ISO, and other styles
9

Chiang, Cheng-Yuan, and 江正元. "Speaker Recognition with Independent Corpus Based on RM Distance Measure." Thesis, 2016. http://ndltd.ncl.edu.tw/handle/9zur32.

Full text
Abstract:
碩士
淡江大學
電機工程學系碩士班
104
The speaker recognition is always a hot topic in the research field. Technologies of speaker recognition under white and color noisy environments have been proposed in recent years. Sparse representation algorithm has been introduced into noise filtering for improving the assessments of speech quality, such as SNR, SNRseg, LLR and PESQ, but the cost time is lengthy. So we employ Label Consistent K-SVD sparse coding (LC-KSVD) to de-noise speech data and decrease processing time. Speaker recognition systems almost use Euclidean distance to compute the distance between features, currently. Our goal is to have short corpus and independent corpus, which makes it more difficult to achieve high recognition accuracy. We propose Riemannian distance replace Euclidean distance, but our experimental results show that Euclidean distance is superior than Riemannian distance. We use waveform, MFCC and MFCC smoothing spectrum with RD and ED for speaker recognition experiment in this paper.
APA, Harvard, Vancouver, ISO, and other styles
10

Liu, Kuan-Liang, and 劉冠良. "A Subset gene selection method based on clustering analysis and distance measure." Thesis, 2007. http://ndltd.ncl.edu.tw/handle/40328330647012774931.

Full text
Abstract:
碩士
國立成功大學
資訊管理研究所
95
After the Human Genome Project, the next challenge for bio-researchers is to understand the meanings of genes and the inter-relationship between them. As the technique of gene expression microarray stores all the gene expression data in a tiny chip, researchers become able to analyze all expression data of genes simultaneously. Nevertheless, compared to the original statistic data, the huge dimensionality and comparatively few sample amounts of gene expression data are still research obstacles. The objective of this research is to screen a representative set of genes according to a specific problem. Although many gene selection methods have been proposed in recent years, problems, such as gene collinearity, lack of consideration for combination genes, and work complexity, are not thoroughly examined and worked out. The gene selection algorithm of this research is tailored to the problems mentioned above. We first distribute the whole data set of genes using density-based clustering technique and screen out genes that are similar and have comparatively lower individual gene rank values. Then we select and substitute combinative genes according to examination of distance measure value, , and cluster similarity index. Considering the characteristic of gene expression data, we introduce relation-based methods and measure similarity between genes. Coupled with the data of tumor classification, the algorithm proposed in this research is tested and the accuracy rate of classification was improved. The gene set of enhanced can really get a higher accuracy rate of classification. In addition, the accuracy rates of gene sets from our selection algorithm are better than the gene sets from individual gene ranking methods.
APA, Harvard, Vancouver, ISO, and other styles
11

Meghdadi, Amir Hossein. "Fuzzy Tolerance Neighborhood Approach to Image Similarity in Content-based Image Retrieval." 2012. http://hdl.handle.net/1993/8094.

Full text
Abstract:
The main contribution of this thesis, is to define similarity measures between two images with the main focus on content-based image retrieval (CBIR). Each image is considered as a set of visual elements that can be described with a set of visual descriptions (features). The similarity between images is then defined as the nearness between sets of elements based on a tolerance and a fuzzy tolerance relation. A tolerance relation is used to describe the approximate nature of the visual perception. A fuzzy tolerance relation is adopted to eliminate the need for a sharp threshold and hence model the gradual changes in perception of similarities. Three real valued similarity measures as well as a fuzzy valued similarity measure are proposed. All of the methods are then used in two CBIR experiments and the results are compared with classical measures of distance (namely, Kantorovich, Hausdorff and Mahalanobis). The results are compared with other published research papers. An important advantage of the proposed methods is shown to be their effectiveness in an unsupervised setting with no prior information. Eighteen different features (based on color, texture and edge) are used in all the experiments. A feature selection algorithm is also used to train the system in choosing a suboptimal set of visual features.
APA, Harvard, Vancouver, ISO, and other styles
12

Mukhopadhyay, Sayan. "Variance of Difference as Distance Like Measure in Time Series Microarray Data Clustering." Thesis, 2014. http://etd.iisc.ernet.in/handle/2005/2986.

Full text
Abstract:
Our intention is to find similarity among the time series expressions of the genes in microarray experiments. It is hypothesized that at a given time point the concentration of one gene’s mRNA is directly affected by the concentration of other gene’s mRNA, and may have biological significance. We define dissimilarity between two time-series data set as the variance of Euclidean distances of each time points. The large numbers of gene expressions make the calculation of variance of distance in each point computationally expensive and therefore computationally challenging in terms of execution time. For this reason we use autoregressive model which estimates nineteen points gene expression to a three point vector. It allows us to find variance of difference between two data sets without point-to-point matching. Previous analysis from the microarray experiments data found that 62 genes are regulated following EGF (Epidermal Growth Factor) and HRG (Heregulin) treatment of the MCF-7 breast cancer cells. We have chosen these suspected cancer-related genes as our reference and investigated which additional set of genes has similar time point expression profiles. Keeping variance of difference as a measure of distance, we have used several methods for clustering the gene expression data, such as our own maximum clique finding heuristics and hierarchical clustering. The results obtained were validated through a text mining study. New predictions from our study could be a basis for further investigations in the genesis of breast cancer. Overall in 84 new genes are found in which 57 genes are related to cancer among them 35 genes are associated with breast cancer.
APA, Harvard, Vancouver, ISO, and other styles
13

Ghobadi, Razieh. "Geostatistical modelling and simulation of karst systems." Thesis, 2016. http://hdl.handle.net/2440/103460.

Full text
Abstract:
Groundwater is a significant water resource and in many parts of the world it occurs in karst aquifers. The modelling of karst systems is a critical component of groundwater resource assessment and flow. Geostatistical techniques have shown useful applications in the area of groundwater research because of their ability to quantify spatial variability, uncertainty and risk. Traditional geostatistical methods, based on variogram models, use only two-point statistics and thus are not capable of modelling the complex and, high-connectivity structures of karst networks. This has led to an increasing focus on spatial multiple-point statistics (MPS) to model these complex systems. In this approach, a training image is used instead of a variogram. Patterns are obtained by scanning and sampling the training image and during the simulation they are reproduced using MPS. There are two implementations of MPS: (i) gridded and (ii) non-gridded. In gridded MPS, the training image, templates and simulations are based on rigid grids, whereas the spatially flexible non-gridded approach does not depend on rigidly specified grids. The non-gridded approach is relatively new (Erzeybek Balan 2012), and applications, especially in hydrogeology are few; however, the method has been used to simulate paleokarsts in petroleum applications. Non-gridded MPS has potential to improve the modelling of karst systems by replacing the fixed gridding procedure, used in the original form of MPS, by a more flexible grid adapted to each specific application. However, there are some weaknesses in the non-gridded approach reported in the literature. For example, the proposed template cannot properly represent the tortuous nature of a network, and the variation of the passage widths is not taken into account. In the case of a simple channelised system with a constant width, sampling the central line of the passages is sufficient; however, most karst systems have networks with significantly varying widths. In addition, the variability among the realisations generated by non-gridded MPS is relatively small, indicating that the realisations do not cover the full space of uncertainty. In practical applications, it is not possible to know the exact extent of the full space of uncertainty, but the observed variability of the geology and geomorphology of similar structures would tell us when the variability among the simulations is too small (or too large). A lack of significant variability among simulated realisations makes the method inapplicable. This thesis presents a modified non-gridded MPS method that increases the variability among realisations and adequately captures the tortuosities of karst networks. To do this, it includes the width and constructs an optimal template based on a representative variety of directions adapted to each network instead of considering only a few major directions using a generic template as applied by Erzeybek Balan (2012). The performance of Erzeybek Balan’s (2012) non-gridded MPS method has only been visually demonstrated, which is not a sufficiently robust measure of performance. In this thesis, a systematic measure is developed to evaluate the variability among the realisations. This provides an objective way of comparing an important feature of the simulations generated by gridded MPS and the proposed modified non-gridded MPS. The research starts with an investigation and modification of non-gridded MPS. A widely used demonstration image, which is based on a channelised system, is used to compare the performances of the original non-gridded MPS (Erzeybek Balan 2012) and the modified version proposed in this thesis. A distance-based measure is used to evaluate and compare pattern reproduction and the variability of the realisations generated by the modified non-gridded MPS and standard gridded MPS methods. This distance measure can be used to compare the multiple-point histograms of the realisations and training images. Gridded MPS and modified non-gridded MPS are then applied to two different karst systems—Olwolgin Cave and Tank Cave—and the realisations generated by each method are evaluated in terms of pattern reproduction and the extent of the uncertainty space. The comparison examples demonstrate that the proposed modified non-gridded MPS generates a larger uncertainty space than that generated by gridded MPS. The results also confirm that modified non-gridded MPS performs significantly better than the original version of non-gridded MPS in terms of a larger (and more realistic) space of uncertainty and pattern reproduction when applied to a complex karst system.
Thesis (M.Phil.) -- University of Adelaide, School of Civil, Environmental and Mining Engineering, 2016.
APA, Harvard, Vancouver, ISO, and other styles
14

Jabłoński, Jędrzej. "Structured population models for predator-prey interactions. The case of Daphnia and size selective planktivorous fish." Doctoral thesis, 2014.

Find full text
Abstract:
In this thesis a model of the dynamics of size-structured population subject to selective predation is built and analyzed. The study is motivated by biological phenomena concerning limnology and oceanography, and in particular diversity of first consumers in aquatic ecosystems. An individual-based model of size-selective visual predator-harvester based on the concept of optimal foraging is proposed. Farther, a simplification of the model, described in terms of operators on the space of measures, is derived based on Holling II-type functional response to eliminate inherent difficulties of individual-based approach. The results are compared against experimental evidence. Considerations involving populations dynamics, namely growth, birth and mortality, are examined in the framework of measure-valued solutions to transport equation and various distances arising from optimal transportation theory. To this end, efficient algorithms for solving transportation problem on a real line are found and finally, numerical schemes based on particle methods for structured population models are improved. Moreover, approximation theory for Radon measures is developed.
APA, Harvard, Vancouver, ISO, and other styles
15

Sodabanlu, Sirawit. "A Likelihood Ratio Based Forensic Text Comparison with Multiple Types of Features." Master's thesis, 2021. http://hdl.handle.net/1885/262301.

Full text
Abstract:
This study aims at further improving forensic text comparison (FTC) under the likelihood ratio (LR) framework. While the use of the LR framework to conclude the strength of evidence is well recognised in forensic science, studies on forensic text evidence within the LR framework are limited, and this study is an attempt of alleviating this situation. There have already been initiatives to obtain LRs for textual evidence by adopting various approaches and using different sets of stylometric features. (Carne & Ishihara, 2020; Ishihara, 2014, 2017a, 2017b, 2021). However, only few features have been tested in the similarity-only score-based approach (Ishihara, 2021), and there are many features left to be further investigated. To achieve the aim of the study, we will investigate some of the features in LR-based FTC and demonstrate how they contribute to the further improvement of the LR-based FTC system. Statistic, word n-gram (n=1,2,3), character n-gram (n=1,2,3,4), and part of speech (POS) n-gram (n=1,2,3) features were separately tested first in this study, and then the separately estimated LRs were fused for overall LRs. The databased used was prepared by Ishihara (2021), and the documents of comparison were modelled into feature vectors using a bag-of-words model. Two groups of documents, which both contained documents of 700, 1,400, and 2,100 words, were concatenated for each author, resulting in the total of 719 same-author comparisons and 516,242 different-author comparisons. The Cosine similarity was used to measure the similarity of texts, and the similarity-only score-based approach was used to estimate the LRs from the scores of similarity (Helper et al., 2012; Bolck et al., 2015). Log-likelihood ratio cost (Cllr) and their composites—Cllrmin and Cllrcal—were used as assessment metrics. Findings indicate that (a) when the LRs of all the feature types are fused, the fused Cllr values are 0.56, 0.30, and 0.19 for 700, 1,400, and 2,100 words, respectively, and (b) feature selection depending on the nature of an FTC task matters to the performance of the FTC system and can contribute to the improvement of LR-based FTC.
APA, Harvard, Vancouver, ISO, and other styles
16

Segoe, Bobo Aaron. "Learner support in the provision of distance teaching programmes for under qualified teachers." Thesis, 2012. http://hdl.handle.net/10500/8559.

Full text
Abstract:
Presently all over the world, there is a great concern among teachers, parents, organisations, community leaders and higher education lecturers about the problems that beset teaching and learning particularly in the teaching programmes, for example, for under-qualified teachers studying at a distance. Most distance education (DE) programmes are concerned with education of adults and it seems fairly obvious that the research plans should be informed by the theories and research about learning in higher education institutions. In terms of teacher education, such studies could, inter alia, focus on teacher development, curriculum planning, learner support programmes, communication and evaluation of DE. DE itself refers to a field of education that focuses on teaching methods and technology with the aim of delivering teaching, often on an individual basis, to learners who are not physically present in a traditional educational setting such as a classroom. Distance learning is becoming an increasingly popular way of studying, and most universities now provide courses using this mode of teaching and learning. Today’s learners, though, are demanding high quality, consumer-focused and flexible courses and learning resources, and active learner support. This means that providers of DE need to reconsider key issues about learner support systems, to ensure that this is delivered appropriately and effectively. This study focuses on learner support in DE education for under-qualified teachers. The concept, learner support, can be traced far back to Vygotsky’s concept of the zone of proximal development, which refers to a learners’ optimal developmental potential if assistance that is timely and appropriate is provided by another person (Vygotsky, 1978). The appeal of the concept of the zone of proximal development lies in the fact that it directs attention to the need for maximum support in the learning process, and does so in a way that emphasises that good teaching is necessarily responsive to the state of understanding achieved by particular learners. Learner support systems may include the resources that the learner can access in order to engage in the learning process, for example, libraries or the resources that relate to the mediation of the communication process such as the media or technology. This study acknowledges that there are different kinds of learner support structures, but argues that there are critical or main components of support services which are registration support, learner support services, contact sessions, technological support and feedback strategies. Therefore, this study focuses on the role played by these five learner support structures as used in the programmes of DE under-qualified teachers at The University of South Africa (UNISA). Supported by empirical data, this study seeks to argue that instructional designers and tutors in the programmes of under-qualified teachers studying through DE need to ensure quality learning support as learning environments are increasingly designed according to the principles of resource-based and independent learning. In a sustainable learning environment, support must be designed according to principles that ensure that learners progress from teacher-directed activity to self-regulated activity. The challenge to deliver a high-quality learner support system, and for tutors and administrators to assume a central educational role in developing effective distance learning environments, the need for teaching and research is increasingly emphasised in research literature. Finally, it is hoped that this piece of work will help to promote more discussion and debate about the use of learner support programmes in DE institutions in particular, and in teaching and learning in general.
Curriculum and Instructional Studies
D. Ed. (Didactics)
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!