Dissertations / Theses on the topic 'Computer animation – Data processing'

To see the other types of publications on this topic, follow the link: Computer animation – Data processing.

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Computer animation – Data processing.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Stainback, Pamela Barth. "Computer animation : the animation capabilities of the Genigraphics 100C /." Online version of thesis, 1990. http://hdl.handle.net/1850/11460.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Yun, Hee Cheol. "Compression of computer animation frames." Diss., Georgia Institute of Technology, 1996. http://hdl.handle.net/1853/13070.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Marshall, Dana T. "The exploitation of image construction data and temporal/image coherence in ray traced animation /." Full text (PDF) from UMI/Dissertation Abstracts International, 2001. http://wwwlib.umi.com/cr/utexas/fullcit?p3008386.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Winnemöller, Holger. "Implementing non-photorealistic rendreing enhancements with real-time performance." Thesis, Rhodes University, 2002. http://hdl.handle.net/10962/d1003135.

Full text
Abstract:
We describe quality and performance enhancements, which work in real-time, to all well-known Non-photorealistic (NPR) rendering styles for use in an interactive context. These include Comic rendering, Sketch rendering, Hatching and Painterly rendering, but we also attempt and justify a widening of the established definition of what is considered NPR. In the individual Chapters, we identify typical stylistic elements of the different NPR styles. We list problems that need to be solved in order to implement the various renderers. Standard solutions available in the literature are introduced and in all cases extended and optimised. In particular, we extend the lighting model of the comic renderer to include a specular component and introduce multiple inter-related but independent geometric approximations which greatly improve rendering performance. We implement two completely different solutions to random perturbation sketching, solve temporal coherence issues for coal sketching and find an unexpected use for 3D textures to implement hatch-shading. Textured brushes of painterly rendering are extended by properties such as stroke-direction and texture, motion, paint capacity, opacity and emission, making them more flexible and versatile. Brushes are also provided with a minimal amount of intelligence, so that they can help in maximising screen coverage of brushes. We furthermore devise a completely new NPR style, which we call super-realistic and show how sample images can be tweened in real-time to produce an image-based six degree-of-freedom renderer performing at roughly 450 frames per second. Performance values for our other renderers all lie between 10 and over 400 frames per second on homePC hardware, justifying our real-time claim. A large number of sample screen-shots, illustrations and animations demonstrate the visual fidelity of our rendered images. In essence, we successfully achieve our attempted goals of increasing the creative, expressive and communicative potential of individual NPR styles, increasing performance of most of them, adding original and interesting visual qualities, and exploring new techniques or existing ones in novel ways.
KMBT_363
Adobe Acrobat 9.54 Paper Capture Plug-in
APA, Harvard, Vancouver, ISO, and other styles
5

Judice, Sicilia Ferreira Ponce Pasini. "Animação de Fluidos via Modelos do Tipo Lattice Gas e Lattice Boltzmann." Laboratório Nacional de Computação Científica, 2009. http://www.lncc.br/tdmc/tde_busca/arquivo.php?codArquivo=181.

Full text
Abstract:
Técnicas baseadas em física têm chamado a atenção da comunidade de computação gráfica, em especial para animação de fluidos (gás ou líquidos). As técnicas tradicionais para animação de fluidos são metodologias top-down baseadas em malhas 2D/3D, tais como Diferenças Finitas e Elementos Finitos, em conjunto com equações de fluidos Navier-Stokes. Entretanto, tais métodos têm um custo computacional alto. Uma alternativa é o uso de técnicas baseadas em Autômatos Celulares do tipo Lattice Gas (LGCA) e o Método de Lattice Boltzmann (LBM). A idéia básica desses métodos consiste em obter a dinâmica macroscópica de um fluido a partir do comportamento coletivo de diversas partículas microscópicas. Em geral, tais metodologias bottom-up são eficientes do ponto de vista computacional. Neste trabalho, são estudados os aspectos teóricos e práticos da animação computacional de fluidos bidimensionais para computação gráfica, usando um método LGCA chamado FHP, e um método LBM chamado D2Q9. É proposto um modelo de fluido 3D baseado nos modelos bidimensionais FHP e D2Q9, bem como em métodos de interpolação. Em seguida, são apresentadas duas aplicações para animação de fluidos através dos métodos mencionados, uma para execução em tempo real e outra para execução off-line. Nos resultados dos experimentos computacionais são enfatizados a simplicidade e o potencial dos modelos propostos quando combinados com técnicas eficientes de rendering.
Physically-based techniques for the animation of fluids (gas or liquids) have taken the attention of the computer graphics community. The traditional fluid animation methods rely on a top down viewpoint that uses 2D/3D mesh based approaches motivated by the Eulerian methods of Finite Element (FE) and Finite Difference (FD), in conjunction with Navier-Stokes equations of fluids. Alternatively, lattice methods comprised by the Lattice Gas Cellular Automata (LGCA) and Lattice Boltzmann (LBM) can be used. The basic idea behind these methods is that the macroscopic dynamics of a fluid is the result of the collective behavior of many microscopic particles. Such bottom-up approaches need low computational resources for both the memory allocation and the computation itself. In this work, we consider animation of fluids for computer graphics applications, using a LGCA method called FHP, and a LBM method called D2Q9, both bidimensional models. We propose 3D fluid animation techniques based on the FHP and D2Q9 as well as interpolation methods. Then, we present two animating frameworks based on the mentioned lattice methods, one for a real time implementation and the other for an off-line implementation. In the experimental results we emphasize the simplicity and power of the presented models when combined with efficient techniques for rendering and compare their efficiency.
APA, Harvard, Vancouver, ISO, and other styles
6

Goes, Fernando Ferrari de. "Analise espectral de superficies e aplicações em computação grafica." [s.n.], 2009. http://repositorio.unicamp.br/jspui/handle/REPOSIP/275916.

Full text
Abstract:
Orientador: Siome Klein Goldenstein
Dissertação (mestrado) - Universidade Estadual de Campinas, Instituto de Computação
Made available in DSpace on 2018-08-14T02:23:56Z (GMT). No. of bitstreams: 1 Goes_FernandoFerraride_M.pdf: 31957234 bytes, checksum: c369081bcbbb5f360184a1f8467839ea (MD5) Previous issue date: 2009
Resumo: Em computação gráfica, diversos problemas consistem na análise e manipulação da geometria de superfícies. O operador Laplace-Beltrami apresenta autovalores e autofunções que caracterizam a geometria de variedades, proporcionando poderosas ferramentas para o processamento geométrico. Nesta dissertação, revisamos as propriedades espectrais do operador Laplace-Beltrami e propomos sua aplicação em computação gráfica. Em especial, introduzimos novas abordagens para os problemas de segmentação semântica e geração de atlas em superfícies
Abstract: Many applications in computer graphics consist of the analysis and manipulation of the geometry of surfaces. The Laplace-Beltrami operator presents eigenvalues and eigenfuncitons which caracterize the geometry of manifolds, supporting powerful tools for geometry processing. In this dissertation, we revisit the spectral properties of the Laplace-Beltrami operator and apply them in computer graphics. In particular, we introduce new approaches for the problems of semantic segmentation and atlas generation on surfaces
Mestrado
Computação Grafica
Mestre em Ciência da Computação
APA, Harvard, Vancouver, ISO, and other styles
7

John, Nigel William. "Techniques for planning computer animation." Thesis, University of Bath, 1989. https://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.329568.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Luther, Kurt. "Supporting and transforming leadership in online creative collaboration." Diss., Georgia Institute of Technology, 2012. http://hdl.handle.net/1853/45822.

Full text
Abstract:
Online creative collaboration is challenging our basic assumptions about how people can create together. Volunteers from around the world who meet and communicate over the Internet have written the world's largest encyclopedia, developed market-leading software products, solved important open problems in mathematics, and produced award-winning films, among many examples. A growing body of research refutes the popular myth that these projects succeed through "self-organization" and instead points to the critical importance of effective leadership. Yet, we know little about what these leaders actually do, the challenges they must manage, and how technology supports or hinders their efforts. In this dissertation, I investigated the role of leadership in online creative collaboration. I first conducted two empirical studies of existing leadership practices, focusing on the domain of online, collaborative animation projects called "collabs." In the first study, I identified the major challenges faced by collab leaders. In the second study, I identified leader traits and behaviors correlated with success. These initial findings suggested that many collab leaders, overburdened and lacking adequate technological support, respond by attempting less ambitious projects and adopting centralized leadership styles. Despite these efforts, leaders frequently become overburdened, and more than 80% of collabs fail. To ease the burden on leaders and encourage more complex, successful projects, I led the development of a web-based, open-source software tool called Pipeline. Pipeline can support leadership by reinforcing a traditional, top-down approach, or transform leadership by redistributing it across many members of a group. This latter approach relies on social processes, rather than technical constraints, to guide behavior. I evaluated Pipeline's ability to effectively support and transform leadership through a detailed case study of Holiday Flood, a six-week collaboration involving nearly 30 artists from around the world. The case study showed that formal leaders remained influential and Pipeline supported their traditional, centralized approach. However, there was also evidence that Pipeline transformed some leadership behaviors, such as clarifying, informing, and monitoring, by redistributing them beyond the project's formal leaders. The result was a significantly more ambitious project which attained its goals and earned high praise from the community. The main contributions of this dissertation include: (1) a rich description of existing leadership practices in online creative collaboration; (2) the development of redistributed leadership as a theoretical framework for analyzing the relationship between leadership and technological support; (3) design implications for supporting and transform leadership; (4) a case study illustrating how technology can support and transform leadership in the real world; and (5) the Pipeline collaboration tool itself, released as open-source software.
APA, Harvard, Vancouver, ISO, and other styles
9

Hunter, Jane Louise. "Integrated sound synchronisation for computer animation." Thesis, University of Cambridge, 1994. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.239569.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Whited, Brian Scott. "Tangent-ball techniques for shape processing." Diss., Atlanta, Ga. : Georgia Institute of Technology, 2009. http://hdl.handle.net/1853/31670.

Full text
Abstract:
Thesis (Ph.D)--Computing, Georgia Institute of Technology, 2010.
Committee Chair: Jarek Rossignac; Committee Member: Greg Slabaugh; Committee Member: Greg Turk; Committee Member: Karen Liu; Committee Member: Maryann Simmons. Part of the SMARTech Electronic Thesis and Dissertation Collection.
APA, Harvard, Vancouver, ISO, and other styles
11

Caplan, Elizabeth A. "The Effects Of Animated Textual Instruction On Learners' Written Production Of German Modal Verb Sentences." Scholar Commons, 2002. http://purl.fcla.edu/fcla/etd/SFE0000042.

Full text
APA, Harvard, Vancouver, ISO, and other styles
12

Loizidou, Stephania M. "Dynamic analysis of anthropomorphic manipulators in computer animation." Thesis, London Metropolitan University, 1992. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.314630.

Full text
APA, Harvard, Vancouver, ISO, and other styles
13

Madhavavapeddy, Neganand. "A methodology for the real-time computer animation of articulated structures." Thesis, Queen's University Belfast, 1998. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.268214.

Full text
APA, Harvard, Vancouver, ISO, and other styles
14

Chuang, Yung-Yu. "New models and methods for matting and compositing /." Thesis, Connect to this title online; UW restricted, 2004. http://hdl.handle.net/1773/6849.

Full text
APA, Harvard, Vancouver, ISO, and other styles
15

Wang, Yi. "Data Management and Data Processing Support on Array-Based Scientific Data." The Ohio State University, 2015. http://rave.ohiolink.edu/etdc/view?acc_num=osu1436157356.

Full text
APA, Harvard, Vancouver, ISO, and other styles
16

Hudson, James. "Processing large point cloud data in computer graphics." Connect to this title online, 2003. http://rave.ohiolink.edu/etdc/view?acc%5Fnum=osu1054233187.

Full text
Abstract:
Thesis (Ph. D.)--Ohio State University, 2003.
Title from first page of PDF file. Document formatted into pages; contains xix, 169 p.; also includes graphics (some col.). Includes bibliographical references (p. 159-169). Available online via OhioLINK's ETD Center
APA, Harvard, Vancouver, ISO, and other styles
17

Fernandez, Noemi. "Statistical information processing for data classification." FIU Digital Commons, 1996. http://digitalcommons.fiu.edu/etd/3297.

Full text
Abstract:
This thesis introduces new algorithms for analysis and classification of multivariate data. Statistical approaches are devised for the objectives of data clustering, data classification and object recognition. An initial investigation begins with the application of fundamental pattern recognition principles. Where such fundamental principles meet their limitations, statistical and neural algorithms are integrated to augment the overall approach for an enhanced solution. This thesis provides a new dimension to the problem of classification of data as a result of the following developments: (1) application of algorithms for object classification and recognition; (2) integration of a neural network algorithm which determines the decision functions associated with the task of classification; (3) determination and use of the eigensystem using newly developed methods with the objectives of achieving optimized data clustering and data classification, and dynamic monitoring of time-varying data; and (4) use of the principal component transform to exploit the eigensystem in order to perform the important tasks of orientation-independent object recognition, and di mensionality reduction of the data such as to optimize the processing time without compromising accuracy in the analysis of this data.
APA, Harvard, Vancouver, ISO, and other styles
18

Nar, Selim. "A Virtual Human Animation Tool Using Motion Capture Data." Master's thesis, METU, 2008. http://etd.lib.metu.edu.tr/upload/12609683/index.pdf.

Full text
Abstract:
In this study, we developed an animation tool to animate 3D virtual characters. The tool offers facilities to integrate motion capture data with a 3D character mesh and animate the mesh by using Skeleton Subsurface Deformation and Dual Quaternion Skinning Methods. It is a compact tool, so it is possible to distribute, install and use the tool with ease. This tool can be used to illustrate medical kinematic gait data for educational purposes. For validation, we obtained medical motion capture data from two separate sources and animated a 3D mesh model by using this data. The animations are presented to physicians for evaluation. The results show that the tool is sufficient in displaying obvious gait patterns of the patients. The tool provides interactivity for inspecting the movements of patient from different angles and distances. We animate anonymous virtual characters which provide anonymity of the patient.
APA, Harvard, Vancouver, ISO, and other styles
19

Golab, Lukasz. "Sliding Window Query Processing over Data Streams." Thesis, University of Waterloo, 2006. http://hdl.handle.net/10012/2930.

Full text
Abstract:
Database management systems (DBMSs) have been used successfully in traditional business applications that require persistent data storage and an efficient querying mechanism. Typically, it is assumed that the data are static, unless explicitly modified or deleted by a user or application. Database queries are executed when issued and their answers reflect the current state of the data. However, emerging applications, such as sensor networks, real-time Internet traffic analysis, and on-line financial trading, require support for processing of unbounded data streams. The fundamental assumption of a data stream management system (DSMS) is that new data are generated continually, making it infeasible to store a stream in its entirety. At best, a sliding window of recently arrived data may be maintained, meaning that old data must be removed as time goes on. Furthermore, as the contents of the sliding windows evolve over time, it makes sense for users to ask a query once and receive updated answers over time.

This dissertation begins with the observation that the two fundamental requirements of a DSMS are dealing with transient (time-evolving) rather than static data and answering persistent rather than transient queries. One implication of the first requirement is that data maintenance costs have a significant effect on the performance of a DSMS. Additionally, traditional query processing algorithms must be re-engineered for the sliding window model because queries may need to re-process expired data and "undo" previously generated results. The second requirement suggests that a DSMS may execute a large number of persistent queries at the same time, therefore there exist opportunities for resource sharing among similar queries.

The purpose of this dissertation is to develop solutions for efficient query processing over sliding windows by focusing on these two fundamental properties. In terms of the transient nature of streaming data, this dissertation is based upon the following insight. Although the data keep changing over time as the windows slide forward, the changes are not random; on the contrary, the inputs and outputs of a DSMS exhibit patterns in the way the data are inserted and deleted. It will be shown that the knowledge of these patterns leads to an understanding of the semantics of persistent queries, lower window maintenance costs, as well as novel query processing, query optimization, and concurrency control strategies. In the context of the persistent nature of DSMS queries, the insight behind the proposed solution is that various queries may need to be refreshed at different times, therefore synchronizing the refresh schedules of similar queries creates more opportunities for resource sharing.
APA, Harvard, Vancouver, ISO, and other styles
20

Vijayakumar, Nithya Nirmal. "Data management in distributed stream processing systems." [Bloomington, Ind.] : Indiana University, 2007. http://gateway.proquest.com/openurl?url_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&res_dat=xri:pqdiss&rft_dat=xri:pqdiss:3278228.

Full text
Abstract:
Thesis (Ph.D.)--Indiana University, Dept. of Computer Science, 2007.
Source: Dissertation Abstracts International, Volume: 68-09, Section: B, page: 6093. Adviser: Beth Plale. Title from dissertation home page (viewed May 9, 2008).
APA, Harvard, Vancouver, ISO, and other styles
21

Battle, Timothy P. "A computer-aided design scheme for drainage and runoff systems." Thesis, University of British Columbia, 1985. http://hdl.handle.net/2429/25077.

Full text
Abstract:
A computer-aided design scheme for both man-made and natural runoff systems is presented. The model uses linear programming to solve Muskingum routing equations through a drainage system, and provides design information through post-optimality (sensitivity) analysis. With the objective of minimizing the peak outflow from the system and using hydrograph ordinates as the decision variables, the output of the linear programming analysis shows the extent that each flow ordinate at every node in the network influences the peak flow at some downstream location. This information can aid the user in speeding up the design process to arrive at an efficient design - i.e., one which either minimizes construction costs or reduces the potential risk of flood damage.
Applied Science, Faculty of
Civil Engineering, Department of
Graduate
APA, Harvard, Vancouver, ISO, and other styles
22

Lau, Yan Nam. "Adaptive approach to triangulation of isosurface in volume data /." View Abstract or Full-Text, 2002. http://library.ust.hk/cgi/db/thesis.pl?COMP%202002%20LAUY.

Full text
Abstract:
Thesis (M. Phil.)--Hong Kong University of Science and Technology, 2002.
Includes bibliographical references (leaves 91-95). Also available in electronic version. Access restricted to campus users.
APA, Harvard, Vancouver, ISO, and other styles
23

Derksen, Timothy J. (Timothy John). "Processing of outliers and missing data in multivariate manufacturing data." Thesis, Massachusetts Institute of Technology, 1996. http://hdl.handle.net/1721.1/38800.

Full text
Abstract:
Thesis (M. Eng.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 1996.
Includes bibliographical references (leaf 64).
by Timothy J. Derksen.
M.Eng.
APA, Harvard, Vancouver, ISO, and other styles
24

Bao, Shunxing. "Algorithmic Enhancements to Data Colocation Grid Frameworks for Big Data Medical Image Processing." Thesis, Vanderbilt University, 2019. http://pqdtopen.proquest.com/#viewpdf?dispub=13877282.

Full text
Abstract:

Large-scale medical imaging studies to date have predominantly leveraged in-house, laboratory-based or traditional grid computing resources for their computing needs, where the applications often use hierarchical data structures (e.g., Network file system file stores) or databases (e.g., COINS, XNAT) for storage and retrieval. The resulting performance for laboratory-based approaches reveal that performance is impeded by standard network switches since typical processing can saturate network bandwidth during transfer from storage to processing nodes for even moderate-sized studies. On the other hand, the grid may be costly to use due to the dedicated resources used to execute the tasks and lack of elasticity. With increasing availability of cloud-based big data frameworks, such as Apache Hadoop, cloud-based services for executing medical imaging studies have shown promise.

Despite this promise, our studies have revealed that existing big data frameworks illustrate different performance limitations for medical imaging applications, which calls for new algorithms that optimize their performance and suitability for medical imaging. For instance, Apache HBases data distribution strategy of region split and merge is detrimental to the hierarchical organization of imaging data (e.g., project, subject, session, scan, slice). Big data medical image processing applications involving multi-stage analysis often exhibit significant variability in processing times ranging from a few seconds to several days. Due to the sequential nature of executing the analysis stages by traditional software technologies and platforms, any errors in the pipeline are only detected at the later stages despite the sources of errors predominantly being the highly compute-intensive first stage. This wastes precious computing resources and incurs prohibitively higher costs for re-executing the application. To address these challenges, this research propose a framework - Hadoop & HBase for Medical Image Processing (HadoopBase-MIP) - which develops a range of performance optimization algorithms and employs a number of system behaviors modeling for data storage, data access and data processing. We also introduce how to build up prototypes to help empirical system behaviors verification. Furthermore, we introduce a discovery with the development of HadoopBase-MIP about a new type of contrast for medical imaging deep brain structure enhancement. And finally we show how to move forward the Hadoop based framework design into a commercialized big data / High performance computing cluster with cheap, scalable and geographically distributed file system.

APA, Harvard, Vancouver, ISO, and other styles
25

Lian, Xiang. "Efficient query processing over uncertain data /." View abstract or full-text, 2009. http://library.ust.hk/cgi/db/thesis.pl?CSED%202009%20LIAN.

Full text
APA, Harvard, Vancouver, ISO, and other styles
26

Da, Yanan. "A Big Spatial Data System for Efficient and Scalable Spatial Data Processing." Thesis, Southern Illinois University at Edwardsville, 2018. http://pqdtopen.proquest.com/#viewpdf?dispub=10682760.

Full text
Abstract:

Today, a large amount of spatial data is generated from a variety of sources, such as mobile devices, sensors, and satellites. Traditional spatial data processing techniques no longer satisfy the efficiency and scalability requirements for large-scale spatial data processing. Existing Big Data processing frameworks such as Hadoop and Spark have been extended to support effective large-scale spatial data processing. In addition to processing data in distributed schemes utilizing computer clusters for efficiency and scalability, single node performance can also be improved by making use of multi-core processors. In this thesis, we investigate approaches to parallelize line segment intersection algorithms for spatial computations on multi-core processors, which can be used as node-level algorithms for distributed spatial data processing. We first provide our design of line segment intersection algorithms and introduce parallelization techniques. Then, we describe experimental results using multiple data sets and speed ups are examined with varying numbers of processing cores. Equipped with the efficient underlying algorithm for spatial computation, we investigate how to build a native big spatial data system from the ground up. We provide a system design for distributed large-scale spatial data management and processing using a two-level hash based Quadtree index as well as algorithms for spatial operations.

APA, Harvard, Vancouver, ISO, and other styles
27

JULARDZIJA, MIRHET. "Processing RAW image data in mobile units." Thesis, Mälardalens högskola, Akademin för innovation, design och teknik, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:mdh:diva-27724.

Full text
APA, Harvard, Vancouver, ISO, and other styles
28

Wang, Jiayin. "Building Efficient Large-Scale Big Data Processing Platforms." Thesis, University of Massachusetts Boston, 2017. http://pqdtopen.proquest.com/#viewpdf?dispub=10262281.

Full text
Abstract:

In the era of big data, many cluster platforms and resource management schemes are created to satisfy the increasing demands on processing a large volume of data. A general setting of big data processing jobs consists of multiple stages, and each stage represents generally defined data operation such as ltering and sorting. To parallelize the job execution in a cluster, each stage includes a number of identical tasks that can be concurrently launched at multiple servers. Practical clusters often involve hundreds or thousands of servers processing a large batch of jobs. Resource management, that manages cluster resource allocation and job execution, is extremely critical for the system performance.

Generally speaking, there are three main challenges in resource management of the new big data processing systems. First, while there are various pending tasks from dierent jobs and stages, it is difficult to determine which ones deserve the priority to obtain the resources for execution, considering the tasks' different characteristics such as resource demand and execution time. Second, there exists dependency among the tasks that can be concurrently running. For any two consecutive stages of a job, the output data of the former stage is the input data of the later one. The resource management has to comply with such dependency. The third challenge is the inconsistent performance of the cluster nodes. In practice, run-time performance of every server is varying. The resource management needs to dynamically adjust the resource allocation according to the performance change of each server.

The resource management in the existing platforms and prior work often rely on fixed user-specic congurations, and assumes consistent performance in each node. The performance, however, is not satisfactory under various workloads. This dissertation aims to explore new approaches to improving the eciency of large-scale big data processing platforms. In particular, the run-time dynamic factors are carefully considered when the system allocates the resources. New algorithms are developed to collect run-time data and predict the characteristics of jobs and the cluster. We further develop resource management schemes that dynamically tune the resource allocation for each stage of every running job in the cluster. New findings and techniques in this dissertation will certainly provide valuable and inspiring insights to other similar problems in the research community.

APA, Harvard, Vancouver, ISO, and other styles
29

Li, Quanzhong. "Indexing and path query processing for XML data." Diss., The University of Arizona, 2004. http://hdl.handle.net/10150/290141.

Full text
Abstract:
XML has emerged as a new standard for information representation and exchange on the Internet. To efficiently process XML data, we propose the extended preorder numbering scheme, which determines the ancestor-descendant relationship between nodes in the hierarchy of XML data in constant time, and adapts to the dynamics of XML data by allocating extra space. Based on this numbering scheme, we propose sort-merge based algorithms, εA-Join and εε-Join, to process ancestor-descendant path expressions. The experimental results showed an order of magnitude performance improvement over conventional methods. We further propose the partition-based algorithms, which can be chosen by a query optimizer according to the characteristics of the input data. For complex path expressions with branches, we propose the Containment B⁺-tree (CB-tree) index and the IndexTwig algorithm. The CB-tree, which is an extension of the B⁺-tree, supports both the containment query and the reverse containment query. It is an effective indexing scheme for XML documents with or without a small number of recursions. The proposed IndexTwig algorithm works with any index supporting containment and reverse containment queries, such as the CB-tree. We also introduce a simplified output model, which outputs only the necessary result of a path expression. The output model enables the Fast Existence Test (FET) optimization to skip unnecessary data and avoid generating unwanted results. Also in this dissertation, we introduce techniques to process the predicates in XML path expressions using the EVR-tree. The EVR-tree combines the advantages of indexing on values or elements individually using B+-trees. It utilizes the high value selectivity and/or high structural selectivity, and provides ordered element access by using a priority queue. At the end of the dissertation, we introduce the XISS/R system, which is an implementation of the XML Indexing and Storage System (XISS) on top of a relational database. The XISS/R includes a web-based user interface and a XPath query engine to translate XPath queries into efficient SQL statements.
APA, Harvard, Vancouver, ISO, and other styles
30

曾偉明 and Wai-ming Peter Tsang. "Computer aided ultrasonic flaw detection and characterization." Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 1987. http://hub.hku.hk/bib/B31231007.

Full text
APA, Harvard, Vancouver, ISO, and other styles
31

Grinman, Alex J. "Natural language processing on encrypted patient data." Thesis, Massachusetts Institute of Technology, 2016. http://hdl.handle.net/1721.1/113438.

Full text
Abstract:
Thesis: M. Eng., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2016.
This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
Cataloged from student-submitted PDF version of thesis.
Includes bibliographical references (pages 85-86).
While many industries can benefit from machine learning techniques for data analysis, they often do not have the technical expertise nor computational power to do so. Therefore, many organizations would benefit from outsourcing their data analysis. Yet, stringent data privacy policies prevent outsourcing sensitive data and may stop the delegation of data analysis in its tracks. In this thesis, we put forth a two-party system where one party capable of powerful computation can run certain machine learning algorithms from the natural language processing domain on the second party's data, where the first party is limited to learning only specific functions of the second party's data and nothing else. Our system provides simple cryptographic schemes for locating keywords, matching approximate regular expressions, and computing frequency analysis on encrypted data. We present a full implementation of this system in the form of a extendible software library and a command line interface. Finally, we discuss a medical case study where we used our system to run a suite of unmodified machine learning algorithms on encrypted free text patient notes.
by Alex J. Grinman.
M. Eng.
APA, Harvard, Vancouver, ISO, and other styles
32

Westlund, Kenneth P. (Kenneth Peter). "Recording and processing data from transient events." Thesis, Massachusetts Institute of Technology, 1988. https://hdl.handle.net/1721.1/129961.

Full text
Abstract:
Thesis (B.S.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 1988.
Includes bibliographical references.
by Kenneth P. Westlund Jr.
Thesis (B.S.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 1988.
APA, Harvard, Vancouver, ISO, and other styles
33

Yuen, Jeanne Y. Y. "Computer Go-Muku." Thesis, McGill University, 1988. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=64063.

Full text
APA, Harvard, Vancouver, ISO, and other styles
34

Parker, Greg. "Robust processing of diffusion weighted image data." Thesis, Cardiff University, 2014. http://orca.cf.ac.uk/61622/.

Full text
Abstract:
The work presented in this thesis comprises a proposed robust diffusion weighted magnetic resonance imaging (DW-MRI) pipeline, each chapter detailing a step designed to ultimately transform raw DW-MRI data into segmented bundles of coherent fibre ready for more complex analysis or manipulation. In addition to this pipeline we will also demonstrate, where appropriate, ways in which each step could be optimized for the maxillofacial region, setting the groundwork for a wider maxillofacial modelling project intended to aid surgical planning. Our contribution begins with RESDORE, an algorithm designed to automatically identify corrupt DW-MRI signal elements. While slower than the closest alternative, RESDORE is also far more robust to localised changes in SNR and pervasive image corruptions. The second step in the pipeline concerns the retrieval of accurate fibre orientation distribution functions (fODFs) from the DW-MRI signal. Chapter 4 comprises a simulation study exploring the application of spherical deconvolution methods to `generic' fibre; finding that the commonly used constrained spherical harmonic deconvolution (CSHD) is extremely sensitive to calibration but, if handled correctly, might be able to resolve muscle fODFs in vivo. Building upon this information, Chapter 5 conducts further simulations and in vivo image experimentation demonstrating that this is indeed the case, allowing us to demonstrate, for the first time, anatomically plausible reconstructions of several maxillofacial muscles. To complete the proposed pipeline, Chapter 6 then introduces a method for segmenting whole volume streamline tractographies into anatomically valid bundles. In addition to providing an accurate segmentation, this shape-based method does not require computationally expensive inter-streamline comparisons employed by other approaches, allowing the algorithm to scale linearly with respect to the number of streamlines within the dataset. This is not often true for comparison based methods which in the best case scale in higher linear time but more often by O(N2) complexity.
APA, Harvard, Vancouver, ISO, and other styles
35

Scheffler, Carl. "Articulated structure from motion." Thesis, University of the Western Cape, 2004. http://etd.uwc.ac.za/index.php?module=etd&action=viewtitle&id=init_2988_1177923873.

Full text
Abstract:
The structure from motion (SfM) problem is that of determining 3-dimensional (3D) information of a scene from sequences of 2-dimensional (2D) images [59]. This information consists of object shape and motion and relative camera motion. In general, objects may undergo complex non-rigid motion and may be occluded by other objects or themselves. These aspects make the general SfM problem under-constrained and the solution subject to missing or incomplete data.
APA, Harvard, Vancouver, ISO, and other styles
36

Tuft, David O. "System for collision detection between deformable models built on axis aligned bounding boxes and GPU based culling /." Diss., CLICK HERE for online access, 2007. http://contentdm.lib.byu.edu/ETD/image/etd1689.pdf.

Full text
APA, Harvard, Vancouver, ISO, and other styles
37

Allander, Karl, Jim Svanberg, and Mattias Klittmark. "Intresseväckande Animation : Utställningsmaterial för Mälsåkerprojektet." Thesis, Mälardalen University, Department of Innovation, Design and Product Development, 2008. http://urn.kb.se/resolve?urn=urn:nbn:se:mdh:diva-703.

Full text
Abstract:

Under sommaren 2006 öppnar Mälsåker Slott för allmänheten och ett av de planerade projekten är en multimedial utställning om norrmän som tränades där i hemlighet under andra världskriget.

Författarna till denna rapport är alla studenter av informationsdesign, med inriktning mot illustration på Mälardalens Högskola. Då vi har ett stort intresse av animation föreföll detta projekt mycket intressant.

Vi ingick i en projektgrupp som innefattade dataloger, textdesigners samt illustratörer. Illustratörernas del av projektet var att skapa det visuella materialet på ett intresseväckande sätt. I rapporten undersöks möjliga lösningar för animation, manér, bilddramaturgi samt de tekniska förutsättningar som krävs för att skapa ett lyckat slutresultat. Rapporten beskriver tillvägagångssättet för att uppnå dessa mål vilket inkluderar metoder som litteraturstudier, diskussioner samt utprovningar. Utställningsformen är experimentell och arbetet har därför givit ett slutresultat som kan ses som en

fallstudie i sig.

Utprovningar av det färdiga materialet visar att vi efter förutsättningarna lyckats uppnå ett gott resultat. Animerat bildspel fungerar i sammanhanget bra som informativt utställningsmaterial.

APA, Harvard, Vancouver, ISO, and other styles
38

Heller, Kelly K. "Satisfiability checking for quality assurance in relational data processing." Thesis, California State University, Long Beach, 2013. http://pqdtopen.proquest.com/#viewpdf?dispub=1524200.

Full text
Abstract:

Developers using the SQL language lack relational SQL debugging tools. SQL also suffers from a lack of sufficient automated frameworks for unit testing and regression testing of queries. This thesis begins with an SQL bug taxonomy. The focus is on Core SQL as defined in the SQL:1999 standard. Features in Core SQL remain virtually unchanged through the latest standard, SQL:2011. Our bug taxonomy illustrates common coding defects related to NULL, LEFT JOIN, selectivity assumptions, and ordering assumptions. We highlight ways that silent failures can impact a databasedependent application. Our proposal for addressing SQL development errors is a verification tool (named POQ) that leverages annotated postconditions attached to each query statement. The formalism of domain relational calculus is the basis for axiomatizing queries. Subsequently, a theorem prover searches for counterexamples that demonstrate when the postcondition fails. Alternative projects using SMT solvers and static analysis methods are also discussed.

APA, Harvard, Vancouver, ISO, and other styles
39

Lin, Wen-Ya. "A Load Balancing Data Allocation for Parallel Query Processing." NSUWorks, 1998. http://nsuworks.nova.edu/gscis_etd/670.

Full text
Abstract:
This paper presents a multidimensional schema, called the multidimensional range tree (MDR-tree), to manage multi dimensionally partitioned relational database tables in parallel database system environments. In order to support the speed-up and scalability for intensive data processing, the parallel data processing paradigm has been proved as being one of the best solutions for handling query processing. One of the most important issues in parallel data processing systems is the management of dynamic load balancing for partitioned relational database tables. In order to support the management of dynamic load balancing, we employ a dynamic multidimensional data structure, the multidimensional range tree (MDR-tree), and its associated operations such as insertion, deletion, split, merge, and tuning. The major purpose of operations such as splitting, merging, and tuning is to balance the distribution data records in each processor or network node. By using the tuning strategy, each processor or network node is initially assigned the partitioned data, with an equal load as the other processors. The equal load in each processor or network node is also maintained when the database undergoes frequent updates such as insertions and deletions. The application of multidimensional partitioning by using the MDR-tree is to support multidimensional query processing, such as OLAP (on-line analytic processing), in large database or data warehouse environments. Due to the nature of intensive data and real-time constraints in database and data warehouse applications, the issues of both speed-up and scalability do require finding good practical solutions to load-balanced partitioning and management of a huge volume of data. In the specific case of star queries in data warehouse applications, multidimensional joins can be processed effectively and in a parallel manner with the support of the MDR-tree. We believe that the multidimensional data partitioning and management can be one of the alternatives for these types of applications. We have performed simulation experiments to validate the effectiveness of the MDR-tree. All experimental results indicate that the MDR-tree is an effective index structure for multidimensional data partitioning.
APA, Harvard, Vancouver, ISO, and other styles
40

Park, Noseong. "Top-K Query Processing in Edge-Labeled Graph Data." Thesis, University of Maryland, College Park, 2016. http://pqdtopen.proquest.com/#viewpdf?dispub=10128677.

Full text
Abstract:

Edge-labeled graphs have proliferated rapidly over the last decade due to the increased popularity of social networks and the Semantic Web. In social networks, relationships between people are represented by edges and each edge is labeled with a semantic annotation. Hence, a huge single graph can express many different relationships between entities. The Semantic Web represents each single fragment of knowledge as a triple (subject, predicate, object), which is conceptually identical to an edge from subject to object labeled with predicates. A set of triples constitutes an edge-labeled graph on which knowledge inference is performed.

Subgraph matching has been extensively used as a query language for patterns in the context of edge-labeled graphs. For example, in social networks, users can specify a subgraph matching query to find all people that have certain neighborhood relationships. Heavily used fragments of the SPARQL query language for the Semantic Web and graph queries of other graph DBMS can also be viewed as subgraph matching over large graphs.

Though subgraph matching has been extensively studied as a query paradigm in the Semantic Web and in social networks, a user can get a large number of answers in response to a query. These answers can be shown to the user in accordance with an importance ranking. In this thesis proposal, we present four different scoring models along with scalable algorithms to find the top-k answers via a suite of intelligent pruning techniques. The suggested models consist of a practically important subset of the SPARQL query language augmented with some additional useful features.

The first model called Substitution Importance Query (SIQ) identifies the top-k answers whose scores are calculated from matched vertices' properties in each answer in accordance with a user-specified notion of importance. The second model called Vertex Importance Query (VIQ) identifies important vertices in accordance with a user-defined scoring method that builds on top of various subgraphs articulated by the user. Approximate Importance Query (AIQ), our third model, allows partial and inexact matchings and returns top-k of them with a user-specified approximation terms and scoring functions. In the fourth model called Probabilistic Importance Query (PIQ), a query consists of several sub-blocks: one mandatory block that must be mapped and other blocks that can be opportunistically mapped. The probability is calculated from various aspects of answers such as the number of mapped blocks, vertices' properties in each block and so on and the most top-k probable answers are returned.

An important distinguishing feature of our work is that we allow the user a huge amount of freedom in specifying: (i) what pattern and approximation he considers important, (ii) how to score answers - irrespective of whether they are vertices or substitution, and (iii) how to combine and aggregate scores generated by multiple patterns and/or multiple substitutions. Because so much power is given to the user, indexing is more challenging than in situations where additional restrictions are imposed on the queries the user can ask.

The proposed algorithms for the first model can also be used for answering SPARQL queries with ORDER BY and LIMIT, and the method for the second model also works for SPARQL queries with GROUP BY, ORDER BY and LIMIT. We test our algorithms on multiple real-world graph databases, showing that our algorithms are far more efficient than popular triple stores.

APA, Harvard, Vancouver, ISO, and other styles
41

李淑儀 and Shuk-yee Wendy Lee. "Computer aided facilities design." Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 1987. http://hub.hku.hk/bib/B31208277.

Full text
APA, Harvard, Vancouver, ISO, and other styles
42

Haley, Brent Kreh. "A Pipeline for the Creation, Compression, and Display of Streamable 3D Motion Capture Based Skeletal Animation Data." The Ohio State University, 2011. http://rave.ohiolink.edu/etdc/view?acc_num=osu1300989069.

Full text
APA, Harvard, Vancouver, ISO, and other styles
43

Ding, Luping. "Metadata-aware query processing over data streams." Worcester, Mass. : Worcester Polytechnic Institute, 2008. http://www.wpi.edu/Pubs/ETD/Available/etd-042208-194826/.

Full text
APA, Harvard, Vancouver, ISO, and other styles
44

Karlsson, Tobias. "Kroppsspråk i 3d-animation : Lögner hos 3d-karaktärer." Thesis, Högskolan i Skövde, Institutionen för kommunikation och information, 2011. http://urn.kb.se/resolve?urn=urn:nbn:se:his:diva-5022.

Full text
Abstract:
Detta arbete undersöker huruvida vi kan uppfatta lögn och nervositet hos en avskalad virtuell karaktär. Utifrån forskning om beteendevetenskap, kroppsspråk och icke-verbala signaler har lögnsignaler och nervositetstecken animerats hos en avskalad virtuell karaktär. Dessa animationer visades utan audiella ledtrådar, såsom tal, för urvalsgrupper och en kvantitativ undersökning genomfördes för att besvara arbetets frågeställning. Undersökningens resultat gav emellertid inget konkret svar på frågeställningen. För att respondenterna skulle uppfatta lögnsignaler hos en avskalad virtuell karaktär krävdes att ett flertal aspekter togs i beaktning, exempelvis behövde lögnsignalerna vara tydliga och inte överskuggas av starkare känslor eller signaler. Nervositetstecken avlästes dock enkelt av respondenterna i undersökningen vilket kan betyda att en avskalad virtuell karaktär kan gestalta ett nervöst tillstånd utan audiella ledtrådar. Resultatet som undersökningen gav har ställts mot insamlad data och tidigare forskning för att diskutera undersökningens brister och lämpliga framtida justeringar för att åtgärda dessa. Slutligen har även möjligheter för framtida arbete diskuterats och spekulationer kring användningsområden för en liknande undersökning i ett branschperspektiv har genomförts.
APA, Harvard, Vancouver, ISO, and other styles
45

Sungoor, Ala M. H. "Genomic signal processing for enhanced microarray data clustering." Thesis, Kingston University, 2009. http://eprints.kingston.ac.uk/20310/.

Full text
Abstract:
Genomic signal processing is a new area of research that combines genomics with digital signal processing methodologies for enhanced genetic data analysis. Microarray is a well known technology for the evaluation of thousands of gene expression profiles. By considering these profiles as digital signals, the power of DSP methods can be applied to produce robust and unsupervised clustering of microarray samples. This can be achieved by transferring expression profiles into spectral components which are interpreted as a measure of profile similarity. This thesis introduces enhanced signal processing algorithms for robust clustering of micro array gene expression samples. The main aim of the research is to design and validate novel genomic signal processing methodologies for micro array data analysis based on different DSP methods. More specifically, clustering algorithms based on Linear prediction coding, Wavelet decomposition and Fractal dimension methods combined with Vector quantisation algorithm are applied and compared on a set of test microarray datasets. These techniques take as an input microarray gene expression samples and produce predictive coefficients arrays associated to the microarray data that are quantised in discrete levels, and consequently used for sample clustering. A variety of standard micro array datasets are used in this work to validate the robustness of these methods compared to conventional methods. Two well known validation approaches, i.e. Silhouette and Davies Bouldin index methods, are applied to evaluate internally and externally the genomic signal processing clustering results. In conclusion, the results demonstrate that genomic signal processing based methods outperform traditional methods by providing more clustering accuracy. Moreover, the study shows that the local features of the gene expression signals are better clustered using wavelets compared to the other DSP methods.
APA, Harvard, Vancouver, ISO, and other styles
46

Chen, Jiawen (Jiawen Kevin). "Efficient data structures for piecewise-smooth video processing." Thesis, Massachusetts Institute of Technology, 2011. http://hdl.handle.net/1721.1/66003.

Full text
Abstract:
Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2011.
Cataloged from PDF version of thesis.
Includes bibliographical references (p. 95-102).
A number of useful image and video processing techniques, ranging from low level operations such as denoising and detail enhancement to higher level methods such as object manipulation and special effects, rely on piecewise-smooth functions computed from the input data. In this thesis, we present two computationally efficient data structures for representing piecewise-smooth visual information and demonstrate how they can dramatically simplify and accelerate a variety of video processing algorithms. We start by introducing the bilateral grid, an image representation that explicitly accounts for intensity edges. By interpreting brightness values as Euclidean coordinates, the bilateral grid enables simple expressions for edge-aware filters. Smooth functions defined on the bilateral grid are piecewise-smooth in image space. Within this framework, we derive efficient reinterpretations of a number of edge-aware filters commonly used in computational photography as operations on the bilateral grid, including the bilateral filter, edgeaware scattered data interpolation, and local histogram equalization. We also show how these techniques can be easily parallelized onto modern graphics hardware for real-time processing of high definition video. The second data structure we introduce is the video mesh, designed as a flexible central data structure for general-purpose video editing. It represents objects in a video sequence as 2.5D "paper cutouts" and allows interactive editing of moving objects and modeling of depth, which enables 3D effects and post-exposure camera control. In our representation, we assume that motion and depth are piecewise-smooth, and encode them sparsely as a set of points tracked over time. The video mesh is a triangulation over this point set and per-pixel information is obtained by interpolation. To handle occlusions and detailed object boundaries, we rely on the user to rotoscope the scene at a sparse set of frames using spline curves. We introduce an algorithm to robustly and automatically cut the mesh into local layers with proper occlusion topology, and propagate the splines to the remaining frames. Object boundaries are refined with per-pixel alpha mattes. At its core, the video mesh is a collection of texture-mapped triangles, which we can edit and render interactively using graphics hardware. We demonstrate the effectiveness of our representation with special effects such as 3D viewpoint changes, object insertion, depthof- field manipulation, and 2D to 3D video conversion.
by Jiawen Chen.
Ph.D.
APA, Harvard, Vancouver, ISO, and other styles
47

Jakubiuk, Wiktor. "High performance data processing pipeline for connectome segmentation." Thesis, Massachusetts Institute of Technology, 2015. http://hdl.handle.net/1721.1/106122.

Full text
Abstract:
Thesis: M. Eng. in Computer Science and Engineering, Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, February 2016.
"December 2015." Cataloged from PDF version of thesis.
Includes bibliographical references (pages 83-88).
By investigating neural connections, neuroscientists try to understand the brain and reconstruct its connectome. Automated connectome reconstruction from high resolution electron miscroscopy is a challenging problem, as all neurons and synapses in a volume have to be detected. A mm3 of a high-resolution brain tissue takes roughly a petabyte of space that the state-of-the-art pipelines are unable to process to date. A high-performance, fully automated image processing pipeline is proposed. Using a combination of image processing and machine learning algorithms (convolutional neural networks and random forests), the pipeline constructs a 3-dimensional connectome from 2-dimensional cross-sections of a mammal's brain. The proposed system achieves a low error rate (comparable with the state-of-the-art) and is capable of processing volumes of 100's of gigabytes in size. The main contributions of this thesis are multiple algorithmic techniques for 2- dimensional pixel classification of varying accuracy and speed trade-off, as well as a fast object segmentation algorithm. The majority of the system is parallelized for multi-core machines, and with minor additional modification is expected to work in a distributed setting.
by Wiktor Jakubiuk.
M. Eng. in Computer Science and Engineering
APA, Harvard, Vancouver, ISO, and other styles
48

Nguyen, Qui T. "Robust data partitioning for ad-hoc query processing." Thesis, Massachusetts Institute of Technology, 2015. http://hdl.handle.net/1721.1/106004.

Full text
Abstract:
Thesis: M. Eng., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2015.
This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
Cataloged from student-submitted PDF version of thesis.
Includes bibliographical references (pages 59-62).
Data partitioning can significantly improve query performance in distributed database systems. Most proposed data partitioning techniques choose the partitioning based on a particular expected query workload or use a simple upfront scheme, such as uniform range partitioning or hash partitioning on a key. However, these techniques do not adequately address the case where the query workload is ad-hoc and unpredictable, as in many analytic applications. The HYPER-PARTITIONING system aims to ll that gap, by using a novel space-partitioning tree on the space of possible attribute values to dene partitions incorporating all attributes of a dataset. The system creates a robust upfront partitioning tree, designed to benet all possible queries, and then adapts it over time in response to the actual workload. This thesis evaluates the robustness of the upfront hyper-partitioning algorithm, describes the implementation of the overall HYPER-PARTITIONING system, and shows how hyper-partitioning improves the performance of both selection and join queries.
by Qui T. Nguyen.
M. Eng.
APA, Harvard, Vancouver, ISO, and other styles
49

Du, Toit André Johan. "Preslab - micro-computer analysis and design of prestressed concrete slabs." Master's thesis, University of Cape Town, 1988. http://hdl.handle.net/11427/17057.

Full text
Abstract:
Bibliography: pages 128-132.
A micro-computer based package for the analysis and design of prestressed flat slabs is presented. The constant strain triangle and the discreet Kirchhoff plate bending triangle are combined to provide an efficient "shell" element. These triangles are used for the finite element analysis of prestressed flat slabs. An efficient out-of-core solver for sets of linear simultaneous equations is presented. This solver was developed especially for micro-computers. Subroutines for the design of prestressed flat slabs include the principal stresses in the top and bottom fibres of the plate, Wood/Armer moments and untensioned steel areas calculated according to Clark's recommendations. Extensive pre- and post-processing facilities are presented. Several plotting routines were developed to aid the user in his understanding of the behaviour of the structure under load and prestressing.
APA, Harvard, Vancouver, ISO, and other styles
50

Samson, Margaret Kingman 1950. "COMPUTER AIDS FOR FACILITY LAYOUT." Thesis, The University of Arizona, 1987. http://hdl.handle.net/10150/276400.

Full text
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography