To see the other types of publications on this topic, follow the link: APACBS.

Dissertations / Theses on the topic 'APACBS'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'APACBS.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Carlisle, Jeffrey D. "Spanish Relations with the Apache Nations East of the Río Grande." Thesis, University of North Texas, 2001. https://digital.library.unt.edu/ark:/67531/metadc2816/.

Full text
Abstract:
This dissertation is a study of the Eastern Apache nations and their struggle to survive with their culture intact against numerous enemies intent on destroying them. It is a synthesis of published secondary and primary materials, supported with archival materials, primarily from the Béxar Archives. The Apaches living on the plains have suffered from a lack of a good comprehensive study, even though they played an important role in hindering Spanish expansion in the American Southwest. When the Spanish first encountered the Apaches they were living peacefully on the plains, although they occasionally raided nearby tribes. When the Spanish began settling in the Southwest they changed the dynamics of the region by introducing horses. The Apaches quickly adopted the animals into their culture and used them to dominate their neighbors. Apache power declined in the eighteenth century when their Caddoan enemies acquired guns from the French, and the powerful Comanches gained access to horses and began invading northern Apache territory. Surrounded by enemies, the Apaches increasingly turned to the Spanish for aid and protection rather than trade. The Spanish-Apache peace was fraught with problems. The Spaniards tended to lump all Apaches into one group even though, in reality, each band operated independently. Thus, when one Apache band raided a Spanish outpost, the Spanish considered the peace broken. On the other hand, since Apaches considered each Spanish settlement a distinct "band" they saw nothing wrong in making peace at one Spanish location while continuing to raid another. Eventually the Spanish encouraged other Indians tribes to launch a campaign of unrelenting war against the Apaches. Despite devastating attacks from their enemies, the Apaches were able to survive. When the Mexican Revolution removed the Spanish from the area, the Apaches remained and still occupied portions of the plains as late as the 1870s. Despite the pressures brought to bear upon them the Apaches prevailed, retaining their freedoms longer than almost any other tribe.
APA, Harvard, Vancouver, ISO, and other styles
2

Wada, Lorena Lai Lin 1963. "Summer habitat use by Apache trout (Oncorhynchus apache) in five streams on the Fort Apache Indian Reservation." Thesis, The University of Arizona, 1991. http://hdl.handle.net/10150/277963.

Full text
Abstract:
In the summer, five creeks on the Fort Apache Indian Reservation were examined. Apache trout generally selected the deepest pools with the greatest amounts of instream cover and bankcuts. Though fish in three creeks may be Apache trout x rainbow trout hybrids, they were found to occupy the same type of pools. Regression models on environmental conditions in two creeks accounted for 83% and 76% of the variability in biomass of Apache trout. There was adequate nursery habitat (and successful reproduction) on two creeks, but such habitat was lacking in the other three creeks, and no evidence of successful reproduction was found. There was little evidence of survival of smaller Apache trout in areas of co-occurrence with non-native trouts but there was evidence of their survival upstream where fewer numbers of non-natives occur. The exclusion of smaller Apache trout may be from elimination or emigration. Greater populations of Apache trout may be supported through eradication of non-native trouts, the addition of instream cover, and structures designed to create longer lasting pools and bankcuts.
APA, Harvard, Vancouver, ISO, and other styles
3

Kitcheyan, David Chris. "Population structure of Apache trout (Oncorhynchus apache) in Flash and Squaw creeks on the Fort Apache Indian Reservation, Arizona." Thesis, The University of Arizona, 1999. http://hdl.handle.net/10150/278700.

Full text
Abstract:
In 1995, Squaw and Flash creeks were renovated with Antimycin-A to eradicate non-native fish. After renovation, 129 Apache trout from Flash Creek were introduced into Squaw Creek. Two years later, all size classes were present. Apache trout above a natural barrier on Flash Creek were allowed to repopulate the renovated section. Three years later, 45 Apache trout were found below the natural barrier. In both streams, adults selected deep, slow moving areas. Juveniles selected shallow areas with fast currents. Both size classes selected open areas exposed to sunlight with surface turbulence and other forms of instream cover. Apache and Gila trout were experimentally PIT-tagged in the: (1) pelvic girdle; (2) abdominal cavity; and (3) dorsal musculature to determine the best tagging location. Fish tagged in the dorsal musculature had 0% tag loss and 98% survival. The minimum size Apache and Gila trout could be tagged was 90 mm TL.
APA, Harvard, Vancouver, ISO, and other styles
4

Kona, Srinand. "Compactions in Apache Cassandra : Performance Analysis of Compaction Strategies in Apache Cassandra." Thesis, Blekinge Tekniska Högskola, Institutionen för kommunikationssystem, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-12885.

Full text
Abstract:
Context: The global communication system is in a tremendous growth, leading to wide range of data generation. The Telecom operators in various Telecom Industries, that generate large amount of data has a need to manage these data efficiently. As the technology involved in the database management systems is increasing, there is a remarkable growth of NoSQL databases in the 20th century. Apache Cassandra is an advanced NoSQL database system, which is popular for handling semi-structured and unstructured format of Big Data. Cassandra has an effective way of compressing data by using different compaction strategies. This research is focused on analyzing the performances of different compaction strategies in different use cases for default Cassandra stress model. The analysis can suggest better usage of compaction strategies in Cassandra, for a write heavy workload. Objectives: In this study, we investigate the appropriate performance metrics to evaluate the performance of compaction strategies. We provide the detailed analysis of Size Tiered Compaction Strategy, Date Tiered Compaction Strategy, and Leveled Compaction Strategy for a write heavy (90/10) work load, using default cassandra stress tool. Methods: A detailed literature research has been conducted to study the NoSQL databases, and the working of different compaction strategies in Apache Cassandra. The performances metrics are considered by the understanding of the literature research conducted, and considering the opinions of supervisors and Ericsson’s Apache Cassandra team. Two different tools were developed for collecting the performances of the considered metrics. The first tool was developed using Jython scripting language to collect the cassandra metrics, and the second tool was developed using python scripting language to collect the Operating System metrics. The graphs have been generated in Microsoft Excel, using the values obtained from the scripts. Results: Date Tiered Compaction Strategy and Size Tiered Compaction strategy showed more or less similar behaviour during the stress tests conducted. Level Tiered Compaction strategy has showed some remarkable results that effected the system performance, as compared to date tiered compaction and size tiered compaction strategies. Date tiered compaction strategy does not perform well for default cassandra stress model. Size tiered compaction can be preferred for default cassandra stress model, but not considerable for big data. Conclusions: With a detailed analysis and logical comparison of metrics, we finally conclude that Level Tiered Compaction Strategy performs better for a write heavy (90/10) workload while using default cassandra stress model, as compared to size tiered compaction and date tiered compaction strategies.
APA, Harvard, Vancouver, ISO, and other styles
5

Gröne, Bernhard, Andreas Knöpfel, Rudolf Kugel, and Oliver Schmidt. "The Apache Modeling Project." Universität Potsdam, 2004. http://opus.kobv.de/ubp/volltexte/2009/3314/.

Full text
Abstract:
This document presents an introduction to the Apache HTTP Server, covering both an overview and implementation details. It presents results of the Apache Modelling Project done by research assistants and students of the Hasso–Plattner–Institute in 2001, 2002 and 2003. The Apache HTTP Server was used to introduce students to the application of the modeling technique FMC, a method that supports transporting knowledge about complex systems in the domain of information processing (software and hardware as well). After an introduction to HTTP servers in general, we will focus on protocols and web technology. Then we will discuss Apache, its operational environment and its extension capabilities— the module API. Finally we will guide the reader through parts of the Apache source code and explain the most important pieces.
APA, Harvard, Vancouver, ISO, and other styles
6

Solem, Knut Magne. "Apache Derby som MMDB." Thesis, Norwegian University of Science and Technology, Department of Computer and Information Science, 2007. http://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-10342.

Full text
Abstract:

Apache Derby er en Open Source-database utviklet i Java. Den er designet som en tradisjonell diskbasert database og er optimalisert for diskaksess. Målet med denne oppgaven er å finne måter å øke ytelsen på databaser der vi antar at hele databasen får plass i minnet. Vi ønsker å optimalisere for aksessering av data i minnet istedenfor på disken og på den måten gjøre databasen mer lik MMDB(Main Memory DataBase)-databaser. Dette har vi gjort ved å identifisere og erstatte flere disk-spesifikke moduler i Derby med nye moduler optimalisert for minneaksessering. Samtidig bruker vi eksisterende moduler for å oppnå persistens av dataene. Resultatene viser at ytelsen på leseaksesser i siste stabile versjon av Derby (10.2.2.0) kan økes med 70-200% og skriveaksesser med 20-100% avhengig av hvor mange klienter som benytter databasen samtidig.

APA, Harvard, Vancouver, ISO, and other styles
7

Le, Kevin. "Pipelined Apache HTTP Server." DigitalCommons@CalPoly, 2009. https://digitalcommons.calpoly.edu/theses/69.

Full text
Abstract:
Web servers often become overloaded with irregular surges in web traffic. Several techniques have been explored to cope with these overloads such as distributing load throughout different servers. This thesis presents Pipelined Apache HTTP Server, a modified version of the Apache Software Foundation’s HTTP Server that utilizes a pipelined execution of Apache’s request cycle. We discuss Apache’s original architecture, the modifications necessary for implementation of pipelined execution, and analyze its run time. Ultimately, we hoped to increase throughput of Apache but fall short because of unbalanced request phases and pipelining overhead.
APA, Harvard, Vancouver, ISO, and other styles
8

Chaskda, Adams Adamanyiwa. "Determinants and consequences of territory quality in the Bar-throated Apalis Apalis thoracica." Doctoral thesis, University of Cape Town, 2011. http://hdl.handle.net/11427/10257.

Full text
Abstract:
Includes bibliographical references.
The underlying principles of territoriality propose that competition for resources compels individuals to defend areas (territories) that give them exclusive access to critical resources. The theory suggests that only high-quality individuals succeed in occupying areas rich in these resources with consequent benefits in terms of higher reproductive success and/or chances of survival. The main aims of this study were: i) to determine the ecological factors influencing variation in territory size and quality among Bar-throated Apalises Apalis thoracica; ii) to investigate the consequences of these factors for reproductive effort and reward, pair stability and survival; and iii) to investigate the signalling properties of the species' black breast band (a prominent 'badge' emblazoned across the white underparts) in relation to both male and territory quality.
APA, Harvard, Vancouver, ISO, and other styles
9

Short, Daniel R. "AH-64 Apache cost reduction." Thesis, Monterey, Calif. : Springfield, Va. : Naval Postgraduate School ; Available from National Technical Information Service, 2000. http://handle.dtic.mil/100.2/ADA377413.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Hall, Kenneth Estes. "Apaches and Comanches on Screen." Digital Commons @ East Tennessee State University, 2012. https://dc.etsu.edu/etsu-works/591.

Full text
Abstract:
Excerpt: A generally accurate appraisal of Western films might claim that Indians as hostiles are grouped into one undifferentiated mass. Popular hostile groups include the Sioux (without much differentiation between tribes or bands, the Apaches, and the Comanches).
APA, Harvard, Vancouver, ISO, and other styles
11

Martins, José Carlos Ferreira da Silva. "QoS em servidores HTTP Apache." Master's thesis, Universidade de Évora, 2015. http://hdl.handle.net/10174/17549.

Full text
Abstract:
Os serviços baseados na Internet têm registado um crescimento contínuo e acelerado nos últimos anos, dependendo o seu sucesso, em larga medida, da qualidade de serviço prestada. A sociedade moderna tornou-se fortemente dependente da Internet e dos vários serviços nela disponibilizados. Nesta dissertação é abordado o problema da qualidade de serviço em servidores HTTP, com particular ênfase no servidor HTTP Apache. Neste trabalho, é definido e implementado um sistema original de controlo em cadeia fechada de QoS, baseado nas metodologias da teoria de controlo, demonstrando-se a sua robustez, estabilidade e capacidade para gerir os recursos de forma dinâmica. Por fim, é comprovada a capacidade do sistema realizar a diferenciação de serviço (DiffServ) entre duas classes de sítios: Premium e Outros. Esta qualidade é comprovada para os objetivos erro nulo, rejeição de perturbações e seguimento de referência; Abstract: QoS for Apache HTTP server Internet-based services have registered a continuous and rapid growth in recent years, depending their success to a large extent on the provided quality of service. Modern society has become heavily dependent on the Internet and the various services it provides. This thesis addresses the issue of quality of service in HTTP servers, with particular emphasis on the Apache HTTP server. In this work, a closed-loop system with QoS, based on control theory methodologies, is defined and implemented. The proposed system robustness, stability and ability to manage resources dynamically is shown. Finally, it is confirmed the system’s ability to provide differentiated services (DiffServ) between two classes of sites: Premium and Others. This quality is proven to the objectives regulatory control, disturbance rejection and time-varying reference.
APA, Harvard, Vancouver, ISO, and other styles
12

Brotánek, Jan. "Apache Hadoop jako analytická platforma." Master's thesis, Vysoká škola ekonomická v Praze, 2017. http://www.nusl.cz/ntk/nusl-358801.

Full text
Abstract:
Diploma Thesis focuses on integrating Hadoop platform into current data warehouse architecture. In theoretical part, properties of Big Data are described together with their methods and processing models. Hadoop framework, its components and distributions are discussed. Moreover, compoments which enables end users, developers and analytics to access Hadoop cluster are described. Case study of batch data extraction from current data warehouse on Oracle platform with aid of Sqoop tool, their transformation in relational structures of Hive component and uploading them back to the original source is being discussed at practical part of thesis. Compression of data and efficiency of queries depending on various storage formats is also discussed. Quality and consistency of manipulated data is checked during all phases of the process. Fraction of practical part discusses ways of storing and capturing stream data. For this purposes tool Flume is used to capture stream data. Further this data are transformed in Pig tool. Purpose of implementing the process is to move part of data and its processing from current data warehouse to Hadoop cluster. Therefore process of integration of current data warehouse and Hortonworks Data Platform and its components, was designed
APA, Harvard, Vancouver, ISO, and other styles
13

Ferreira, Viviane Gonçalves. "Governança colaborativa na prática : uma análise das experiências nas APACs." reponame:Repositório Institucional do FGV, 2015. http://hdl.handle.net/10438/13627.

Full text
Abstract:
Submitted by Viviane Oliveira (vivianegferreira@gmail.com) on 2015-04-06T17:42:53Z No. of bitstreams: 1 Dissertacao_VivianeFerreira2015.pdf: 3501523 bytes, checksum: 4c3301dfd705132092878838204cab6e (MD5)
Approved for entry into archive by ÁUREA CORRÊA DA FONSECA CORRÊA DA FONSECA (aurea.fonseca@fgv.br) on 2015-04-08T14:36:11Z (GMT) No. of bitstreams: 1 Dissertacao_VivianeFerreira2015.pdf: 3501523 bytes, checksum: 4c3301dfd705132092878838204cab6e (MD5)
Approved for entry into archive by Marcia Bacha (marcia.bacha@fgv.br) on 2015-04-14T12:50:07Z (GMT) No. of bitstreams: 1 Dissertacao_VivianeFerreira2015.pdf: 3501523 bytes, checksum: 4c3301dfd705132092878838204cab6e (MD5)
Made available in DSpace on 2015-04-14T12:50:19Z (GMT). No. of bitstreams: 1 Dissertacao_VivianeFerreira2015.pdf: 3501523 bytes, checksum: 4c3301dfd705132092878838204cab6e (MD5) Previous issue date: 2015-02-25
O ciclo das políticas públicas brasileiro é, tradicionalmente, traçado a partir do modelo descendente ou top-down. Neste modelo, as decisões políticas são tomadas a partir de concepções limitadas de um grupo seleto de pessoas. Estas decisões, a pesar de afetarem diretamente à sociedade, não se detêm exclusivamente à sua demanda. O conceito de Governança Colaborativa repensa o verdadeiro papel da sociedade, sugerindo uma posição mais ativa da sociedade, que deixa de se limitar à posição apenas de 'beneficiária' das políticas públicas. Dentre as diversas formas de participação da sociedade (participação na elaboração de políticas através de comitês, reivindicações através de movimentos, manifestações, etc.) este trabalho tem como foco a participação na aplicação de uma política já existente, como é o caso da APAC (Associação de Proteção e Assistência aos Condenados). Diante dos problemas identificados pela ineficácia da ressocialização nos presídios comuns do Brasil, as APACs surgem como proposta de parceria, visando à humanização do cumprimento das penas, e oferecendo ao indivíduo maiores condições de recuperar-se, resultando em uma melhor inserção na sociedade. Desta forma, o presente trabalho tem, como objetivo final, a análise comparativa do método APAC. Para tanto, foi utilizado o método qualitativo através da análise comparativa entre dois presídios semelhantes em características, tais como regimes existentes, sexo dos detentos, Estado em que se localizam, sendo que, um destes apresenta o método APAC e outro não. A partir de entrevistas, foram identificadas as características dos detentos, seus comportamentos e perspectivas. Em um primeiro momento, foram comparados índices de reincidência penal de ambos os casos. Em segundo momento, foram analisados os índices de fugas em saída temporária com e sem a presença da metodologia da APAC. Diante dos dados levantados, observou-se aspectos positivos no método APAC, portanto, apesar de ser clara a aceitação e aprovação dos envolvidos, conclui-se que os dados ainda são escassos, ou seja, ainda existe uma carência de informações capazes de embasar uma 'justificativa estratégica' do projeto APAC. Desta forma, ao final da pesquisa, são definidas dimensões de indicadores com objetivo de propor futuras pesquisas sobre a metodologia de avaliação de impacto das APACs nas políticas públicas de segurança.
APA, Harvard, Vancouver, ISO, and other styles
14

Laluk, Nicholas. "An Integrative Approach to Interpretations of an Historical-Period Apache Scout Camp at Fort Apache, Arizona." Thesis, The University of Arizona, 2006. http://hdl.handle.net/10150/193260.

Full text
Abstract:
With the encroachment of the United States military onto Apache lands many Apache men joined the military due to intolerable reservation conditions and the unique economic opportunity of enlisting as scouts for the military. This thesis attempts to better understand the relationships among military personnel, Apache scouts, and nonmilitary Apache people. By examining the material remains of a scout camp located on the Fort Apache Indian Reservation (FAIR), and integrating these findings with oral history and information collected from White Mountain Apache consultants, a better understanding of historical Western Apache life can be delineated. This thesis examines these lifeways and interactions by applying a theoretical framework adopted from Steven Silliman's practical politics, Richard White's concept of the middle ground, and Western Apache landscape knowledge and stories.
APA, Harvard, Vancouver, ISO, and other styles
15

Visnes, Snorre. "Skalering av leseoperasjoner med Apache Derby." Thesis, Norwegian University of Science and Technology, Department of Computer and Information Science, 2007. http://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-10330.

Full text
Abstract:

Oppgaven tar for seg hvordan det er mulig å lage et cluster basert på Derby som støtter et høyt volum av lesetransaksjoner. Skriving er ikke i fokus ytelsesmessig, men er mulig gjennom Derbys støtte for XA. Det faktum at XA er et verktøy for å gjennomføre 2-fase commit, ikke replisering, gjør at skriving kun er mulig for administrator. Hovedsakelig grunnet manglende sekvensering av transaksjoner, samt mangel på automatisk opprydning etter feilede transaksjoner. Testing viser at skaleringsgraden for et slikt system er på 100%. Det er ingen sammenkobling mellom servernoder, og dermed ingen øvre grense for antall noder. Det at det ikke er noen sammenkobling mellom servernoder gjør at disse kan spres geografisk. Sammen med en fail-over mekanisme i klienten kan dette systemet oppnå høy tilgjengelighet ved lesing.

APA, Harvard, Vancouver, ISO, and other styles
16

McCoy, Jan. "Apache Cows Get the Once-Over." College of Agriculture, University of Arizona (Tucson, AZ), 1990. http://hdl.handle.net/10150/295617.

Full text
APA, Harvard, Vancouver, ISO, and other styles
17

Lind, Simon. "Distributed Ensemble Learning With Apache Spark." Thesis, Uppsala universitet, Institutionen för biologisk grundutbildning, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-274323.

Full text
APA, Harvard, Vancouver, ISO, and other styles
18

Yu, Lucy M. Eng Massachusetts Institute of Technology. "Work-sharing framework for Apache Spark." Thesis, Massachusetts Institute of Technology, 2016. http://hdl.handle.net/1721.1/113441.

Full text
Abstract:
Thesis: M. Eng., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2016.
This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
Cataloged from student-submitted PDF version of thesis.
Includes bibliographical references (page 39).
Apache Spark is a popular framework for distributed data processing that generalizes the MapReduce model and significantly improves the performance of many use cases. People can use Spark to query enormous data sets faster than before to gain insights for a competitive edge in industry. Often these ad-hoc queries perform similar work, and there is an opportunity to share the work of different queries. This can reduce the total computation time even more. We have developed a Wrapper class which performs such optimizations. In particular, its strategy of lazy evaluation allows duplicate computation to be avoided and multiple related Spark jobs to be executed at the same time, reducing the scheduling overhead. Overall, the system demonstrates significant efficiency gains when compared to default Spark.
by Lucy Yu.
M. Eng.
APA, Harvard, Vancouver, ISO, and other styles
19

Brundin, Michelle. "Data Stream Queries to Apache SPARK." Thesis, Uppsala universitet, Institutionen för informationsteknologi, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-301326.

Full text
Abstract:
Many fields have a need to process and analyze data streams in real-time. In industrial applications the data can come from big sensor networks, where the processing of the data streams can be used for performance monitoring and fault detection in real time. Another example is in social media where data stream processing can be used to detect and prevent spam. A data stream management system (DSMS) is a system that can be used to manage and query continuously received data streams. The queries a DSMS executes are called continuous queries (CQs). In contrast to regular database queries they execute continuously until canceled. SCSQ is a DSMS developed at Uppsala university. Apache Spark is a large scale general data processing engine. It has, among other things, a component for data stream processing, Spark Streaming. In this project a system called SCSQ Spark Streaming Interface (SSI) was implemented that allows Spark Streaming applications to be called from CQs in SCSQ. It allows the Spark Streaming applications to receive input streams from SCSQ as well as emitting resulting stream elements back to SCSQ. To demonstrate SSI, two examples are shown where it is used for stream clustering in CQs using the streaming k-means implementation in Spark Streaming.
APA, Harvard, Vancouver, ISO, and other styles
20

AL, Jorani Salam. "Performance assessment of Apache Spark applications." Thesis, Linnéuniversitetet, Institutionen för datavetenskap och medieteknik (DM), 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:lnu:diva-80181.

Full text
Abstract:
This thesis addresses the challenges of large software and data-intensive systems. We will discuss a Big Data software that consists of quite a bit of Linux configuration, some Scala coding and a set of frameworks that work together to achieve the smooth performance of the system. Moreover, the thesis focuses on the Apache Spark framework and the challenging of measuring the lazy evaluation of the transformation operations of Spark. Investigating the challenges are essential for the performance engineers to increase their ability to study how the system behaves and take decisions in early design iteration. Thus, we made some experiments and measurements to achieve this goal. In addition to that, and after analyzing the result we could create a formula that will be useful for the engineers to predict the performance of the system in production.
APA, Harvard, Vancouver, ISO, and other styles
21

Cassales, Guilherme Weigert. "Escalonamento adaptativo para o Apache Hadoop." Universidade Federal de Santa Maria, 2016. http://repositorio.ufsm.br/handle/1/12025.

Full text
Abstract:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior - CAPES
Many alternatives have been employed in order to process all the data generated by current applications in a timely manner. One of these alternatives, the Apache Hadoop, combines parallel and distributed processing with the MapReduce paradigm in order to provide an environment that is able to process a huge data volume using a simple programming model. However, Apache Hadoop has been designed for dedicated and homogeneous clusters, a limitation that creates challenges for those who wish to use the framework in other circumstances. Often, acquiring a dedicated cluster can be impracticable due to the cost, and the acquisition of reposition parts can be a threat to the homogeneity of a cluster. In these cases, an option commonly used by the companies is the usage of idle computing resources in their network, however the original distribution of Hadoop would show serious performance issues in these conditions. Thus, this study was aimed to improve Hadoop’s capacity of adapting to pervasive and shared environments, where the availability of resources will undergo variations during the execution. Therefore, context-awareness techniques were used in order to collect information about the available capacity in each worker node and distributed communication techniques were used to update this information on scheduler. The joint usage of both techniques aimed at minimizing and/or eliminating the overload that would happen on shared nodes, resulting in an improvement of up to 50% on performance in a shared cluster, when compared to the original distribution, and indicated that a simple solution can positively impact the scheduling, increasing the variety of environments where the use of Hadoop is possible.
Diversas alternativas têm sido empregadas para o processamento, em tempo hábil, da grande quantidade de dados que é gerada pelas aplicações atuais. Uma destas alternativas, o Apache Hadoop, combina processamento paralelo e distribuído com o paradigma MapReduce para fornecer um ambiente capaz de processar um grande volume de informações através de um modelo de programação simplificada. No entanto, o Apache Hadoop foi projetado para utilização em clusters dedicados e homogêneos, uma limitação que gera desafios para aqueles que desejam utilizá-lo sob outras circunstâncias. Muitas vezes um cluster dedicado pode ser inviável pelo custo de aquisição e a homogeneidade pode ser ameaçada devido à dificuldade de adquirir peças de reposição. Em muitos desses casos, uma opção encontrada pelas empresas é a utilização dos recursos computacionais ociosos em sua rede, porém a distribuição original do Hadoop apresentaria sérios problemas de desempenho nestas condições. Sendo assim, este estudo propôs melhorar a capacidade do Hadoop em adaptar-se a ambientes, pervasivos e compartilhados, onde a disponibilidade de recursos sofrerá variações no decorrer da execução. Para tanto, utilizaram-se técnicas de sensibilidade ao contexto para coletar informações sobre a capacidade disponível nos nós trabalhadores e técnicas de comunicação distribuída para atualizar estas informações no escalonador. A utilização conjunta dessas técnicas teve como objetivo a minimização e/ou eliminação da sobrecarga que seria causada em nós com compartilhamento, resultando em uma melhora de até 50% no desempenho em um cluster compartilhado, quando comparado com a distribuição original, e indicou que uma solução simples pode impactar positivamente o escalonamento, aumentando a variedade de ambientes onde a utilização do Hadoop é possível.
APA, Harvard, Vancouver, ISO, and other styles
22

Mohyedin, Kermani Ehsan. "Distributed linear programming with Apache Spark." Thesis, University of British Columbia, 2016. http://hdl.handle.net/2429/59990.

Full text
Abstract:
For this thesis project, we have implemented Mehrotra's predictor-corrector interior point algorithm on top of Apache Spark for solving large-scale linear programming problems. Our large-scale solver (Spark-LP) is unique because it is open-source, fault-tolerant and can be used on commodity cluster of machines. As a result, Spark-LP provides an opportunity to solve large-scale problems at the lowest possible cost. We have assessed the performance and convergent results of our solver on self-generated, sparse and dense large-scale problems over small to medium-sized clusters, composed of 16 to 64 Amazon's Elastic Computing Cloud r3.xlarge instances. In conclusions, we have made important suggestions for breaking the current structural limitations so that our solver can be used on heterogeneous clusters containing CPUs and GPUs on JVM environment without the usual numerical limitations and overheads.
Science, Faculty of
Computer Science, Department of
Graduate
APA, Harvard, Vancouver, ISO, and other styles
23

Svantesson, Carlhåkan. "Webbserverprogram: Öppen källkods-alternativ till Apache." Thesis, Högskolan i Skövde, Institutionen för kommunikation och information, 2012. http://urn.kb.se/resolve?urn=urn:nbn:se:his:diva-6136.

Full text
Abstract:
Det har blivit allt vanligare för företag att marknadsföra sig via Internet vilket oftast innebär att företaget behöver en webbplats. Denna webbplats använder ett webbserverprogram för att hantera kundernars förfrågningar och det webbserverprogram som är störst på marknaden med god marginal är Apache. Apache har existerat i över 15 år och är öppen-källkod. Det här examensarbetet undersöker om det finns några öppen källkods-alternativ till det marknadsledande webbserverprogrammet Apache genom att titta på funktionalitet och prestanda. Prestandatesterna har genomförs både med statiska och dynamiska webbsidor. De alternativ som undersöks är Nginx och Lighttpd. Resultaten visar på att både Nginx och Lighttpd i det stora hela presterar bättre än Apache. Det här syns främst i de statiska prestandatesterna där Nginx och Lighttpd presterar mer än dubbelt så bra som Apache. I de dynamiska prestandatesterna så har Nginx och Apache jämförbar prestanda medan Lighttpd inte riktigt kommer upp i samma prestanda. Nginx saknar viss funktionalitet i jämförelse med de andra två, det är dock inga kritiska funktioner som saknas.
APA, Harvard, Vancouver, ISO, and other styles
24

Beligianni, Foteini. "Streaming Predictive Analytics on Apache Flink." Thesis, KTH, Skolan för informations- och kommunikationsteknik (ICT), 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-171355.

Full text
Abstract:
Data analysis and predictive analytics today are driven by large scale dis- tributed deployments of complex pipelines, guiding data cleaning, model training and evaluation. A wide range of systems and tools provide the basic abstractions for building such complex pipelines for offline data processing, however, there is an increasing demand for providing support for incremental models over unbounded streaming data. In this work, we focus on the prob- lem of modelling such a pipeline framework and providing algorithms that build on top of basic abstractions, fundamental to stream processing. We design a streaming machine learning pipeline as a series of stages such as model building, concept drift detection and continuous evaluation. We build our prototype on Apache Flink, a distributed data processing system with streaming capabilities along with a state-of-the-art implementation of a varia- tion of Vertical Hoeffding Tree (VHT), a distributed decision tree classification algorithm as a proof of concept. Furthermore, we compare our version of VHT with the current state-of- the-art implementations on distributed data processing systems in terms of performance and accuracy. Our experimental results on real-world data sets show significant performance benefits of our pipeline while maintaining low classification error. We believe that this pipeline framework can offer a good baseline for a full-fledged implementation of various streaming algorithms that can work in parallel.
Dataanalys och predictive analytics drivs idag av storskaliga distribuerade distributioner av komplexa pipelines, guiding data cleaning, model training och utvärdering. Ett brett utbud av system och verktyg ger endast grundläggande abstractions (struktur) för att bygga sådana komplexa pipelines för databehandling i off-line läge, men det finns en ökande efterfrågan att tillhandahålla stöd för stegvis modell över unbounded streaming data. I detta arbete fokuserar vi på problemet med modellering som ramverket för pipeline och ger algoritmer som bygger på grundläggande abstraktioner för stream processing. Vi konstruerar en streaming maskininlärnings pipeline som innehåller steg som model building, concept drift detection och kontinuerlig utvärdering. Vi bygger vår prototyp på Apache Flink, ett distribuerat databehandlingssystem med strömnings kapacitet tillsammans med den bästa tillgängliga implementation av en Vertical Hoeffding Tree (VHT) variant och ett distribuerat beslutsträd algoritm som koncepttest. Dessutom jämför vi vår version av VHT med den senaste tekniken inom destributed data processing systems i termer av prestanda och precision. Vårt experimentella resultaten visar betydande fördelarna med vår pipeline och samtidigt bibehållen låg klassificerat felet. Vi anser att detta ramverk kan erbjuda en bra utgångspunkt vid genomförandet av olika streaming algoritmer som kan arbeta parallellt.
APA, Harvard, Vancouver, ISO, and other styles
25

Rexa, Denis. "Výpočetní úlohy pro řešení paralelního zpracování dat." Master's thesis, Vysoké učení technické v Brně. Fakulta elektrotechniky a komunikačních technologií, 2019. http://www.nusl.cz/ntk/nusl-400899.

Full text
Abstract:
The goal of this diploma thesis was to create four laboratory exercises for the subject "Parallel Data Processing", where students will try on the options and capabilities of Apache Spark as a parallel computing platform. The work also includes basic setup and use of Apache Kafka technology and NoSQL Apache Cassandra database. The other two lab assignments focus on working with a Travelling Salesman Problem. The first lab was designed to demonstrate the difficulty of a task where the student will face an exponential increase in complexity. The second task consists of an optimization algorithm to solve the problem in cluster. This algorithm is subjected to performance measurements in clusters. The conclusion of the thesis contains recommendations for optimization as well as comparison of running with different number of computing devices.
APA, Harvard, Vancouver, ISO, and other styles
26

Righi, Massimo. "apache cassandra: studio ed analisi di prestazioni." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2018. http://amslaurea.unibo.it/16713/.

Full text
Abstract:
La tesi persegue due scopi: il primo è veri�care sperimentalmente l'incremento di performance che si può avere nell'utilizzare un database di tipo NoSql (MongoDB e Cassandra) rispetto a uno relazionale (MYSQL) lavorando con grandi quantità di dati. Il secondo, invece, si propone di confrontare (sempre sperimentalmente) l'utilizzo di due tipi di database della famiglia dei NoSql, ovvero MongoDB, basato su un architettura Document-oriented, e Cassandra (in forte sviluppo), della famiglia dei Column-oriented. Nella prima parte della tesi verranno esposte le caratteristiche tecniche delle due famiglie di database utilizzate: relazionale e NoSql. In particolare, per il gruppo NoSql, verrà fornita una descrizione più dettagliata, classi�ficando i sottogruppi che ne fanno parte in base al tipo di modello che utilizzano per la memorizzazione dei dati. Verrà posta molta attenzione sul database Cassandra. Nella seconda parte verranno effettuate le analisi tecniche di performance per ogni tipo di database basandosi sulle 4 operazioni CRUD, partendo da un caso d'uso da me scelto. Al termine delle analisi verranno stilate delle conclusioni relativamente a pregi e difetti di ognuna delle tre tecnologie.
APA, Harvard, Vancouver, ISO, and other styles
27

Holum, Henrik, and Svein Erik Reknes Løvland. "Joining in Apache Derby: Removing the Obstacles." Thesis, Norwegian University of Science and Technology, Department of Computer and Information Science, 2008. http://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-9717.

Full text
Abstract:

Over the last decade, the amount of commercial interest in Open Source has been growing rapidly. This has led to commercially driven Open Source projects. Those projects have problems keeping their newcomers and needs ways to ease the joining process. Therefore we ask these research questions: RQ1: Which obstacles are encountered by Newcomers to Apache Derby when Joining? RQ2: What can be done to ease the Joining process? There has been very little research on what the OSS projects can do in this area. As a consequence it is hard to find good reliable theory to cross-reference this research. If the research is successful, it can contribute to the literature on joining in OSS projects. This literature will then contain all obstacles encountered by newcomers to OSS projects and ways to mitigate these. In this master's thesis Canonical Action Research was used to study the Open Source project Apache Derby. Canonical Action Research is a qualitative research method where the researchers enters the environment they are researching to extract the data needed. We have three contributions in this thesis. The first contribution is a list of obstacles in the joining process of Apache Derby. The second contribution is suggestions on how a project can mitigate the contribution barriers we found. The third contribution is a refined version of CAR to use when studying Open Source Software Development. The list of obstacles is a contribution specific to the Apache Derby project, and it is very unlikely that other non Apache projects will benefit from it. Our suggestions on how a project can mitigate contribution barriers are potentially generalizable. Different projects have different structures, and some of the contribution barriers might therefore not apply to them all. The refined CAR model is general for all research on OSS projects. This is the result we think can have the biggest impact on the research community if proven successful.

APA, Harvard, Vancouver, ISO, and other styles
28

Tuttle, Sabrina. "The San Carlos Apache Reservation Quick Facts." College of Agriculture and Life Sciences, University of Arizona (Tucson, AZ), 2008. http://hdl.handle.net/10150/144740.

Full text
APA, Harvard, Vancouver, ISO, and other styles
29

Hsu, Shyh-Shyan 1963. "Laboratory shear tests on Apache Leap tuff." Thesis, The University of Arizona, 1990. http://hdl.handle.net/10150/291721.

Full text
Abstract:
The laboratory shear tests on Apache Leap tuff have been conducted to investigate the joint dilatant behavior and the influence of surface roughness, sample size, water and shear velocity on joint shear behavior, and to compare the fit between Coulomb's linear and Archard's non-linear model. Experimental results indicate that the dilation of rock joint is influenced by joint surface roughness and normal stress. Surface roughness also determines joint shear strength. Sample scale influences the following joint properties, e.g. peak shear strength, peak dilation angle, peak shear displacement and mode of shear failure. This effect depends on normal stress. The shear strength of rock joint also varies with the presence of water and shear velocity. Archard's power law generally gives a better fit on the experimental data than does Coulomb's linear law.
APA, Harvard, Vancouver, ISO, and other styles
30

Kolb, Pascal. "Realization of EAI patterns with Apache Camel." Stuttgart : Universitätsbibliothek der Universität Stuttgart, 2008. http://nbn-resolving.de/urn:nbn:de:bsz:93-opus-35200.

Full text
APA, Harvard, Vancouver, ISO, and other styles
31

Dahal, Janak. "Assessing Apache Spark Streaming with Scientific Data." ScholarWorks@UNO, 2018. https://scholarworks.uno.edu/td/2506.

Full text
Abstract:
Processing real-world data requires the ability to analyze data in real-time. Data processing engines like Hadoop come short when results are needed on the fly. Apache Spark's streaming library is increasingly becoming a popular choice as it can stream and analyze a significant amount of data. To showcase and assess the ability of Spark various metrics were designed and operated using data collected from the USGODAE data catalog. The latency of streaming in Apache Spark was measured and analyzed against many nodes in the cluster. Scalability was monitored by adding and removing nodes in the middle of a streaming job. Fault tolerance was verified by stopping nodes in the middle of a job and making sure that the job was rescheduled and completed on other node/s. A full stack application was designed that would automate data collection, data processing and visualizing the results. Google Maps API was used to visualize results by color coding the world map with values from various analytics.
APA, Harvard, Vancouver, ISO, and other styles
32

Gonzalez, Lopez Jorge. "Distributed multi-label learning on Apache Spark." VCU Scholars Compass, 2019. https://scholarscompass.vcu.edu/etd/5775.

Full text
Abstract:
This thesis proposes a series of multi-label learning algorithms for classification and feature selection implemented on the Apache Spark distributed computing model. Five approaches for determining the optimal architecture to speed up multi-label learning methods are presented. These approaches range from local parallelization using threads to distributed computing using independent or shared memory spaces. It is shown that the optimal approach performs hundreds of times faster than the baseline method. Three distributed multi-label k nearest neighbors methods built on top of the Spark architecture are proposed: an exact iterative method that computes pair-wise distances, an approximate tree-based method that indexes the instances across multiple nodes, and an approximate local sensitive hashing method that builds multiple hash tables to index the data. The results indicated that the predictions of the tree-based method are on par with those of an exact method while reducing the execution times in all the scenarios. The aforementioned method is then used to evaluate the quality of a selected feature subset. The optimal adaptation for a multi-label feature selection criterion is discussed and two distributed feature selection methods for multi-label problems are proposed: a method that selects the feature subset that maximizes the Euclidean norm of individual information measures, and a method that selects the subset of features maximizing the geometric mean. The results indicate that each method excels in different scenarios depending on type of features and the number of labels. Rigorous experimental studies and statistical analyses over many multi-label metrics and datasets confirm that the proposals achieve better performances and provide better scalability to bigger data than the methods compared in the state of the art.
APA, Harvard, Vancouver, ISO, and other styles
33

Béder, Michal. "Zpracování síťové komunikace v prostředí Apache Spark." Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2018. http://www.nusl.cz/ntk/nusl-385893.

Full text
Abstract:
The aim of this thesis is to show how to design and implement an application for network traces analysis using Apache Spark distributed system. Implementation can be divided into three parts - loading data from a distributed HDFS storage, supported network protocols analysis and distributed data processing. As a data visualization tool is used web-based notebook Apache Zeppelin. The resulting application is able to analyze individual packets as well as the entire flows. It supports JSON and pcap as input data formats. The goal of the application is to allow Big Data processing. The greatest impact on its performance has the input data format and allocation of the available cores.
APA, Harvard, Vancouver, ISO, and other styles
34

Safari, Arash. "Matrix Multiplications on Apache Spark through GPUs." Thesis, KTH, Skolan för datavetenskap och kommunikation (CSC), 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-210891.

Full text
Abstract:
In this report, we consider the distribution of large scale matrix multiplications across a group of systems through Apache Spark, where each individual system utilizes Graphical Processor Units (GPUs) in order to perform the matrix multiplication. The purpose of this thesis is to research whether the GPU's advantage in performing parallel work can be applied to a distributed environment, and whether it scales noticeably better than a CPU implementation in a distributed environment. This question was resolved by benchmarking the different implementations at their peak. Based on these benchmarks, it was concluded that GPUs indeed do perform better as long as single precision support is available in the distributed environment. When single precision operations are not supported, GPUs perform much worse due to the low double precision performance of most GPU devices.
I denna rapport betraktar vi fördelningen av storskaliga matrismultiplikationeröver ett Apache Spark kluster, där varje system i klustret delegerar beräkningarnatill grafiska processorenheter (GPU). Syftet med denna avhandling är attundersöka huruvida GPU:s fördel vid parallellt arbete kan tillämpas på en distribuerad miljö, och om det skalar märkbart bättre än en CPU-implementationi en distribuerad miljö. Detta gjordes genom att testa de olika implementationerna i en miljö däroptimal prestanda kunde förväntas. Baserat på resultat ifrån dessa tester drogsslutsatsen att GPU-enheter preseterar bättre än CPU-enheter så länge ramverkethar stöd för single precision beräkningar. När detta inte är fallet så presterar deflesta GPU-enheterna betydligt sämre på grund av deras låga double-precisionprestanda.
APA, Harvard, Vancouver, ISO, and other styles
35

Furini, Michele. "Tecnologie per l’analisi in tempo reale di Big Data: prestazioni a confronto." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2016.

Find full text
Abstract:
Lo scopo di questo l'elaborato è l'analisi,lo studio e il confronto delle tecnologie per l'analisi in tempo reale di Big Data: Apache Spark Streaming, Apache Storm e Apache Flink. Per eseguire un adeguato confronto si è deciso di realizzare un sistema di rilevamento e riconoscimento facciale all’interno di un video, in maniera da poter parallelizzare le elaborazioni necessarie sfruttando le potenzialità di ogni architettura. Dopo aver realizzato dei prototipi realistici, uno per ogni architettura, si è passati alla fase di testing per misurarne le prestazioni. Attraverso l’impiego di cluster appositamente realizzati in ambiente locale e cloud, sono state misurare le caratteristiche che rappresentavano, meglio di altre, le differenze tra le architetture, cercando di dimostrarne quantitativamente l’efficacia degli algoritmi utilizzati e l’efficienza delle stesse. Si è scelto quindi il massimo input rate sostenibile e la latenza misurate al variare del numero di nodi. In questo modo era possibile osservare la scalabilità di architettura, per analizzarne l’andamento e verificare fino a che limite si potesse giungere per mantenere un compromesso accettabile tra il numero di nodi e l’input rate sostenibile. Gli esperimenti effettuati hanno mostrato che, all’aumentare del numero di worker le prestazioni del sistema migliorano, rendendo i sistemi studiati adatti all’utilizzo su larga scala. Inoltre sono state rilevate sostanziali differenze tra i vari framework, riportando pro e contro di ognuno, cercando di evidenziarne i più idonei al caso di studio.
APA, Harvard, Vancouver, ISO, and other styles
36

Tuttle, Sabrina. "The San Carlos Apache Reservation and Extension Programs." College of Agriculture and Life Sciences, University of Arizona (Tucson, AZ), 2008. http://hdl.handle.net/10150/144739.

Full text
Abstract:
4 pp.
This fact sheet describes the socioeconomic and cultural aspects of the San Carlos Apache reservation, as well as the history of extension and effective extension programs and collaborations conducted on this reservation.
APA, Harvard, Vancouver, ISO, and other styles
37

Casey, Walker Evan. "Scalable Collaborative Filtering Recommendation Algorithms on Apache Spark." Scholarship @ Claremont, 2014. http://scholarship.claremont.edu/cmc_theses/873.

Full text
Abstract:
Collaborative filtering based recommender systems use information about a user's preferences to make personalized predictions about content, such as topics, people, or products, that they might find relevant. As the volume of accessible information and active users on the Internet continues to grow, it becomes increasingly difficult to compute recommendations quickly and accurately over a large dataset. In this study, we will introduce an algorithmic framework built on top of Apache Spark for parallel computation of the neighborhood-based collaborative filtering problem, which allows the algorithm to scale linearly with a growing number of users. We also investigate several different variants of this technique including user and item-based recommendation approaches, correlation and vector-based similarity calculations, and selective down-sampling of user interactions. Finally, we provide an experimental comparison of these techniques on the MovieLens dataset consisting of 10 million movie ratings.
APA, Harvard, Vancouver, ISO, and other styles
38

Pontius, Brandon H. "Information security considerations for applications using Apache Accumulo." Thesis, Monterey, California: Naval Postgraduate School, 2014. http://hdl.handle.net/10945/43980.

Full text
Abstract:
Approved for public release; distribution is unlimited
NoSQL databases are gaining popularity due to their ability to store and process large heterogeneous data sets more efficiently than relational databases. Apache Accumulo is a NoSQL database that introduced a unique information security feature—cell-level access control. We study Accumulo to examine its cell-level access control policy enforcement mechanism. We survey existing Accumulo applications, focusing on Koverse as a case study to model the interaction between Accumulo and a client application. We conclude with a discussion of potential security concerns for Accumulo applications. We argue that Accumulo’s cell-level access control can assist developers in creating a stronger information security policy, but Accumulo cannot provide security—particularly enforcement of information flow policies—on its own. Furthermore, popular patterns for interaction between Accumulo and its clients require diligence on the part of developers, which may otherwise lead to unexpected behavior that undermines system policy. We highlight some undesirable but reasonable confusions stemming from the semantic gap between cell-level and table-level policies, and between policies for end-users and Accumulo clients.
APA, Harvard, Vancouver, ISO, and other styles
39

Heifetz, Alexander G. (Alexander Garon). "Shade : a differentially private wrapper around Apache Spark." Thesis, Massachusetts Institute of Technology, 2017. http://hdl.handle.net/1721.1/119522.

Full text
Abstract:
Thesis: M. Eng., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2017.
This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
Cataloged from student-submitted PDF version of thesis.
Includes bibliographical references (pages 85-88).
Enterprises usually provide strong controls to prevent external cyberattacks and inadvertent leakage of data to external entities. However, in the case where employees and data scientists have legitimate access to analyze and derive insights from the data, there are insufficient controls and employees are usually permitted access to all information about the customers of the enterprise including sensitive and private information. Though it is important to be able to identify useful patterns of one's customers for better customization and service, customers' privacy must not be sacrificed to do so. We propose an alternative - a framework that will allow privacy preserving data analytics over big data. In this paper, we present an efficient and scalable framework for Apache Spark, a cluster computing framework, that provides strong privacy guarantees for users even in the presence of an informed adversary, while still providing high utility for analysts in an interactive wrapper. The framework, titled Shade, includes two mechanisms - SparkLAP, which provides Laplacian perturbation based on a user's query and SparkSAM, which uses the contents of the database itself in order to calculate the perturbation. We show that performance of Shade is substantially better than earlier differential privacy systems without loss of accuracy, particularly when run on datasets small enough to fit in memory, and find that SparkSAM can even exceed performance of an identical non-private Spark query.
by Alexander G. Heifetz.
M. Eng.
APA, Harvard, Vancouver, ISO, and other styles
40

Kotlář, Pavel. "Možnosti optimalizace výkonu LAMP (linux/apache/mysql/php)." Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2009. http://www.nusl.cz/ntk/nusl-236749.

Full text
Abstract:
This work deals with topic of LAMP software bundle performance optimalization. Step by step, it tries to discover performance problems in all four parts of LAMP (in Linux, HTTP server Apache, MySQL database and PHP language interpreter). A model web application is created for these testing purposes. When a problem is found, a change in configuration files is done or a performance improving technology is applied to the corresponding part. A set of optimalization recommendations is compiled and verified on server running real web application.
APA, Harvard, Vancouver, ISO, and other styles
41

Kannan, Vijayasarathy. "A Distributed Approach to EpiFast using Apache Spark." Thesis, Virginia Tech, 2015. http://hdl.handle.net/10919/55272.

Full text
Abstract:
EpiFast is a parallel algorithm for large-scale epidemic simulations, based on an interpretation of the stochastic disease propagation in a contact network. The original EpiFast implementation is based on a master-slave computation model with a focus on distributed memory using message-passing-interface (MPI). However, it suffers from few shortcomings with respect to scale of networks being studied. This thesis addresses these shortcomings and provides two different implementations: Spark-EpiFast based on the Apache Spark big data processing engine and Charm-EpiFast based on the Charm++ parallel programming framework. The study focuses on exploiting features of both systems that we believe could potentially benefit in terms of performance and scalability. We present models of EpiFast specific to each system and relate algorithm specifics to several optimization techniques. We also provide a detailed analysis of these optimizations through a range of experiments that consider scale of networks and environment settings we used. Our analysis shows that the Spark-based version is more efficient than the Charm++ and MPI-based counterparts. To the best of our knowledge, ours is one of the preliminary efforts of using Apache Spark for epidemic simulations. We believe that our proposed model could act as a reference for similar large-scale epidemiological simulations exploring non-MPI or MapReduce-like approaches.
Master of Science
APA, Harvard, Vancouver, ISO, and other styles
42

Geirsson, Sigurdur. "Parallel Bayesian Additive Regression Trees, using Apache Spark." Thesis, Uppsala universitet, Institutionen för informationsteknologi, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-322247.

Full text
Abstract:
New methods have been developed to find patterns and trends in order to gainknowledge from large datasets in various disciplines, such as bioinformatics, consumer behavior in advertising and weather forecasting.The goal of many of these new methods is to construct prediction models from the data. Linear regression, which is widely used for analyzing data, is very powerful ford etecting simple patterns, but higher complexity requires a more sophisticated solution. Regression trees split up the problem into numerous parts but they do not generalizewell as they tend to have high variance. Ensemble methods, a collection of regressiontrees, solves that problem by spreading the model over numerous trees. Ensemble methods such as Random Forest, Gradient Boosted Trees and Bayesian Additive Regression Trees, all have different ways to constructing prediction modelfrom data. Using these models for large datasets are computationally demanding.The aim of this work is to explore a parallel implementation of Bayesian Additive Regression Trees (BART) using Apache Spark framework. Spark is ideal in this case asit is great for iterative and data intensive jobs.We show that our parallel implementation is about 35 times faster for a dataset of pig's genomes. Most of the speed improvement is due to serial code modification that minimizes scanning of the data.The gain from parallelization is a speedup of 2.2x, gained by using four cores on aquad core system. Measurements on a computer clusters consisting of four computers resulted in a maximum speedup of 2.1x for eight cores.We should emphasize that these gains are heavily dependent on size of datasets.
APA, Harvard, Vancouver, ISO, and other styles
43

Majewski, Steven R., Ricardo P. Schiavon, Peter M. Frinchaboy, Carlos Allende Prieto, Robert Barkhouser, Dmitry Bizyaev, Basil Blank, et al. "The Apache Point Observatory Galactic Evolution Experiment (APOGEE)." IOP PUBLISHING LTD, 2017. http://hdl.handle.net/10150/625493.

Full text
Abstract:
The Apache Point Observatory Galactic Evolution Experiment (APOGEE), one of the programs in the Sloan Digital Sky Survey III (SDSS-III), has now completed its systematic, homogeneous spectroscopic survey sampling all major populations of the Milky Way. After a three-year observing campaign on the Sloan 2.5 m Telescope, APOGEE has collected a half million high-resolution (R similar to 22,500), high signal-to-noise ratio (>100), infrared (1.51-1.70 mu m) spectra for 146,000 stars, with time series information via repeat visits to most of these stars. This paper describes the motivations for the survey and its overall design-hardware, field placement, target selection, operations-and gives an overview of these aspects as well as the data reduction, analysis, and products. An index is also given to the complement of technical papers that describe various critical survey components in detail. Finally, we discuss the achieved survey performance and illustrate the variety of potential uses of the data products by way of a number of science demonstrations, which span from time series analysis of stellar spectral variations and radial velocity variations from stellar companions, to spatial maps of kinematics, metallicity, and abundance patterns across the Galaxy and as a function of age, to new views of the interstellar medium, the chemistry of star clusters, and the discovery of rare stellar species. As part of SDSS-III Data Release 12 and later releases, all of the APOGEE data products are publicly available.
APA, Harvard, Vancouver, ISO, and other styles
44

Cuevas, Vannesa, Michael Diehl, and Tab Wilcox. "INSTRUMENTING AN ETHERNET AVIONICS BUS FOR APACHE TESTING." International Foundation for Telemetering, 2017. http://hdl.handle.net/10150/626967.

Full text
Abstract:
Last year, the U.S. Army Yuma Proving Ground (YPG) faced a short suspense requirement to instrument an Apache's Ethernet bus. YPG was able to implement an effective method to satisfy this requirement; however, more capable and effective methodologies could not be developed and utilized due to time constraints. While continuing to support ongoing Apache testing, YPG is working to implement more efficient methods to capture and utilize the Ethernet data. This paper will discuss YPG’s initial implementation and the follow-on efforts being pursued.
APA, Harvard, Vancouver, ISO, and other styles
45

Gammill, Troy. "McDonnell Douglas Helicopter Company Apache Telemetry Antenna Analysis." International Foundation for Telemetering, 1989. http://hdl.handle.net/10150/614738.

Full text
Abstract:
International Telemetering Conference Proceedings / October 30-November 02, 1989 / Town & Country Hotel & Convention Center, San Diego, California
The analysis of the AH-64 Apache helicopter telemetry antenna was performed to improve the present telemetry antenna patterns. This telemetry antenna was used by the Mesa, Arizona, branch of McDonnell Douglas helicopter division to transmit real-time data gathered during test flights to a Fixed Base Data System developed by the University of New Mexico's Physical Science Laboratory (PSL) in 1988. This analysis was necessary due to the high data dropout encountered when the helicopter was performing maneuvers on the test range, a potentially dangerous situation. Since the telemetry antenna patterns had never been studied, the PSL was awarded a contract to conduct a study of the telemetry antenna and suggest ways to improve the performance of the present telemetry system. Extensive use of software developed by Ohio State University for the analysis of antennas on aircraft was applied to finding a solution to the problems encountered with the present system. This included analysis of several different configurations of antennas and antenna placement. The result of this analysis was to present McDonnell Douglas with a recommendation of telemetry antenna and telemetry antenna placement that would satisfy the requirements of the current test system in use. This represents a three-month effort by the PSL utilizing advanced antenna analysis techniques.
APA, Harvard, Vancouver, ISO, and other styles
46

Barbone, Paul Joseph. ""We Were Recruited From the Warriors of Many Famous Nations," Cultural Preservation: U.S. Army Western Apache Scouts, 1871-1947." Thesis, The University of Arizona, 2010. http://hdl.handle.net/10150/193387.

Full text
Abstract:
The Western Apache Scouts of the 1870s who assisted the United States Army in tracking down the Chiricahua Apaches that had escaped from the federal reservations in the Arizona Territory laid the foundation for what became seventy-six years of military service in the U.S. Army. Consolidated and reassigned to Ft. Huachuca, Arizona in 1922, these scouts continued to serve with distinction long after the Army needed their skills as trackers. In 1947, the final four scouts retired from United States military service, each having served for over twenty-five years. This thesis explores how these men used their military service in order to survive, serving with honor while maintaining their cultural traditions within a changing world.
APA, Harvard, Vancouver, ISO, and other styles
47

Berni, Mila. "Inclusione di Apache Samza e Kafka nel framework RAM3S." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2021.

Find full text
Abstract:
La rapida diffusione di dispositivi connessi ad Internet e il conseguente aumento della generazione di dati ha portato le piattaforme di data processing a voler sempre di più diminuire i tempi di latenza dell'elaborazione delle informazioni. Esistono vari framework dedicati al real-time processing, tutti con vari pro e contro, dipendenti anche dal tipo di applicazione che si vuole sviluppare. In particolare, il framework RAM3S si basa su Flink, Storm e Spark, tre piattaforme di Apache con caratteristiche differenti ma tutte aderenti al paradigma dello stream processing. Tramite RAM3S lo sviluppatore viene sgravato dall'impegno di dover conoscere approfonditamente i framework prima citati, mettendo a disposizione delle interfacce per semplificare lo sviluppo delle applicazioni. In questo lavoro di tesi verranno descritti i procedimenti per includere Apache Samza e Kafka all'interno di RAM3S. Samza è un framework per lo stream processing da affiancare a Flink, Storm e Spark mentre Kafka mantiene la coda di messaggi che, al momento, viene amministrata da RabbitMQ. Verranno inoltre svolte alcune analisi di prestazioni per valutare velocità e throughput del sistema in seguito ai cambiamenti prima citati.
APA, Harvard, Vancouver, ISO, and other styles
48

Donepudi, Harinivesh. "An Apache Hadoop Framework for Large-Scale Peptide Identification." TopSCHOLAR®, 2015. http://digitalcommons.wku.edu/theses/1527.

Full text
Abstract:
Peptide identification is an essential step in protein identification, and Peptide Spectrum Match (PSM) data set is huge, which is a time consuming process to work on a single machine. In a typical run of the peptide identification method, PSMs are positioned by a cross correlation, a statistical score, or a likelihood that the match between the trial and hypothetical is correct and unique. This process takes a long time to execute, and there is a demand for an increase in performance to handle large peptide data sets. Development of distributed frameworks are needed to reduce the processing time, but this comes at the price of complexity in developing and executing them. In distributed computing, the program may divide into multiple parts to be executed. The work in this thesis describes the implementation of Apache Hadoop framework for large-scale peptide identification using C-Ranker. The Apache Hadoop data processing software is immersed in a complex environment composed of massive machine clusters, large data sets, and several processing jobs. The framework uses Apache Hadoop Distributed File System (HDFS) and Apache Mapreduce to store and process the peptide data respectively.The proposed framework uses a peptide processing algorithm named CRanker which takes peptide data as an input and identifies the correct PSMs. The framework has two steps: Execute the C-Ranker algorithm on Hadoop cluster and compare the correct PSMs data generated via Hadoop approach with the normal execution approach of C-Ranker. The goal of this framework is to process large peptide datasets using Apache Hadoop distributed approach.
APA, Harvard, Vancouver, ISO, and other styles
49

Case, Shane. "Remora aggressive power management for Apache HTTPD web server /." Diss., Online access via UMI:, 2009.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
50

Babcock, Matthew M. "Turning Apaches into Spaniards North America's forgotten Indian reservations /." Ann Arbor, Mich. : ProQuest, 2008. http://gateway.proquest.com/openurl?url_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&res_dat=xri:pqdiss&rft_dat=xri:pqdiss:3307175.

Full text
Abstract:
Thesis (Ph.D. in History)--S.M.U.
Title from PDF title page (viewed Mar. 16, 2009). Source: Dissertation Abstracts International, Volume: 69-04, Section: A, page: 1506. Adviser: David J. Weber. Includes bibliographical references.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography